Open dineshkh opened 1 month ago
Thanks for your interest! In an initial experiment, training with fp16 resulted in instabilities. Consequently, we adopted bf16 for training while continuing to use fp16 for inference. This approach has not led to any significant issues to date.
In the gsm8k script (link), the distilled dataset is generated using fp16 precision, while the model is trained on this dataset using bf16.
Shouldn't the precision format be consistent throughout the process?