Closed veritas9872 closed 7 months ago
Hi, you can check this to enable the FP32 training.
Thank you for the help! Unfortunately, I have found that this still requires manual editing of the model because parts such as the rotary embedding expect BF16.
Hello, I am currently trying out pre-training and I was curious if the data type used for pre-training could be configured from using
torch.bfloat16
totorch.float32
. I would like to try this out in some unstable phases of model pre-training where the extra precision might be useful. I am aware that TF32 has been enabled in the repository, so I think that the training will not become too slow. However, I was not able to find a configurable option for using FP32. Is there a location where this can be done? If not, do I have to manually change the code manually to use FP32? Many thanks in advance!