Open smhd001 opened 5 months ago
Hey, is this behavior consistent? Does this happen with other datasets or on another retry? How's your eval loss?
It is consistent through reruns and across different subsets of this dataset. However, I currently don't have access to test it with a totally different dataset my eval loss seems normal
You may try the datasets in the example configs for testing though they're a bit small.
Please check that this issue hasn't been reported before.
Expected Behavior
no spike in loss
Current behaviour
Steps to reproduce
train a model with following config
Config yaml
Possible solution
Which Operating Systems are you using?
Python Version
3.10/docker
axolotl branch-commit
main/decb66e17013
Acknowledgements