Open lucyknada opened 4 months ago
Zamba is not officially supported in transformers. I expect simply adding a catchall **kwargs
in the forward would help. see https://github.com/Zyphra/transformers_zamba2/blob/main/src/transformers/models/zamba2/modeling_zamba2.py#L786
Thanks for pointing this out! The error is related to the trainer passing an additional positional argument, so we added *args in this commit https://github.com/Zyphra/transformers_zamba2/commit/8af28567d3e8913730243167ec5c70c1a656bbcd. Please let us know if you're having more issues with this.
Please check that this issue hasn't been reported before.
Expected Behavior
no error
Current behaviour
Steps to reproduce
install the zamb2 transformers:
then simply run the config on 8x a6000, it'll crash.
Config yaml
Possible solution
No response
Which Operating Systems are you using?
Python Version
Python 3.10.14
axolotl branch-commit
c5587b45accdc50e0be7b2ec9ed3a66879d68156
Acknowledgements