Closed jimburtoft closed 3 months ago
This model generation config is invalid (always has been), and the latest version of transformers
rejects it when we try to save it during export.
https://github.com/huggingface/transformers/blob/96eb06286b63c9c93334d507e632c175d6ba8b28/src/transformers/generation/configuration_utils.py#L720
What could be done is to detect the invalid generation_config during export and skip it (in that case a new one would be created from the model).
I created a pull-request to update the model generation config:
https://huggingface.co/NousResearch/Llama-2-7b-chat-hf/discussions/9
And here is a pull-request with a workaround for optimum-neuron
to be included in the next release:
https://github.com/huggingface/optimum-neuron/pull/618
@jimburtoft since the Nous models have been updated, can we close this ?
System Info
Who can help?
@dacorvo @JingyaHuang
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction (minimal, reproducible, runnable)
The following code works on optimum 0.0.22, but failed on 0.0.23. Fails on HF DLAMI 20240531 on inf2.8xlarge. (worked on the previous HF DLAMI).
Code:
On 0.0.23:
Working on 0.0.22:
Expected behavior
Included above. Expected behavior should be the same success as on 0.0.22