Open fake-name opened 1 week ago
@fake-name Is this similar? https://github.com/huggingface/transformers/blob/1c1aec2ef1d6822fae3ffbb973b4c941f65f4ddf/docs/source/en/model_sharing.md?plain=1#L42-L48
It's not complicated to implement, just pass in a parameter, and by default it will be the default branch and the latest commit.
You may want this https://github.com/InternLM/lmdeploy/blob/main/lmdeploy/messages.py#L151, but it can't be set by cli currently. The cli should be updated.
With https://github.com/InternLM/lmdeploy/pull/1814 you are specify a model revision. But we don't support bartowski/Yi-34B-200K-RPMerge-exl2. It use ExLlamaV2 to quant the model, we only support awq quantization method.
For awq quantization, you can refer to this
Checklist
Describe the bug
From what I can tell, this uses the various Transformers
*.from_pretrained()
calls to download models from hugging face.I'm using some models where the the actual model files are only present in non-main branches of the repository. For an example, see https://huggingface.co/bartowski/Yi-34B-200K-RPMerge-exl2
This fails if you try to simply refer to the model.
I cannot see any way to specify a specific branch.
Reproduction
Environment