Closed dequeueing closed 2 weeks ago
That's because you are using AutoModel
instead of AutoModelForCausalLM
to load and save your model checkpoint from remote.
You can use this code to output the checkpoint:
# Download model and tokenizer
model = AutoModelForCausalLM.from_pretrained(MODEL)
tokenizer = AutoTokenizer.from_pretrained(MODEL)
# Save model and tokenizer
tokenizer.save_pretrained(model_save_path)
model.save_pretrained(model_save_path)
print(f"Model and tokenizer saved to {model_save_path}")
BTW, I recommend you to use huggingface-cli download
to download the model repo instead of load/save from checkpoint, because loading model to memory is unnecessary for downloading:
huggingface-cli download mistralai/Mistral-7B-Instruct-v0.3 --local-dir <Your_model_save_path>
Great, that solves the problem. Thank you!
Your current environment
How would you like to use vllm
I want to run a model "mistralai/Mistral-7B-Instruct-v0.3". Here is how I download the model:
The inference python script:
And my config file:
The result I got:
Why vllm does not support MistralModel?
Before submitting a new issue...