Closed heylamourding closed 1 year ago
Please check out the updated instruction for loading lora weights. You need to pass a corresponding --model-base
, and in your case llama-2-7b-chat
.
请问可以生成类似“liuhaotian/llava-llama-2-13b-chat-lightning-preview”的,7b版本的文件吗?
Hi @haotian-liu thanks for your reply!
I tried to send
python -m llava.serve.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40000 --worker http://localhost:40000 --model-path liuhaotian/llava-llama-2-7b-chat-lightning-lora-preview --model-base llama-2-7b-chat
I got following errors:
OSError: llama-2-7b-chat is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
May I know what value should I put under model-base?
Alternatively, is it possible to generate liuhaotian/llava-llama-2-7b-chat-lightning-preview?
Thanks!
Hi @haotian-liu , Can we generate a 7b version file like “liuhaotian/llava-llama-2-13b-chat-lightning-preview” by ourselves?
Updated: error is solved. Need to get llama2 access in HF first. Then
python merge_lora_weights.py --model-path liuhaotian/llava-llama-2-7b-chat-lightning-lora-preview --model-base meta-llama/Llama-2-7b-chat-hf --save-model-path ./checkpoints/llava-7b-llama-2-7b-chat
Use ./checkpoints/llava-7b-llama-2-7b-chat in inference.
Hi @heylamourding ,I have got the llava-2 access in HF, but how do I use it on the server?
Describe the issue
Issue:
Command:
Log:
Screenshots: You may attach screenshots if it better explains the issue.