Open jcushman opened 1 year ago
For loading from file system, hf_model_id
should be set to a path pointing to a directory containing a model in huggingface format (config.json, bin files, etc.), not to the bin file itself. Can you try that?
Hi, i got the same error when trying to load the model from S3. I've followed the instructions on the /model/README.md. Here are my engine configs:
engine_config:
model_id: meta/llama-2-7b
s3_mirror_config:
bucket_uri: s3://bucket_name/llama-2-7b
type: VLLMEngine
...
and inside the bucket there are:
- config.json
- generation_config.json
- model.safetensors
- quant_config.json
- special_tokens_map.json
- tokenizer_config.json
- tokenizer.json
- tokenizer.model
models/README.md says "For loading a model from file system, set engine_config.hf_model_id to an absolute filesystem path accessible from every node in the cluster."
I ran:
with /models/myconfig.yaml having
hf_model_id: /models/llama-2-13b-chat.ggmlv3.q4_1.bin
The output was:
This looks like
hf_model_id
is being validated as a Hugging Face repo name and can't be an absolute path.