Are there docs on best practices for using vllm hosted models?
I create a model using
python -m vllm.entrypoints.openai.api_server --model model_path
and try running it as
lm_eval --model local-chat-completions --model_args model=model_path,base_url=http://localhost:8000/v1 --tasks /home/darshshah/lm_eval/tasks/financebench_inference_binary --batch_size 12 --output_path /home/darshshah/lm_eval/tasks/financebench_inference_binary/outputs --log_samples
But get the following errors
openai.OpenAIError: The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable
Are there docs on best practices for using vllm hosted models?
I create a model using python -m vllm.entrypoints.openai.api_server --model model_path
and try running it as lm_eval --model local-chat-completions --model_args model=model_path,base_url=http://localhost:8000/v1 --tasks /home/darshshah/lm_eval/tasks/financebench_inference_binary --batch_size 12 --output_path /home/darshshah/lm_eval/tasks/financebench_inference_binary/outputs --log_samples
But get the following errors openai.OpenAIError: The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable