Open preethivenkatesh opened 2 months ago
https://github.com/opea-project/GenAIComps/tree/main/comps/llms#1-start-microservice-with-python-option-1
docker run -it --name vllm_service -p 8008:80 -e HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -v ./data:/data vllm:cpu /bin/bash -c "cd / && export VLLM_CPU_KVCACHE_SPACE=40 && python3 -m vllm.entrypoints.openai.api_server --model ${your_hf_llm_model} --port 80"
vllm:cpu is not in docker hub, instruction missing to build from source
add the command bash build_docker_vllm.sh
bash build_docker_vllm.sh
https://github.com/opea-project/GenAIComps/tree/main/comps/llms#1-start-microservice-with-python-option-1
docker run -it --name vllm_service -p 8008:80 -e HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -v ./data:/data vllm:cpu /bin/bash -c "cd / && export VLLM_CPU_KVCACHE_SPACE=40 && python3 -m vllm.entrypoints.openai.api_server --model ${your_hf_llm_model} --port 80"
vllm:cpu is not in docker hub, instruction missing to build from source