Open alexhegit opened 17 hours ago
the same error
The problem is that the server runs on IPv6, but I don't know how to disable this using docker run
use docker run --rm -it -p 5000:5000 -v ~/.llama:/root/.llama --gpus=all llamastack/llamastack-local-gpu --disable-ipv6
Step1: Start the server with docker docker run --rm -it -p 5000:5000 -v ~/.llama:/root/.llama --gpus=all llamastack/llamastack-local-gpu
Step2: run the client python -m llama_stack.apis.inference.client localhost 5000
Logs: