Closed doug-wade closed 6 months ago
I followed the instructions for setting it up via docker compose from the community, and I am still getting gRPC service errors:
» curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{
"model": "codellama-7b-gguf",
"messages": [{"role": "user", "content": "Please write a function that calculates the first n prime numbers."}],
"temperature": 0.9
}'
{"error":{"code":500,"message":"could not load model - all backends returned error: 24 errors occurred:\n\t* grpc service not ready\n\t* could not load model: rpc error: code = Unavailable desc = error reading from server: EOF\n\t* grpc service not ready\n\t* could not load model: rpc error: code = Unknown desc = failed loading model\n\t* could not load model: rpc error: code = Unknown desc = failed loading model\n\t* could not load model: rpc error: code = Unavailable desc = error reading from server: EOF\n\t* grpc service not ready\n\t* could not load model: rpc error: code = Unknown desc = stat /models/codellama-7b-gguf: no such file or directory\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/tinydream. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* could not load model: rpc error: code = Unknown desc = unsupported model type /models/codellama-7b-gguf (should end with .onnx)\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/bark/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/sentencetransformers/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/coqui/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/exllama2/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/vllm/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/diffusers/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/petals/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/mamba/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/transformers-musicgen/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/autogptq/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/transformers/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/exllama/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/sentencetransformers/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\t* grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/build/backend/python/vall-e-x/run.sh. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS\n\n","type":""}}%
This seems to be the error that is output when you request a model that does not exist -- I had my model file in the root directory of the project instead of /models
, which is where it is configured to search by default. Not sure whether to leave this open for a better error message in this case, or to close it because it seems to be caused by user error.
You solved it?, i have the same error
@sesirbu I managed to work around it by building from source. I think the trouble is that I'm on Apple Silicon and pulling Docker image for linux/amd64
, and there is some incompatibility being surfaced. I followed this guide and the binary runs without these errors.
I am also experiencing this issue.
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:30PM INF Loading model 'all-MiniLM-L6-v2' with backend sentencetransformers
8:31PM ERR Failed starting/connecting to the gRPC service: rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:45031: connect: connection refused"
8:31PM ERR Failed starting/connecting to the gRPC service: rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:36203: connect: connection refused"
...
using docker compose with config
version: "3.9"
name: emesabot
networks:
emesabot:
driver: host
services:
localai:
container_name: localai
image: localai/localai:v2.6.1-cublas-cuda12-core
command: llava phi-2 all-minilm-l6-v2
volumes:
- "./models:/build/models:cached"
environment:
- 'ADDRESS=0.0.0.0:8080'
- 'GALLERIES=[{"name":"model-gallery", "url":"github:go-skynet/model-gallery/index.yaml"}, {"url": "github:go-skynet/model-gallery/huggingface.yaml","name":"huggingface"}]'
ports:
- "8080:8080"
networks:
- emesabot
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: 1
capabilities: [gpu]
anything-llm:
container_name: anything-llm
image: mintplexlabs/anythingllm:latest
cap_add:
- SYS_ADMIN
volumes:
- "./.env:/app/server/.env"
- ".anything-llm/server/storage:/app/server/storage"
- ".anything-llm/collector/hotdir/:/app/collector/hotdir"
- ".anything-llm/collector/outputs/:/app/collector/outputs"
user: "${UID:-1000}:${GID:-1000}"
ports:
- "3001:3001"
- "8888:8888"
env_file:
- .env
networks:
- emesabot
extra_hosts:
- "host.docker.internal:host-gateway"
you have to make the gRPC service first if you want to use, exemple:
make GO_TAGS=stablediffusion,tts CUDA_LIBPATH=/usr/local/cuda FMPEG=true BUILD_API_ONLY=false BUILD_TYPE=cuBLAS BUILD_GRPC_FOR_BACKEND_LLAMA=true build
I have the same error while using Docker image and just a single model (docker-compose):
services:
api:
image: localai/localai:v2.12.4-ffmpeg-core
command:
- whisper-base
LocalAI version: 2.8.2
Environment, CPU architecture, OS, and Version:
Describe the bug
To Reproduce
Expected behavior To return a code snippet
Logs
Additional context This is my first time trying to start the project.