getumbrel / llama-gpt

A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device. New: Code Llama support!
https://apps.umbrel.com/app/llama-gpt
MIT License
10.59k stars 673 forks source link

M2 macbook air Internal Error #73

Open akai07 opened 11 months ago

akai07 commented 11 months ago
Screenshot 2023-08-28 at 12 29 26 AM

Terminal - INFO: Started server process [2986] INFO: Waiting for application startup. INFO: Application startup complete. INFO: Uvicorn running on http://localhost:3001/ (Press CTRL+C to quit) llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models INFO: 127.0.0.1:49485 - "GET /v1/models HTTP/1.1" 200 OK llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models INFO: 127.0.0.1:49488 - "GET /v1/models HTTP/1.1" 200 OK llama-gpt-llama-gpt-ui-mac-1 | { llama-gpt-llama-gpt-ui-mac-1 | id: '/models/llama-2-7b-chat.bin', llama-gpt-llama-gpt-ui-mac-1 | name: 'Llama 2 7B', llama-gpt-llama-gpt-ui-mac-1 | maxLength: 12000, llama-gpt-llama-gpt-ui-mac-1 | tokenLimit: 4000 llama-gpt-llama-gpt-ui-mac-1 | } 'You are a helpful and friendly AI assistant. Respond very concisely.' 0.5 '' [ { role: 'user', content: 'hi' } ] ggml_metal_graph_compute: command buffer 0 failed with status 5 GGML_ASSERT: /private/var/folders/nw/8b8162fj3sq3667m79wm49cw0000gn/T/pip-install-hcpw_ie5/llama-cpp-python_7f3b8275091343838f2dd60c58213caf/vendor/llama.cpp/ggml-metal.m:1094: false llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [SocketError: other side closed] { llama-gpt-llama-gpt-ui-mac-1 | name: 'SocketError', llama-gpt-llama-gpt-ui-mac-1 | code: 'UND_ERR_SOCKET', llama-gpt-llama-gpt-ui-mac-1 | socket: { llama-gpt-llama-gpt-ui-mac-1 | localAddress: '172.18.0.2', llama-gpt-llama-gpt-ui-mac-1 | localPort: 58896, llama-gpt-llama-gpt-ui-mac-1 | remoteAddress: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | remotePort: 3001, llama-gpt-llama-gpt-ui-mac-1 | remoteFamily: 'IPv4', llama-gpt-llama-gpt-ui-mac-1 | timeout: undefined, llama-gpt-llama-gpt-ui-mac-1 | bytesWritten: 587, llama-gpt-llama-gpt-ui-mac-1 | bytesRead: 0 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | }

Full Docker Version Terminal Log - https://ctxt.io/2/AABQSslxFQ

SIKtt commented 10 months ago

I got the same HTTP 500 error at front-end and server llama-gpt-llama-gpt-api exit with 139 error code. That occured with incomplete model file under /model folder. Try download complete model from HuggingFace and replace *.gguf file with it.

second2falcon commented 10 months ago

@SIKtt even then it does not work for me

automatumkilian commented 8 months ago

same here