getumbrel / llama-gpt

A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device. New: Code Llama support!
https://apps.umbrel.com/app/llama-gpt
MIT License
10.59k stars 673 forks source link

Code Llama 7b does not work on MacBook M1 #95

Open AbhiPawar5 opened 10 months ago

AbhiPawar5 commented 10 months ago

Hi team, awesome work making these models run locally :)

I see the following connection refused error when I try to run Code Llama 7b on MacBook M1 Pro. Command: ./run-mac.sh --model code-7b

Output: Xcode is installed at /Library/Developer/CommandLineTools Conda is installed. Conda environment 'llama-gpt' already exists. Activating the conda environment 'llama-gpt'... ERROR: Pipe to stdout was broken Exception ignored in: <_io.TextIOWrapper name='' mode='w' encoding='utf-8'> BrokenPipeError: [Errno 32] Broken pipe llama-cpp-python is installed. llama-cpp-python version is 0.1.80. ./models/code-llama-7b-chat.gguf model found. Initializing server with: Batch size: 2096 Number of CPU threads: 8 Number of GPU layers: 1 Context window: 8192 GQA: [+] Building 0.3s (2/3) docker:desktop-linux => [llama-gpt-ui-mac internal] load .dockerignore 0.0s => => transferring context: 82B 0.0s => [llama-gpt-ui-mac internal] load build definition from no-wait.Docker 0.0s => => transferring dockerfile: 753B 0.0s => [llama-gpt-ui-mac internal] load metadata for docker.io/library/node: 0.3s /Users/abhishek.pawar/miniconda3/envs/llama-gpt/lib/python3.11/site-packages/pyd[+] Building 1.0s (17/17) FINISHED docker:desktop-linuxi => [llama-gpt-ui-mac internal] load .dockerignore 0.0s => => transferring context: 82B 0.0s => [llama-gpt-ui-mac internal] load build definition from no-wait.Docker 0.0se => => transferring dockerfile: 753B 0.0s => [llama-gpt-ui-mac internal] load metadata for docker.io/library/node: 0.9s => [llama-gpt-ui-mac base 1/3] FROM docker.io/library/node:19-alpine@sha 0.0s => [llama-gpt-ui-mac internal] load build context 0.0s => => transferring context: 13.73kB 0.0s => CACHED [llama-gpt-ui-mac base 2/3] WORKDIR /app 0.0s => CACHED [llama-gpt-ui-mac base 3/3] COPY package*.json ./ 0.0s => CACHED [llama-gpt-ui-mac dependencies 1/1] RUN npm ci 0.0s => CACHED [llama-gpt-ui-mac production 3/8] COPY --from=dependencies /ap 0.0s => CACHED [llama-gpt-ui-mac build 1/2] COPY . . 0.0s => CACHED [llama-gpt-ui-mac build 2/2] RUN npm run build 0.0s => CACHED [llama-gpt-ui-mac production 4/8] COPY --from=build /app/.next 0.0s => CACHED [llama-gpt-ui-mac production 5/8] COPY --from=build /app/publi 0.0s => CACHED [llama-gpt-ui-mac production 6/8] COPY --from=build /app/packa 0.0s => CACHED [llama-gpt-ui-mac production 7/8] COPY --from=build /app/next. 0.0s => CACHED [llama-gpt-ui-mac production 8/8] COPY --from=build /app/next- 0.0s => [llama-gpt-ui-mac] exporting to image 0.0s => => exporting layers 0.0s => => writing image sha256:4187e01c6c862f9db9160be158ee74bb0151edec0a84a 0.0s => => naming to docker.io/library/llama-gpt-llama-gpt-ui-mac 0.0s [+] Running 2/1 ✔ Network llama-gpt_default Created 0.0s ✔ Container llama-gpt-llama-gpt-ui-mac-1 Created 0.0s Attaching to llama-gpt-llama-gpt-ui-mac-1 llama-gpt-llama-gpt-ui-mac-1 | llama-gpt-llama-gpt-ui-mac-1 | > ai-chatbot-starter@0.1.0 start llama-gpt-llama-gpt-ui-mac-1 | > next start llama-gpt-llama-gpt-ui-mac-1 | llama-gpt-llama-gpt-ui-mac-1 | llama-gpt-llama-gpt-ui-mac-1 | ready - started server on 0.0.0.0:3000, url: http://localhost:3000 llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | making request to http://host.docker.internal:3001/v1/models llama-gpt-llama-gpt-ui-mac-1 | [TypeError: fetch failed] { llama-gpt-llama-gpt-ui-mac-1 | cause: [Error: connect ECONNREFUSED 192.168.65.254:3001] { llama-gpt-llama-gpt-ui-mac-1 | errno: -111, llama-gpt-llama-gpt-ui-mac-1 | code: 'ECONNREFUSED', llama-gpt-llama-gpt-ui-mac-1 | syscall: 'connect', llama-gpt-llama-gpt-ui-mac-1 | address: '192.168.65.254', llama-gpt-llama-gpt-ui-mac-1 | port: 3001 llama-gpt-llama-gpt-ui-mac-1 | } llama-gpt-llama-gpt-ui-mac-1 | } ^C./run-mac.sh: line 250: 32853 Segmentation fault: 11 python3 -m llama_cpp.server --n_ctx $n_ctx --n_threads $n_threads --n_gpu_layers $n_gpu_layers --n_batch $n_batch --model $MODEL --port 3001 Stopping docker-compose... Gracefully stopping... (press Ctrl+C again to force) Aborting on container exit... [+] Stopping 1/1 ✔ Container llama-gpt-llama-gpt-ui-mac-1 Stopped 0.1s canceled [+] Running 2/0 ✔ Container llama-gpt-llama-gpt-ui-mac-1 Removed 0.0s ✔ Network llama-gpt_default Removed 0.0s Stopping python server...

Nisse123 commented 10 months ago

I think it is the same error, the repo is not working right now due to this issue. https://github.com/abetlen/llama-cpp-python/issues/520

Daniel1989 commented 9 months ago

same issue

fangdajiang commented 9 months ago

same issue

100tomer commented 8 months ago

any solution?

automatumkilian commented 8 months ago

same issue here

dylan-sh commented 8 months ago

same