Open ProgrammingLife opened 8 months ago
nvidia-ctk runtime configure
systemctl restart docker
This worked for me, thanks! Now I'm at "forward compatibility was attempted on non supported HW"
llama-gpt-api-cuda-ggml-1 | llama-gpt-api-cuda-ggml-1 | /models/llama-2-7b-chat.bin model found. llama-gpt-api-cuda-ggml-1 | make: *** No rule to make target 'build'. Stop. llama-gpt-api-cuda-ggml-1 | Initializing server with: llama-gpt-api-cuda-ggml-1 | Batch size: 2096 llama-gpt-api-cuda-ggml-1 | Number of CPU threads: 24 llama-gpt-api-cuda-ggml-1 | Number of GPU layers: 10 llama-gpt-api-cuda-ggml-1 | Context window: 4096 llama-gpt-api-cuda-ggml-1 | CUDA error 804 at /tmp/pip-install-7rxfzzup/llama-cpp-python_c62cf07cbfa449a7b268f9102316d6db/vendor/llama.cpp/ggml-cuda.cu:4883: forward compatibility was attempted on non supported HW
Setup:
System:
Host: TowerPC Kernel: 6.1.0-18-amd64 arch: x86_64 bits: 64
Desktop: KDE Plasma v: 5.27.5 Distro: Debian GNU/Linux 12 (bookworm)
Machine:
Type: Desktop Mobo: ASUSTeK model: PRIME X299-A II v: Rev 1.xx
serial:
I'm seeing this same error in a 1080 Ti, but when issuing nvidia-smi
, the result is
❯ nvidia-smi
Failed to initialize NVML: Driver/library version mismatch
NVML library version: 550.54
I followed the instructions from nVidia's CUDA page here to install the CUDA drivers, but I'm guessing there's a driver mismatch somewhere..
Edit Miraculously, a reboot didn't break my system and I now see similar output in the nvidia-smi
command. However, I don't have an nvidia-ctk
command that I can run, so I remain stuck(ish)
Edit 2 - A quick search indicated I needed to install the nVidia Container Toolkit, and restart docker. The errors have now gone away.
I was able to bring up the docker-compose-cuda-ggml.yml
file using the command docker compose -f docker-compose-cuda-ggml.yml up -d
- however, the other cuda compose (gguf) did not work for me
I've successfully use my RTX 3080ti with Stable Diffusion, Fooocus, Stable Cascade and my system is ready to work with GPU. Arch Linux.
What should I check?