Closed AceBeaker2 closed 1 year ago
Can you show the output of nvidia-smi
?
(llama) orion@skynet:~$ nvidia-smi
Fri Mar 17 20:30:47 2023
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 530.30.02 Driver Version: 530.30.02 CUDA Version: 12.1 |
|-----------------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 NVIDIA GeForce RTX 3060 Ti On | 00000000:09:00.0 Off | N/A |
| 0% 38C P8 18W / 200W| 157MiB / 8192MiB | 3% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
+---------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=======================================================================================|
| 0 N/A N/A 2019 G /usr/lib/xorg/Xorg 71MiB |
| 0 N/A N/A 2309 G cinnamon 33MiB |
| 0 N/A N/A 3444 G ...53317852,5022907119068916348,131072 51MiB |
+---------------------------------------------------------------------------------------+
Can you show the output of
nvidia-smi
?
I managed to fix it by rebooting, but I'm getting a new error now:
(llama) orion@skynet:~/AI-Horde-Worker/llama.cpp$ python3 pyllama/llama/llama_quant.py models/7B/ c4 --wbits 16 --save pyllama-7B8b.pt
Traceback (most recent call last):
File "/home/orion/AI-Horde-Worker/llama.cpp/pyllama/llama/llama_quant.py", line 6, in <module>
from gptq import (
File "/home/orion/.local/lib/python3.10/site-packages/gptq/__init__.py", line 9, in <module>
from .gptq import GPTQ
File "/home/orion/.local/lib/python3.10/site-packages/gptq/gptq.py", line 5, in <module>
from .quant import quantize
File "/home/orion/.local/lib/python3.10/site-packages/gptq/quant.py", line 4, in <module>
from quant_cuda import matvmul2, matvmul3, matvmul4, matvmul8, matvmul16
ModuleNotFoundError: No module named 'quant_cuda'
Environment is ubuntu 22, cuda 12.1, rtx 3060 ti