Open Bedoshady opened 6 months ago
This isn't a bug. You don't have torch + cuda installed.
conda install pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidia
Obviously install latest CUDA toolkit.
https://developer.nvidia.com/cuda-downloads?target_os=Windows&target_arch=x86_64&target_version=10&target_type=exe_local
Also you dont need to download .pth and convert to safetensors. You can grab safetensors from:
https://huggingface.co/meta-llama/Meta-Llama-3-8B/tree/main
I downloaded cuba toolkit from https://developer.nvidia.com/cuda-downloads?target_os=Windows&target_arch=x86_64&target_version=10&target_type=exe_local
and I already downloaded torch from this command
conda install pytorch torchvision torchaudio pytorch-cuda=11.7 -c pytorch -c nvidia
instead of the one you provided, but it still gives the same error.
I downloaded Cuba and mini conda to custom files, is this the problem. Do I need to restart computer after downloading cuba or the error is something else. Thanks in advance for help.
this is the automatic choices for loading model is this correct
Your snapshot shows your GPUs arent detected. Grab a clean copy of text-generation-webui
and run start_windows.bat
. Go through the installation process and follow the instructions. It should install everything you need, including torch+cuda. So long as you have Nvidia toolkit installed properly, this should work.
Keep in mind, latest webui
uses torch 2.2.1+cu121
. So your Nvidia toolkit should be at least 12.1
. Get rid of any other versions you have installed, and make sure your environmental variables CUDA_HOME
and CUDA_PATH
point to your installed location like C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.1
I am getting Retrieving notices: ...working... OPENSSL_Uplink(00007FFFB6BABD50,08): no OPENSSL_Applink
When I used the converted safetensor with web text ui, it didnt work. I tried to convert to gguf but it didnt work then I discovered that for some reason when converting .pth file to hf safetensor files were correct while the other files werent so I downloaded them of hugging face and tried again with web text ui and again it didnt work, but I converted the model to gguf successfully and now can run the model as gguf
I'm having a similar problem on Apple M2. Even when I direct the install script to use the CPU, any attempt to load a model fails because CUDA_HOME environment variable is not set.
Describe the bug
I tried to load llama 3 8b safetensor model but it doesnt work.
Is there an existing issue for this?
Reproduction
Download .pth file from meta website of llama 3 8B model then use the convert_llama_weight_to_hf provided by hugging face to turn it into safetensor model then try to load it using web text ui.
Screenshot
No response
Logs
System Info