Open tomgelu opened 1 year ago
I could not run it on GPU either
ngl i recommend you just buy a second hard drive and slap linux on it as a second boot and do all of your cuda work there. It makes your life infinitely easier.
Unfortunately I'm using virtual machine running on Windows with a A4500 GC, but Windows is without virtualization enabled...
I'm able to run it on a Windows machine using cuda 11.8.
Unfortunately I'm using virtual machine running on Windows with a A4500 GC, but Windows is without virtualization enabled...
If you are not using a Windows Host machine, maybe you have No GPU Passthrough: Without virtualization extensions, utilizing GPU passthrough (allocating the physical GPU to the VM) might not be possible or could be challenging in your virtual machine. Maybe it would someone help to get more information which virtualization software, which version you are using, what you have everything tried...., if it is your virtualization issue, then you should open a ticket for your virtualization software, if they have support for GPU Passthrough.
But if you try it out on your Windows Hostmachine, this worked for me on Windows 11 I assume A4500 GC is the the nvidia RTX 4500 graphic card and should work with cuda
If your Windows Driver can't be updated over >=527.41 , checkout which cuda version you should install here
git clone https://github.com/PromtEngineer/localGPT.git
conda create -n localGPT python=3.10.0
conda activate localGPT
conda install virtualenv
virtualenv YourNewVirtualEnvironment
.\YourNewVirtualEnvironment\Scripts\activate
pip install -r requirements.txt
set CMAKE_ARGS="-DLLAMA_CUBLAS=on"
set FORCE_CMAKE=1
pip install llama-cpp-python --no-cache-dir
pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
On my troubleshooting journey where chatGPT and also following posts usefull
Hello, I've been trying to run it on Windows with gpu and different models. I tried Cuda 11.8 and 12.2 with pytorch compiled with matching version but wasn't able to even chat with a model.
Is it even possible ?