Closed Aunxfb closed 1 year ago
I got it runing on windows 11 so its definitly possible not sure. Maybe ask chatGPT4 in chat.openai website. There are many things it had me do. Also it depends if i rightclick open command promt inside the folder or where i initiate it from
I got it runing on windows 11 so its definitly possible not sure. Maybe ask chatGPT4 in chat.openai website. There are many things it had me do. Also it depends if i rightclick open command promt inside the folder or where i initiate it from
What did you ask chatGPT?
I also got it running on Windows 11 with the following hardware: Intel(R) Core(TM) i5-6500 CPU @ 3.20GHz 3.19 GHz and Installed RAM 15.9 GB
I used the Visual Studio download, put the model in the chat folder and voila, I was able to run it. This was even before I had python installed (required for the GPT4All-UI). The model I used was gpt4all-lora-quantized.bin ... it worked out of the box for me. My setup took about 10 minutes. Maybe try deleting everything and starting over from scratch and don't do anything other than follow the instructions exactly? You probably did but who knows? Maybe something went wrong during the download. I'd try again.
However note: on my hardware the model works, but after a few minutes pegs my CPU at 100% and then gets very sketchy after that. Finally it will crash powershell and I have to start again. So keep that in mind for lower end windows machines... like mine.
I also got it running on Windows 11 with the following hardware: Intel(R) Core(TM) i5-6500 CPU @ 3.20GHz 3.19 GHz and Installed RAM 15.9 GB
I used the Visual Studio download, put the model in the chat folder and voila, I was able to run it. This was even before I had python installed (required for the GPT4All-UI). The model I used was gpt4all-lora-quantized.bin ... it worked out of the box for me. My setup took about 10 minutes. Maybe try deleting everything and starting over from scratch and don't do anything other than follow the instructions exactly? You probably did but who knows? Maybe something went wrong during the download. I'd try again.
However note: on my hardware the model works, but after a few minutes pegs my CPU at 100% and then gets very sketchy after that. Finally it will crash powershell and I have to start again. So keep that in mind for lower end windows machines... like mine.
That did not sound like you ran it on GPU tbh... (the use of gpt4all-lora-quantized.bin gave it away...). Thanks for trying to help but that's not what I'm trying to do.. and I did follow the instructions exactly, specifically the "GPU Interface" section.
Edit: I did manage to run it the normal / CPU way, but it's quite slow so i want to utilize my GPU instead.
Anyone figure this out?
Tried https://huggingface.co/nomic-ai/gpt4all-j and got:
RuntimeError: "LayerNormKernelImpl" not implemented for 'Half'
full ouput: error.txt
I also got it running on Windows 11 with the following hardware: Intel(R) Core(TM) i5-6500 CPU @ 3.20GHz 3.19 GHz and Installed RAM 15.9 GB I used the Visual Studio download, put the model in the chat folder and voila, I was able to run it. This was even before I had python installed (required for the GPT4All-UI). The model I used was gpt4all-lora-quantized.bin ... it worked out of the box for me. My setup took about 10 minutes. Maybe try deleting everything and starting over from scratch and don't do anything other than follow the instructions exactly? You probably did but who knows? Maybe something went wrong during the download. I'd try again. However note: on my hardware the model works, but after a few minutes pegs my CPU at 100% and then gets very sketchy after that. Finally it will crash powershell and I have to start again. So keep that in mind for lower end windows machines... like mine.
That did not sound like you ran it on GPU tbh... (the use of gpt4all-lora-quantized.bin gave it away...). Thanks for trying to help but that's not what I'm trying to do.. and I did follow the instructions exactly, specifically the "GPU Interface" section.
Edit: I did manage to run it the normal / CPU way, but it's quite slow so i want to utilize my GPU instead.
I'm also in the same situation, similar specs to you, want to take advantage of my GPU having successfully run on CPU.
Did you end up finding a solution? Which model have yo used? I'm trying with llama-7b-hf.
Hi, I am a total newbie and I ma having the same issue. I followed all instructions in the nomic repo but can't find a way to fix this. where do you change the datatype to float32? I am using the sample py script in the repo to start and the model "decapoda-research/llama-7b-hf"
I tried on a 24Gb A5500 and an AMD Radeon Pro W6800 32Gb
Thx
Tried https://huggingface.co/nomic-ai/gpt4all-j and got:
RuntimeError: "LayerNormKernelImpl" not implemented for 'Half'
full ouput: error.txt
i am facing the same issue. Did you find any solution?
I haven't found a working model for GPU yet. I haven't troubleshooted much more than that though, Only trying to download various models gpt4all and lora/llama models hoping to get lucky. It's probably a simple solution to update the code, or to convert a model for it, but I haven't taken the time to understand what the model config requirements are and what ggml, quantized, etc mean in this context.
I'd look into seeing how the peft and transformers wheels were compiled, it might be as simple as updating those. Guessing because of:
\.venv\lib\site-packages\peft\tuners\lora.py:357 in forward
The other thing I've thought about doing is checking the code when the GPU instructions were added to the readme, and see what changes have been made in the code and dependency versions, etc since. Earliest mention of GPU I've found so far is here: https://github.com/nomic-ai/gpt4all/tree/e8c6aeeea27fe786b2bc6c3c32c2720c9660660e
I tried changing flotas16 to floats32 in the transformers json config file and in gpt4all...but no luck, same error. Also other models are not working for the same or other errors. I am a newbie in coding so haven't got the time now tocatch with 10 years of knowledge and debug 😄
Stale, please open a new issue if this is still relevant.
Summary: Can't get pass #RuntimeError: "addmm_implcpu" not implemented for 'Half'#
Since the error seems to be due to things not being run on GPU(?), I tried:
I tried to ran gpt4all with GPU with the following code from the readMe:
And downloaded some models from: https://huggingface.co/models
I tried llama-7b-hf and alpaca-lora-7b, both hit with the same error; The following is the full console output:
Not sure if it matters, but my specs: 64GB RAM RTX 3080 10GB i7 12700k