henk717 / KoboldAI

KoboldAI is generative AI software optimized for fictional use, but capable of much more!
http://koboldai.com
GNU Affero General Public License v3.0
359 stars 130 forks source link

Error with 4bit-128g model loading #439

Open Magnaderra opened 1 year ago

Magnaderra commented 1 year ago

Hi to all!

When I load this model (from directory): https://huggingface.co/OccamRazor/pygmalion-6b-gptq-4bit, I get error:

FileNotFoundError: [Errno 2] No such file or directory: 'C:\\KoboldAI\\models\\pygmalion-6b-gptq-4bit\\quantize_config.json'

Here full bug report: Bug_Report.txt

I previously used KoboldAI's fork by 0cc4m and there was no such error.

Here the debug dump: kobold_debug.zip

henk717 commented 1 year ago

Did you select occam gptq as the GPTQ backend? Because for autogptq you need an extra file.

Magnaderra commented 1 year ago

Did you select occam gptq as the GPTQ backend? Because for autogptq you need an extra file.

Yes, I have occam gptq selected

henk717 commented 1 year ago

Traced this back to the newer transformers version we are using compared to occam's old branch. We will have to find a way to make the changes required to handle this old model, for the time being you have two options. You can either downgrade to the old transformers by opening the KoboldAI command prompt and typing this : pip install transformers == 4.28 This method will harm your model compatibility but restores this old model, you can also use a different version of the model such as the one in the chat model menu.

Magnaderra commented 1 year ago

pip install transformers == 4.28

Okay, so far that's worked.