Open madhack2142 opened 1 year ago
Same issue here...
Try to put the .bin files directly into the models/llama
directory, the script doesn't support walking through folders yet
Try to put the .bin files directly into the
models/llama
directory, the script doesn't support walking through folders yet
We tried several models such as ggml-model-q4_0.bin by Drararara.
After a little more than a minute a simple request "Hi!" gives out nonsense:
We only checked the text chat. The bot is silent in the voice chat. All of requirements was installed correctly.
System:
This may be due to the temperature / frequency penalty / presence penalty in the config, I find that a higher frequency penalty (1.1) is better for LLaMA models. Personally I use these settings for LLaMA:
I've tested it with pygmalion-7b, wizardlm-7b/13b and this works pretty well.
It also helps LLaMA if you provide a short chat example in your initial prompt, like so:
...your initial prompt
{user_name}: Hello!
{bot_name}: Hi, how's it going today?
{user_name}: Fine, how about you?
{bot_name}: I'm doing well too.
Please let me know if this works out for you, if not I will be happy to help you out some more.
models that tried to load, same error log when_trying_to_request.txt