LostRuins / koboldcpp

Run GGUF models easily with a KoboldAI UI. One File. Zero Install.
https://github.com/lostruins/koboldcpp
GNU Affero General Public License v3.0
5.35k stars 363 forks source link

Getting a looping error during inference reading [control_76][control_76][control_76]... #1056

Open ReMeDy-TV opened 3 months ago

ReMeDy-TV commented 3 months ago

Describe the Issue All the Mistral-Large GGUF models result in Koboldcpp looping a strange error msg during inference reading [control_76].

Additional Information:

split inst default settings

Vast pod hardware details: pod details

Here is the vast log, although there's no errors and nothing looks unusual to me: vast log.txt

LostRuins commented 3 months ago

Could you try a different model file? That might be a bad quant.

henk717 commented 3 months ago

Tested it with the following model : https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00001-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00002-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00003-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00004-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00005-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00006-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00007-of-00007.gguf

This works for me on the https://koboldai.org/runpodcpp A100. Runpod and VastAI share the same docker image so I expect this model to also work for you on vast.