Open ReMeDy-TV opened 3 months ago
Could you try a different model file? That might be a bad quant.
Tested it with the following model : https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00001-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00002-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00003-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00004-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00005-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00006-of-00007.gguf,https://huggingface.co/MaziyarPanahi/Mistral-Large-Instruct-2407-GGUF/resolve/main/Mistral-Large-Instruct-2407.Q4_K_S.gguf-00007-of-00007.gguf
This works for me on the https://koboldai.org/runpodcpp A100. Runpod and VastAI share the same docker image so I expect this model to also work for you on vast.
Describe the Issue All the Mistral-Large GGUF models result in Koboldcpp looping a strange error msg during inference reading [control_76].
Additional Information:
Vast pod hardware details:
Here is the vast log, although there's no errors and nothing looks unusual to me: vast log.txt