The most no-nonsense, locally or API-hosted AI code completion plugin for Visual Studio Code - like GitHub Copilot but completely free and 100% private.
Describe the bug
any call to the chat completion api in llama.cpp returns a string of "undefined" with the DeepSeek-Coder-V2-Lite-Instruct-Q6_K.gguf model
To Reproduce
Download the model, run with the following command:
./llama-server -ngl 99 --ctx-size 16384 -m models/DeepSeek-Coder-V2-Lite-Instruct/DeepSeek-Coder-V2-Lite-Instruct-Q6_K.gguf
, add a chat provider for llamacpp at localhost:8080/v1/chat/completions.
Expected behavior
undefined is NOT repeated and the generated text makes at least some level of sense
Describe the bug any call to the chat completion api in llama.cpp returns a string of "undefined" with the DeepSeek-Coder-V2-Lite-Instruct-Q6_K.gguf model
To Reproduce Download the model, run with the following command:
./llama-server -ngl 99 --ctx-size 16384 -m models/DeepSeek-Coder-V2-Lite-Instruct/DeepSeek-Coder-V2-Lite-Instruct-Q6_K.gguf
, add a chat provider for llamacpp at localhost:8080/v1/chat/completions.Expected behavior undefined is NOT repeated and the generated text makes at least some level of sense
Screenshots![image](https://github.com/twinnydotdev/twinny/assets/11528421/7790e163-0334-4c23-966f-9e4ea2fedd19)
Logging logs.txt
API Provider as above, llama.cpp default port on localhost:8080/v1/chat/completions
Chat or Auto Complete? chat
Model Name DeepSeek-Coder-V2-Lite-Instruct-Q6_K.gguf
Desktop (please complete the following information):
Additional context