Closed YXTR closed 2 months ago
Hello, you have some issues with your configurations.
FIM: llama3:latest
does not support FIM completions, you need to use a different model, or customise the fim template (results will be bad with llama3). For FIM codellama:7b-code
or a deepseek
base model is recommended. The rest of the configuration is correct for FIM in your screenshot.
CHAT: You are using the wrong endpoint; you should be using /v1/chat/completions
Please check https://github.com/rjmacarthy/twinny/blob/main/docs/providers.md for more information and recommended configurations.
Regards,
Hello, you have some issues with your configurations.
FIM:
llama3:latest
does not support FIM completions, you need to use a different model, or customise the fim template (results will be bad with llama3). For FIMcodellama:7b-code
or adeepseek
base model is recommended. The rest of the configuration is correct for FIM in your screenshot.CHAT: You are using the wrong endpoint; you should be using
/v1/chat/completions
Please check https://github.com/rjmacarthy/twinny/blob/main/docs/providers.md for more information and recommended configurations.
Regards,
Thank you for the quick response! I'll try out your suggestions right away.
Best regards,
CHAT: You are using the wrong endpoint; you should be using
/v1/chat/completions
Thank you for your help with the chat endpoint; it's working great now!
FIM:
llama3:latest
does not support FIM completions, you need to use a different model, or customise the fim template (results will be bad with llama3). For FIMcodellama:7b-code
or adeepseek
base model is recommended. The rest of the configuration is correct for FIM in your screenshot.Regarding FIM, I've switched to using
codellama:7b-code
with thecodellama
template as you suggested. The extension provided code suggestions correctly the first time, but then the same issue occurred again. Could you please advise on what might be causing this and how to resolve it?
In the screenshot below, you can see that the model generates pd.read_csv(...)
correctly, but then it fails to proceed further.
The configuration looks correct. Please try a restart of vscode.
Edit: I just noticed that there is an issue with completions without a tree-sitter parser, please try the latest version and let me know how you get on.
Many thanks,
The latest version works great!
Describe the bug The code autocompletion feature does not work and no responses are received in chat; the UI continuously shows a loading spinner.
To Reproduce
curl http://localhost:11434/api/generate -d '{ "model": "llama3", "prompt": "Why is the sky blue?" }'
(see screenshot below).Expected behavior The system should complete code snippets and reply to dialogues.
Screenshots
API Provider:
Successfully execute the following test via command line:
Logging
API Provider Label: LLama 3 Code Provider: ollama Type: fim Fim Template: automatic Hostname: localhost Path: /api/generate Protocol: http Port: 11434
Label: LLama 3 Chat Provider: ollama Type: chat Hostname: localhost Path: /api/chat Protocol: http Port: 11434
Chat or Auto Complete? Both Auto Complete and Chat.
Model Name llama3:latest
Desktop (please complete the following information):
Additional context This issue occurs on the Windows version of Ollama.
VSCode Version: