-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue]…
-
Thank you for your outstanding work. I need to evaluate my base model after fine-tuning it, but it seems that our fine-tuning is different, leading to mistakes in the evaluation process. Could you ple…
-
The "/api/generate" is not functioning and display 404 on the Windows version (not WSL), despite the Ollama server running and "/" being accessible. The same code works on the Ollama server on m…
-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue](ht…
-
### What happened?
Since version 2.3.1 it is not sending the message history of a chat tab, but only the system prompt and the latest message.
This is the request of the second message:
```
[2…
-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue]…
-
I'm running all controller, model_worker and openai-server together with CodeLlama-7b-Instruct-hf model. When I send the request with `stream: true` it fails to answer. The request is the following.
…
-
Has there been a mix up with the files uploaded to HF and some of the `CodeLlama-70b-Instruct-hf` files have been used instead of `CodeLlama-70b-hf`:
https://huggingface.co/openbmb/Eurus-70b-nca/di…
-
There are several projects aiming to make inference on CPU efficient.
The first part is research:
- Which project works better,
- And compatible with Refact license,
- And doesn't bloat the dock…
-
### Motivation
In the code-llama's deploy tutorial, quantization chapter remains to be done, when will this feature finished?
### Related resources
_No response_
### Additional context
_No respon…