-
## Description
The ollama package (or llama2) could be interesting to demonstrate a CHAT command in mlhub.
Once we a re ready for this task, Dawei Chen may be able to give us a quick start with olla…
-
### What are you trying to do?
I would like to propose the addition of tokenize and detokenize endpoints to the Ollama server. This feature is crucial for the Ollama client interfaces (such as loll…
-
### Environment
🪟 Windows
### System
Windows 10
### Version
1.12.3
### Desktop Information
* Generation API - llama-cpp
### Describe the problem
Very recently, issue [#8402](https://github.co…
-
Is there a way to do so-called legacy completions?
- https://platform.openai.com/docs/guides/text-generation/completions-api
- https://platform.openai.com/docs/api-reference/completions/create
``…
-
The latest version of Ollama (0.3.4) is not detecting or utilizing AMD GPUs (Instinct MI210) when run in a Docker container.
I tried to run it with docker because I didn't want to update the Ollama…
-
Would like to be able to run this with local llm stacks like litellm or ollama etc.
Could you provide a parameter to specify llm and base url
-
Would be great if we can configure to add an OpenAI API.
Ollama recently added support for it. https://ollama.com/blog/openai-compatibility
EdgeAI supports OpenAI API. https://edgen.co
-
### What happened?
Why did I set up the Ollama model, but it doesn't use the locally deployed model on Ollama to answer? The Ollama link is valid because the Ollama embedding model can work. However…
-
What's the correct way to stand up this container so it's only dependent on OpenAI?
Here's what I tried:
In /frontend, I ran:
```
npm install
npm run build -- --mode hosted
```
Then I …
-
https://huggingface.co/vonjack/bge-m3-gguf
from: https://github.com/ggerganov/llama.cpp/issues/6007
I am looking for recommendations on a high-quality multilingual embedder that includes support for…