-
It will be great if baml works with some batch inference tools like vLLM or add its own
-
### Description
I have the following error while uploading and indexing file (both in docker in direct installation)
![image](https://github.com/user-attachments/assets/85b1717d-2228-41ec-9518-79bdb…
-
![微信截图_20240626152830](https://github.com/ollama/ollama/assets/129468439/23e316b8-cb87-4783-81af-96b94690a61a)
-
The wrong occurred when I use new llama3.1,and it never occurred when I use other models
![download](https://github.com/user-attachments/assets/5bd93a1b-c70e-42d4-a5de-c4e3f752ceae)
-
This is a little more complicated as it will require creating an Ollama Modelfile / manifest in addition to linking the models.
- lm-studio (mostly) parses the filename and the GGML/GGUF metadata t…
-
It would be great to have some sort of authentication in front of the ollama api. Currently I´m using Nginx Proxy Manager to add a Access List to prevent unauthorized access but a standard way impleme…
-
when comparing local llama2 model with GPT4, got error:
ERROR Failed to send a test chat completion to the GPT deployment due to error: …
-
I'm seeing this error when trying to translate selection. API calls are proxied via `litellm` so I can see there was no outgoing call. Seems to be happening when content is parse before the API call..…
-
Add Gradio UI
Add ngrok support if gardio is running on server
-
Is it possible to support MLX as backend? That would be great for local development environment.