-
Hello,
Great work with Stella!
I was just wondering if you've created a GGUF version of [stella_en_400M_v5](https://huggingface.co/dunzhang/stella_en_400M_v5) at all?
This would allow everyone …
-
May I use Local OpenAI-Compatible like llama.cpp server or TabbyAPI instead of Ollama?
-
I would like to run Llama-3.2 11B Vision in KoboldCPP. Ollama recently added the support, so I guess it is possible to pull it here :)
Also, there is support needed for GGUF conversion!
-
Thanks for the work。
how to build ollama with t-mac ? just replace the llama.cpp submodule ? anything to config ?
-
### What happened?
I’m experiencing an issue when using litellm proxy to communicate with the qwen-vl-plus model for multimodal interactions. When I send an image URL directly to qwen-vl-plus, it pr…
-
## Issue
* `console.log(...)` is only bound when setting up `events.ts` found in `utils`.
* It would be nice to implement some kind of universal logging for all types of interactions.
* Events
…
-
**Problem Description**
I have different ollama endpoints and I would like to choose from them. Right now I can only configure one. I run smaller models locally and larger models on inference server.…
-
### What is the issue?
I installed the latest Ollama for Windows (ARM64 build) on my 2023 Windows Dev Kit, which has an 8-core ARM processor, a Snapdragon 8cx Gen 3. It's running Windows 11 Pro.
…
-
### What is the issue?
Running llama 3.1 70b q3km on 2x4090 when there is already a colbert retriever loaded (takes up ~2800MiB VRAM) should work, but doesn't - ollama ps reports that the model is ru…
-
### Bug Description
When using AI nodes with local model, sometimes they take more than 5 mins to respond, which causes fetch to timeout (i looked through the code and past changes, and it seems to b…