-
### What is the issue?
```
Jun 27 12:06:15 ollama[11759]: INFO [main] build info | build=1 commit="7c26775" tid="124734763667456" timestamp=1719504375
Jun 27 12:06:15 ollama[11759]: INFO [main] s…
-
**rust info:**
```
stable-x86_64-pc-windows-msvc (default)
rustc 1.76.0 (07dca489a 2024-02-04)
```
**run:**
```bash
cargo build --release
./target/release/mistralrs-server.exe --port 8080 -…
-
### MediaPipe Solution (you are using)
Version: 0.10.14
### Programming language
_No response_
### Are you willing to contribute it
None
### Describe the feature and the current behaviour/state
…
-
### What is the issue?
When I want to download new models, for example I run: ollama run gemma2:27b
The model download will stuck, and according to my task manager, my C: drive SSD is always 100…
-
### Do you need to file an issue?
- [X] I have searched the existing issues and this bug is not already filed.
- [ ] My model is hosted on OpenAI or Azure. If not, please look at the "model providers…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
## Goal
- Cortex's Built-in Libraries have a "Cortex Model repo" format
- Cortex Model Repos are a critical data structure to support `cortex pull ` and `cortex run `
## High-level Structure
…
-
### What is the issue?
I'm using OLLAMA with Anything LLM. It's taking a lot of time to respond to the prompts
The following error I'm getting from the docker logs of anything llm
[OllamaEmbedd…
-
尝试用lmdeloy进行gemma2的离线推理:
```
from lmdeploy import pipeline, GenerationConfig, PytorchEngineConfig
backend_config = PytorchEngineConfig(tp=2)
gen_config = GenerationConfig(top_p=0.8,
…
-
### What do you need?
After some experimentation, q8 llama3 and mistral **run locally** are performing **much** better after adjusting the prompt order. I propose restructuring the patterns so tha…