-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
In the inference widget for chat models e.g. https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1, we should send the example back, otherwise if a user sends a new message, they'll get this err…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
I use torch.profiler.profile() to profile mixtral based on vllm. And I found lots of blank before each runing step.
![S85Z22{PW)GZ0(E)4AH4AF1](https://uploads.linear.app/342cff15-f40f-4cf7-8bee-343d2…
-
There's a problem when the Mistral client is deleted.
Minimal example:
```
from mistralai.client import MistralClient
from mistralai.models.chat_completion import ChatMessage
import os
# Retri…
-
### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
Go into your Zed config file and set a model for the assistant that is not an official OpenAI mode…
-
Got error "Error Building Component
Error building vertex Hugging Face API: Failed to resolve model_id:Could not find model id for inference server: https://api-inference.huggingface.co/models/mi…
-
### The model to consider.
https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407
### The closest model vllm already supports.
- https://github.com/vllm-project/vllm/blob/main/vllm/model…
-
As an app developer who wants to add the AI Assist feature via VZCode, I want to use CodeLlama, so that I'm not locked into OpenAI.
See https://replicate.com/meta/codellama-34b/api?tab=node
```j…
-
### What is the issue?
While running mixtral:8x7b-instruct-v0.1-q4_K_M on my physical machine with x3 7900 XTX I got this error:
```
[root@5dc6ecf27031 /]# ollama run mixtral:8x7b-instruct-v0.1-…