-
## ⚙️ Request New Models
- Link to an existing implementation (e.g. Hugging Face/Github):
## Additional context
-
# Language Model Overview
## OpenAI
| | gpt-4o | gpt-4o-mini …
-
### Your current environment
The output of `python collect_env.py`
```text
PyTorch version: 2.4.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N…
-
Please make a note of this on the readme document here and on the library page at https://ollama.com/library.
Ollama gave me this error message when I tried to run mistral-large. It's huge.
-
On `"@slack/bolt": "^3.19.0"`, I got a strange bug where a `static_select` element from Block Builder do not display the `initial_option` correctly in my Slack app.
I'm using `"slack-block-builder"…
-
# Motivation
I wanted to participate more in solving the listed issues, but I already spent more than $30 on debugging with the ChatGPT API, lol.
Recently, Mistral announced that they have reduced…
-
### What happened?
Was running Mistral Large 2 with partial offload with AMD 5600X + RTX 3090.
Provided the same ~28k prompt to each, llama.cpp produced output that was coherent and similar to non q…
-
I am working with “magnum-v2-123b-Q4_K_L” model (I also tried “magnum-v2-123b-iQ4_K_M” - no difference). I've noticed that the context shift mechanism with this model works somehow wrong, if not to sa…
-
### Misc discussion on performance
Hi all, I'm having trouble with maximizing the performance of batch inference of big models on vllm 0.6.3
(Llama 3.1 70b, 405b, Mistral large)
My command…
-
### Have you searched for similar requests?
Yes
### Is your feature request related to a problem? If so, please describe.
According to the console logs, Mistral Large 2 support in current ST:
- is…