-
## Bug report
### Describe the bug
Why I toggle mpd playback twice in a short time, a second of the song is skipped. This can be most easily reproduced when running the following commands when mpd…
-
## Bug report
### Describe the bug
I'm using a HifiBerry DAC2 HD on a Raspberry Pi 3B+ running Raspberry Pi OS.
```
uname -a
Linux music 5.15.32-v7+ #1538 SMP Thu Mar 31 19:38:48 BST 2022 a…
-
# 1. Ollama
## 1. use Ollama CLI:
```
ollama serve
ollama run llama2:7b, llama3, llama3:70b, mistral, dophin-phi, phi, neural-chat, codellama, llama2:13b, llama2:70b
ollama list
ollama show
…
-
### System Info
I tried the following systems, both with the same exception:
- ghcr.io/huggingface/text-generation-inference:sha-6aebf44 locally with docker on nvidia rtx 3600
- ghcr.io/huggingface…
-
https://lmsys.org/blog/2023-06-29-longchat/
https://arxiv.org/abs/2305.07185
https://www.reddit.com/r/LocalLLaMA/comments/14fgjqj/a_simple_way_to_extending_context_to_8k/
https://github.com/epfml…
-
### System Info
latest tgi docker image (I think 2.0)
ubuntu 20
### Information
- [X] Docker
- [ ] The CLI directly
### Tasks
- [ ] An officially supported command
- [ ] My own modifications
##…
onel updated
3 months ago
-
## Describe the bug
Prompt Tuning model generates low-quality output
## Platform
Please provide details about the environment you are using, including the following:
- Interpreter version:…
-
### System Info
Compared with **v1.2.2-release** tgi-gaudi, sending the query to the **v2.0.0-release** tgi-server will hit the input_token_length + output_token_length assertion.
Especially, when…
-
https://github.com/huggingface/text-generation-inference/blob/d0225b10156320f294647ac676c130d03626473d/server/text_generation_server/layers/fp8.py#L4
@Narsil what do you think about enabling torch.…
-
### Model description
If I uploaded the model to TGI, should I prepare the text in the required template for the generate query? Or does TGI itself perform this operation at the time of receiving the…