-
I'm trying to run the example in `inference/notebooks/01_single_image_understanding.ipynb`, but I get this error:
```
text = processor.apply_chat_template(messages, add_generation_prompt=True)
…
-
### System Info
- `transformers` version: 4.45.0.dev0
- Platform: macOS-14.6.1-arm64-arm-64bit
- Python version: 3.12.4
- Huggingface_hub version: 0.24.6
- Safetensors version: 0.4.5
- Acceler…
-
- [ ] OpenAI
- [ ] Anthropic
- [ ] Groq
- [ ] Cohere
- [ ] Llama some how (ollama & groq are fine)
-
### 📚 The doc issue
I am new to TorchServe and was looking for some features that I need to be able to consider using TorchServe for LLM text generation.
Today, there are a couple inference servin…
-
### System Info
```shell
AWS EC2 instance: trn1.32xlarge
OS: Ubuntu 22.04.4 LTS
Platform:
- Platform: Linux-6.5.0-1023-aws-x86_64-with-glibc2.35
- Python version: 3.10.12
Python packages:
…
cszhz updated
3 weeks ago
-
I have been trying to fix this error for a while now, and the ongoing threads are of NO help.
I have checked these (and ALL issue on the HF community page for this model):
* https://github.com/Qwe…
-
使用Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int4,遇到报错RuntimeError: expected mat1 and mat2 to have the same dtype, but got: float != c10::BFloat16
-
Hi there,
First thank you for unsloth, it's great!
I've finetuned a llama-3-8b-Instruct-bnb-4bit and pushed it to hf hub. When I try to deploy it using [hf Inference Endpoints](https://huggingfa…
-
### Describe the bug
Intel Mac.
```
tts=TTS(model_name='multi-dataset/xtts_v2/en',progress_bar=True).to('mps')
tts.tts_to_file(text='testing', file_path='out.wav', speaker='Craig Gutsy', language=…
-
### System Info
Command Causing Issue:
```
model=microsoft/Phi-3-mini-4k-instruct
volume=$PWD/data # share a volume with the Docker container to avoid downloading weights every run
docker run -…