-
### System Info
- `transformers` version: 4.38.1
- Platform: Linux-5.15.146.1-microsoft-standard-WSL2-x86_64-with-glibc2.31
- Python version: 3.10.13
- Huggingface_hub version: 0.20.3
- Safeten…
-
Dear all,
Thank you so much for sharing the llama3.2 vision model fine-tuning script so fast!
I got the following error when running the demo
```
The model weights are not tied. Please use t…
-
I ran the first command provided (to do some sanity checking of my setup since I usually get very high output errors for larger models like LLMs) and I get an output validation error.
I've made sur…
-
### Your current environment
PyTorch version: 2.4.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.3 LTS (x86_64)
GCC version: (U…
-
## Bug Description
https://github.com/pytorch/TensorRT/blob/main/examples/dynamo/mutable_torchtrt_module_example.py
I replaced hugging face whisper model instead of diffusion model
## To Repr…
-
### 🚀 The feature, motivation and pitch
Enable support for Flash Attention Memory Efficient and SDPA kernels for AMD GPUs.
At present using these gives below warning with latest nightlies (torch==…
-
**Environment:**
1. Framework: (TensorFlow, Keras, PyTorch, MXNet) Pytorch
2. Framework version: latest code from huggingface: https://github.com/huggingface/pytorch-transformers
3. Horovo…
-
## 问题原因
在加载模型时,from_pretrained 方法会调用 transformers/modeling_utils.py 中的 _autoset_attn_implementation 方法自动开启 sdpa (新版 pytorch 中 Scaled Dot-Product Attention 的高效实现):
![_autoset_attn_implementation](htt…
-
https://github.com/NVIDIA/apex/blob/a0f5f3ac0f6bf39feee6e60eee66ec873dc299ab/apex/transformer/pipeline_parallel/p2p_communication.py#L271 might be able to be removed after confirming https://github.co…
-
issue about the fact generate is not possible with fp16 (deepspeed)
introduced when fp16 feature was introduced https://github.com/lucidrains/DALLE-pytorch/pull/157 :
```
/pytorch/aten/src/THC/THC…