-
### Feature request
Optimize Transformers' image_processors to decrease image processing time, and reduce inference latency for vision models and vlms.
### Motivation
The Transformers library relie…
-
### System Info
- `transformers` version: 4.45.2
- Platform: Linux-5.4.0-187-generic-x86_64-with-glibc2.31
- Python version: 3.11.5
- Huggingface_hub version: 0.23.4
- Safetensors version: 0.4.…
-
Hey guys, I am trying to use MiniCPMV6 with Outlines using https://huggingface.co/openbmb/MiniCPM-V-2_6
I am using the outlines.models.transformers_vision api to load the model, but I can't find t…
-
- Link: https://arxiv.org/abs/2104.11227
-
### Feature request
I am trying to train off-line RL using decision transformer, convert to .onnx.
```
from pathlib import Path
from transformers.onnx import FeaturesManager
feature = "seq…
-
MMDetection includes both SWIN and DETR, if I understand the concept correctly, both could be fine-tuned with LORA in a fast and memory efficient manner.
Support for training with LORA in object d…
-
## 論文リンク
- [arXiv](https://arxiv.org/abs/2105.07926)
- [CVF](https://openaccess.thecvf.com/content/CVPR2022/html/Mao_Towards_Robust_Vision_Transformer_CVPR_2022_paper.html)
- [GitHub](https://githu…
-
Hi,
I have finetuned Qwen2-VL using Llama-Factory.
I successfully quantized the fine-tuned model as given
```
from transformers import Qwen2VLProcessor
from auto_gptq import BaseQuantizeC…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
- `llamafactory` version: 0.9.1.dev0
- Platform: Linux-5.15.120.bsk.2-amd64-x86_64-with-glibc2.31
…
-
Hi I wanted to know if there is a version of FullGrad which could be applied on Vision Transformers like ViT or the Swin Transformer, or if there are some small changes that could be done in the code …