-
[ ] I checked the [documentation](https://docs.ragas.io/) and related resources and couldn't find an answer to my question.
** Facing error with using Langchain wrapped hugging face models**
I am …
-
Hello,
I am not able to load `sentence-transformers/distiluse-base-multilingual-cased-v2` model. Script example:
```python
from sentence_transformers import SentenceTransformer
model_name = …
-
**Describe the bug**
When I use llm-compressor to quantize llava model, but at the begining, it failed. (Unrecognized configuration class: 'transformers.models.llava.configuration_llava.LlavaConfig'…
-
In modeling_qwen2_vl.py https://github.com/huggingface/transformers/blob/main/src/transformers/models/qwen2_vl/modeling_qwen2_vl.py#L343
The attention_mask is set for each frame, when not set the f…
-
[Scalable Diffusion Models with Transformers](https://arxiv.org/pdf/2212.09748)
Due to the remarkable achievements of Google AlphaFold 3, it also uses DiT, which combines Diffusion and Transformers…
-
> Please provide us with the following information:
> ---------------------------------------------------------------
### This issue is for Phi-3.5-vision-instruct: (mark with an `x`)
```
- [ …
-
### Feature request
`bias` of linear layers in `qwen2` model is hard coded as following:
- https://github.com/huggingface/transformers/blob/85345bb439652d3f03bb4e123cef7a440f2ba95b/src/transformers/…
-
Are there any plans to add decimated Mamba-2? Also is there a chance we will see some implementations based on transformers package implementations of mamba? https://github.com/huggingface/transformer…
-
Cannot import this from the library,
optimum-1.21.4
transformers-4.43.4
how to fix this issue the code i am running is from the quick start documentation of `Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4…
-
**Describe the bug**
When deploying LLaVA-NeXT-Video-34B-hf, I find that the configuration key passed to transformers is "llava_next_video", while the accurate key in tranformers is "llava-next-video…