-
**Describe the bug**
When I use llm-compressor to quantize llava model, but at the begining, it failed. (Unrecognized configuration class: 'transformers.models.llava.configuration_llava.LlavaConfig'…
-
### System Info / 系統信息
cuda:12.4
centeos7.9
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [X] docker / docker
- [ ] pip install / 通过 pip install 安装
- [ ] installation from sou…
-
### System Info
```shell
python 3.10.14
torch 2.4.0+cu121
optimum 1.21.4
onnx 1.16.2
onnxruntime 1.19.0
transformers 4.43.4
optim…
-
### Feature request
Recently, we have added the ability to load `gguf` files within [transformers](https://huggingface.co/docs/hub/en/gguf).
The goal was to offer the possibility to users …
-
Traceback ( most recent call last):
File "C:\ Users\15729\ Downloads\Qwen2- Boundless- main\Qwen2- Boundless- main\ basic _ usage. py", line 10, in < module>
device _ map=" auto"
File "C:\ Users…
-
In modeling_qwen2_vl.py https://github.com/huggingface/transformers/blob/main/src/transformers/models/qwen2_vl/modeling_qwen2_vl.py#L343
The attention_mask is set for each frame, when not set the f…
-
!pip install transformers datasets
from transformers import GPT2Tokenizer, GPT2LMHeadModel, Trainer, TrainingArguments
from datasets import load_dataset, load_metric
tokenizer = GPT2Tokenizer.from_…
-
Hi,
I'm trying to constrain the generation of my VLMs using this repo; however i can't figure out the way to personalize the pipeline for handling inputs (query+image). Whereas it is documented as …
-
SentenceTransformerEmbeddingFunction is erroring out during initializing the model.dense.SentenceTransformerEmbeddingFunction.
Code :
from pymilvus import model
sentence_transformer_ef = model.d…
-
Hi :) I have an issue with how to calculate the relevance scores w.r.t two (or more) output values for lxt.model.llama.
Initially, following your demo in llama, I write the following code.
```py…