-
### Describe the issue as clearly as possible:
During the generate process, the tokenizer is being sent to the transformers model.generate() function as a kwarg, which is then being caught during t…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### How would you like to use vllm
I need to extend the context length of gemma2-9b model along also with other mo…
-
# ComfyUI Error Report
## Error Details
- **Node ID:** 224
- **Node Type:** IPAdapterFluxLoader
- **Exception Type:** OSError
- **Exception Message:** We couldn't connect to 'https://huggingface.…
-
### What is the issue?
When using different context sizes (`num_ctx`) with the Ollama embedding model, I noticed big differences in the cosine similarity of the embeddings. Specifically, when I set t…
-
Hi @cfahlgren1 and @davidberenstein1957,
really nice work, i like the idea!
while going through the code, I had a refactoring idea to make it easier to add support for other text generation/chat co…
-
### Model Series
Qwen2.5
### What are the models used?
Qwen2.5-7B
### What is the scenario where the problem happened?
transformers
### Is this a known issue?
- [X] I have followed [the GitHub …
-
I am trying to evaluate llm4decompile-6.7b-v1.5 using the methods you provided. The model weights were downloaded from the Hugging Face repository of the same name. However, I keep encountering an err…
-
### Feature request
I would like to request that BetterTransformer not be deprecated.
### Motivation
I have come to rely on BetterTransformer significantly for accelerating RoBERTa and BERT models.…
-
我将chatglm2下载到如下图的目录
![image](https://github.com/user-attachments/assets/cabf6b33-475a-4783-b74e-7bb26fdc1088)
但报错说没有指定的模型
-
从https://hf-mirror.com/下载的模型 `./hfd.sh IDEA-CCNL/Taiyi-Stable-Diffusion-1B-Chinese-v0.1 --tool wget -x 6`.
使用 MNN/transformers/diffusion/export/onnx_export.py 转换出错:
```
[root@localhost export]$ pyt…