-
## Description:
Hello! I’ve been following the development of this repository and appreciate the efforts to benchmark various efficient Transformer variants. I’d like to propose the implementation of…
-
File: ...\ComfyUI\custom_nodes\ComfyUI-AniPortrait\src\models\unet_2d_blocks.py
Line 9: `from diffusers.models.dual_transformer_2d import DualTransformer2DModel`
Throw Exception: No module named '…
Aljnk updated
1 month ago
-
**Is your feature request related to a problem? Please describe.**
It would be nice to intergrate https://llama-cpp-python.readthedocs.io/en/stable/#embeddings because of the speed of default `senten…
-
Hi TensorRT-LLM team, Your work is incredible.
By following the READme file for [multi-modeling](https://github.com/NVIDIA/TensorRT-LLM/blob/main/examples/multimodal/README.md), we were sucess to run…
-
Perhaps we can use clamping, as per:
> https://github.com/huggingface/transformers/blob/main/src/transformers/models/t5/modeling_t5.py#L748-L755
Using BF16 works on CUDA.
-
### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits of both this extension and the webui
### What happened?
I encountered an err…
-
### System Info
not relevant here
### Who can help?
@stevhliu
### Information
- [ ] The official example scripts
- [ ] My own modified scripts
### Tasks
- [ ] An officially suppo…
-
```
lm-format-enforcer==0.10.7
torch==2.4.1+cu121
transformers==4.45.0
```
When using the library together with the newly released Llama3.2-11B-Instruct we get a CUDA error.
```model_id = …
-
What should I specify as the `model_type` in the JSON file?
from transformers import AutoModel
model = AutoModel.from_pretrained("zxhezexin/openlrm-obj-base-1.1")
ValueError: Unrecogniz…
-
The loading of Hugging Face models adheres to the following logic: models using device_map must forcibly enable low_cpu_mem_usage. (Or a user might have manually enabled low_cpu_mem_usage in from_pret…