-
In order to reduce the memory usage, I use optimize.quanto to quantize transformer, controlnet, and t5encoder in fp8, but I encounter an error
```
File "/home/yongfang/miniconda3/envs/diffusers/l…
-
### System Info
I am using cuda_12.2, torch 2.1.0a0+29c30b1, bitsandbytes 0.43.3, python 3.10
Driver Version: 535.113.01
NVIDIA GeForce RTX 2080 Ti
### Reproduction
```python
import gc
impo…
-
CAUTION: The operator 'aten::_transformer_encoder_layer_fwd' is not currently supported on the NPU backend and will fall back to run on the CPU. This may have performance implications.
torch 2.1.0…
-
the flow model in cosyvoice, its encoder Conformers contains position embeds while in its decoder transformers, I see no such addition. is that means no benifit here in flow-matching? sorry for it if…
-
Hello, I saw a paragraph in the paper that simply stated that the attention operation was omitted in the encoder module, so the encoder only consists of FFN layers.
Here, we omit the attention mech…
-
### Describe the bug
train_dreambooth_lora_flux.py when running with --train_text_encoder --optimizer="prodigy" causes IndexError: list index out of range because of this:
09/18/2024 20:06:33 - WA…
-
I`m not quite familiar with the Transformer model. There are more steps to do than other model with the Encoder and Decoder. Such as the last encoder block output needs to be as the input for the nex…
-
### Question
您好。在摘要中提到的relation encoder我没有在代码中找到对应的部分,请问能告诉我关于这部分的内容吗?
### 补充信息
_No response_
-
![model](https://github.com/user-attachments/assets/2d027209-e144-48de-a76d-594c7632bed0)
您好,我有两个疑惑。
1:请问论文中图2的model overview里面,X和Xk-1两个有什么区别?
2:不同的transformer层是怎么连接的?比如我看您的代码里面只有encoder,没有decoder,…
-
Support to convert model black-forest-labs/FLUX.1-schnell, receive this error:
after running:
`python -m python_coreml_stable_diffusion.torch2coreml --convert-unet --convert-text-encoder --convert…