-
Hi, thanks for the nice work and great repo!
I changed config to train_with_clip=1 to include ClipLoss.
Then, I am getting the following error in the eval step:
![image](https://user-images.githu…
-
[Scalable Diffusion Models with Transformers](https://arxiv.org/pdf/2212.09748)
Due to the remarkable achievements of Google AlphaFold 3, it also uses DiT, which combines Diffusion and Transformers…
-
![image](https://github.com/user-attachments/assets/00501d0e-a886-4a15-b2fd-29b09eee99aa)
-
It'd be great to have XLabs ControlNets supported in `diffusers`. We already support their LoRAs.
Code: https://github.com/XLabs-AI/x-flux/
Checkpoint: https://huggingface.co/XLabs-AI/flux-contro…
-
### System Info
```Shell
- `Accelerate` version: 0.33.0
- Platform: Windows-10-10.0.22631-SP0
- `accelerate` bash location: C:\Users\Nech\anaconda3\envs\transformer-multi-device\Scripts\accelera…
-
Models envisioned:
- TCN / LSTM / GRU as sequential models;
- Decided to abandon HLSTM, indie LSTM, transformer for various reasons;
- Attention schemes:
- Average pooling
- Plain self-attent…
-
normalization seems different from the paper #attention is all you need#
in paper, normalization layer stays after mha and feed forward layer, in torchnlp, it stays before them
x…
-
I have build my own demo file. after uploading one video, it gives blank output. Could anyone help me out?
-------------------------------Here's the demo file-------------------------
from argpars…
-
Dear Lee,
Awesome job and congratulation!
It seems that their is only multi-head self attention edition SpatialNet here. Will you release the online mamba edition in the future?
Best!
-
**/tmp/tmppngxpwds.obj
Traceback (most recent call last):**
File "/home/jkx/anaconda3/envs/InstantMesh/lib/python3.10/site-packages/gradio/queueing.py", line 536, in process_events
response =…