-
### Describe your use-case.
There are multiple simple models used in this repository: Blip, Clip and WD-taggers. However, when it comes to detailed description, they are all dwarfed by modern multi…
-
Can we have node supported for https://github.com/Alpha-VLLM/Lumina-T2X?
-
Hi,
Thank you for your nice work.
You use video and image data in the training process. How much does adding image data help the final generated result? I found that adding image data did not im…
-
I know this feels a bit _anal retentive_, please bear with me…
It seems that if the last changed branch is not a tip, the whole line graph for that branch is thin and dashed.
![screen shot 2017-…
-
my script is below
```
torchrun --nproc_per_node=2 examples/pixartalpha_example.py \
--model PixArt-alpha/PixArt-XL-2-1024-MS \
--height 2048 \
--width 2048 \
--pipefusion_parallel_degree 1 \
-…
-
![lumina](https://github.com/kijai/ComfyUI-LuminaWrapper/assets/173285092/79f2dfad-363a-43b9-832c-2c52c2aaaa8c)
got prompt
[rgthree] Using rgthree's optimized recursive execution.
[rgthree] First…
-
I have a question, I see the Pixart Sigma train by default is 600M, I want to try finetune with 900M PixArt Sigma like [PixArt Sigma 900M: Enhanced Text-to-Image Model
](https://civitai.com/models/57…
-
I use kvcompress to train, but when I infer, I need to convert PixArt to diffusers, there are some errors when I run convert_pixart_to_diffusers.py, How to solve it?
AssertionError: State dict is…
-
ampler:
Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!
File "E:\AI\execution.py", line 151, in recursive_execute
output_data, output_ui = get_out…
-
# Lumina-T2X: Transforming Text into Any Modality, Resolution, and Duration via Flow-based Large Diffusion Transformers
> Sora unveils the potential of scaling Diffusion Transformer (DiT) for gener…