PixArt-alpha / PixArt-sigma

PixArt-Σ: Weak-to-Strong Training of Diffusion Transformer for 4K Text-to-Image Generation
https://pixart-alpha.github.io/PixArt-sigma-project/
GNU Affero General Public License v3.0
1.63k stars 77 forks source link

Inquiry About Hardware Requirements and Training Details for PIXART-Σ #24

Closed sohowj closed 5 months ago

sohowj commented 5 months ago

I am planning to train the model using the training code provided in https://github.com/PixArt-alpha/PixArt-sigma and am currently gathering information to understand the necessary hardware and data requirements. I would appreciate if you could provide some insights based on your experiences with PIXART-Σ. Specifically, I have the following questions:

  1. What hardware configurations were used for training the model, e.g., how many Nvidia A100/H100 GPUs? (or other types of GPUs)
  2. Could you specify the size of the dataset used during training?
  3. How long did the training process take?

Any details you provide will be incredibly helpful as I am trying to estimate the resources and time needed for my own training setup.

Thank you in advance for your assistance!

lawrence-cj commented 5 months ago

1.All of our experiments are done using maximum 32 V100 GPUs. 2-3. As we talked about in both PixArt-α: 15M dataset and PixArt-Σ:30M dataset. For details, please refer to the Paper

image image
kingkong135 commented 4 months ago

what is mean "32x32" in Batch size, it is use battch size 32 in 32 computer ?