masashi-hatano / MM-CDFSL

[ECCV 2024] Official code release for "Multimodal Cross-Domain Few-Shot Learning for Egocentric Action Recognition"
https://masashi-hatano.github.io/MM-CDFSL/
11 stars 1 forks source link

How many GPU are needed for the training of the whole model? #2

Open Saunak626 opened 2 weeks ago

Saunak626 commented 2 weeks ago

config_pretrain.yaml accelerator: gpu devices: [0, 1]

Hello! Is this only using two graphics cards? How much video memory?

masashi-hatano commented 2 weeks ago

Yes, we use two GPUs with 24GB during training due to the resource limit. But, it won't take much time to train as we initialize with the pre-trained weight trained on the Kinetics dataset.

xgiaogiao commented 1 week ago

Hello! I noticed you mentioned that "we use two GPUs with 24GB during training due to the resource limit. But, it won't take much time to train as we initialize with the pre-trained weight trained on the Kinetics dataset." I am reaching out to inquire about the approximate time required per epoch when training on the EPIC dataset under similar settings. With my setup of two 24GB GPUs (NVIDIA 3090), each epoch is taking approximately 3 hours to complete. Could you kindly share how long it typically took for you per epoch?

masashi-hatano commented 5 days ago

The training time is approximately half a day for the pretraining stage on the Ego4D-EPIC scenario and 1.5 days for the multimodal distillation stage. We used two RTX 4090 GPUs for these trainings.