Closed weiaicunzai closed 8 months ago
halo, thanks for your interest!
Four GPUs with smaller memories also can be used for pre-training, you can just decrease the batch_size_per_gpu
to fit your GPUs.
halo, thanks for your interest! Four GPUs with smaller memories also can be used for pre-training, you can just decrease the
batch_size_per_gpu
to fit your GPUs.
Thank you, I have watched the presentation video of your work on VALSE. And I have noticed that you have used eight 3090 GPUs in your experiment. Could please share with us how many days you have used for pre-training?
It should take about 1 day for pre-training.
It should take about 1 day for pre-training.
Thanks.
Hi, thanks for your great work. I have noticed that you have set
--nproc_per_node=4
in training scripttrain_clips32k.sh
, so could you please tell us if four GPUs with smaller GPU memories (2080ti, Titan V) can be used for training? (without the OOM error)