Closed truetone2022 closed 4 years ago
Thanks a lot for the question. Without knowing other information, like batch size, GPU type I don't think I can answer this question.
Just to let you know, for V-COCO with a batch size of 8 in a RTX 2080TI each epoch takes close to 10 mints and as mentioned in the repo for hico it takes close to 40 mints in 4 RTX 2080TIs with a batch size of 64. One thing to point out, we work with a flexible number of humans and objects, so the usage of GPU changes heavily during different iterations.
Thanks a lot for your kind reply! It's very helpful! I also set a batch size of 64 in 4 RTX 2080TIs for HICO-DET, the GPU memory is almost run out, but the gpu util rate is less than 30% all the time.Were you facing the same situation when training the VSGNet? Very thanks for any helpful reply~Merci beaucoup~
I cant recall the gpu use actually. But if it takes close to 40 mints then I would say there is nothing to get worried.
Very thanks for your kind reply!
Should the learning rate be set to 0.001*8 when the batch size is 64 in the procedure of training VSGNet on HICO-DET ? Very thanks for any helpful reply!
Without running the model I can't say anything in particular but as per as I remember for HICO our model was converging fine with different learning rates.