Closed XiaoHaoPan closed 11 months ago
Until now it has happened this problem, it seems to be out of memory, I have 40G of memory.
Its memory footprint keeps going up and stops at 40.
Hi, @XiaoHaoPan, which module do you train? The stage-one LFAE or stage-two DM?
The stage-one LFAE
I never had this issue before... Could you try a very small batch size (2 or 4) and see what happens?
Thanks for your advice. I'll give it a try.
Hi, How small can the size of the batch_size in the training phase be? At the size you set, I found that my machine didn't have enough video memory, and then I adjusted it to 30 to train.I'm just using a piece of 2080ti (11G)