Closed Bian-666 closed 1 year ago
Hi, Thank you for your interest in our work! With the efficient version, you can train the model with rtx3090 with a batch size of 2, we were using RTX Titan with the same memory size, it did work (torch.backends.cudnn.benchmark=True can further reduce the memeory footprint for some gpus). If you want to have a bigger batch size, maybe try with gradient accumulation .
Another solution would be use another PVT_v2 backbone. We used B2 for TransCenter(-Dual) and B0 for TransCenter-Lite. You have other options from B0 to B5.
Have fun!
Thank you for your prompt reply,thanks for your great work!
I'm interested in your work. but i only have a single rtx3090 machine.