PatrickHua / SimSiam

A pytorch implementation for paper 'Exploring Simple Siamese Representation Learning'
MIT License
814 stars 135 forks source link

Consumes a lot GPU memory than standard? #51

Open mmw0909 opened 2 years ago

mmw0909 commented 2 years ago

Hi, I used a ResNet34 backbone to train on (1, 128, 128) images with a batch size of 128. The total allocated memory is >35GB. According to the post, a ResNet50 on (3,256,256) images with a batch size of 96 only consumes 10GB. I am wondering if anyone else experiences the same issue and if there is any clue as to why this network takes such a lot of memory.