Vanint / SADE-AgnosticLT

This repository is the official Pytorch implementation of Self-Supervised Aggregation of Diverse Experts for Test-Agnostic Long-Tailed Recognition (NeurIPS 2022).
MIT License
146 stars 20 forks source link

n_gpus vs batch_size #8

Closed rahulvigneswaran closed 2 years ago

rahulvigneswaran commented 2 years ago

Are you offsetting the batch_size for the number n_gpus in the config itself?

CIFAR100:

ImageNet-LT:

iNaturalist18:

Vanint commented 2 years ago

In fact, we did not consider this point. Usually, the larger batch size or larger image resolution (e.g., CIFAR 32X32, others 224x224) requires more GPU memories (number). Moreover, the learning rate is positively related to the batch size. For example, considering lr=0.025 for batch size 64, then lr=0.05 for batch size 128 or lr=0.1 for batch size 256.

rahulvigneswaran commented 2 years ago

Okay, so you are saying that you changed lr, batch_size, according to the n_gpus.

Vanint commented 2 years ago

Nope, n_gpus does not have a fixed equation with lr and batch_size.

rahulvigneswaran commented 2 years ago

Also, how did you reproduce the results of other papers?

  1. Did you reimplement all the methods or just downloaded their GitHub code?
  2. Did you use your own config or used their config?
Vanint commented 2 years ago

In this work, we reproduced existing methods by running existing GitHub codes. If you want to do experiments, you can try their released config and your own config, and select the better one.