Closed GWwangshuo closed 3 years ago
This happens because the model was based on 8 cards of 32G Titan V100 with a batch size 2. (to my recall, GTX 2080Ti has an 11G memory) You can either cut down the batch size to 1 or perform a smaller augmentation for the input size of the image (permalink below). https://github.com/kakaobrain/HOTR/blob/a2e19511d6131220f430ba2d76a95a03a0f7556e/hotr/data/datasets/vcoco.py#L440 Although this might make the model run, note that both will harm the original performance of the model.
Thanks for your nice work. When evaluating HOTR on vcoco dataset (vcoco_multi_train) on a server with 8 GeForce RTX 2080Ti Cards, I encountered the
CUDA out of memory
problem.This problem seems happens on line 53 of evaluator_vcoco.py,
utils.all_gather(res)
. Any suggestions how to solve this problem? Thanks a lot.