Closed ngfuong closed 2 years ago
Hi @ngfuong ,
Thanks for your interest in LSeg!
As has been reported, we only train a few epochs of LSeg on zero-shot for PASCAL and COCO experiments. I haven't met such a problem, you might need to check your system for details. Also, we have released our checkpoints, please feel free to use them.
Hope this helps.
Best, Boyi
Hi @Boyiliee Thank you for replying! I'm trying to reproduce your results so whether your reported results matche mine is of critical importance. Could you verify if my arguments and hyperparameters are reasonable? I'm training with 01 Tesla A100 80GB GPU.
Thank you for your great paper.
I tried to train a zero-shot model (vitl16_384) and tested on PASCAL fold 0 and got the following problems:
This is my training script:
How does the
max_epochs
argument take part in the training process since there are only 4 epochs logged out? Apart from changing the model fromvitl16_384
tovitb32_384
, is there anything wrong with my training script?While training, DPP is enabled and I only used 01 GPU with
batch_size = 4
. I am not sure if this damages training. Does the argumentaccumulate_grad_batches
probably make this happen?