Hi I'm trying to reproduce the the performance on HICO dataset. Using ResNet only you reported the performance of 30.2 mAP. But I only got 26.3. Could you please share the parameters you use?
These parameters are:
How the image is pre-processed for training/validating
Are there any other information used (pose?text?)
Is the loss function softmax cross entropy or sigmod binary cross entropy for multi-class?
what's the optimizer? (adam or SGD?)
what's the training scheduler? (learning rate change, freezing any layers?)
Same as what I do for MPII, random crops, flips etc. Input image size is 480px
No
Softmax cross-entropy. I randomly pick a class the image belongs to each time it is seen at train time.
Trained over 4 GPUs using sync-SGD, I use LR=0.001, batch size=16 per GPU, LR decay by 0.1 every 5000 iterations, and I train a total of 12000 iterations. The model is initialized from ImageNet pre-trained weights.
Hi I'm trying to reproduce the the performance on HICO dataset. Using ResNet only you reported the performance of 30.2 mAP. But I only got 26.3. Could you please share the parameters you use? These parameters are:
Thanks a lot!