OpenMMLab Video Perception Toolbox. It supports Video Object Detection (VID), Multiple Object Tracking (MOT), Single Object Tracking (SOT), Video Instance Segmentation (VIS) with a unified framework.
Thanks for your error report and we appreciate it a lot.
Checklist
I have searched related issues but cannot get the expected help.
The bug has not been fixed in the latest version.
Describe the bug
A clear and concise description of what the bug is.
I have trained selsa and fgfa in mmtracking, but got worse results compared with results in model zoo.
Reproduction
You may add addition that may be helpful for locating the problem, such as
How you installed PyTorch [e.g., pip, conda, source]
conda
Other environment variables that may be related (such as $PATH, $LD_LIBRARY_PATH, $PYTHONPATH, etc.)
Error traceback
If applicable, paste the error trackback here.
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.467
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.745
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.512
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.063
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.205 Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.528 Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.470 Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.599 Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.604 Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.108
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.401
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.659
{'bbox_mAP': 0.467, 'bbox_mAP_50': 0.745, 'bbox_mAP_75': 0.512, 'bbox_mAP_s': 0.063, 'bbox_mAP_m': 0.205, 'bbox_mAP_l': 0.528, 'bbox_mAP_copypaste': '0.467 0.745 0.512 0.063 0.205 0.528'}
Bug fix
If you have already identified the reason, you can provide the information here. If you are willing to create a PR to fix it, please also leave a comment here and that would be much appreciated!
Did you test fgfa and selsa method with models training by yourself?
If so, the fgfa model with r50 as backbone achieves 74.5 mAP@0.5, while the corresponding model in model zoo achieve 74.7 mAP@0.5.
As described in here, the difference of 0.2 points is normal.
Thanks for your error report and we appreciate it a lot.
Checklist
Describe the bug A clear and concise description of what the bug is. I have trained selsa and fgfa in mmtracking, but got worse results compared with results in model zoo. Reproduction
Did you make any modifications on the code or config? Did you understand what you have modified? nothing change
What dataset did you use and what task did you run? ILSVRC, VID & DET__train_30classes Environment
Please run
python mmtrack/utils/collect_env.py
to collect necessary environment information and paste it here.TorchVision: 0.9.0 OpenCV: 4.5.4-dev MMCV: 1.3.17 MMCV Compiler: GCC 7.3 MMCV CUDA Compiler: 11.1 MMTracking: 0.8.0+fab6942
$PATH
,$LD_LIBRARY_PATH
,$PYTHONPATH
, etc.)Error traceback If applicable, paste the error trackback here.
Bug fix If you have already identified the reason, you can provide the information here. If you are willing to create a PR to fix it, please also leave a comment here and that would be much appreciated!