Closed HuangJunJie2017 closed 3 years ago
Thanks for your interest!
We use the same data augmentation strategy (following the HRNet code) for all our models. The long training schedule is a necessary setting for training our transformer-based models. Thank your suggestion. We would consider controlling the data augmentation strategy to be the same when comparing with SimpleBaseline.
emmm, under this situation, an ablation study is appealing and necessary for making this fair. This also make me curious that do the transformer-based models perform very poor when using the aug-schedule configuration in the original SimpleBaseline? (drop more than 1.5 AP, may be 2 or 3 AP) As the application of transformer in DERT also requires a longer schedule~
Hi, @HuangJunJie2017 . We did not train our models using the aug-schedule configuration of the original SimpleBaseline.
We provide the convergence curves and log of the mAP for TransPose-R-A4. The curves are as below:
We also use the same data augmentation strategy of TransPose-R,
DATASET:
COLOR_RGB: true
DATASET: 'coco'
DATA_FORMAT: jpg
FLIP: true
NUM_JOINTS_HALF_BODY: 8
PROB_HALF_BODY: 0.3
ROOT: 'data/coco/'
ROT_FACTOR: 45
SCALE_FACTOR: 0.35
TEST_SET: 'val2017'
TRAIN_SET: 'train2017'
and a longer training schedule (203 epochs), to train the original SimpleBaseline-Res50+DarkPose, it now brings 0.4 AP gain (72.0->72.4) on coco val datset with the same detected boxes (56.4AP).
Arch | AP | Ap .5 | AP .75 | AP (M) | AP (L) | AR | AR .5 | AR .75 | AR (M) | AR (L) |
---|---|---|---|---|---|---|---|---|---|---|
pose_resnet | 0.724 | 0.894 | 0.796 | 0.685 | 0.795 | 0.779 | 0.932 | 0.841 | 0.734 | 0.844 |
emmm, i have not try SimpleBaseline-Res50+darkpose+hrnetaugschedule. But i have tried another similar configuration SimpleBaseline-Res50+UDP+hrnetaugschedule. This configuration scores 73.4 AP on coco val datset with the same detected boxes (56.4AP). 72.0 for SimpleBaseline-Res50+darkpose+hrnetaugschedule is too low. There may be some problem in your experiment which may has negative impact on both SimpleBaseline and TransPose? TransPose can perform better?
We run the official DarkPose code on our machine. Here we provide the history log files.
orignal_config_baseline_res50_256x192_d256x3_adam_lr1e-3_2020-11-10-13-01_train.log
orignal_config_baseline_res50_256x192_d256x3_adam_lr1e-3_2020-11-11-20-52_valid.log
emm, it it seem that you don't train SimpleBaseline-Res50+darkpose+hrnetaugschedule this with 210 epoch, but 140 epoch instead. This make this configuration perform poor (72.0) ~, if you train this with 210 epoch, SimpleBaseline-Res50+darkpose+hrnetaugschedule will score more than 73.0 AP coco val datset with the same detected boxes (56.4AP).
A longer scheduler or stronger augmentation indeed increases the AP. But, the original SimpleBaseline-Res50 and SimpleBaseline-Res50+DarkPose are trained with 140 epochs. Due to that we use the data augmentation of HRNet for TransPose-R, so we trained SimpleBaseline-Res50+DarkPose with stronger data augmentation (HRNet) for a fair comparison. In addition, the current result I reported (204 epochs) is 72.4 AP now. Maybe a better result will be achieved by adding training steps, but we don't think it is a necessity that we should use the same training schedule to train such two different models (FCN vs CNN+Transformer) for comparison.
Point taken, the comparison between SimpleBaseline, SimpleBaseline-darkpose and TransPose-R is unfair, but the comparison between TransPose-R and TransPose-H is fair.
Nice pioneer~ But it is not fair to compare TransPose-R with the original SimpleBaseline-Res, as the original SimpleBaseline use a weaker augmentation strategy and a shorter training schedule. This will lead to a large gap of 1.5 AP. I look forward to a fair comparison between them for reference.