open-mmlab / mmpretrain

OpenMMLab Pre-training Toolbox and Benchmark
https://mmpretrain.readthedocs.io/en/latest/
Apache License 2.0
3.38k stars 1.05k forks source link

TopK accuracy greater than 1.0 on ImageNet #742

Closed BebDong closed 2 years ago

BebDong commented 2 years ago

Checklist

Describe the question you meet

I am training the T2T-ViT model for ImageNet classification following exactly the provided config file. However, after several epochs, the TopK accuracy during validation becomes greater than 1.0.

Post related information

  1. Environments image image

  2. Your config file if you modified it or created a new one. No modifications to the provided config file.

  3. Your train log file if you meet the problem during training.

    2022-03-23 10:30:07,011 - mmcls - INFO - workflow: [('train', 1)], max: 310 epochs
    2022-03-23 10:30:07,030 - mmcls - INFO - Checkpoints will be saved to /home/user/modelarts/outputs/train-url_0 by HardDiskBackend.
    2022-03-23 10:30:53,826 - mmcls - INFO - Epoch [1][100/2503]    lr: 1.978e-06, eta: 4 days, 4:34:05, time: 0.467, data_time: 0.038, memory: 12880, loss: 6.9772
    2022-03-23 10:31:36,302 - mmcls - INFO - Epoch [1][200/2503]    lr: 3.976e-06, eta: 4 days, 0:02:11, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.9475
    2022-03-23 10:32:18,814 - mmcls - INFO - Epoch [1][300/2503]    lr: 5.973e-06, eta: 3 days, 22:32:50, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.9307
    2022-03-23 10:33:01,284 - mmcls - INFO - Epoch [1][400/2503]    lr: 7.971e-06, eta: 3 days, 21:46:32, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.9210
    2022-03-23 10:33:43,711 - mmcls - INFO - Epoch [1][500/2503]    lr: 9.969e-06, eta: 3 days, 21:17:11, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.9126
    2022-03-23 10:34:26,177 - mmcls - INFO - Epoch [1][600/2503]    lr: 1.197e-05, eta: 3 days, 20:58:20, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.9060
    2022-03-23 10:35:08,629 - mmcls - INFO - Epoch [1][700/2503]    lr: 1.396e-05, eta: 3 days, 20:44:22, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8971
    2022-03-23 10:35:51,097 - mmcls - INFO - Epoch [1][800/2503]    lr: 1.596e-05, eta: 3 days, 20:33:58, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8902
    2022-03-23 10:36:33,594 - mmcls - INFO - Epoch [1][900/2503]    lr: 1.796e-05, eta: 3 days, 20:26:03, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8821
    2022-03-23 10:37:16,064 - mmcls - INFO - Epoch [1][1000/2503]   lr: 1.996e-05, eta: 3 days, 20:19:24, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8725
    2022-03-23 10:37:58,525 - mmcls - INFO - Epoch [1][1100/2503]   lr: 2.195e-05, eta: 3 days, 20:13:36, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8675
    2022-03-23 10:38:40,963 - mmcls - INFO - Epoch [1][1200/2503]   lr: 2.395e-05, eta: 3 days, 20:08:28, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.8637
    2022-03-23 10:39:23,436 - mmcls - INFO - Epoch [1][1300/2503]   lr: 2.595e-05, eta: 3 days, 20:04:19, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8560
    2022-03-23 10:40:05,937 - mmcls - INFO - Epoch [1][1400/2503]   lr: 2.795e-05, eta: 3 days, 20:00:56, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8535
    2022-03-23 10:40:48,491 - mmcls - INFO - Epoch [1][1500/2503]   lr: 2.994e-05, eta: 3 days, 19:58:23, time: 0.426, data_time: 0.001, memory: 12880, loss: 6.8454
    2022-03-23 10:41:30,941 - mmcls - INFO - Epoch [1][1600/2503]   lr: 3.194e-05, eta: 3 days, 19:55:12, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.8330
    2022-03-23 10:42:13,420 - mmcls - INFO - Epoch [1][1700/2503]   lr: 3.394e-05, eta: 3 days, 19:52:34, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8261
    2022-03-23 10:42:55,879 - mmcls - INFO - Epoch [1][1800/2503]   lr: 3.594e-05, eta: 3 days, 19:49:58, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8161
    2022-03-23 10:43:38,318 - mmcls - INFO - Epoch [1][1900/2503]   lr: 3.793e-05, eta: 3 days, 19:47:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.8092
    2022-03-23 10:44:20,785 - mmcls - INFO - Epoch [1][2000/2503]   lr: 3.993e-05, eta: 3 days, 19:45:15, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.8026
    2022-03-23 10:45:03,231 - mmcls - INFO - Epoch [1][2100/2503]   lr: 4.193e-05, eta: 3 days, 19:43:07, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.7921
    2022-03-23 10:45:45,716 - mmcls - INFO - Epoch [1][2200/2503]   lr: 4.393e-05, eta: 3 days, 19:41:18, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.7822
    2022-03-23 10:46:28,194 - mmcls - INFO - Epoch [1][2300/2503]   lr: 4.593e-05, eta: 3 days, 19:39:35, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.7741
    2022-03-23 10:47:10,646 - mmcls - INFO - Epoch [1][2400/2503]   lr: 4.792e-05, eta: 3 days, 19:37:47, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.7744
    2022-03-23 10:47:53,113 - mmcls - INFO - Epoch [1][2500/2503]   lr: 4.992e-05, eta: 3 days, 19:36:09, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.7601
    2022-03-23 10:47:54,117 - mmcls - INFO - Saving checkpoint at 1 epochs
    2022-03-23 10:48:15,246 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_1.pth.
    2022-03-23 10:48:15,246 - mmcls - INFO - Best accuracy_top-1 is 0.1000 at 1 epoch.
    2022-03-23 10:48:15,271 - mmcls - INFO - Epoch(val) [1][98] accuracy_top-1: 0.1000, accuracy_top-5: 0.5520
    2022-03-23 10:49:00,549 - mmcls - INFO - Epoch [2][100/2503]    lr: 5.198e-05, eta: 3 days, 19:42:06, time: 0.453, data_time: 0.028, memory: 12880, loss: 6.7430
    2022-03-23 10:49:43,036 - mmcls - INFO - Epoch [2][200/2503]    lr: 5.397e-05, eta: 3 days, 19:40:25, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.7240
    2022-03-23 10:50:25,495 - mmcls - INFO - Epoch [2][300/2503]    lr: 5.597e-05, eta: 3 days, 19:38:40, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.7251
    2022-03-23 10:51:07,989 - mmcls - INFO - Epoch [2][400/2503]    lr: 5.797e-05, eta: 3 days, 19:37:10, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.7043
    2022-03-23 10:51:50,446 - mmcls - INFO - Epoch [2][500/2503]    lr: 5.997e-05, eta: 3 days, 19:35:33, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.6939
    2022-03-23 10:52:32,926 - mmcls - INFO - Epoch [2][600/2503]    lr: 6.196e-05, eta: 3 days, 19:34:05, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.6708
    2022-03-23 10:53:15,368 - mmcls - INFO - Epoch [2][700/2503]    lr: 6.396e-05, eta: 3 days, 19:32:31, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.6535
    2022-03-23 10:53:57,826 - mmcls - INFO - Epoch [2][800/2503]    lr: 6.596e-05, eta: 3 days, 19:31:03, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.6509
    2022-03-23 10:54:40,269 - mmcls - INFO - Epoch [2][900/2503]    lr: 6.796e-05, eta: 3 days, 19:29:35, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.6416
    2022-03-23 10:55:22,729 - mmcls - INFO - Epoch [2][1000/2503]   lr: 6.995e-05, eta: 3 days, 19:28:13, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.6273
    2022-03-23 10:56:05,185 - mmcls - INFO - Epoch [2][1100/2503]   lr: 7.195e-05, eta: 3 days, 19:26:53, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.6184
    2022-03-23 10:56:47,625 - mmcls - INFO - Epoch [2][1200/2503]   lr: 7.395e-05, eta: 3 days, 19:25:32, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.6069
    2022-03-23 10:57:30,061 - mmcls - INFO - Epoch [2][1300/2503]   lr: 7.595e-05, eta: 3 days, 19:24:11, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.5661
    2022-03-23 10:58:12,534 - mmcls - INFO - Epoch [2][1400/2503]   lr: 7.794e-05, eta: 3 days, 19:22:59, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.5708
    2022-03-23 10:58:54,944 - mmcls - INFO - Epoch [2][1500/2503]   lr: 7.994e-05, eta: 3 days, 19:21:38, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.5633
    2022-03-23 10:59:37,400 - mmcls - INFO - Epoch [2][1600/2503]   lr: 8.194e-05, eta: 3 days, 19:20:27, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.5599
    2022-03-23 11:00:19,853 - mmcls - INFO - Epoch [2][1700/2503]   lr: 8.394e-05, eta: 3 days, 19:19:16, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.5668
    2022-03-23 11:01:02,349 - mmcls - INFO - Epoch [2][1800/2503]   lr: 8.593e-05, eta: 3 days, 19:18:14, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.5445
    2022-03-23 11:01:44,804 - mmcls - INFO - Epoch [2][1900/2503]   lr: 8.793e-05, eta: 3 days, 19:17:07, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.5000
    2022-03-23 11:02:27,252 - mmcls - INFO - Epoch [2][2000/2503]   lr: 8.993e-05, eta: 3 days, 19:15:59, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.5198
    2022-03-23 11:03:09,725 - mmcls - INFO - Epoch [2][2100/2503]   lr: 9.193e-05, eta: 3 days, 19:14:57, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.5199
    2022-03-23 11:03:52,170 - mmcls - INFO - Epoch [2][2200/2503]   lr: 9.393e-05, eta: 3 days, 19:13:50, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.5073
    2022-03-23 11:04:34,653 - mmcls - INFO - Epoch [2][2300/2503]   lr: 9.592e-05, eta: 3 days, 19:12:51, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4719
    2022-03-23 11:05:17,103 - mmcls - INFO - Epoch [2][2400/2503]   lr: 9.792e-05, eta: 3 days, 19:11:48, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4549
    2022-03-23 11:05:59,587 - mmcls - INFO - Epoch [2][2500/2503]   lr: 9.992e-05, eta: 3 days, 19:10:50, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4562
    2022-03-23 11:06:00,589 - mmcls - INFO - Saving checkpoint at 2 epochs
    2022-03-23 11:06:20,918 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_1.pth was removed
    2022-03-23 11:06:21,648 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_2.pth.
    2022-03-23 11:06:21,648 - mmcls - INFO - Best accuracy_top-1 is 0.1460 at 2 epoch.
    2022-03-23 11:06:21,718 - mmcls - INFO - Epoch(val) [2][98] accuracy_top-1: 0.1460, accuracy_top-5: 0.5960
    2022-03-23 11:07:06,979 - mmcls - INFO - Epoch [3][100/2503]    lr: 1.020e-04, eta: 3 days, 19:13:37, time: 0.452, data_time: 0.028, memory: 12880, loss: 6.4256
    2022-03-23 11:07:49,407 - mmcls - INFO - Epoch [3][200/2503]    lr: 1.040e-04, eta: 3 days, 19:12:27, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.4549
    2022-03-23 11:08:31,818 - mmcls - INFO - Epoch [3][300/2503]    lr: 1.060e-04, eta: 3 days, 19:11:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.4416
    2022-03-23 11:09:14,290 - mmcls - INFO - Epoch [3][400/2503]    lr: 1.080e-04, eta: 3 days, 19:10:16, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4367
    2022-03-23 11:09:56,753 - mmcls - INFO - Epoch [3][500/2503]    lr: 1.100e-04, eta: 3 days, 19:09:14, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4105
    2022-03-23 11:10:39,228 - mmcls - INFO - Epoch [3][600/2503]    lr: 1.120e-04, eta: 3 days, 19:08:15, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4391
    2022-03-23 11:11:21,707 - mmcls - INFO - Epoch [3][700/2503]    lr: 1.140e-04, eta: 3 days, 19:07:17, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4273
    2022-03-23 11:12:04,184 - mmcls - INFO - Epoch [3][800/2503]    lr: 1.159e-04, eta: 3 days, 19:06:20, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.4026
    2022-03-23 11:12:46,625 - mmcls - INFO - Epoch [3][900/2503]    lr: 1.179e-04, eta: 3 days, 19:05:18, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.3889
    2022-03-23 11:13:29,069 - mmcls - INFO - Epoch [3][1000/2503]   lr: 1.199e-04, eta: 3 days, 19:04:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.3889
    2022-03-23 11:14:11,528 - mmcls - INFO - Epoch [3][1100/2503]   lr: 1.219e-04, eta: 3 days, 19:03:18, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3871
    2022-03-23 11:14:53,959 - mmcls - INFO - Epoch [3][1200/2503]   lr: 1.239e-04, eta: 3 days, 19:02:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.3959
    2022-03-23 11:15:36,396 - mmcls - INFO - Epoch [3][1300/2503]   lr: 1.259e-04, eta: 3 days, 19:01:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.3745
    2022-03-23 11:16:18,866 - mmcls - INFO - Epoch [3][1400/2503]   lr: 1.279e-04, eta: 3 days, 19:00:22, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3735
    2022-03-23 11:17:01,329 - mmcls - INFO - Epoch [3][1500/2503]   lr: 1.299e-04, eta: 3 days, 18:59:26, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3725
    2022-03-23 11:17:43,782 - mmcls - INFO - Epoch [3][1600/2503]   lr: 1.319e-04, eta: 3 days, 18:58:29, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3825
    2022-03-23 11:18:26,270 - mmcls - INFO - Epoch [3][1700/2503]   lr: 1.339e-04, eta: 3 days, 18:57:37, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3081
    2022-03-23 11:19:08,728 - mmcls - INFO - Epoch [3][1800/2503]   lr: 1.359e-04, eta: 3 days, 18:56:42, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3687
    2022-03-23 11:19:51,210 - mmcls - INFO - Epoch [3][1900/2503]   lr: 1.379e-04, eta: 3 days, 18:55:50, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3503
    2022-03-23 11:20:33,717 - mmcls - INFO - Epoch [3][2000/2503]   lr: 1.399e-04, eta: 3 days, 18:55:00, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3340
    2022-03-23 11:21:16,195 - mmcls - INFO - Epoch [3][2100/2503]   lr: 1.419e-04, eta: 3 days, 18:54:08, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.3334
    2022-03-23 11:21:58,633 - mmcls - INFO - Epoch [3][2200/2503]   lr: 1.439e-04, eta: 3 days, 18:53:12, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.3142
    2022-03-23 11:22:41,080 - mmcls - INFO - Epoch [3][2300/2503]   lr: 1.459e-04, eta: 3 days, 18:52:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.3473
    2022-03-23 11:23:23,498 - mmcls - INFO - Epoch [3][2400/2503]   lr: 1.479e-04, eta: 3 days, 18:51:19, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2653
    2022-03-23 11:24:05,951 - mmcls - INFO - Epoch [3][2500/2503]   lr: 1.499e-04, eta: 3 days, 18:50:26, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.2913
    2022-03-23 11:24:06,953 - mmcls - INFO - Saving checkpoint at 3 epochs
    2022-03-23 11:24:26,914 - mmcls - INFO - Epoch(val) [3][98] accuracy_top-1: 0.1400, accuracy_top-5: 0.8340
    2022-03-23 11:25:12,200 - mmcls - INFO - Epoch [4][100/2503]    lr: 1.519e-04, eta: 3 days, 18:52:07, time: 0.453, data_time: 0.029, memory: 12880, loss: 6.2569
    2022-03-23 11:25:54,676 - mmcls - INFO - Epoch [4][200/2503]    lr: 1.539e-04, eta: 3 days, 18:51:14, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.2797
    2022-03-23 11:26:37,123 - mmcls - INFO - Epoch [4][300/2503]    lr: 1.559e-04, eta: 3 days, 18:50:19, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2431
    2022-03-23 11:27:19,561 - mmcls - INFO - Epoch [4][400/2503]    lr: 1.579e-04, eta: 3 days, 18:49:23, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2062
    2022-03-23 11:28:01,977 - mmcls - INFO - Epoch [4][500/2503]    lr: 1.599e-04, eta: 3 days, 18:48:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2374
    2022-03-23 11:28:44,413 - mmcls - INFO - Epoch [4][600/2503]    lr: 1.619e-04, eta: 3 days, 18:47:30, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2185
    2022-03-23 11:29:26,814 - mmcls - INFO - Epoch [4][700/2503]    lr: 1.639e-04, eta: 3 days, 18:46:32, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2191
    2022-03-23 11:30:09,232 - mmcls - INFO - Epoch [4][800/2503]    lr: 1.659e-04, eta: 3 days, 18:45:36, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1896
    2022-03-23 11:30:51,666 - mmcls - INFO - Epoch [4][900/2503]    lr: 1.679e-04, eta: 3 days, 18:44:41, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2025
    2022-03-23 11:31:34,087 - mmcls - INFO - Epoch [4][1000/2503]   lr: 1.699e-04, eta: 3 days, 18:43:46, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1794
    2022-03-23 11:32:16,504 - mmcls - INFO - Epoch [4][1100/2503]   lr: 1.719e-04, eta: 3 days, 18:42:50, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.2062
    2022-03-23 11:32:58,903 - mmcls - INFO - Epoch [4][1200/2503]   lr: 1.739e-04, eta: 3 days, 18:41:53, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1968
    2022-03-23 11:33:41,321 - mmcls - INFO - Epoch [4][1300/2503]   lr: 1.759e-04, eta: 3 days, 18:40:58, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1645
    2022-03-23 11:34:23,740 - mmcls - INFO - Epoch [4][1400/2503]   lr: 1.779e-04, eta: 3 days, 18:40:04, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1673
    2022-03-23 11:35:06,179 - mmcls - INFO - Epoch [4][1500/2503]   lr: 1.799e-04, eta: 3 days, 18:39:11, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1351
    2022-03-23 11:35:48,630 - mmcls - INFO - Epoch [4][1600/2503]   lr: 1.819e-04, eta: 3 days, 18:38:20, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1459
    2022-03-23 11:36:31,063 - mmcls - INFO - Epoch [4][1700/2503]   lr: 1.839e-04, eta: 3 days, 18:37:27, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1667
    2022-03-23 11:37:13,502 - mmcls - INFO - Epoch [4][1800/2503]   lr: 1.859e-04, eta: 3 days, 18:36:36, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1458
    2022-03-23 11:37:55,927 - mmcls - INFO - Epoch [4][1900/2503]   lr: 1.879e-04, eta: 3 days, 18:35:43, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1541
    2022-03-23 11:38:38,354 - mmcls - INFO - Epoch [4][2000/2503]   lr: 1.899e-04, eta: 3 days, 18:34:51, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.0991
    2022-03-23 11:39:20,788 - mmcls - INFO - Epoch [4][2100/2503]   lr: 1.919e-04, eta: 3 days, 18:33:59, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1121
    2022-03-23 11:40:03,207 - mmcls - INFO - Epoch [4][2200/2503]   lr: 1.939e-04, eta: 3 days, 18:33:06, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.0911
    2022-03-23 11:40:45,648 - mmcls - INFO - Epoch [4][2300/2503]   lr: 1.959e-04, eta: 3 days, 18:32:16, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.1186
    2022-03-23 11:41:28,111 - mmcls - INFO - Epoch [4][2400/2503]   lr: 1.979e-04, eta: 3 days, 18:31:27, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.0845
    2022-03-23 11:42:10,548 - mmcls - INFO - Epoch [4][2500/2503]   lr: 1.999e-04, eta: 3 days, 18:30:36, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.0249
    2022-03-23 11:42:11,554 - mmcls - INFO - Saving checkpoint at 4 epochs
    2022-03-23 11:42:31,707 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_2.pth was removed
    2022-03-23 11:42:32,413 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_4.pth.
    2022-03-23 11:42:32,414 - mmcls - INFO - Best accuracy_top-1 is 0.2500 at 4 epoch.
    2022-03-23 11:42:32,437 - mmcls - INFO - Epoch(val) [4][98] accuracy_top-1: 0.2500, accuracy_top-5: 1.1240
    2022-03-23 11:43:17,657 - mmcls - INFO - Epoch [5][100/2503]    lr: 2.019e-04, eta: 3 days, 18:31:37, time: 0.452, data_time: 0.028, memory: 12880, loss: 6.0452
    2022-03-23 11:44:00,118 - mmcls - INFO - Epoch [5][200/2503]    lr: 2.039e-04, eta: 3 days, 18:30:47, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.0335
    2022-03-23 11:44:42,522 - mmcls - INFO - Epoch [5][300/2503]    lr: 2.059e-04, eta: 3 days, 18:29:54, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9925
    2022-03-23 11:45:24,994 - mmcls - INFO - Epoch [5][400/2503]    lr: 2.079e-04, eta: 3 days, 18:29:05, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.0757
    2022-03-23 11:46:07,425 - mmcls - INFO - Epoch [5][500/2503]    lr: 2.099e-04, eta: 3 days, 18:28:13, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.0613
    2022-03-23 11:46:49,865 - mmcls - INFO - Epoch [5][600/2503]    lr: 2.119e-04, eta: 3 days, 18:27:23, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9896
    2022-03-23 11:47:32,294 - mmcls - INFO - Epoch [5][700/2503]    lr: 2.139e-04, eta: 3 days, 18:26:32, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9823
    2022-03-23 11:48:14,771 - mmcls - INFO - Epoch [5][800/2503]    lr: 2.159e-04, eta: 3 days, 18:25:44, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.0308
    2022-03-23 11:48:57,192 - mmcls - INFO - Epoch [5][900/2503]    lr: 2.179e-04, eta: 3 days, 18:24:52, time: 0.424, data_time: 0.001, memory: 12880, loss: 6.0407
    2022-03-23 11:49:39,606 - mmcls - INFO - Epoch [5][1000/2503]   lr: 2.199e-04, eta: 3 days, 18:24:01, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9634
    2022-03-23 11:50:22,050 - mmcls - INFO - Epoch [5][1100/2503]   lr: 2.219e-04, eta: 3 days, 18:23:11, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9886
    2022-03-23 11:51:04,502 - mmcls - INFO - Epoch [5][1200/2503]   lr: 2.239e-04, eta: 3 days, 18:22:22, time: 0.425, data_time: 0.001, memory: 12880, loss: 6.0200
    2022-03-23 11:51:46,910 - mmcls - INFO - Epoch [5][1300/2503]   lr: 2.259e-04, eta: 3 days, 18:21:30, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9780
    2022-03-23 11:52:29,380 - mmcls - INFO - Epoch [5][1400/2503]   lr: 2.278e-04, eta: 3 days, 18:20:43, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.9368
    2022-03-23 11:53:11,847 - mmcls - INFO - Epoch [5][1500/2503]   lr: 2.298e-04, eta: 3 days, 18:19:55, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.9987
    2022-03-23 11:53:54,291 - mmcls - INFO - Epoch [5][1600/2503]   lr: 2.318e-04, eta: 3 days, 18:19:06, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.8718
    2022-03-23 11:54:36,732 - mmcls - INFO - Epoch [5][1700/2503]   lr: 2.338e-04, eta: 3 days, 18:18:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.8597
    2022-03-23 11:55:19,219 - mmcls - INFO - Epoch [5][1800/2503]   lr: 2.358e-04, eta: 3 days, 18:17:31, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8847
    2022-03-23 11:56:01,680 - mmcls - INFO - Epoch [5][1900/2503]   lr: 2.378e-04, eta: 3 days, 18:16:43, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8749
    2022-03-23 11:56:44,138 - mmcls - INFO - Epoch [5][2000/2503]   lr: 2.398e-04, eta: 3 days, 18:15:56, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8870
    2022-03-23 11:57:26,586 - mmcls - INFO - Epoch [5][2100/2503]   lr: 2.418e-04, eta: 3 days, 18:15:07, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.9272
    2022-03-23 11:58:09,045 - mmcls - INFO - Epoch [5][2200/2503]   lr: 2.438e-04, eta: 3 days, 18:14:20, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8992
    2022-03-23 11:58:51,488 - mmcls - INFO - Epoch [5][2300/2503]   lr: 2.458e-04, eta: 3 days, 18:13:31, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.8747
    2022-03-23 11:59:33,932 - mmcls - INFO - Epoch [5][2400/2503]   lr: 2.478e-04, eta: 3 days, 18:12:43, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.8151
    2022-03-23 12:00:16,393 - mmcls - INFO - Epoch [5][2500/2503]   lr: 2.498e-04, eta: 3 days, 18:11:56, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8157
    2022-03-23 12:00:17,396 - mmcls - INFO - Saving checkpoint at 5 epochs
    2022-03-23 12:00:37,504 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_4.pth was removed
    2022-03-23 12:00:38,212 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_5.pth.
    2022-03-23 12:00:38,212 - mmcls - INFO - Best accuracy_top-1 is 0.3660 at 5 epoch.
    2022-03-23 12:00:38,233 - mmcls - INFO - Epoch(val) [5][98] accuracy_top-1: 0.3660, accuracy_top-5: 1.4640
    2022-03-23 12:01:23,493 - mmcls - INFO - Epoch [6][100/2503]    lr: 2.518e-04, eta: 3 days, 18:12:39, time: 0.452, data_time: 0.028, memory: 12880, loss: 5.8835
    2022-03-23 12:02:05,950 - mmcls - INFO - Epoch [6][200/2503]    lr: 2.538e-04, eta: 3 days, 18:11:51, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8280
    2022-03-23 12:02:48,432 - mmcls - INFO - Epoch [6][300/2503]    lr: 2.558e-04, eta: 3 days, 18:11:05, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.8135
    2022-03-23 12:03:30,867 - mmcls - INFO - Epoch [6][400/2503]    lr: 2.578e-04, eta: 3 days, 18:10:16, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7560
    2022-03-23 12:04:13,270 - mmcls - INFO - Epoch [6][500/2503]    lr: 2.598e-04, eta: 3 days, 18:09:25, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7943
    2022-03-23 12:04:55,685 - mmcls - INFO - Epoch [6][600/2503]    lr: 2.618e-04, eta: 3 days, 18:08:35, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7484
    2022-03-23 12:05:38,079 - mmcls - INFO - Epoch [6][700/2503]    lr: 2.638e-04, eta: 3 days, 18:07:44, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7534
    2022-03-23 12:06:20,521 - mmcls - INFO - Epoch [6][800/2503]    lr: 2.658e-04, eta: 3 days, 18:06:55, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7400
    2022-03-23 12:07:02,935 - mmcls - INFO - Epoch [6][900/2503]    lr: 2.678e-04, eta: 3 days, 18:06:06, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7088
    2022-03-23 12:07:45,343 - mmcls - INFO - Epoch [6][1000/2503]   lr: 2.698e-04, eta: 3 days, 18:05:16, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7690
    2022-03-23 12:08:27,751 - mmcls - INFO - Epoch [6][1100/2503]   lr: 2.718e-04, eta: 3 days, 18:04:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7042
    2022-03-23 12:09:10,178 - mmcls - INFO - Epoch [6][1200/2503]   lr: 2.738e-04, eta: 3 days, 18:03:37, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7129
    2022-03-23 12:09:52,610 - mmcls - INFO - Epoch [6][1300/2503]   lr: 2.758e-04, eta: 3 days, 18:02:49, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7093
    2022-03-23 12:10:35,104 - mmcls - INFO - Epoch [6][1400/2503]   lr: 2.778e-04, eta: 3 days, 18:02:04, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.7566
    2022-03-23 12:11:17,529 - mmcls - INFO - Epoch [6][1500/2503]   lr: 2.798e-04, eta: 3 days, 18:01:15, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.7343
    2022-03-23 12:11:59,964 - mmcls - INFO - Epoch [6][1600/2503]   lr: 2.818e-04, eta: 3 days, 18:00:27, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.6821
    2022-03-23 12:12:42,380 - mmcls - INFO - Epoch [6][1700/2503]   lr: 2.837e-04, eta: 3 days, 17:59:39, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.6761
    2022-03-23 12:13:24,785 - mmcls - INFO - Epoch [6][1800/2503]   lr: 2.857e-04, eta: 3 days, 17:58:49, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5884
    2022-03-23 12:14:07,193 - mmcls - INFO - Epoch [6][1900/2503]   lr: 2.877e-04, eta: 3 days, 17:58:00, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5798
    2022-03-23 12:14:49,602 - mmcls - INFO - Epoch [6][2000/2503]   lr: 2.897e-04, eta: 3 days, 17:57:11, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.6753
    2022-03-23 12:15:32,047 - mmcls - INFO - Epoch [6][2100/2503]   lr: 2.917e-04, eta: 3 days, 17:56:24, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5436
    2022-03-23 12:16:14,503 - mmcls - INFO - Epoch [6][2200/2503]   lr: 2.937e-04, eta: 3 days, 17:55:38, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.6113
    2022-03-23 12:16:56,940 - mmcls - INFO - Epoch [6][2300/2503]   lr: 2.957e-04, eta: 3 days, 17:54:50, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.6247
    2022-03-23 12:17:39,411 - mmcls - INFO - Epoch [6][2400/2503]   lr: 2.977e-04, eta: 3 days, 17:54:05, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.6226
    2022-03-23 12:18:21,866 - mmcls - INFO - Epoch [6][2500/2503]   lr: 2.997e-04, eta: 3 days, 17:53:19, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.4714
    2022-03-23 12:18:22,875 - mmcls - INFO - Saving checkpoint at 6 epochs
    2022-03-23 12:18:42,557 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_5.pth was removed
    2022-03-23 12:18:43,228 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_6.pth.
    2022-03-23 12:18:43,228 - mmcls - INFO - Best accuracy_top-1 is 0.4420 at 6 epoch.
    2022-03-23 12:18:43,248 - mmcls - INFO - Epoch(val) [6][98] accuracy_top-1: 0.4420, accuracy_top-5: 1.8180
    2022-03-23 12:19:28,602 - mmcls - INFO - Epoch [7][100/2503]    lr: 3.017e-04, eta: 3 days, 17:53:52, time: 0.453, data_time: 0.029, memory: 12880, loss: 5.6051
    2022-03-23 12:20:11,003 - mmcls - INFO - Epoch [7][200/2503]    lr: 3.037e-04, eta: 3 days, 17:53:03, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5935
    2022-03-23 12:20:53,398 - mmcls - INFO - Epoch [7][300/2503]    lr: 3.057e-04, eta: 3 days, 17:52:13, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5488
    2022-03-23 12:21:35,842 - mmcls - INFO - Epoch [7][400/2503]    lr: 3.077e-04, eta: 3 days, 17:51:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5793
    2022-03-23 12:22:18,264 - mmcls - INFO - Epoch [7][500/2503]    lr: 3.097e-04, eta: 3 days, 17:50:38, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.6229
    2022-03-23 12:23:00,701 - mmcls - INFO - Epoch [7][600/2503]    lr: 3.117e-04, eta: 3 days, 17:49:51, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5269
    2022-03-23 12:23:43,146 - mmcls - INFO - Epoch [7][700/2503]    lr: 3.137e-04, eta: 3 days, 17:49:04, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.6257
    2022-03-23 12:24:25,580 - mmcls - INFO - Epoch [7][800/2503]    lr: 3.157e-04, eta: 3 days, 17:48:16, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5215
    2022-03-23 12:25:07,988 - mmcls - INFO - Epoch [7][900/2503]    lr: 3.177e-04, eta: 3 days, 17:47:28, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5657
    2022-03-23 12:25:50,445 - mmcls - INFO - Epoch [7][1000/2503]   lr: 3.196e-04, eta: 3 days, 17:46:41, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5037
    2022-03-23 12:26:32,881 - mmcls - INFO - Epoch [7][1100/2503]   lr: 3.216e-04, eta: 3 days, 17:45:55, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5610
    2022-03-23 12:27:15,304 - mmcls - INFO - Epoch [7][1200/2503]   lr: 3.236e-04, eta: 3 days, 17:45:07, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5020
    2022-03-23 12:27:57,711 - mmcls - INFO - Epoch [7][1300/2503]   lr: 3.256e-04, eta: 3 days, 17:44:19, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5258
    2022-03-23 12:28:40,126 - mmcls - INFO - Epoch [7][1400/2503]   lr: 3.276e-04, eta: 3 days, 17:43:31, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5536
    2022-03-23 12:29:22,550 - mmcls - INFO - Epoch [7][1500/2503]   lr: 3.296e-04, eta: 3 days, 17:42:43, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4569
    2022-03-23 12:30:04,959 - mmcls - INFO - Epoch [7][1600/2503]   lr: 3.316e-04, eta: 3 days, 17:41:55, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4433
    2022-03-23 12:30:47,364 - mmcls - INFO - Epoch [7][1700/2503]   lr: 3.336e-04, eta: 3 days, 17:41:07, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.5185
    2022-03-23 12:31:29,789 - mmcls - INFO - Epoch [7][1800/2503]   lr: 3.356e-04, eta: 3 days, 17:40:20, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4619
    2022-03-23 12:32:12,228 - mmcls - INFO - Epoch [7][1900/2503]   lr: 3.376e-04, eta: 3 days, 17:39:34, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4917
    2022-03-23 12:32:54,663 - mmcls - INFO - Epoch [7][2000/2503]   lr: 3.396e-04, eta: 3 days, 17:38:47, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4863
    2022-03-23 12:33:37,099 - mmcls - INFO - Epoch [7][2100/2503]   lr: 3.416e-04, eta: 3 days, 17:38:01, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3654
    2022-03-23 12:34:19,556 - mmcls - INFO - Epoch [7][2200/2503]   lr: 3.436e-04, eta: 3 days, 17:37:15, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.3662
    2022-03-23 12:35:01,992 - mmcls - INFO - Epoch [7][2300/2503]   lr: 3.456e-04, eta: 3 days, 17:36:29, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3584
    2022-03-23 12:35:44,432 - mmcls - INFO - Epoch [7][2400/2503]   lr: 3.476e-04, eta: 3 days, 17:35:42, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4054
    2022-03-23 12:36:26,857 - mmcls - INFO - Epoch [7][2500/2503]   lr: 3.496e-04, eta: 3 days, 17:34:56, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4745
    2022-03-23 12:36:27,862 - mmcls - INFO - Saving checkpoint at 7 epochs
    2022-03-23 12:36:47,272 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_6.pth was removed
    2022-03-23 12:36:47,982 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_7.pth.
    2022-03-23 12:36:47,982 - mmcls - INFO - Best accuracy_top-1 is 0.6260 at 7 epoch.
    2022-03-23 12:36:48,000 - mmcls - INFO - Epoch(val) [7][98] accuracy_top-1: 0.6260, accuracy_top-5: 2.3580
    2022-03-23 12:37:33,139 - mmcls - INFO - Epoch [8][100/2503]    lr: 3.515e-04, eta: 3 days, 17:35:09, time: 0.451, data_time: 0.027, memory: 12880, loss: 5.3889
    2022-03-23 12:38:15,538 - mmcls - INFO - Epoch [8][200/2503]    lr: 3.535e-04, eta: 3 days, 17:34:21, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3353
    2022-03-23 12:38:57,965 - mmcls - INFO - Epoch [8][300/2503]    lr: 3.555e-04, eta: 3 days, 17:33:34, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3097
    2022-03-23 12:39:40,378 - mmcls - INFO - Epoch [8][400/2503]    lr: 3.575e-04, eta: 3 days, 17:32:46, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3686
    2022-03-23 12:40:22,791 - mmcls - INFO - Epoch [8][500/2503]    lr: 3.595e-04, eta: 3 days, 17:31:59, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3871
    2022-03-23 12:41:05,189 - mmcls - INFO - Epoch [8][600/2503]    lr: 3.615e-04, eta: 3 days, 17:31:11, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2959
    2022-03-23 12:41:47,636 - mmcls - INFO - Epoch [8][700/2503]    lr: 3.635e-04, eta: 3 days, 17:30:25, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4423
    2022-03-23 12:42:30,076 - mmcls - INFO - Epoch [8][800/2503]    lr: 3.655e-04, eta: 3 days, 17:29:39, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3492
    2022-03-23 12:43:12,491 - mmcls - INFO - Epoch [8][900/2503]    lr: 3.675e-04, eta: 3 days, 17:28:52, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3208
    2022-03-23 12:43:54,906 - mmcls - INFO - Epoch [8][1000/2503]   lr: 3.695e-04, eta: 3 days, 17:28:05, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3191
    2022-03-23 12:44:37,328 - mmcls - INFO - Epoch [8][1100/2503]   lr: 3.715e-04, eta: 3 days, 17:27:18, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2006
    2022-03-23 12:45:19,764 - mmcls - INFO - Epoch [8][1200/2503]   lr: 3.735e-04, eta: 3 days, 17:26:32, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3412
    2022-03-23 12:46:02,227 - mmcls - INFO - Epoch [8][1300/2503]   lr: 3.755e-04, eta: 3 days, 17:25:47, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.3219
    2022-03-23 12:46:44,636 - mmcls - INFO - Epoch [8][1400/2503]   lr: 3.774e-04, eta: 3 days, 17:25:00, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.4217
    2022-03-23 12:47:27,055 - mmcls - INFO - Epoch [8][1500/2503]   lr: 3.794e-04, eta: 3 days, 17:24:13, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2380
    2022-03-23 12:48:09,452 - mmcls - INFO - Epoch [8][1600/2503]   lr: 3.814e-04, eta: 3 days, 17:23:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3091
    2022-03-23 12:48:51,871 - mmcls - INFO - Epoch [8][1700/2503]   lr: 3.834e-04, eta: 3 days, 17:22:39, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2439
    2022-03-23 12:49:34,286 - mmcls - INFO - Epoch [8][1800/2503]   lr: 3.854e-04, eta: 3 days, 17:21:52, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3191
    2022-03-23 12:50:16,764 - mmcls - INFO - Epoch [8][1900/2503]   lr: 3.874e-04, eta: 3 days, 17:21:08, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.2014
    2022-03-23 12:50:59,182 - mmcls - INFO - Epoch [8][2000/2503]   lr: 3.894e-04, eta: 3 days, 17:20:22, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.3611
    2022-03-23 12:51:41,622 - mmcls - INFO - Epoch [8][2100/2503]   lr: 3.914e-04, eta: 3 days, 17:19:36, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2972
    2022-03-23 12:52:24,078 - mmcls - INFO - Epoch [8][2200/2503]   lr: 3.934e-04, eta: 3 days, 17:18:51, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.2609
    2022-03-23 12:53:06,517 - mmcls - INFO - Epoch [8][2300/2503]   lr: 3.954e-04, eta: 3 days, 17:18:05, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2051
    2022-03-23 12:53:48,940 - mmcls - INFO - Epoch [8][2400/2503]   lr: 3.974e-04, eta: 3 days, 17:17:19, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2481
    2022-03-23 12:54:31,389 - mmcls - INFO - Epoch [8][2500/2503]   lr: 3.994e-04, eta: 3 days, 17:16:34, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.2086
    2022-03-23 12:54:32,392 - mmcls - INFO - Saving checkpoint at 8 epochs
    2022-03-23 12:54:51,744 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_7.pth was removed
    2022-03-23 12:54:52,411 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_8.pth.
    2022-03-23 12:54:52,411 - mmcls - INFO - Best accuracy_top-1 is 0.8340 at 8 epoch.
    2022-03-23 12:54:52,444 - mmcls - INFO - Epoch(val) [8][98] accuracy_top-1: 0.8340, accuracy_top-5: 2.9860
    2022-03-23 12:55:37,606 - mmcls - INFO - Epoch [9][100/2503]    lr: 4.013e-04, eta: 3 days, 17:16:41, time: 0.451, data_time: 0.028, memory: 12880, loss: 5.1619
    2022-03-23 12:56:20,007 - mmcls - INFO - Epoch [9][200/2503]    lr: 4.033e-04, eta: 3 days, 17:15:54, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2232
    2022-03-23 12:57:02,429 - mmcls - INFO - Epoch [9][300/2503]    lr: 4.053e-04, eta: 3 days, 17:15:08, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2716
    2022-03-23 12:57:44,860 - mmcls - INFO - Epoch [9][400/2503]    lr: 4.073e-04, eta: 3 days, 17:14:22, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1478
    2022-03-23 12:58:27,279 - mmcls - INFO - Epoch [9][500/2503]    lr: 4.093e-04, eta: 3 days, 17:13:35, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1532
    2022-03-23 12:59:09,706 - mmcls - INFO - Epoch [9][600/2503]    lr: 4.113e-04, eta: 3 days, 17:12:49, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2009
    2022-03-23 12:59:52,122 - mmcls - INFO - Epoch [9][700/2503]    lr: 4.133e-04, eta: 3 days, 17:12:03, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1359
    2022-03-23 13:00:34,544 - mmcls - INFO - Epoch [9][800/2503]    lr: 4.152e-04, eta: 3 days, 17:11:17, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1830
    2022-03-23 13:01:16,923 - mmcls - INFO - Epoch [9][900/2503]    lr: 4.172e-04, eta: 3 days, 17:10:29, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1473
    2022-03-23 13:01:59,330 - mmcls - INFO - Epoch [9][1000/2503]   lr: 4.192e-04, eta: 3 days, 17:09:42, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2444
    2022-03-23 13:02:41,772 - mmcls - INFO - Epoch [9][1100/2503]   lr: 4.212e-04, eta: 3 days, 17:08:57, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1273
    2022-03-23 13:03:24,221 - mmcls - INFO - Epoch [9][1200/2503]   lr: 4.232e-04, eta: 3 days, 17:08:12, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1167
    2022-03-23 13:04:06,630 - mmcls - INFO - Epoch [9][1300/2503]   lr: 4.252e-04, eta: 3 days, 17:07:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1668
    2022-03-23 13:04:49,038 - mmcls - INFO - Epoch [9][1400/2503]   lr: 4.272e-04, eta: 3 days, 17:06:39, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.2332
    2022-03-23 13:05:31,444 - mmcls - INFO - Epoch [9][1500/2503]   lr: 4.292e-04, eta: 3 days, 17:05:53, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0396
    2022-03-23 13:06:13,858 - mmcls - INFO - Epoch [9][1600/2503]   lr: 4.312e-04, eta: 3 days, 17:05:07, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1504
    2022-03-23 13:06:56,272 - mmcls - INFO - Epoch [9][1700/2503]   lr: 4.332e-04, eta: 3 days, 17:04:20, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1715
    2022-03-23 13:07:38,707 - mmcls - INFO - Epoch [9][1800/2503]   lr: 4.352e-04, eta: 3 days, 17:03:35, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1593
    2022-03-23 13:08:21,154 - mmcls - INFO - Epoch [9][1900/2503]   lr: 4.372e-04, eta: 3 days, 17:02:50, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1322
    2022-03-23 13:09:03,577 - mmcls - INFO - Epoch [9][2000/2503]   lr: 4.392e-04, eta: 3 days, 17:02:04, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0678
    2022-03-23 13:09:46,013 - mmcls - INFO - Epoch [9][2100/2503]   lr: 4.412e-04, eta: 3 days, 17:01:19, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0796
    2022-03-23 13:10:28,445 - mmcls - INFO - Epoch [9][2200/2503]   lr: 4.432e-04, eta: 3 days, 17:00:34, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0929
    2022-03-23 13:11:10,933 - mmcls - INFO - Epoch [9][2300/2503]   lr: 4.452e-04, eta: 3 days, 16:59:50, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.1142
    2022-03-23 13:11:53,396 - mmcls - INFO - Epoch [9][2400/2503]   lr: 4.472e-04, eta: 3 days, 16:59:06, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.1410
    2022-03-23 13:12:35,810 - mmcls - INFO - Epoch [9][2500/2503]   lr: 4.491e-04, eta: 3 days, 16:58:20, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1710
    2022-03-23 13:12:36,812 - mmcls - INFO - Saving checkpoint at 9 epochs
    2022-03-23 13:12:56,453 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_8.pth was removed
    2022-03-23 13:12:57,145 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_9.pth.
    2022-03-23 13:12:57,145 - mmcls - INFO - Best accuracy_top-1 is 0.9320 at 9 epoch.
    2022-03-23 13:12:57,164 - mmcls - INFO - Epoch(val) [9][98] accuracy_top-1: 0.9320, accuracy_top-5: 3.3960
    2022-03-23 13:13:42,347 - mmcls - INFO - Epoch [10][100/2503]   lr: 4.510e-04, eta: 3 days, 16:58:22, time: 0.452, data_time: 0.028, memory: 12880, loss: 5.1018
    2022-03-23 13:14:24,748 - mmcls - INFO - Epoch [10][200/2503]   lr: 4.530e-04, eta: 3 days, 16:57:36, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9857
    2022-03-23 13:15:07,166 - mmcls - INFO - Epoch [10][300/2503]   lr: 4.550e-04, eta: 3 days, 16:56:50, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0640
    2022-03-23 13:15:49,619 - mmcls - INFO - Epoch [10][400/2503]   lr: 4.570e-04, eta: 3 days, 16:56:05, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.1239
    2022-03-23 13:16:32,052 - mmcls - INFO - Epoch [10][500/2503]   lr: 4.590e-04, eta: 3 days, 16:55:20, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0758
    2022-03-23 13:17:14,475 - mmcls - INFO - Epoch [10][600/2503]   lr: 4.610e-04, eta: 3 days, 16:54:34, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0414
    2022-03-23 13:17:56,885 - mmcls - INFO - Epoch [10][700/2503]   lr: 4.630e-04, eta: 3 days, 16:53:48, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0110
    2022-03-23 13:18:39,276 - mmcls - INFO - Epoch [10][800/2503]   lr: 4.649e-04, eta: 3 days, 16:53:02, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0853
    2022-03-23 13:19:21,679 - mmcls - INFO - Epoch [10][900/2503]   lr: 4.669e-04, eta: 3 days, 16:52:15, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0756
    2022-03-23 13:20:04,105 - mmcls - INFO - Epoch [10][1000/2503]  lr: 4.689e-04, eta: 3 days, 16:51:30, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0635
    2022-03-23 13:20:46,500 - mmcls - INFO - Epoch [10][1100/2503]  lr: 4.709e-04, eta: 3 days, 16:50:44, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9883
    2022-03-23 13:21:28,923 - mmcls - INFO - Epoch [10][1200/2503]  lr: 4.729e-04, eta: 3 days, 16:49:58, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0592
    2022-03-23 13:22:11,337 - mmcls - INFO - Epoch [10][1300/2503]  lr: 4.749e-04, eta: 3 days, 16:49:12, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0517
    2022-03-23 13:22:53,750 - mmcls - INFO - Epoch [10][1400/2503]  lr: 4.769e-04, eta: 3 days, 16:48:27, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.1607
    2022-03-23 13:23:36,180 - mmcls - INFO - Epoch [10][1500/2503]  lr: 4.789e-04, eta: 3 days, 16:47:42, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0675
    2022-03-23 13:24:18,607 - mmcls - INFO - Epoch [10][1600/2503]  lr: 4.809e-04, eta: 3 days, 16:46:56, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0869
    2022-03-23 13:25:01,048 - mmcls - INFO - Epoch [10][1700/2503]  lr: 4.829e-04, eta: 3 days, 16:46:12, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0254
    2022-03-23 13:25:43,474 - mmcls - INFO - Epoch [10][1800/2503]  lr: 4.849e-04, eta: 3 days, 16:45:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0226
    2022-03-23 13:26:25,906 - mmcls - INFO - Epoch [10][1900/2503]  lr: 4.869e-04, eta: 3 days, 16:44:41, time: 0.424, data_time: 0.001, memory: 12880, loss: 5.0220
    2022-03-23 13:27:08,344 - mmcls - INFO - Epoch [10][2000/2503]  lr: 4.889e-04, eta: 3 days, 16:43:56, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9867
    2022-03-23 13:27:50,782 - mmcls - INFO - Epoch [10][2100/2503]  lr: 4.909e-04, eta: 3 days, 16:43:12, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9525
    2022-03-23 13:28:33,201 - mmcls - INFO - Epoch [10][2200/2503]  lr: 4.929e-04, eta: 3 days, 16:42:26, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9255
    2022-03-23 13:29:15,639 - mmcls - INFO - Epoch [10][2300/2503]  lr: 4.948e-04, eta: 3 days, 16:41:42, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9990
    2022-03-23 13:29:58,068 - mmcls - INFO - Epoch [10][2400/2503]  lr: 4.968e-04, eta: 3 days, 16:40:56, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.8986
    2022-03-23 13:30:40,530 - mmcls - INFO - Epoch [10][2500/2503]  lr: 4.988e-04, eta: 3 days, 16:40:13, time: 0.425, data_time: 0.001, memory: 12880, loss: 5.0543
    2022-03-23 13:30:41,536 - mmcls - INFO - Saving checkpoint at 10 epochs
    2022-03-23 13:31:01,121 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_9.pth was removed
    2022-03-23 13:31:01,814 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_10.pth.
    2022-03-23 13:31:01,814 - mmcls - INFO - Best accuracy_top-1 is 1.0060 at 10 epoch.
    2022-03-23 13:31:01,832 - mmcls - INFO - Epoch(val) [10][98]    accuracy_top-1: 1.0060, accuracy_top-5: 3.6420
    2022-03-23 13:31:47,050 - mmcls - INFO - Epoch [11][100/2503]   lr: 4.987e-04, eta: 3 days, 16:40:11, time: 0.452, data_time: 0.028, memory: 12880, loss: 5.0122
    2022-03-23 13:32:29,469 - mmcls - INFO - Epoch [11][200/2503]   lr: 4.987e-04, eta: 3 days, 16:39:26, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8856
    2022-03-23 13:33:11,877 - mmcls - INFO - Epoch [11][300/2503]   lr: 4.987e-04, eta: 3 days, 16:38:40, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8784
    2022-03-23 13:33:54,298 - mmcls - INFO - Epoch [11][400/2503]   lr: 4.987e-04, eta: 3 days, 16:37:55, time: 0.424, data_time: 0.000, memory: 12880, loss: 5.0453
    2022-03-23 13:34:36,709 - mmcls - INFO - Epoch [11][500/2503]   lr: 4.987e-04, eta: 3 days, 16:37:09, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9763
    2022-03-23 13:35:19,105 - mmcls - INFO - Epoch [11][600/2503]   lr: 4.987e-04, eta: 3 days, 16:36:23, time: 0.424, data_time: 0.000, memory: 12880, loss: 5.0149
    2022-03-23 13:36:01,493 - mmcls - INFO - Epoch [11][700/2503]   lr: 4.987e-04, eta: 3 days, 16:35:37, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8400
    2022-03-23 13:36:43,928 - mmcls - INFO - Epoch [11][800/2503]   lr: 4.987e-04, eta: 3 days, 16:34:52, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9379
    2022-03-23 13:37:26,330 - mmcls - INFO - Epoch [11][900/2503]   lr: 4.987e-04, eta: 3 days, 16:34:06, time: 0.424, data_time: 0.000, memory: 12880, loss: 5.0567
    2022-03-23 13:38:08,770 - mmcls - INFO - Epoch [11][1000/2503]  lr: 4.987e-04, eta: 3 days, 16:33:21, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8969
    2022-03-23 13:38:51,170 - mmcls - INFO - Epoch [11][1100/2503]  lr: 4.987e-04, eta: 3 days, 16:32:36, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9219
    2022-03-23 13:39:33,588 - mmcls - INFO - Epoch [11][1200/2503]  lr: 4.987e-04, eta: 3 days, 16:31:51, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9480
    2022-03-23 13:40:16,042 - mmcls - INFO - Epoch [11][1300/2503]  lr: 4.987e-04, eta: 3 days, 16:31:06, time: 0.425, data_time: 0.000, memory: 12880, loss: 4.7931
    2022-03-23 13:40:58,450 - mmcls - INFO - Epoch [11][1400/2503]  lr: 4.987e-04, eta: 3 days, 16:30:21, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9114
    2022-03-23 13:41:40,879 - mmcls - INFO - Epoch [11][1500/2503]  lr: 4.987e-04, eta: 3 days, 16:29:36, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7671
    2022-03-23 13:42:23,308 - mmcls - INFO - Epoch [11][1600/2503]  lr: 4.987e-04, eta: 3 days, 16:28:51, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9768
    2022-03-23 13:43:05,768 - mmcls - INFO - Epoch [11][1700/2503]  lr: 4.987e-04, eta: 3 days, 16:28:07, time: 0.425, data_time: 0.000, memory: 12880, loss: 4.9730
    2022-03-23 13:43:48,184 - mmcls - INFO - Epoch [11][1800/2503]  lr: 4.987e-04, eta: 3 days, 16:27:22, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7879
    2022-03-23 13:44:30,600 - mmcls - INFO - Epoch [11][1900/2503]  lr: 4.987e-04, eta: 3 days, 16:26:37, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8925
    2022-03-23 13:45:13,025 - mmcls - INFO - Epoch [11][2000/2503]  lr: 4.987e-04, eta: 3 days, 16:25:52, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8880
    2022-03-23 13:45:55,447 - mmcls - INFO - Epoch [11][2100/2503]  lr: 4.987e-04, eta: 3 days, 16:25:07, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.9461
    2022-03-23 13:46:37,884 - mmcls - INFO - Epoch [11][2200/2503]  lr: 4.987e-04, eta: 3 days, 16:24:23, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8730
    2022-03-23 13:47:20,327 - mmcls - INFO - Epoch [11][2300/2503]  lr: 4.987e-04, eta: 3 days, 16:23:38, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8595
    2022-03-23 13:48:02,732 - mmcls - INFO - Epoch [11][2400/2503]  lr: 4.987e-04, eta: 3 days, 16:22:53, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8118
    2022-03-23 13:48:45,166 - mmcls - INFO - Epoch [11][2500/2503]  lr: 4.987e-04, eta: 3 days, 16:22:08, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.9018
    2022-03-23 13:48:46,168 - mmcls - INFO - Saving checkpoint at 11 epochs
    2022-03-23 13:49:05,956 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_10.pth was removed
    2022-03-23 13:49:06,696 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_11.pth.
    2022-03-23 13:49:06,696 - mmcls - INFO - Best accuracy_top-1 is 1.1200 at 11 epoch.
    2022-03-23 13:49:06,716 - mmcls - INFO - Epoch(val) [11][98]    accuracy_top-1: 1.1200, accuracy_top-5: 3.8460
    2022-03-23 13:49:51,908 - mmcls - INFO - Epoch [12][100/2503]   lr: 4.984e-04, eta: 3 days, 16:22:02, time: 0.452, data_time: 0.027, memory: 12880, loss: 4.7691
    2022-03-23 13:50:34,323 - mmcls - INFO - Epoch [12][200/2503]   lr: 4.984e-04, eta: 3 days, 16:21:17, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8273
    2022-03-23 13:51:16,762 - mmcls - INFO - Epoch [12][300/2503]   lr: 4.984e-04, eta: 3 days, 16:20:33, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8532
    2022-03-23 13:51:59,164 - mmcls - INFO - Epoch [12][400/2503]   lr: 4.984e-04, eta: 3 days, 16:19:47, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8709
    2022-03-23 13:52:41,574 - mmcls - INFO - Epoch [12][500/2503]   lr: 4.984e-04, eta: 3 days, 16:19:02, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7656
    2022-03-23 13:53:23,978 - mmcls - INFO - Epoch [12][600/2503]   lr: 4.984e-04, eta: 3 days, 16:18:17, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8151
    2022-03-23 13:54:06,417 - mmcls - INFO - Epoch [12][700/2503]   lr: 4.984e-04, eta: 3 days, 16:17:32, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7869
    2022-03-23 13:54:48,845 - mmcls - INFO - Epoch [12][800/2503]   lr: 4.984e-04, eta: 3 days, 16:16:47, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8620
    2022-03-23 13:55:31,273 - mmcls - INFO - Epoch [12][900/2503]   lr: 4.984e-04, eta: 3 days, 16:16:02, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8394
    2022-03-23 13:56:13,714 - mmcls - INFO - Epoch [12][1000/2503]  lr: 4.984e-04, eta: 3 days, 16:15:18, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7983
    2022-03-23 13:56:56,119 - mmcls - INFO - Epoch [12][1100/2503]  lr: 4.984e-04, eta: 3 days, 16:14:33, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8821
    2022-03-23 13:57:38,512 - mmcls - INFO - Epoch [12][1200/2503]  lr: 4.984e-04, eta: 3 days, 16:13:47, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8515
    2022-03-23 13:58:20,930 - mmcls - INFO - Epoch [12][1300/2503]  lr: 4.984e-04, eta: 3 days, 16:13:02, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8303
    2022-03-23 13:59:03,336 - mmcls - INFO - Epoch [12][1400/2503]  lr: 4.984e-04, eta: 3 days, 16:12:17, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7587
    2022-03-23 13:59:45,763 - mmcls - INFO - Epoch [12][1500/2503]  lr: 4.984e-04, eta: 3 days, 16:11:33, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7310
    2022-03-23 14:00:28,215 - mmcls - INFO - Epoch [12][1600/2503]  lr: 4.984e-04, eta: 3 days, 16:10:49, time: 0.425, data_time: 0.000, memory: 12880, loss: 4.7011
    2022-03-23 14:01:10,651 - mmcls - INFO - Epoch [12][1700/2503]  lr: 4.984e-04, eta: 3 days, 16:10:04, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8352
    2022-03-23 14:01:53,079 - mmcls - INFO - Epoch [12][1800/2503]  lr: 4.984e-04, eta: 3 days, 16:09:20, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7838
    2022-03-23 14:02:35,516 - mmcls - INFO - Epoch [12][1900/2503]  lr: 4.984e-04, eta: 3 days, 16:08:35, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7827
    2022-03-23 14:03:17,938 - mmcls - INFO - Epoch [12][2000/2503]  lr: 4.984e-04, eta: 3 days, 16:07:51, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.6722
    2022-03-23 14:04:00,380 - mmcls - INFO - Epoch [12][2100/2503]  lr: 4.984e-04, eta: 3 days, 16:07:07, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.8223
    2022-03-23 14:04:42,805 - mmcls - INFO - Epoch [12][2200/2503]  lr: 4.984e-04, eta: 3 days, 16:06:22, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7261
    2022-03-23 14:05:25,257 - mmcls - INFO - Epoch [12][2300/2503]  lr: 4.984e-04, eta: 3 days, 16:05:38, time: 0.424, data_time: 0.000, memory: 12880, loss: 4.7646
    2022-03-23 14:06:07,721 - mmcls - INFO - Epoch [12][2400/2503]  lr: 4.984e-04, eta: 3 days, 16:04:55, time: 0.425, data_time: 0.000, memory: 12880, loss: 4.7331
    2022-03-23 14:06:50,161 - mmcls - INFO - Epoch [12][2500/2503]  lr: 4.984e-04, eta: 3 days, 16:04:10, time: 0.424, data_time: 0.001, memory: 12880, loss: 4.7305
    2022-03-23 14:06:51,164 - mmcls - INFO - Saving checkpoint at 12 epochs
    2022-03-23 14:07:10,529 - mmcls - INFO - The previous best checkpoint /home/ma-user/modelarts/outputs/train-url_0/best_accuracy_top-1_epoch_11.pth was removed
    2022-03-23 14:07:11,208 - mmcls - INFO - Now best checkpoint is saved as best_accuracy_top-1_epoch_12.pth.
    2022-03-23 14:07:11,208 - mmcls - INFO - Best accuracy_top-1 is 1.3100 at 12 epoch.
    2022-03-23 14:07:11,229 - mmcls - INFO - Epoch(val) [12][98]    accuracy_top-1: 1.3100, accuracy_top-5: 4.3860
  4. Other code you modified in the mmcls folder. No extra modifications.

Ezra-Yu commented 2 years ago

Have you tested the ImageNet-1k dataset by using the given checkpoints? Make sure that your dataset can get given reported accuracy.

BebDong commented 2 years ago

Yes, I have tested with the given checkpoints and got consistent results as the repo. So the dataset may be OK.

Ezra-Yu commented 2 years ago

can you provide result.pkl get by your checkpoint using test.py,

./tools/dist_test.sh ${CONFIG_FILE} ${CHECKPOINT_FILE} ${GPU_NUM} [--metrics ${METRICS}] [--out ${RESULT_FILE}]
BebDong commented 2 years ago

External file sending may not be convenient. But I noticed that when using tools/test.py to do evaluation the log showed The model and loaded state dict do not match exactly. Details:

unexpected key in source state_dict: ema_backbone_cls_token, ema_backbone_tokens_to_token_attention1_ln1_weight, ema_backbone_tokens_to_token_attention1_ln1_bias, ema_backbone_tokens_to_token_attention1_attn_qkv_weight, ema_backbone_tokens_to_token_attention1_attn_proj_weight, ema_backbone_tokens_to_token_attention1_attn_proj_bias, ema_backbone_tokens_to_token_attention1_ln2_weight, ema_backbone_tokens_to_token_attention1_ln2_bias, ema_backbone_tokens_to_token_attention1_ffn_layers_0_0_weight, ema_backbone_tokens_to_token_attention1_ffn_layers_0_0_bias, ema_backbone_tokens_to_token_attention1_ffn_layers_1_weight, ema_backbone_tokens_to_token_attention1_ffn_layers_1_bias, ema_backbone_tokens_to_token_attention2_ln1_weight, ema_backbone_tokens_to_token_attention2_ln1_bias, ema_backbone_tokens_to_token_attention2_attn_qkv_weight, ema_backbone_tokens_to_token_attention2_attn_proj_weight, ema_backbone_tokens_to_token_attention2_attn_proj_bias, ema_backbone_tokens_to_token_attention2_ln2_weight, ema_backbone_tokens_to_token_attention2_ln2_bias, ema_backbone_tokens_to_token_attention2_ffn_layers_0_0_weight, ema_backbone_tokens_to_token_attention2_ffn_layers_0_0_bias, ema_backbone_tokens_to_token_attention2_ffn_layers_1_weight, ema_backbone_tokens_to_token_attention2_ffn_layers_1_bias, ema_backbone_tokens_to_token_project_weight, ema_backbone_tokens_to_token_project_bias, ema_backbone_encoder_0_ln1_weight, ema_backbone_encoder_0_ln1_bias, ema_backbone_encoder_0_attn_qkv_weight, ema_backbone_encoder_0_attn_proj_weight, ema_backbone_encoder_0_attn_proj_bias, ema_backbone_encoder_0_ln2_weight, ema_backbone_encoder_0_ln2_bias, ema_backbone_encoder_0_ffn_layers_0_0_weight, ema_backbone_encoder_0_ffn_layers_0_0_bias, ema_backbone_encoder_0_ffn_layers_1_weight, ema_backbone_encoder_0_ffn_layers_1_bias, ema_backbone_encoder_1_ln1_weight, ema_backbone_encoder_1_ln1_bias, ema_backbone_encoder_1_attn_qkv_weight, ema_backbone_encoder_1_attn_proj_weight, ema_backbone_encoder_1_attn_proj_bias, ema_backbone_encoder_1_ln2_weight, ema_backbone_encoder_1_ln2_bias, ema_backbone_encoder_1_ffn_layers_0_0_weight, ema_backbone_encoder_1_ffn_layers_0_0_bias, ema_backbone_encoder_1_ffn_layers_1_weight, ema_backbone_encoder_1_ffn_layers_1_bias, ema_backbone_encoder_2_ln1_weight, ema_backbone_encoder_2_ln1_bias, ema_backbone_encoder_2_attn_qkv_weight, ema_backbone_encoder_2_attn_proj_weight, ema_backbone_encoder_2_attn_proj_bias, ema_backbone_encoder_2_ln2_weight, ema_backbone_encoder_2_ln2_bias, ema_backbone_encoder_2_ffn_layers_0_0_weight, ema_backbone_encoder_2_ffn_layers_0_0_bias, ema_backbone_encoder_2_ffn_layers_1_weight, ema_backbone_encoder_2_ffn_layers_1_bias, ema_backbone_encoder_3_ln1_weight, ema_backbone_encoder_3_ln1_bias, ema_backbone_encoder_3_attn_qkv_weight, ema_backbone_encoder_3_attn_proj_weight, ema_backbone_encoder_3_attn_proj_bias, ema_backbone_encoder_3_ln2_weight, ema_backbone_encoder_3_ln2_bias, ema_backbone_encoder_3_ffn_layers_0_0_weight, ema_backbone_encoder_3_ffn_layers_0_0_bias, ema_backbone_encoder_3_ffn_layers_1_weight, ema_backbone_encoder_3_ffn_layers_1_bias, ema_backbone_encoder_4_ln1_weight, ema_backbone_encoder_4_ln1_bias, ema_backbone_encoder_4_attn_qkv_weight, ema_backbone_encoder_4_attn_proj_weight, ema_backbone_encoder_4_attn_proj_bias, ema_backbone_encoder_4_ln2_weight, ema_backbone_encoder_4_ln2_bias, ema_backbone_encoder_4_ffn_layers_0_0_weight, ema_backbone_encoder_4_ffn_layers_0_0_bias, ema_backbone_encoder_4_ffn_layers_1_weight, ema_backbone_encoder_4_ffn_layers_1_bias, ema_backbone_encoder_5_ln1_weight, ema_backbone_encoder_5_ln1_bias, ema_backbone_encoder_5_attn_qkv_weight, ema_backbone_encoder_5_attn_proj_weight, ema_backbone_encoder_5_attn_proj_bias, ema_backbone_encoder_5_ln2_weight, ema_backbone_encoder_5_ln2_bias, ema_backbone_encoder_5_ffn_layers_0_0_weight, ema_backbone_encoder_5_ffn_layers_0_0_bias, ema_backbone_encoder_5_ffn_layers_1_weight, ema_backbone_encoder_5_ffn_layers_1_bias, ema_backbone_encoder_6_ln1_weight, ema_backbone_encoder_6_ln1_bias, ema_backbone_encoder_6_attn_qkv_weight, ema_backbone_encoder_6_attn_proj_weight, ema_backbone_encoder_6_attn_proj_bias, ema_backbone_encoder_6_ln2_weight, ema_backbone_encoder_6_ln2_bias, ema_backbone_encoder_6_ffn_layers_0_0_weight, ema_backbone_encoder_6_ffn_layers_0_0_bias, ema_backbone_encoder_6_ffn_layers_1_weight, ema_backbone_encoder_6_ffn_layers_1_bias, ema_backbone_encoder_7_ln1_weight, ema_backbone_encoder_7_ln1_bias, ema_backbone_encoder_7_attn_qkv_weight, ema_backbone_encoder_7_attn_proj_weight, ema_backbone_encoder_7_attn_proj_bias, ema_backbone_encoder_7_ln2_weight, ema_backbone_encoder_7_ln2_bias, ema_backbone_encoder_7_ffn_layers_0_0_weight, ema_backbone_encoder_7_ffn_layers_0_0_bias, ema_backbone_encoder_7_ffn_layers_1_weight, ema_backbone_encoder_7_ffn_layers_1_bias, ema_backbone_encoder_8_ln1_weight, ema_backbone_encoder_8_ln1_bias, ema_backbone_encoder_8_attn_qkv_weight, ema_backbone_encoder_8_attn_proj_weight, ema_backbone_encoder_8_attn_proj_bias, ema_backbone_encoder_8_ln2_weight, ema_backbone_encoder_8_ln2_bias, ema_backbone_encoder_8_ffn_layers_0_0_weight, ema_backbone_encoder_8_ffn_layers_0_0_bias, ema_backbone_encoder_8_ffn_layers_1_weight, ema_backbone_encoder_8_ffn_layers_1_bias, ema_backbone_encoder_9_ln1_weight, ema_backbone_encoder_9_ln1_bias, ema_backbone_encoder_9_attn_qkv_weight, ema_backbone_encoder_9_attn_proj_weight, ema_backbone_encoder_9_attn_proj_bias, ema_backbone_encoder_9_ln2_weight, ema_backbone_encoder_9_ln2_bias, ema_backbone_encoder_9_ffn_layers_0_0_weight, ema_backbone_encoder_9_ffn_layers_0_0_bias, ema_backbone_encoder_9_ffn_layers_1_weight, ema_backbone_encoder_9_ffn_layers_1_bias, ema_backbone_encoder_10_ln1_weight, ema_backbone_encoder_10_ln1_bias, ema_backbone_encoder_10_attn_qkv_weight, ema_backbone_encoder_10_attn_proj_weight, ema_backbone_encoder_10_attn_proj_bias, ema_backbone_encoder_10_ln2_weight, ema_backbone_encoder_10_ln2_bias, ema_backbone_encoder_10_ffn_layers_0_0_weight, ema_backbone_encoder_10_ffn_layers_0_0_bias, ema_backbone_encoder_10_ffn_layers_1_weight, ema_backbone_encoder_10_ffn_layers_1_bias, ema_backbone_encoder_11_ln1_weight, ema_backbone_encoder_11_ln1_bias, ema_backbone_encoder_11_attn_qkv_weight, ema_backbone_encoder_11_attn_proj_weight, ema_backbone_encoder_11_attn_proj_bias, ema_backbone_encoder_11_ln2_weight, ema_backbone_encoder_11_ln2_bias, ema_backbone_encoder_11_ffn_layers_0_0_weight, ema_backbone_encoder_11_ffn_layers_0_0_bias, ema_backbone_encoder_11_ffn_layers_1_weight, ema_backbone_encoder_11_ffn_layers_1_bias, ema_backbone_encoder_12_ln1_weight, ema_backbone_encoder_12_ln1_bias, ema_backbone_encoder_12_attn_qkv_weight, ema_backbone_encoder_12_attn_proj_weight, ema_backbone_encoder_12_attn_proj_bias, ema_backbone_encoder_12_ln2_weight, ema_backbone_encoder_12_ln2_bias, ema_backbone_encoder_12_ffn_layers_0_0_weight, ema_backbone_encoder_12_ffn_layers_0_0_bias, ema_backbone_encoder_12_ffn_layers_1_weight, ema_backbone_encoder_12_ffn_layers_1_bias, ema_backbone_encoder_13_ln1_weight, ema_backbone_encoder_13_ln1_bias, ema_backbone_encoder_13_attn_qkv_weight, ema_backbone_encoder_13_attn_proj_weight, ema_backbone_encoder_13_attn_proj_bias, ema_backbone_encoder_13_ln2_weight, ema_backbone_encoder_13_ln2_bias, ema_backbone_encoder_13_ffn_layers_0_0_weight, ema_backbone_encoder_13_ffn_layers_0_0_bias, ema_backbone_encoder_13_ffn_layers_1_weight, ema_backbone_encoder_13_ffn_layers_1_bias, ema_backbone_norm_weight, ema_backbone_norm_bias, ema_head_layers_head_weight, ema_head_layers_head_bias
BebDong commented 2 years ago

Part of the printed result.pkl info:

{'accuracy_top-1': 0.9320000410079956, 'accuracy_top-5': 3.3960001468658447, 'class_scores': array([[0.00042726, 0.00068343, 0.0071601 , ..., 0.00039209, 0.00057438,
        0.00125795],
       [0.00038586, 0.00074933, 0.00519408, ..., 0.00047776, 0.00053544,
        0.00120644],
       [0.00067862, 0.0008437 , 0.00210273, ..., 0.00102068, 0.0007887 ,
        0.00138641],
       ...,
       [0.00043719, 0.00092473, 0.00403855, ..., 0.00063519, 0.0005136 ,
        0.00120862],
       [0.00047263, 0.00100704, 0.00258351, ..., 0.00073512, 0.0006326 ,
        0.00279792],
       [0.00056749, 0.00111329, 0.00241045, ..., 0.00080397, 0.00081281,
        0.00100305]], dtype=float32), 'pred_score': array([0.00787427, 0.00642915, 0.00520228, ..., 0.00558689, 0.00464927,
       0.00547685], dtype=float32), 'pred_label': array([  6, 802, 500, ..., 500, 619, 116]), 'pred_class': ['stingray', 'snowmobile', 'cliff dwelling', 'water snake', 'chiton, coat-of-mail shell, sea cradle, polyplacophore', 'stingray', 'diamondback, diamondback rattlesnake, Crotalus adamanteus', 'stingray', 'cliff dwelling', 'lampshade, lamp shade',
Ezra-Yu commented 2 years ago

Can you show your config? we will run to check.

BebDong commented 2 years ago

Thanks a lot. Config is as follows:

2022-03-23 10:29:51,329 - mmcls - INFO - Distributed training: True
2022-03-23 10:29:51,855 - mmcls - INFO - Config:
embed_dims = 384
num_classes = 1000
model = dict(
    type='ImageClassifier',
    backbone=dict(
        type='T2T_ViT',
        img_size=224,
        in_channels=3,
        embed_dims=384,
        t2t_cfg=dict(token_dims=64, use_performer=False),
        num_layers=14,
        layer_cfgs=dict(num_heads=6, feedforward_channels=1152),
        drop_path_rate=0.1,
        init_cfg=[
            dict(type='TruncNormal', layer='Linear', std=0.02),
            dict(type='Constant', layer='LayerNorm', val=1.0, bias=0.0)
        ]),
    neck=None,
    head=dict(
        type='VisionTransformerClsHead',
        num_classes=1000,
        in_channels=384,
        loss=dict(
            type='LabelSmoothLoss', label_smooth_val=0.1, mode='original'),
        topk=(1, 5),
        init_cfg=dict(type='TruncNormal', layer='Linear', std=0.02)),
    train_cfg=dict(augments=[
        dict(type='BatchMixup', alpha=0.8, prob=0.5, num_classes=1000),
        dict(type='BatchCutMix', alpha=1.0, prob=0.5, num_classes=1000)
    ]))
rand_increasing_policies = [
    dict(type='AutoContrast'),
    dict(type='Equalize'),
    dict(type='Invert'),
    dict(type='Rotate', magnitude_key='angle', magnitude_range=(0, 30)),
    dict(type='Posterize', magnitude_key='bits', magnitude_range=(4, 0)),
    dict(type='Solarize', magnitude_key='thr', magnitude_range=(256, 0)),
    dict(
        type='SolarizeAdd',
        magnitude_key='magnitude',
        magnitude_range=(0, 110)),
    dict(
        type='ColorTransform',
        magnitude_key='magnitude',
        magnitude_range=(0, 0.9)),
    dict(type='Contrast', magnitude_key='magnitude', magnitude_range=(0, 0.9)),
    dict(
        type='Brightness', magnitude_key='magnitude',
        magnitude_range=(0, 0.9)),
    dict(
        type='Sharpness', magnitude_key='magnitude', magnitude_range=(0, 0.9)),
    dict(
        type='Shear',
        magnitude_key='magnitude',
        magnitude_range=(0, 0.3),
        direction='horizontal'),
    dict(
        type='Shear',
        magnitude_key='magnitude',
        magnitude_range=(0, 0.3),
        direction='vertical'),
    dict(
        type='Translate',
        magnitude_key='magnitude',
        magnitude_range=(0, 0.45),
        direction='horizontal'),
    dict(
        type='Translate',
        magnitude_key='magnitude',
        magnitude_range=(0, 0.45),
        direction='vertical')
]
dataset_type = 'ImageNet'
img_norm_cfg = dict(
    mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True)
train_pipeline = [
    dict(type='LoadImageFromFile'),
    dict(
        type='RandomResizedCrop',
        size=224,
        backend='pillow',
        interpolation='bicubic'),
    dict(type='RandomFlip', flip_prob=0.5, direction='horizontal'),
    dict(
        type='RandAugment',
        policies=[
            dict(type='AutoContrast'),
            dict(type='Equalize'),
            dict(type='Invert'),
            dict(
                type='Rotate', magnitude_key='angle', magnitude_range=(0, 30)),
            dict(
                type='Posterize', magnitude_key='bits',
                magnitude_range=(4, 0)),
            dict(
                type='Solarize', magnitude_key='thr',
                magnitude_range=(256, 0)),
            dict(
                type='SolarizeAdd',
                magnitude_key='magnitude',
                magnitude_range=(0, 110)),
            dict(
                type='ColorTransform',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.9)),
            dict(
                type='Contrast',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.9)),
            dict(
                type='Brightness',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.9)),
            dict(
                type='Sharpness',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.9)),
            dict(
                type='Shear',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.3),
                direction='horizontal'),
            dict(
                type='Shear',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.3),
                direction='vertical'),
            dict(
                type='Translate',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.45),
                direction='horizontal'),
            dict(
                type='Translate',
                magnitude_key='magnitude',
                magnitude_range=(0, 0.45),
                direction='vertical')
        ],
        num_policies=2,
        total_level=10,
        magnitude_level=9,
        magnitude_std=0.5,
        hparams=dict(pad_val=[104, 116, 124], interpolation='bicubic')),
    dict(
        type='RandomErasing',
        erase_prob=0.25,
        mode='rand',
        min_area_ratio=0.02,
        max_area_ratio=0.3333333333333333,
        fill_color=[103.53, 116.28, 123.675],
        fill_std=[57.375, 57.12, 58.395]),
    dict(
        type='Normalize',
        mean=[123.675, 116.28, 103.53],
        std=[58.395, 57.12, 57.375],
        to_rgb=True),
    dict(type='ImageToTensor', keys=['img']),
    dict(type='ToTensor', keys=['gt_label']),
    dict(type='Collect', keys=['img', 'gt_label'])
]
test_pipeline = [
    dict(type='LoadImageFromFile'),
    dict(
        type='Resize',
        size=(248, -1),
        backend='pillow',
        interpolation='bicubic'),
    dict(type='CenterCrop', crop_size=224),
    dict(
        type='Normalize',
        mean=[123.675, 116.28, 103.53],
        std=[58.395, 57.12, 57.375],
        to_rgb=True),
    dict(type='ImageToTensor', keys=['img']),
    dict(type='Collect', keys=['img'])
]
data = dict(
    samples_per_gpu=64,
    workers_per_gpu=4,
    train=dict(
        type='ImageNet',
        data_prefix='/cache/data/imagenet/train',
        pipeline=[
            dict(type='LoadImageFromFile'),
            dict(
                type='RandomResizedCrop',
                size=224,
                backend='pillow',
                interpolation='bicubic'),
            dict(type='RandomFlip', flip_prob=0.5, direction='horizontal'),
            dict(
                type='RandAugment',
                policies=[
                    dict(type='AutoContrast'),
                    dict(type='Equalize'),
                    dict(type='Invert'),
                    dict(
                        type='Rotate',
                        magnitude_key='angle',
                        magnitude_range=(0, 30)),
                    dict(
                        type='Posterize',
                        magnitude_key='bits',
                        magnitude_range=(4, 0)),
                    dict(
                        type='Solarize',
                        magnitude_key='thr',
                        magnitude_range=(256, 0)),
                    dict(
                        type='SolarizeAdd',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 110)),
                    dict(
                        type='ColorTransform',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.9)),
                    dict(
                        type='Contrast',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.9)),
                    dict(
                        type='Brightness',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.9)),
                    dict(
                        type='Sharpness',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.9)),
                    dict(
                        type='Shear',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.3),
                        direction='horizontal'),
                    dict(
                        type='Shear',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.3),
                        direction='vertical'),
                    dict(
                        type='Translate',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.45),
                        direction='horizontal'),
                    dict(
                        type='Translate',
                        magnitude_key='magnitude',
                        magnitude_range=(0, 0.45),
                        direction='vertical')
                ],
                num_policies=2,
                total_level=10,
                magnitude_level=9,
                magnitude_std=0.5,
                hparams=dict(pad_val=[104, 116, 124],
                             interpolation='bicubic')),
            dict(
                type='RandomErasing',
                erase_prob=0.25,
                mode='rand',
                min_area_ratio=0.02,
                max_area_ratio=0.3333333333333333,
                fill_color=[103.53, 116.28, 123.675],
                fill_std=[57.375, 57.12, 58.395]),
            dict(
                type='Normalize',
                mean=[123.675, 116.28, 103.53],
                std=[58.395, 57.12, 57.375],
                to_rgb=True),
            dict(type='ImageToTensor', keys=['img']),
            dict(type='ToTensor', keys=['gt_label']),
            dict(type='Collect', keys=['img', 'gt_label'])
        ]),
    val=dict(
        type='ImageNet',
        data_prefix='/cache/data/imagenet/val',
        ann_file='/cache/data/imagenet/meta/val.txt',
        pipeline=[
            dict(type='LoadImageFromFile'),
            dict(
                type='Resize',
                size=(248, -1),
                backend='pillow',
                interpolation='bicubic'),
            dict(type='CenterCrop', crop_size=224),
            dict(
                type='Normalize',
                mean=[123.675, 116.28, 103.53],
                std=[58.395, 57.12, 57.375],
                to_rgb=True),
            dict(type='ImageToTensor', keys=['img']),
            dict(type='Collect', keys=['img'])
        ]),
    test=dict(
        type='ImageNet',
        data_prefix='/cache/data/imagenet/val',
        ann_file='/cache/data/imagenet/meta/val.txt',
        pipeline=[
            dict(type='LoadImageFromFile'),
            dict(
                type='Resize',
                size=(248, -1),
                backend='pillow',
                interpolation='bicubic'),
            dict(type='CenterCrop', crop_size=224),
            dict(
                type='Normalize',
                mean=[123.675, 116.28, 103.53],
                std=[58.395, 57.12, 57.375],
                to_rgb=True),
            dict(type='ImageToTensor', keys=['img']),
            dict(type='Collect', keys=['img'])
        ]))
evaluation = dict(interval=1, metric='accuracy', save_best='auto')
checkpoint_config = dict(interval=1)
log_config = dict(interval=100, hooks=[dict(type='TextLoggerHook')])
dist_params = dict(backend='nccl')
log_level = 'INFO'
load_from = None
resume_from = None
workflow = [('train', 1)]
paramwise_cfg = dict(
    norm_decay_mult=0.0,
    bias_decay_mult=0.0,
    custom_keys=dict(cls_token=dict(decay_mult=0.0)))
optimizer = dict(
    type='AdamW',
    lr=0.0005,
    weight_decay=0.05,
    paramwise_cfg=dict(
        norm_decay_mult=0.0,
        bias_decay_mult=0.0,
        custom_keys=dict(cls_token=dict(decay_mult=0.0))))
optimizer_config = dict(grad_clip=None)
lr_config = dict(
    policy='CosineAnnealingCooldown',
    min_lr=1e-05,
    cool_down_time=10,
    cool_down_ratio=0.1,
    by_epoch=True,
    warmup_by_epoch=True,
    warmup='linear',
    warmup_iters=10,
    warmup_ratio=1e-06)
custom_hooks = [dict(type='EMAHook', momentum=4e-05, priority='ABOVE_NORMAL')]
runner = dict(type='EpochBasedRunner', max_epochs=310)
work_dir = '/home/ma-user/modelarts/outputs/train-url_0/'
gpu_ids = range(0, 8)
mzr1996 commented 2 years ago

Well, our maximum accuracy is 100, instead of 1.0

BebDong commented 2 years ago

@mzr1996 Well then, it is reasonable. Thanks!

BebDong commented 2 years ago

But I noticed that when using tools/test.py to do evaluation the log showed The model and loaded state dict do not match exactly. Details:

unexpected key in source state_dict: ema_backbone_cls_token, ema_backbone_tokens_to_token_attention1_ln1_weight, ema_backbone_tokens_to_token_attention1_ln1_bias, ema_backbone_tokens_to_token_attention1_attn_qkv_weight, ema_backbone_tokens_to_token_attention1_attn_proj_weight, ema_backbone_tokens_to_token_attention1_attn_proj_bias, ema_backbone_tokens_to_token_attention1_ln2_weight, ema_backbone_tokens_to_token_attention1_ln2_bias, ema_backbone_tokens_to_token_attention1_ffn_layers_0_0_weight, ema_backbone_tokens_to_token_attention1_ffn_layers_0_0_bias, ema_backbone_tokens_to_token_attention1_ffn_layers_1_weight, ema_backbone_tokens_to_token_attention1_ffn_layers_1_bias, ema_backbone_tokens_to_token_attention2_ln1_weight, ema_backbone_tokens_to_token_attention2_ln1_bias, ema_backbone_tokens_to_token_attention2_attn_qkv_weight, ema_backbone_tokens_to_token_attention2_attn_proj_weight, ema_backbone_tokens_to_token_attention2_attn_proj_bias, ema_backbone_tokens_to_token_attention2_ln2_weight, ema_backbone_tokens_to_token_attention2_ln2_bias, ema_backbone_tokens_to_token_attention2_ffn_layers_0_0_weight, ema_backbone_tokens_to_token_attention2_ffn_layers_0_0_bias, ema_backbone_tokens_to_token_attention2_ffn_layers_1_weight, ema_backbone_tokens_to_token_attention2_ffn_layers_1_bias, ema_backbone_tokens_to_token_project_weight, ema_backbone_tokens_to_token_project_bias, ema_backbone_encoder_0_ln1_weight, ema_backbone_encoder_0_ln1_bias, ema_backbone_encoder_0_attn_qkv_weight, ema_backbone_encoder_0_attn_proj_weight, ema_backbone_encoder_0_attn_proj_bias, ema_backbone_encoder_0_ln2_weight, ema_backbone_encoder_0_ln2_bias, ema_backbone_encoder_0_ffn_layers_0_0_weight, ema_backbone_encoder_0_ffn_layers_0_0_bias, ema_backbone_encoder_0_ffn_layers_1_weight, ema_backbone_encoder_0_ffn_layers_1_bias, ema_backbone_encoder_1_ln1_weight, ema_backbone_encoder_1_ln1_bias, ema_backbone_encoder_1_attn_qkv_weight, ema_backbone_encoder_1_attn_proj_weight, ema_backbone_encoder_1_attn_proj_bias, ema_backbone_encoder_1_ln2_weight, ema_backbone_encoder_1_ln2_bias, ema_backbone_encoder_1_ffn_layers_0_0_weight, ema_backbone_encoder_1_ffn_layers_0_0_bias, ema_backbone_encoder_1_ffn_layers_1_weight, ema_backbone_encoder_1_ffn_layers_1_bias, ema_backbone_encoder_2_ln1_weight, ema_backbone_encoder_2_ln1_bias, ema_backbone_encoder_2_attn_qkv_weight, ema_backbone_encoder_2_attn_proj_weight, ema_backbone_encoder_2_attn_proj_bias, ema_backbone_encoder_2_ln2_weight, ema_backbone_encoder_2_ln2_bias, ema_backbone_encoder_2_ffn_layers_0_0_weight, ema_backbone_encoder_2_ffn_layers_0_0_bias, ema_backbone_encoder_2_ffn_layers_1_weight, ema_backbone_encoder_2_ffn_layers_1_bias, ema_backbone_encoder_3_ln1_weight, ema_backbone_encoder_3_ln1_bias, ema_backbone_encoder_3_attn_qkv_weight, ema_backbone_encoder_3_attn_proj_weight, ema_backbone_encoder_3_attn_proj_bias, ema_backbone_encoder_3_ln2_weight, ema_backbone_encoder_3_ln2_bias, ema_backbone_encoder_3_ffn_layers_0_0_weight, ema_backbone_encoder_3_ffn_layers_0_0_bias, ema_backbone_encoder_3_ffn_layers_1_weight, ema_backbone_encoder_3_ffn_layers_1_bias, ema_backbone_encoder_4_ln1_weight, ema_backbone_encoder_4_ln1_bias, ema_backbone_encoder_4_attn_qkv_weight, ema_backbone_encoder_4_attn_proj_weight, ema_backbone_encoder_4_attn_proj_bias, ema_backbone_encoder_4_ln2_weight, ema_backbone_encoder_4_ln2_bias, ema_backbone_encoder_4_ffn_layers_0_0_weight, ema_backbone_encoder_4_ffn_layers_0_0_bias, ema_backbone_encoder_4_ffn_layers_1_weight, ema_backbone_encoder_4_ffn_layers_1_bias, ema_backbone_encoder_5_ln1_weight, ema_backbone_encoder_5_ln1_bias, ema_backbone_encoder_5_attn_qkv_weight, ema_backbone_encoder_5_attn_proj_weight, ema_backbone_encoder_5_attn_proj_bias, ema_backbone_encoder_5_ln2_weight, ema_backbone_encoder_5_ln2_bias, ema_backbone_encoder_5_ffn_layers_0_0_weight, ema_backbone_encoder_5_ffn_layers_0_0_bias, ema_backbone_encoder_5_ffn_layers_1_weight, ema_backbone_encoder_5_ffn_layers_1_bias, ema_backbone_encoder_6_ln1_weight, ema_backbone_encoder_6_ln1_bias, ema_backbone_encoder_6_attn_qkv_weight, ema_backbone_encoder_6_attn_proj_weight, ema_backbone_encoder_6_attn_proj_bias, ema_backbone_encoder_6_ln2_weight, ema_backbone_encoder_6_ln2_bias, ema_backbone_encoder_6_ffn_layers_0_0_weight, ema_backbone_encoder_6_ffn_layers_0_0_bias, ema_backbone_encoder_6_ffn_layers_1_weight, ema_backbone_encoder_6_ffn_layers_1_bias, ema_backbone_encoder_7_ln1_weight, ema_backbone_encoder_7_ln1_bias, ema_backbone_encoder_7_attn_qkv_weight, ema_backbone_encoder_7_attn_proj_weight, ema_backbone_encoder_7_attn_proj_bias, ema_backbone_encoder_7_ln2_weight, ema_backbone_encoder_7_ln2_bias, ema_backbone_encoder_7_ffn_layers_0_0_weight, ema_backbone_encoder_7_ffn_layers_0_0_bias, ema_backbone_encoder_7_ffn_layers_1_weight, ema_backbone_encoder_7_ffn_layers_1_bias, ema_backbone_encoder_8_ln1_weight, ema_backbone_encoder_8_ln1_bias, ema_backbone_encoder_8_attn_qkv_weight, ema_backbone_encoder_8_attn_proj_weight, ema_backbone_encoder_8_attn_proj_bias, ema_backbone_encoder_8_ln2_weight, ema_backbone_encoder_8_ln2_bias, ema_backbone_encoder_8_ffn_layers_0_0_weight, ema_backbone_encoder_8_ffn_layers_0_0_bias, ema_backbone_encoder_8_ffn_layers_1_weight, ema_backbone_encoder_8_ffn_layers_1_bias, ema_backbone_encoder_9_ln1_weight, ema_backbone_encoder_9_ln1_bias, ema_backbone_encoder_9_attn_qkv_weight, ema_backbone_encoder_9_attn_proj_weight, ema_backbone_encoder_9_attn_proj_bias, ema_backbone_encoder_9_ln2_weight, ema_backbone_encoder_9_ln2_bias, ema_backbone_encoder_9_ffn_layers_0_0_weight, ema_backbone_encoder_9_ffn_layers_0_0_bias, ema_backbone_encoder_9_ffn_layers_1_weight, ema_backbone_encoder_9_ffn_layers_1_bias, ema_backbone_encoder_10_ln1_weight, ema_backbone_encoder_10_ln1_bias, ema_backbone_encoder_10_attn_qkv_weight, ema_backbone_encoder_10_attn_proj_weight, ema_backbone_encoder_10_attn_proj_bias, ema_backbone_encoder_10_ln2_weight, ema_backbone_encoder_10_ln2_bias, ema_backbone_encoder_10_ffn_layers_0_0_weight, ema_backbone_encoder_10_ffn_layers_0_0_bias, ema_backbone_encoder_10_ffn_layers_1_weight, ema_backbone_encoder_10_ffn_layers_1_bias, ema_backbone_encoder_11_ln1_weight, ema_backbone_encoder_11_ln1_bias, ema_backbone_encoder_11_attn_qkv_weight, ema_backbone_encoder_11_attn_proj_weight, ema_backbone_encoder_11_attn_proj_bias, ema_backbone_encoder_11_ln2_weight, ema_backbone_encoder_11_ln2_bias, ema_backbone_encoder_11_ffn_layers_0_0_weight, ema_backbone_encoder_11_ffn_layers_0_0_bias, ema_backbone_encoder_11_ffn_layers_1_weight, ema_backbone_encoder_11_ffn_layers_1_bias, ema_backbone_encoder_12_ln1_weight, ema_backbone_encoder_12_ln1_bias, ema_backbone_encoder_12_attn_qkv_weight, ema_backbone_encoder_12_attn_proj_weight, ema_backbone_encoder_12_attn_proj_bias, ema_backbone_encoder_12_ln2_weight, ema_backbone_encoder_12_ln2_bias, ema_backbone_encoder_12_ffn_layers_0_0_weight, ema_backbone_encoder_12_ffn_layers_0_0_bias, ema_backbone_encoder_12_ffn_layers_1_weight, ema_backbone_encoder_12_ffn_layers_1_bias, ema_backbone_encoder_13_ln1_weight, ema_backbone_encoder_13_ln1_bias, ema_backbone_encoder_13_attn_qkv_weight, ema_backbone_encoder_13_attn_proj_weight, ema_backbone_encoder_13_attn_proj_bias, ema_backbone_encoder_13_ln2_weight, ema_backbone_encoder_13_ln2_bias, ema_backbone_encoder_13_ffn_layers_0_0_weight, ema_backbone_encoder_13_ffn_layers_0_0_bias, ema_backbone_encoder_13_ffn_layers_1_weight, ema_backbone_encoder_13_ffn_layers_1_bias, ema_backbone_norm_weight, ema_backbone_norm_bias, ema_head_layers_head_weight, ema_head_layers_head_bias

It seems the saved state_dict dose not match the model. I can not figure it out temporarily.

mzr1996 commented 2 years ago

That's about the implementation of EMAHook, and these keys are used to resume EMA status. That warning can be ignored.