Open huang229 opened 4 years ago
In addition, before getting the precision here, I used my own environment, so the code was modified as follows: pytorch1.2.0 syncbatchnorm----->nn.batchnorm batch_size =8 gpu_nums =1,Only one GPU train_data_szie =640*480 I worked for half a month. No matter how I train, the highest accuracy is only 70.0%. Therefore, I would like to remind you to deploy an environment similar to that of the author as much as possible to ensure that the downloaded code can be trained directly without any modification. After getting the author's training accuracy, then gradually change each influencing factor, get the accuracy mentioned by the author, and then slowly change towards their own environment. Only in this way can we know which factors cause the accuracy can not be reproduced.
Thanks for verifying!! I am happy that you can train your model well.
evaluating the model ... setup and restore model compute the mIOU 100%|█████████████████████████████████████████| 250/250 [17:49<00:00, 4.28s/it] mIOU is: 0.780227 mIOU = 0.7802269045024711 (pytorch)BiSeNet_syncbn$ CUDA_VISIBLE_DEVICES=0 python -m torch.distributed.launch --npr oc_per_node=1 train.py 100%|█████████████████████████████████████████| 250/250 [03:28<00:00, 1.20it/s] mIOU = 0.7599654703378823
Hi, how long did your training process take with batch_size=6? It appears to more than 2 days with batch size=16, and my gpu_nums=4 (2080Ti). Is it normal?
Although this is not the first time for me to hang out with the author, I would like to thank the author again for the code. I've almost reproduced the accuracy of the open source code here. Accuracy of single scale test: 76.17% Multi scale accuracy test: 77.90% It's very easy to get this precision. I downloaded the code directly here, and then the training environment is similar to the one mentioned in the author's code. That is, I can run the code directly without any modification. Be sure to remember to train directly without any modification. To add up, the capacity of my GPU graphics card is still a little small. Each GPU's batch_size is 6, and the sum of the two graphics cards is 12.So it's normal that the accuracy here is a little bit poor.