Closed Zigars closed 1 year ago
I switch to main branch, and this problem review again, also I change the backbone to regnetx, it doesn't work either, train step was still stucking in 12000 iter, which the default starting vos iter.
hi, @Zigars , thanks for the interest in our codebase. not sure about your machine, i think the speed of vos will be slower than the vanilla training, but it will not stuck.
could you double check whether it is caused by the slower training speed?
@Zigars , hi! do you figure out this issue? i actually encountered the same question when I try to reproduce my results. is it related to the version of detectron2 and torch, or cuda version?
never mind, i found that we need to use torch 1.7.1 will work.
Using main-MLP branch, and training command:
python train_net_gmm.py --dataset-dir /data/haitaozhang/dataset/object_detection/VOC --num-gpus 4 --config-file VOC-Detection/faster-rcnn/vos.yaml --random-seed 0 --resume
it's stucking in 12000 iter, and gpus usage do not work.
how can I solve this problem, maybe need use the main branch?