Megvii-BaseDetection / DenseTeacher

DenseTeacher: Dense Pseudo-Label for Semi-supervised Object Detection
Apache License 2.0
119 stars 12 forks source link

Need some help for experiments on other dataset. #15

Closed heiyuxiaokai closed 2 years ago

heiyuxiaokai commented 2 years ago

I change some configs and use other dataset. The size of my dataset is relatively small. So I reduce the total iterations to 10k. For test results, the mAP of denseteacher is much lower than corresponding supervised model(relative to the coco dataset). There may be several factors:

  1. The total iterations are too small for model convergence.
  2. The size of dataset is too small.
  3. The strategy of unlabel image sampling affects a lot.

I do not kown which is the main factor. Can you give me some advice? Here are some logs.

ZRandomize commented 2 years ago

thanks, we do have some experience on data scaling. Firstly, if your sup dataset is small and training time is short, consider using smaller burn_in_step so the teacher shouldn't overfit when SSL begin. If you want to check whether training has converged, check if the mAP is stable, a converged mAP curve is like: image

Secondly, the TRAINER.DISTILL.RATIO should be changed for different dataset, this hyper-param is sensitive to dataset as introduced in our paper, and its optimal value can be estimated using dataset. Besides, we don't see the effect of different unsup data sampling on coco, it won't matter if you have enough unsup data.

heiyuxiaokai commented 2 years ago

Thanks for your sincere advice~ I tried several seeds of sampling sets and got some normal results(similar with experiments on coco). It seems my small dataset and the imbalance category distribution lead extreme sampling sets. For example, results appear normal with seed5 and appear strange with seed1.