ku21fan / STR-Fewer-Labels

Scene Text Recognition (STR) methods trained with fewer real labels (CVPR 2021)
MIT License
173 stars 27 forks source link

About experiment with pseudo label #17

Closed KhaLee2307 closed 1 year ago

KhaLee2307 commented 1 year ago

Thank you for your gorgeous research. I have some questions about your experiment (focus on Semi-Supervised Learning):

  1. When you use pseudo labels to retrain your model, do you select all of them or only choose data having high confidence scores?
  2. The model you used for the pseudo labels step was fixed?
  3. You tried to use unlabeled real-world data for Semi-Supervised Learning on Synthetic data and fall. Can you provide me with some detail about this experiment (setting, hyperparameter, protocol)?
  4. Finally, I will be very grateful if you give me some of your comments about the failed experiment mentioned in the question above.

From the bottom of my heart, I'd be grateful if I could get a response from you. It will help my project a lot.

KhaLee2307 commented 1 year ago

Additionally, one more question, when semi-supervised learning, did you freeze any layer?

Thank you.

ku21fan commented 1 year ago

Hi, Thank you for your interest in our work :)

  1. We do not choose data having high confidence scores. We do slight filtering here though. After this filtering, we used all of them. https://github.com/ku21fan/STR-Fewer-Labels/blob/e6aa817e2eacbf29b3fcd11390d78b1a8f96bf78/modules/semi_supervised.py#L54-L60

  2. Yes, it is fixed.

  3. and 4. Do you mean the fine-tuning experiments in Table 5? If so, it was not semi-supervised learning on synthetic data. For the experiments, after pretraining a model on synthetic data, we fine-tuned the model on real data. Thus semi-supervised learning is conducted with real data.

Additional question: No freeze. Although we made the freeze option here, we did not use it for our paper.

Best,