MichalBusta / DeepTextSpotter

285 stars 101 forks source link

how to pretrain? #63

Open mowkee opened 6 years ago

mowkee commented 6 years ago

@MichalBusta Hope this message finds you well.

I have been reading your PhD thesis and you have mentioned about pre-training your models using SynthText and Synthetic Word datasets. There's no script for performing that. I can only see training in train.py. Is it necessary to pre-train further? does it make any difference?

Please help.

MichalBusta commented 6 years ago

Hi, we have trained modes from scratch. It is about the speed. (even in cpp implementation pooling of proposals is slow - GPU is just waiting for job).

For the fine tuning OCR network the optimizer state is important. (since provided script creates small batches - expected thing is that it will destroy even "good" model in first few iterations)

if you are interested in simple working solution, just train 2 stand alone networks - the extra augmentation with inperfect proposals give you just 2-5% on ICDAR tasks (good for competition mode, wasting of resources otherwise).