Closed felixdittrich92 closed 2 years ago
Hello @felixdittrich92. Thanks for your interest and initiative for integration.
The results for ViTSTR for 128x32 images are in the paper. arXiv link in the README. :)
Yes, all models were trained from scratch using the same exact datasets and training pipeline and strategy (in contrast to the original ViTSTR which used DeiT weights for initialization).
@baudm Thanks 👍
Hi @baudm 👋 ,
thanks a lot for this great repository 👍 I saw you have run some experiments with ViTSTR and an input of 32x128.
Could you maybe share the results ? :) And do you run the experiment from scratch ?
I have planned to integrate both ViTSTR and ParSeq in https://github.com/mindee/doctr in the next time 👍
Best regards