opconty / Transformer_STR

PyTorch implementation of my new method for Scene Text Recognition (STR) based on Transformer,Equipped with Transformer, this method outperforms the best model of the aforementioned deep-text-recognition-benchmark by 7.6% on CUTE80.
https://mp.weixin.qq.com/s/a_ahIwxiCaO7Bxmj81HUTw
224 stars 53 forks source link

Why do you use "self.seq" (which means two Bilstm) before transeformer in model/__init.py ? #4

Open yoyoshuang opened 4 years ago

yoyoshuang commented 4 years ago

Thanks for you great job.

I am not very understand this part "Why do you use "self.seq" (which means two Bilstm) before transeformer in model/__init.py ?"

What is the net learning in this part?

Would you please help me to understand this part?

And in my work , I have try the same way to combine the CNN's feature map to the transeformer part , but the results looks like accoured something wrong, after adding this "self.seq" part, the work seems right , I can't explain it .

Thank you for you helping!!!

opconty commented 4 years ago

Hi, Thanks for your attention, according to the paper https://arxiv.org/abs/1904.01906,

 self.seq = nn.Sequential(BiLSTM(output_channel, hidden_size, hidden_size),
                                 BiLSTM(hidden_size, hidden_size, hidden_size))

is the sequence module, bilstm is not necessary and you can remove it,