summerlvsong / Aggregation-Cross-Entropy

Aggregation Cross-Entropy for Sequence Recognition. CVPR 2019.
303 stars 60 forks source link

I would like to ask you how to accurately predict the character order of a word. #8

Open Meicsu199345 opened 5 years ago

Meicsu199345 commented 5 years ago

I recreated your project and found that the input GT was converted into a word list, which had lost its order, and your prediction only provided the number of characters. Only through the two-dimensional matrix position of the network output can barely judge the order, I would like to ask you how to accurately predict the character order of a word.

lamhoangtung commented 5 years ago

+1 upvote

summerlvsong commented 5 years ago

As mentioned in the paper, to decode the 2D prediction, we flattened the 2D prediction by concatenating each column in order from left to right and top to bottom and then decoded the flattened 1D prediction following the general procedure.

chenjun2hao commented 5 years ago

@summerlvsong ,the target label don't need a fixed order?

chenjun2hao commented 5 years ago

and i am confused about the 2D example, the label texts don't have a fixed order. if so, how to solve the 1D problem. waiting your reply.

summerlvsong commented 5 years ago

During training, we don't need a fixed order for supervision. When testing, for the 2D scene text recogntion problem, we use the hypothesis that character distribute form left to right in the 2D output. Therefore, we can decode the 2D prediction by flattening the 2D prediction by concatenating each column in order from left to right and top to bottom and then decoding the flattened 1D prediction following the general procedure.

yinghuozijin commented 4 years ago

@summerlvsong Thanks for your reply.