Open moonist opened 7 years ago
@moonist I am debugging and testing this code now, even it looks like working properly. What hardware you used for training and how many hours you used? According to the author of paper, they spend 50 hours for training with K40.
@wcy940418 I'am using a GTX1070 GPU card for nearly 10 hours, but the accuracy(not the printed edit distance) remains zero. When I use torch to train the same model, the accuracy is 91% after 20 hours, when using caffe, the accuracy will also be a positive value. Have you had a successful training? maybe a pretrained model will be helpful.
plus. I wonder if this helps network convergence, https://github.com/OlavHN/bnlstm.git, which adds batch normalization in lstm
@moonist I have the same situation.I'am using a GTX1080TI GPU for nearly 40hours,the accuracy is 0.000000,but the loss is nearly 0.001(the loss is 20 at the start).
Hi, I was confused by the following code: images = np.add(images, -128.0) Why does the image subtract 128?
@songwendong I think it is because the author wanted to remap 8bit gray scale image to float number [-1, 1]. I just followed the original implementation, but this code does not work at all.
Hi @wcy940418
你应该是中国人吧,可以说汉语吧?
我用了你的代码,卷积层自己重新定义的
lstm和ctc根据你的这个来做,发现完全不收敛。。loss起初17变成16后就不再下降了,
有没有什么建议呢。。还是哪里有问题
@liu6381810 你好,抱歉这个代码完全不工作。我是完全按照crnn的来做的,但是debug了很长时间完全不知道是为什么不工作。根据我的猜测,可能是因为有以下问题造成不工作的:
我用的pytorch, resnet 预训练的CNN加lstm,用ctc一样是loss到16左右就不再降了,毫无头绪。
I only changed the dataset, when training more than 30W steps, the network cannot convergence, the edit lengths are always large(almost 1). Any suggestions?