Closed chaiyixuan closed 6 years ago
很抱歉, 我自己執行時沒有這個錯誤 不知道您在執行時版本是否正確呢?
之前Adam由于版本原因无法使用,我改用GradientDecent训练,导致无法收敛。后来做代码兼容性调整,还是用Adam训练,模型loss收敛到7.8左右。(我用batch80 和epoch50)但是test得出的结果好像不太好,您知道是什么原因么?
附结果: question => Have you heard about 'machine learning and having it deep and structured'? generated_sentence => President's poem. question => How are you? generated_sentence => Street cemetery doe. question => What's your name? generated_sentence => Scum baltimore result casino. Casino casino doyle. question => Hello generated_sentence => Scum baltimore result landed. Street couldn't zira ruin ha. question => Thank God! If I had to hear one more story about your coiffure generated_sentence => End. question => You never wanted to go out with me, did you?
可以試試看 RMSprop GradientDecent 收斂太慢。
您好,我重新训练SEQ2SEQ模型后用 text.py 但是得到以下结果: Concentration concentration concentration concentration planning priest. Concentration concentration concentration concentration planning concentration planning priest. Concentration concentration concentration concentration planning priest. Concentration concentration concentration concentration planning concentration planning priest. Concentration concentration concentration concentration planning priest. Concentration concentration concentration concentration planning concentration planning priest. 您知道是什么原因么?