Closed perprit closed 5 years ago
Your lambdas of epoch 11 is [0.1, 0.0, 0.0]
, in that case, your generated sequence totally depends on your model during training, so it does make sense that your model will collapse.
Our initial lambdas is [0.04, 0.96, 0.0]
(here). Your setting of lambdas is different from ours.
Hi, thanks for the comment.
I set the lambda as [0.04, 0.06, 0.0]
as the README says.
Sorry for that I didn't understand what the initial lambda values mean when I first ran this code, which makes me not notice a trivial error like this..
I think the README needs to be fixed anyway.
Oh..I'm sorry for that. It's my fault. I will fix the typo soon.
Thank you very much for pointing this out.
Hi, thanks for the great work. I've tried training an NMT model on IWSLT 14 with interpolation algorithm, (https://github.com/asyml/texar/tree/master/examples/seq2seq_exposure_bias) but while training I found that the BLEU suddenly dropped to 0.0000 at about 11 epoch.
Below is the training log I ran into:
And the test_results10.txt is like:
And the test_results11.txt (when the BLEU dropped) is like:
I guess it's something to do with the lambda value that changed, but I have no idea right now. I've only modified configs to set batch_size as 32 (from 64), and using python v3.5.2 with tensorflow-gpu v1.8.0. Could you guess any reason why? Thanks.