qhduan / just_another_seq2seq

Just another seq2seq repo
329 stars 97 forks source link

loss sometimes jumpes to 200 #11

Open yzho0907 opened 6 years ago

yzho0907 commented 6 years ago

my params are { "cell_type": "lstm", "depth": 2, "attention_type": "Luong", "bidirectional": true, "use_residual": true, "use_dropout": false, "time_major": true, "hidden_units": 1024, "optimizer": "adam", "learning_rate": 0.001 } and i am using fasttext pre-trained word2vec as well batch size is 128 which i haven't changed it. But the loss is always fluctuating and sometimes jumps to a huge number (200), it usually started as 10-20. It's cuz this version of seq2seq add a part of reinforcement learning? plz tell u how to fix it?

yzho0907 commented 6 years ago

我把词向量去掉,把参数调到默认,loss有时候也会跳到一个很大的数字,请问下怎么解决啊?训练出来的对话也完全不对。

yzho0907 commented 6 years ago

训练轮数增加后loss就会变的很巨大,请问有什么办法可以解决嘛?