Open Jakkque opened 5 years ago
Hi,
Sorry for this late reply. I guess it is because the training has already failed. The perplexity has already become -0.004 , which is a really large number. Maybe you can print the results and see if the outputs are apparently fake.
During the training, I have to trace if the values of "perp", "D-loss", etc. are in a good way. If not, the training should be early stopped. Or I have to start training a new model.
If you want to train longer, the hyper-parameters might have to be fine-tuned and the initial random seed would also influence the results a lot.
Hey,
I try to use your algorithm for my own data. First, however, I try to get used to your code by training the network with the counting task, i.e.
Unfortunately, I get an error while training. When I restart the training after the error, sometimes it works for another few steps, then the error occurs once again, e.g. after step
41600
, I got this message:The
argmax
results are still not convincing, so I think it should train longer. Did you ever had this issue and know how to solve it?Cheers!