Closed LRParser closed 6 years ago
Hi @LRParser, Thanks for reporting this. I think this is the natural of neural network works, which largely depend on the initialization of the parameters in RNN. The loss increases mean that learning rate should be changed to be smaller(code will be updated soon). If you want to obtain fast convergence, please reduce the hidden-dimensions of GRU(e.g. reduce 512 to 256).
I have changed the code of train_RNN.py, please re-run: python train_RNN.py. I think you can obtain very good accuracy and smaller loss.
Thanks for a great paper and project! Just a quick question - is it expected to see the validation loss increasing during training? Here's an excerpt of my train_RNN.py run:
I would have expected validation loss to go down during training, but perhaps I'm not letting training run long enough? Any guidance here is appreciated.