Open hanskrupakar opened 8 years ago
I can't reproduce this on the latest revision.
I didn't make any changes except specify the epoch to start the loading from. I have attached a log file specifying the train and load from commands, which remain the same except me specifying the load from file.
Something is not right. According to your log file, you always run the same command:
th train.lua -data_file data/demo-train.hdf5 -val_data_file data/demo-val.hdf5 -savefile demo-model
Is it the case?
I ran it again from the beginning again after you said it was strange. Attached is the log file for that. Also attached is the train.lua
and preprocess.py
I used.
preprocess.py.docx
train.lua.docx
error.txt
It seems that AdaGrad does not play nicely with the train_from
option at the moment. I would advise you to stick with the default SGD which works well.
Also, please don't set your option within the code. It is error prone and harder for whoever might assist you to know what you are doing.
Will remember not to inline changes from now.
I implemented SGD and the train_from
works as expected.
Thanks.
I am trying to resume training from checkpoint file and even though it says loaded model, the perplexity restarts at weight initialization level and the accuracy of translation when I use evaluate.lua also seems to indicate that the model is simply reinitializing the vectors instead of loading from checkpoint.
Is this an issue with the API? What am I doing wrong?