Closed soheilesm closed 5 years ago
options["is_debugging"] is always false, I think it is only for debugging, not for real experiment, at training time, options["is_predicting"] = False, options["model_selection"] = Flase, at tuning time, options["is_predicting"] = True, options["model_selection"] = True, options["model_selection"] = True it means that only use valid set to choose the model.
Shall I run tuning.sh to get the rouge scores? I do so, but I get error on perl /home/pijili/tools/ROUGE-1.5.5/ROUGE-1.5.5.pl -n 4 -w 1.2 -m -2 4 -u -c 95 -r 1000 -f A -p 0.5 -t 0 myROUGE_Config.xml C
As the original GitHub repo that contains ROUGE-1.5 is not active anymore - >https://github.com/andersjo/pyrouge/tree/master/tools/ROUGE-1.5.5, can you put a wrapper for this so that we can calculate the rouge scores by just running the tunning.sh file?
Thanks. I will integrate the pyrouge package later.
Just to make sure I am doing things the right way: So I trained the model with options["is_debugging"] = False options["is_predicting"] = False options["model_selection"] = True in the main.py, and got some models saved in the model folder.
Then for creating the validation summaries and running decoder, I need to set options["is_debugging"] = True options["is_predicting"] = True options["model_selection"] = False in the main.py and it will run for the saved models. However, doing that I only get 13 files generated as the summary and ground_truth pairs, shouldn't it be more?
when options["is_debugging"] is true: only use several (16) samples to conduct the training.
Training: options["is_debugging"] = False options["is_predicting"] = False options["model_selection"] = False
Validation on a subset: options["is_debugging"] = False options["is_predicting"] = True options["model_selection"] = True
Evaluation on test set: options["is_debugging"] = False options["is_predicting"] = True options["model_selection"] = False
Could you also please mention what combination of parameters (i.e. coverage, copy) you ran and for how many epochs and you got your reported results?
coverage=true, copy=true Number of epochs: https://github.com/lipiji/TranSummar/blob/master/README.md
Shall I run tuning.sh to get the rouge scores? I do so, but I get error on perl /home/pijili/tools/ROUGE-1.5.5/ROUGE-1.5.5.pl -n 4 -w 1.2 -m -2 4 -u -c 95 -r 1000 -f A -p 0.5 -t 0 myROUGE_Config.xml C
As the original GitHub repo that contains ROUGE-1.5 is not active anymore - >https://github.com/andersjo/pyrouge/tree/master/tools/ROUGE-1.5.5, can you put a wrapper for this so that we can calculate the rouge scores by just running the tunning.sh file?
Just to make sure I am doing things the right way: So I trained the model with options["is_debugging"] = False options["is_predicting"] = False options["model_selection"] = True in the main.py, and got some models saved in the model folder.
Then for creating the validation summaries and running decoder, I need to set options["is_debugging"] = True options["is_predicting"] = True options["model_selection"] = False in the main.py and it will run for the saved models. However, doing that I only get 13 files generated as the summary and ground_truth pairs, shouldn't it be more?
Could you also please mention what combination of parameters (i.e. coverage, copy) you ran and for how many epochs and you got your reported results?