reproducibility-challenge / iclr_2019

ICLR Reproducibility Challenge 2019
https://reproducibility-challenge.github.io/iclr_2019/
219 stars 40 forks source link

Submission for issue #47 #143

Open lcalem opened 5 years ago

lcalem commented 5 years ago

47

reproducibility-org commented 5 years ago

Hi, please find below a review submitted by one of the reviewers:

Score: 6 Reviewer 1 comment : This work tries to reproduce soft-Q learning with MI regularization and compare results with entropy regularization. Although the results in reproduction looks much worse than the original paper, the reproduction details are stated clear in the report. A major problem is in the report, which is less-organized and contains uninformative paragraphs. For example, (1) you don't need to introduce the importance of reproduction everywhere, and (2) I would suggest to include reproduction summarization in abstract, instead of introducing the challenge. Hyper-parameters are less-searched either. Even though, I would rate a borderline score as an encouragement to participate this reproduction challenge. Confidence : 3

reproducibility-org commented 5 years ago

Hi, please find below a review submitted by one of the reviewers:

Score: 5 Reviewer 2 comment : Context: choosing a proper baseline proved difficult and time consuming, and the report authors ended up implementing everything from scratch. After clarifying all the missing pieces of information and fixing some numerical stability problems, the authors did not have the time to launch the proper experiments to reproduce either the tabular setting or the Atari one.


Conclusion: given that the needed experiments could not be conducted due to lack of time, little conclusions about the reproducibility of the original article can be drawn.


Reviewers are asked to go through a list of items to evaluate the report. Here I include my list with some comments that may help understand how I perceived the information in the report as well as its specific strengths and weaknesses:

reproducibility-org commented 5 years ago

Hi, please find below a review submitted by one of the reviewers:

Score: 7 Reviewer 3 comment : This work tries to reproduce the paper "Soft Q-learning with mutual information regularization" and is able to try out some of the experiments of the paper. I feel it's a good initial work and future work could build upon the submitted report to test the reproducibility of the original paper. However, this work is not complete in itself and I feel there are certain parts where the authors of the report could improve upon to make this submission more useful to the community in general. The authors could describe the paper and the associated reproducible aspects of it in the abstract as opposed to describing the challenge. They do a good job in introducing the problem and providing a brief summary of all the key concepts and theoretical contributions of the paper (including the blog post). The effort of the authors to write the code from scratch is absolutely commendable. It would be great if the authors could have communicated with the authors via the OpenReview forum. I feel the authors could improve upon the hyperparameter search done. Although I feel time was a considerable bottleneck, it would have been helpful to observe how the MIRL performance varies with different hyperparameter values. Had the authors shown such a dependence and commented on possible reasons for the same, it would have given better insights into the reproducibility aspect of the algorithm as well as its robustness in other applications. The authors could have also done an ablation study by keeping beta constant. This could have helped them study the performance without running into numerical overflows. The authors describe various roadblocks faced in the result reproduction pipeline. From the report, I could gather that the results could not be reproduced and the authors believe the major differences lie in hyperparameter values and minor implementation tweaks. It would be more helpful if they could layout a formal set of guidelines or recommendations for the authors of the original paper to improve reproducibility. Overall, I would like to appreciate the efforts of the authors in participating and implementing the paper. However, the report could have been more organized to clarify their contributions and highlight necessary tweaks to aid future work.
Confidence : 3