Open XuperX opened 1 year ago
This paper presents a language model framework for code and comment inconsistency evaluation and correction. It utilized cutting-edge NLP technologies and provided working implementations for code improvements, which is a good fit for the EMNLP system demonstration track.
I am still concerned about whether the model resolution is sufficient to explain code-comment mismatch generated during different iterations. For instance, when I tested it on functions related to JSON file validation, the results largely depended on whether the word “validation” was there or not. But when the basic structure of the code doesn’t change much between versions, can the model handle such differences?
About the methods:
I tested its performance in python. If comments are marked with # and span multiple lines, the model only recognizes the first line as a valid comment. Seems to be a small bug.
Rule of Thumb
If I were the author, would I find the review helpful
My template
From sharky6000
For the love of everything that is good in this world, read the rebuttals! Participate in the discussion. Update your reviews post discussion. This varies from conference to conference, but in my experience, probably 35% of reviewers do not react at all after submitting the review and I wonder how they keep getting asked to review.
From Mateusz Buda
Brief summary
Structure
Novelty
Criticism of methodology
Criticism of results
Strong points
Conclusion
Examples from ACL