microsoft / CodeXGLUE

CodeXGLUE
MIT License
1.5k stars 363 forks source link

Text-Code Reproduction #145

Open MT010104 opened 1 year ago

MT010104 commented 1 year ago

Hi! Thanks for the great work. 1668151645125 Compared to the results of the model finetuned(60000 steps) by myself, the results above are much better. I wanna know the epochs of your finetuning and does "GPT-2" refers to "gpt2" on Huggingface which is not pretrained by code? Thanks in advance.

celbree commented 1 year ago

Yes. "GPT-2" is NOT pre-trained on code. As for the different results, the hyper-parameters in this repo are used in fine-tuning CodeGPT. For GPT-2, you may try to finetune with more steps and select the best checkpoint with highest BLEU score in dev set. It's not always the latest checkpoint that performs the best in this task.

MT010104 commented 1 year ago

Thanks for your reply.