Open chunhualiao opened 2 months ago
We need two more tables similar to Tables 1 and 2: fine-tune (instruct models C ) using our paired dataset to get model D.
The new two tables will show model D's improvements over model C (instruct models).
When using instruct models fine-tuned with our data, the prompt asking for translation can have more instructions to specify using the markdown syntax to enclosed output code, avoid natural text explanation etc. So the postprocessing may be easier.
Please do the experiments and add the two tables into the paper. Thanks. @bin123apple
I think the argument that "fine-tuning instruct models will get worse translation performance" is settled. Your existing experiments show that "deepseek-coder-33b-instruct shows the greatest improvement". Please confirm this so we do not need to debate this anymore. @bin123apple
Three models
And in the overleaf, I compared the performance of these instruct-tuned models (model C) with the model fine-tuned with our dataset (model B) in the previous version.