Open fillassuncao opened 1 year ago
Hi there, <extra_id_1>
is also a special token and here this is a kind of unexpected output. We would suggest to use codet5p-220m
and codet5p-770m
models in the finetuning setting.
For zero-shot setting, some truncation strategies should be used to obtain your desired output. This is because that the model did not see such exact input and output pair during pretraining and was difficult for it to learn to when to stop the generation. For instance in HumanEval evalution, it is a common practice to use truncation based on some stop tokens to generate a desired clean output.
Given the code bellow:
The output I get is:
I was not expecting to get multiple
<extra_id_1>
tokens. Is this known or expected?