Thank you so much for this nice repo for running finetuning.
I have one question and did not find a better way to communicate, so feel free to answer and then close this issue.
In the following code, input_ids and labels are the same for supervised fine tuning.
Is there somewhere in the model training parameter that knows it is a causal LM training, so it will shift the labels by one, so that input_ids and labels become a next token prediction task?
Hi @loubnabnl
Thank you so much for this nice repo for running finetuning.
I have one question and did not find a better way to communicate, so feel free to answer and then close this issue.
In the following code, input_ids and labels are the same for supervised fine tuning. Is there somewhere in the model training parameter that knows it is a causal LM training, so it will shift the labels by one, so that input_ids and labels become a next token prediction task?