Hi, thanks for your good job! I have some doubts about the training labels.
The preprocess function in toolbench/train/train.py just copies the input_ids as the target and then masks the target. However, it seems that there is no position shift operation for target as the LLM should perform next token prediction.
Hi, thanks for your good job! I have some doubts about the training labels. The
preprocess
function in toolbench/train/train.py just copies the input_ids as the target and then masks the target. However, it seems that there is no position shift operation for target as the LLM should perform next token prediction.Can you help me explain the reasons or there are any details I might have overlooked? Thank you very much.