Open ZZZYYYLL opened 1 week ago
Thanks for the interest! Here's an example: https://github.com/orionw/FollowIR/issues/5#issuecomment-2330372098
Your format looks correct offhand, but I would probably do a diff to be certain. EDIT: ah I think you're adding the Mistral tokens, but llama factory does that with the --template
flag. There are probably a few other small differences like that between the two.
Hi, thanks for the great work. I have a question for how to transforming the training dataset to fit llama_factory format
I'd like to ask for advice on how to properly construct the training data format for llama_factory fine-tuning. I found FollowIR-7B's training set on huggingface, and the format is as follows:
For fitting the llama_factory 's format, Should the format I build for fine-tuning look like this:
I will appreciate it if you can give me an example for it.