Hi,
"We fine-tune the 7B and 13B models with 80k and 18k conversations, respectively."
Could you provide more details about the training data? How the 80k data are prepared? Are they all with length of 16k?
Is the data used for training longchat-v1.5 the same as previous version?
Hi, "We fine-tune the 7B and 13B models with 80k and 18k conversations, respectively." Could you provide more details about the training data? How the 80k data are prepared? Are they all with length of 16k?
Is the data used for training longchat-v1.5 the same as previous version?