DachengLi1 / LongChat

Official repository for LongChat and LongEval
Apache License 2.0
504 stars 29 forks source link

dummy conversations seem to be short #40

Closed Arist12 closed 1 year ago

Arist12 commented 1 year ago

It seems like dummy_converstions in the repository is far shorter than 16k, I'm wondering how are these short conversation data able to fine-tune 16k long context models?

DachengLi1 commented 1 year ago

@Arist12 Dummy data is not for training; rather for testing whether you setup the pipeline correctly. In actual training, you will put long conversations, codes, etc.

Arist12 commented 1 year ago

Thank you!