tianyi-lab / Cherry_LLM

[NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other models
306 stars 21 forks source link

'The training of pre-experienced models is discarded for more efficient usage': that means we can only use base model to do cherry analysis and selection? #15

Closed Labmem009 closed 11 months ago

Labmem009 commented 11 months ago

I noticed your update in README. Did that mean we can only use base model to do cherry analysis and selection? Directly use base model or use pre-experienced model, which performance is better?

MingLiiii commented 11 months ago

Thanks for your interest in our work.

For the first question, it would be yes. It is fine to do cherry analysis and selection directly on base Llama2 models. For the second question, I would say that from the Research Viewpoint, using pre-experienced models is more reasonable and performs better. However from the Real-world Implementation Viewpoint, directly using the base model is more easy and efficient. It's kind of a tradeoff.