RyanSelesnik / AI-Toy

0 stars 0 forks source link

Fine-tuned response generation #13

Open racheledelstein opened 1 year ago

RyanSelesnik commented 1 year ago

Could potentially fine-tune on daily dialogues data or any data I guess. I did find this recent paper on fine-tuning DialoGPT with promotes which is supposed to generate "more novel and diverse responses with a higher BLEU score while training only a small portion of the DialoGPT/GPT-2 parameters". So this could save on training time. I'd be keen to work on this together

racheledelstein commented 1 year ago

That sounds good

RyanSelesnik commented 1 year ago

Another potentially useful thread here

racheledelstein commented 1 year ago

And here which they linked in the DialoGPT repo

RyanSelesnik commented 1 year ago

Oh cool yeah I've seen that. Though from what I've been reading now fine-tuning actually doesn't solve the problem since the "fine-tuning process robs [GPT2] of the ability to produce diverse responses, eventually reducing them to dull conversation partners.