Open Nanayeb34 opened 3 months ago
Thanks for raising this, something to look into. Could you print out the data inputs that are fed to the LLM to better see the issue? On that note, also, I think not all of the entries have multi-turn answers
Hi @rasbt . Thanks for following up on this. Yes, not all the entries have multi turn answers. The last 30 entries in the dataset are the ones with multi turn answers according to the LIMA paper.
when you mention print out data inputs fed to the LLM, do you mean print out some samples when i am running the code below?
litgpt finetune lora \
--data LIMA \
--data.include_multiturn_conversations True
--checkpoint_dir "/content/lima"
It was stated to use the follow up questions in the multi turn dialogues for LIMA, you would have to set
--data.include_multiturn_conversations True
. I included that and compared it with the original data. It seems only the first instruction-response pair is selected. The follow up pairs are not included in the generated json.steps to reproduce the dataset creation
you can find the generated file here-new_lima_ds.json.
i am curious to know if the
--data.include_multiturn_conversations True
actually works and the expected output because i don't think it includes the follow up response-pairs.