Open utterances-bot opened 1 year ago
Really glad I came across your articles, John! I am just beginning my journey with Open LLMs. I have two questions. How could I use the finetuned model (after I train it with my own data) to make an interactive chatbot? And is there any evaluation method?
Glad you enjoyed it. As you can see from the notebook the model has been fine tuned to take in a prompt (or instruction) and respond so you could leverage that to have a back and forth dialogue. If you want to train the model to have more chatlike dialog where the prompts and the responses can refer further back into the context you could look at the open assistant dataset and finetune your model with an augmented version of that.
@johnrobinsn , Thanks for your reply, John. Appreciate it.
I just got a chance to read this and I really love this article. It was very helpful
Fantastic job, John, very clear and informative. Thanks a lot!
@johnrobinsn , Thanks for writing such a great site; I really enjoy it. Is it possible to finetune openlm-research/open_llama_3b in a similar way?or Could we consider togethercomputer/RedPajama-INCITE-Base-3B-v1 as openlm-research/open_llama_3b? Thank you.
yes training should be very similar for the openlm-research/open_llama_3b. They are two different models trained by different teams both models were initially trained on the redpajama dataset. Both should work well.
Finetuning Redpajama (OpenLlama)
Finetuning Redpajama
https://www.storminthecastle.com/posts/finetune_redpajama/