Closed aoezis closed 1 month ago
I'd like to train a similar router for a language other than English, And I'd like to know the exact format and where the data came from. Is it like 80k samples from chatbot-arena and 110k from nectar-gpt-4?
Hi there, you are correct! The GPT-4 augmented MF router was trained using ~60k battles from Chatbot Arena and ~110k from Nectar using a GPT-4 judge.
Chatbot Arena data is close to: https://huggingface.co/datasets/routellm/lmsys-arena-human-preference-55k-thresholds
Nectar dataset: https://huggingface.co/datasets/routellm/gpt4_judge_battles
Alright, got it. thanks a lot.
Hello,
I have not clearly understood the format and source of the datasets used to train these routers. It's said to be published in huggingface. but for example I can't find the dataset that is used to train: routellm/mf_gpt4_augmented. As I understood from the code: train_matrix_factorization.py there has to be a json dataset with keys idx, model_a, model_b, winner. But there is no such dataset in the huggingface. Could you clarify the format and the creation of the dataset that is used for mf training?