Thank you for maintaining the benchmarks. Currently, I am evaluating models with Japanese MT-Bench, and I need the prompts “single-v1-multi-turn” and “single-math-v1-multi-turn” to execute the function 'make_judge_simgle' in gen_judgement.py, but I can't find them. There seems to be only the prompts “single-v1” and “single-math-v1” in fastchat/llm_judge/data/judge_ja_prompts.jsonl. I would appreciate it if you could tell me where to find them or how to evaluate them.
Thank you for maintaining the benchmarks. Currently, I am evaluating models with Japanese MT-Bench, and I need the prompts “single-v1-multi-turn” and “single-math-v1-multi-turn” to execute the function 'make_judge_simgle' in gen_judgement.py, but I can't find them. There seems to be only the prompts “single-v1” and “single-math-v1” in fastchat/llm_judge/data/judge_ja_prompts.jsonl. I would appreciate it if you could tell me where to find them or how to evaluate them.