Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110). This framework is also used to evaluate text-to-image models in Holistic Evaluation of Text-to-Image Models (HEIM) (https://arxiv.org/abs/2311.04287).
I have been following this project and was very impressed by the comprehensive framework it provides for benchmarking language models. Our team is eager to submit our model to the HELM leaderboard website to evaluate its performance relative to other state-of-the-art models.
Despite thoroughly reviewing the GitHub repository, including the issues and docs file, we are still unclear on the exact steps required to make a submission to the leaderboard website.
Our model is compatible with OpenAI's API format, could you kindly provide some guidance on how we can proceed with our submission?
Any additional details on the process or requirements would be greatly appreciated.
Hello HELM Team,
I have been following this project and was very impressed by the comprehensive framework it provides for benchmarking language models. Our team is eager to submit our model to the HELM leaderboard website to evaluate its performance relative to other state-of-the-art models.
Despite thoroughly reviewing the GitHub repository, including the issues and docs file, we are still unclear on the exact steps required to make a submission to the leaderboard website.
Our model is compatible with OpenAI's API format, could you kindly provide some guidance on how we can proceed with our submission? Any additional details on the process or requirements would be greatly appreciated.
Thank you for your time and assistance.