Closed abhay-shete closed 2 weeks ago
My overall feedback is that most of the contents of this pull request can live in its own Python package instead of being merged into the main branch. Here's how you would do this:
myproject
, and run pip install -e .
to install your package in editable mode.src/helm/benchmark/scenarios/*_scenario.py
src/helm/clients/azure_*_client.py
src/helm/benchmark/run_specs/myproject_run_specs.py
file in your package and your run spec functions from lite_run_specs.py
to there. Revert the changes in lite_run_specs.py
.auto_client.py
.schema_classic.yaml
to your working directory and rename it to schema_myproject.yaml
.model_metadata.yaml
and model_deployments.yaml
to the prod_env
folder in your working directory.deployment
arguments to client_spec.args
in model_deployments.yaml
(example).You should now be able to run helm-run
with your scenarios and clients. You should run helm-summarize
with the --schema-path schema_myproject.yaml
flag.
1) Added support for the Llama3 model and OpenAI gpt-4 model hosted on Azure 2) Modified instructions in lite-specs for multiple choice qa questions to only output a character based index for the answer 3) API keys and endpoints can be provided from credentials.conf instead of using environment variables Sample format is as follows:
{ azureLlama3ApiKey: "<azure llama 3 api key>" azureLlama3Endpoint: "<azure llama 3 endpoint>" azureLlama3Deployment: ""
azureOpenAIApiKey: ""
azureOpenAIEndpoint: ""
azureOpenAIDeployment: "t"
}
Note that the api keys, endpoints and deployment keys for a particular model should have the same prefix. These variables are then provided in the constructor of respective client classes