instadeepai / Mava

🦁 A research-friendly codebase for fast experimentation of multi-agent reinforcement learning in JAX
Apache License 2.0
737 stars 90 forks source link

Add return metrics for experiments and corresponding Hydra config option #1022

Closed liamclarkza closed 9 months ago

liamclarkza commented 9 months ago

What?

We now return a performance evaluation metric from our experiment.

Why?

Allows for us to hyperparameter tune using Hydra & OpTuna.

How?

Modifies the run_experiment method and hydra_entry_point in the various systems to return a metric to be used for hyperparameter tuning through Hydra (using something like OpTuna). The Hydra config option, config.env.eval_metric,is used to select the metric (episode_return or win_rate) for an environment.

CLAassistant commented 9 months ago

CLA assistant check
All committers have signed the CLA.

EdanToledo commented 9 months ago

LGTM