DLR-RM / rl-baselines3-zoo

A training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included.
https://rl-baselines3-zoo.readthedocs.io
MIT License
2.01k stars 510 forks source link

[Question] Correct way of hyper parameter optimization for new algorithm? #384

Closed tyler-ingebrand closed 1 year ago

tyler-ingebrand commented 1 year ago

❓ Question

Hello,

I am writing a new algorithm based on the sb3 implementation of PPO. However, I need to run hyper parameter optimization since I changed some parts of the algorithm and the defaults no longer work for me. What is the best way to test a new algorithm, which implements BasePolicy, using this package? Do I need to just clone the repo and modify utils.py, hyperparams_opt.py, etc? Or is there a cleaner way?

Thank you for your time

Checklist

araffin commented 1 year ago

Hello, yes, a fork would be the way to go. If you just want to add/replace an algorithm, you can take a look at https://rl-baselines3-zoo.readthedocs.io/en/master/guide/sbx.html