-
Building off the jsons of the best hyperparameters saved in https://github.com/DLR-RM/rl-baselines3-zoo/pull/140, I have a script that takes each and runs trains it for 10 times (by default) to determ…
-
run:pip install -e .
return:ERROR: Could not build wheels for detectron2, which is required to install pyproject.toml-based projects
detail:
39\detectron2\model_zoo\configs\new_baselines
…
-
**Is your feature request related to a problem? Please describe.**
The current version of smarts contains a lot of different components but these components can conflict and cause problems with other…
-
Hi there,
When doing hyper-parameter training with rl-zoo I often accidentally test a parameter that produces invalid or explosive policies for a particular algorithm/environment.
Occasionally t…
-
Hello,
I've tried in vain to find suitable hyperparameters for SAC in order to solve MountainCarContinuous-v0.
Even with hyperparameter tuning (see "add-trpo" branch of [rl baselines zoo](https:…
-
Hi,
Currently, I have used 4 algorithms from stable-baselines for the task of Roboschool HumanoidFlagrunHarder. My evaluation metric is the mean reward of 100 episodes. Basically: PPO2 is perfect, A2…
-
I'm trying to resume the model training and I'm getting some strange results. Using SubProcVecEnv and VecNormalize on a custom environment:
```
from stable_baselines.common.policies import MlpPoli…
-
### ❓ Question
Hello, I'm trying to solve the Frozenlake-v1 environment with is_slippery = True (non-deterministic) with the stable baselines 3 A2C algorithm. I can solve the 4x4 version but I can't …
-
Hi I'm using PPO1 in my experiments. I scale the action between [-1, 1] and do the rescale in the environment. The result is relatively good but for one issue, the actions are very unstable and have …
-
Thank you for opensourcing such a great code!
I have questions about your IQN implementation, especially on how it can reproduce the scores reported by th paper.
First, your config file https://…