-
`baselines/deepq/experiments/train_mountaincar.py` doesn't converge and remains at -200 mean episode reward. I'm assuming these experiments have been added because they are working (rather than just u…
-
Hi,
Apologies if this question doesn't below here. I've noticed while training mountain car with a dqn using the sample scripts the agent while training never converges to produce the same set of a…
-
**Describe the bug**
$ python run_MountainCar.py
Discrete(3)
Box(2,)
[0.6 0.07]
[-1.2 -0.07]
2019-02-11 17:32:47.428166: I tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports …
-
If there's something wrong with openai gym and chat didn't resolve it in 10 minutes, feel free to complain here.
-
-
Hello.
Dose PPO experiments work in environments like CartPole or Atari?
Such action spaces are discrete, defined in type but not .
If it works in discrete action space envs, then how can I …
-
I tried running the code for PongDeterministic-v4 for 1.75 M steps but there was no improvement in score. It started with avg -21 but even after 1.75 M step the avg score was around -20.4. In the last…
-
@smilesun
What do you think is better for the user interface? Having one function `makeEnvironment` which then calls the corresponding environment subclass as in 1. or exposing all the subclasses …
-
I started doing this class this week and I really like how well everything was thought out for self-learning. I really like the notebooks. I feel like I got a good grasp on Crossentropy, completing th…
-