-
Hello,
Thanks for your great work, I am trying to learn tensorflow and keras from your tutorial code. When I used this A2C code to tackle the "MountainCar-v0" problem, I found that the RAM occupation…
-
I've implemented Cross-Entropy method and passed CartPole-v0 and MountainCarContinuous-v0 with the same hyperparameters and default reward definition from gym. But it didn't work on MountainCar-v0 eve…
-
**Summary:** I've noticed that the Spinning Up algorithm implementations don't seem to support discrete **observation** spaces defined with [`gym.spaces.Discrete`](https://github.com/openai/gym/blob/m…
-
I tried your `run_mountain_car.py`, but the accumulated rewards do not change at all.
Are there any hyper-parameters that I need to change? And how many episodes are needed in general?
-
I've had a bunch of gym tabs open over the last couple days, and finding the one I want is really hard because they're all called "OpenAI Gym". It'd be great if pages had more descriptive titles like,…
-
I have been playing with `basic-rs` replacing `CartPole-v0` with envs from [OpenAI](https://gym.openai.com/envs/). Some work some do not. The ones that fail seem to fail here:
I had errors with:
…
-
I'd like to implement Hindsight Experience Replay (HER). This can be based on a whatever goal-parameterized RL off-policy algorithm.
**Goal-parameterized architectures**: it requires a variable for…
-
when the rllab trpo code is applied to the mountain-car env., it does not climb the mountain well until 500 iterations.
It is strange since the TRPO algorithm implemented by OpenAI (https://github.…
-
* Weights and Biases version: 0.6.32
* Python version: 2.7
* Operating System: OSX
### Description
Error because of how unicode is handled in python 2. I had a unicode character in the descrip…
-
Checklist:
- [x] Parallelized training for experts (Adam finished this via #57).
- [x] Get good experts for Humanoid and Ant [which are not doing well right now](https://github.com/HumanCompatib…