-
Hi, thanks for your implementation of TRPO.
In [https://github.com/wojzaremba/trpo/blob/master/main.py#L128-L132](url) you normalize an advantage function.
I couldn't find any description about th…
-
https://github.com/ikostrikov/pytorch-trpo/blob/e200eb8a23b3c7941a0091efb9750dafa4b23cbb/main.py#L108-L119
The fixed log prob part of the line and the "get_loss" function part are exactly the sam…
-
Im just curious if there is any effort to add these policy gradient methods ?
-
I successfully installed rllab, and the trpo_cartpole.py example worked fine, but I'm getting an error with trpo_gym.py:
```
Traceback (most recent call last):
File "trpo_gym.py", line 6, in
…
-
Hi, I am running codes without changes.
when running "time python3 -m baselines_energyplus.trpo_mpi.run_energyplus --num-timesteps 1000000000" the program stops at iteration 31.
the problem is abo…
-
Apparently, classic control environments in Gym have a different key for render modes in `env.metadata`.
In fact:
```python
>>> import gym
>>> env = gym.make('CartPole-v1')
>>> env.metadata
{…
-
https://github.com/openai/baselines/blob/f2729693253c0ef4d4086231d36e0a4307ec1cb3/baselines/gail/dataset/mujoco_dset.py#L53
When I run run_mujoco.py, the code merged in #447 is currently erroring o…
-
Hi, I shifted from TRPO to PPO2, and I know how to train the model and save it and then re-train the saved model in TRPO.
But, in PPO2, I don't know the code for it.
@antoine-galataud you had on…
-
I've tried to use TRPO to create a model for `CartPole-v0` by following the instructions on your [OpenAI Gym page](https://gym.openai.com/evaluations/eval_4QXCRAATTDqakJV0YZlJ4g#reproducibility), chan…
-
Hello,
I'm sorry for asking so many questions. But would you know how to go about saving and loading the subclassed model? I have tried everything I can think of including trying to change the subc…