-
```
В файле задание на второй релиз.
Пожалуйста, задавайте вопросы, если что
непонятно или с Вашей точки зрения может
иметь различное толкование.
И.Г.
```
Original issue reported on code.google.co…
-
Hi, I've been trying to replicate some of your results and compare against Reptile, for the reinforcement learning case. Specifically, I'm trying to run the point mass experiments.
After installing…
-
Hi, I try to run the code. Last day, I success. But today I meet a strange problem.
I don't modify any code, just use setup_mujoco.sh script from [](https://github.com/rll/rllab/blob/master/scripts/s…
-
Thank you for the wonderful tool.
(1) I was able to complete training using Cartpole expert data and GAIL.
(2) Next, I added a wrapper to Cartpole to discretize observations and was able to comple…
-
Hi, I am new to RL, new to rllab, and new to python. I am trying to train a policy to do continuous control for a custom environment.
I implement the environment in `my_sim_env.py` according to thi…
-
hello developers,
I want to use my own expert datas, but I don't know how to make expert datas. Could you tell me about how to make? For example: gail's deterministic.trpo.Hopper.0.00.npz
-
Noting these down for the [neurips bbo challenge](http://bbochallenge.com/leaderboard)
- idea 1: generate more suggestions and only send the top
`n_suggestions` ranked by value.
- idea 2: gener…
-
I tried to benchmark the follwing environments ['BipedalWalker-v2', 'BipedalWalkerHardcore-v2', 'CarRacing-v0', 'LunarLander-v2', 'LunarLanderContinuous-v2'] using ['A3C', 'DDPG', 'TD3', 'SAC', 'PG',…
-
I write a simple robot arm model as the following pic
![image](https://user-images.githubusercontent.com/31737143/34651123-309b05d0-f3cc-11e7-9dba-0360ae108b7f.png)
I use trpo to train the arm, to l…
-
See https://github.com/pytorch/pytorch/issues/975 for more info
PyTorch TRPO appears 50% slower than TF. Not sure about PPO, but I expect the wall-clock time gap will be the same.
To fix this is…