-
Hi Vince, many thanks to your fantastic work! I would like to know if there is a plant to support the TRPO algorithm? Thanks a lot!
-
Hi,
I am wondering whether chainerrl supports TRPO to run atari? I tried to do so by following the code for training PPO on atari, but I am faced with the following error:
> Traceback (most rec…
-
Training is very slow, need to check the original paper hyper parameters.
-
When I run wgail.py, I find that it seems like missing trpo module? Can you provide some details about running this file? Thanks very much!
-
I want to replace the TRPO with DDPG + HER and am having difficulties. The combination only works with a task that is registered with Gym. How did TRPO avoid that?
-
The TRPO paper (Appendix C) claims that "we use a line search to ensure improvement of the surrogate objective and satisfaction of the KL divergence constraint". However, in the current codebase, the …
-
Hello, thank you for sharing your code.
May I ask a paper question? Since ppo is the upgrade of trpo. Have you considered to use ppo instead of trpo? I am facing this question in my thesis. I wonder…
-
On executing trpo_continous.py, I get the following error:
> [2017-07-01 23:52:58,375] Making new env: CartPole-v0
> [TL] InputLayer continous_shared/continous_input_layer: (?, 3)
> [TL…
-
Запили сука новый проект в Visual Studio в том же решении. Назови его TRPO.Database. В него перенеси класс Database из проекта TRPO.Services
-
Apologies if I am misunderstanding something but it seems that the direction of the KL divergence calculations used throughout the TRPO code seems to be at odds with the TRPO paper. Instead of KL[new …