-
- [ ] I have marked all applicable categories:
+ [x] exception-raising bug
+ [x] RL algorithm bug
+ [ ] documentation request (i.e. "X is missing from the documentation.")
+ [ ] ne…
-
Sorry, if this appears to be a stupid question. I am trying to implement gradient inversion using PyTorch based on the paper but I would like to ask for some clarifications. Is the inversion done on a…
-
`Agent`s are entities with a `sample_action` and `update` method, in potence.
We exclude from the list exploration strategies and curricula.
_Implement_ means either to produce new code from the pape…
-
Hello Patrick,
I am doing an implementation of the PPO algorithm for a custom environment and first wanted to test things out with a standard example and I choose CartPole-v1 implemented with [`gym…
-
Hello,
Thanks for sharing this great code of chainer based DQN.
I recently started to use chainer. The code works great for me and I would like to implement a critic-actor architecture based on your …
-
### Question
I tried hard to train an agent to solve any of the AntMaze environments. I tried the stable baselines 3 implementations of SAC (dense and sparse) and PPO, but could not solve even a sm…
meppe updated
9 months ago
-
Link: [Arxiv](https://arxiv.org/pdf/1906.08253.pdf)
Code: https://github.com/JannerM/mbpo
This paper is very similar to "Benchmarking Model-Based Reinforcement Learning" #5 .
This paper is publis…
-
If I increase both the HEIGHT and WIDTH from 5 to 10 keeping the obstacles and the final goal at the same position, Deep SARSA network doesn't seem to converge. What do you think is the problem? Shoul…
-
How to improve the success rate, my goal is to use BAXTER robot to push the object to the target point in MUJOCO, my GYM environment has been completed, but his training success rate has been very low…
-
Hello, I can't search your paper, has it been published yet? If possible, can you send a link to the paper. thank you.