-
This is a work of art, it would be wonderful to find examples inspired by you with PPO and other agents.
I uncommented and tried to test the DQN_Agent in https://github.com/druce/rl/blob/master/Tra…
-
Hello, I have read your paper and open-source code and feel that I have benefited a lot. But as I have just started, there are still some areas that I don't understand that I would like to ask you for…
-
Thanks for offering this wonderful code. But I have a question.
1. Why in the combination part of the equation, the advantage A need to subtract it's average? I've already refer to the paper but sti…
-
I managed to run the code, but during the process, I realized that the maximum STEP for each batch is only 50,
`steps: 50, loss_val: 0.1930, action_spread: tensor([26, 24], device='cuda:0'): 18%|█▊…
-
- [Distributed Prioritized Experience Replay](https://arxiv.org/abs/1803.00933)
- Current implementation works only for DDPG variants, so extends it to work with DQN like agent
-
[A Distributional Perspective on Reinforcement Learning](https://arxiv.org/abs/1707.06887)
-
## Problem Description
Hi I would like to add the double DQN algorithm to cleanrl. Can someone give me the go-ahead?
-
-
there is no file of 230415_Pure_BattPwrOpt_orig_comp.mat
Can you tell me how to create it?
thanks
-
[Paper](https://papers.nips.cc/paper/6501-deep-exploration-via-bootstrapped-dqn.pdf)
Tensorflow Implementation: https://github.com/rrmenon10/Bootstrapped-DQN
zer0n updated
6 years ago