-
-
Thank you so much that you program the Double DQN algorithm. However when I run this algorithm I faced a memory increase consistantly when trainning. Do you have any idea where the memory leak could h…
-
Hi,
I've been using some of the excellent pre-trained models from DQN and its variants. However, looking at more recent algorithms (PPO, A2C, TRPO, etc.) it seems like [we're now using a different …
-
{
"base_config": "configs/HighwayEnv/agents/DQNAgent/ddqn.json",
"model": {
"type": "EgoAttentionNetwork",
"embedding_layer": {
"type": "MultiLayerPerceptron",…
-
Is there any plan to support Random Ensemble Mixture(REM) variant of DQN . It helps for robust offline training and results are much better compared to other variants of DQN.
Please refer Google De…
-
- [Distributed Prioritized Experience Replay](https://arxiv.org/abs/1803.00933)
- Current implementation works only for DDPG variants, so extends it to work with DQN like agent
-
## 🚀 Feature
The RL section of bolts currently only includes variants of DQN and VPG and lacks some of the more modern RL algorithms. Adding PPO, A2C, curiosity exploration etc. might be prudent
-…
-
The default target_update_interval for DQN based algorithms is set as 312 and not changed for any of the variants in the configs (except for R2D1, which seems to be correctly set to 2500). I don't thi…
-
A feature request proposal to add support of Duel DQN, as suggested in [paper](https://arxiv.org/pdf/1511.06581.pdf) [Dueling Network Architectures for Deep Reinforcement Learning] , which is describe…
-
Hi,
while working on a PyTorch DQN agent for BSuite experiments, I noticed quite bad results on the mnist and mountain car experiments. I see that a similar question was addressed [here](https://gi…