-
DQN: Y = r + gamma * max(Q'(s'))
Double DQN: Y = r + gamma * Q'(s', argmax(Q(s')))
-
## Problem Description
Hi I would like to add the double DQN algorithm to cleanrl. Can someone give me the go-ahead?
-
**Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email.
Please post your question on the [RL Discord](https://discord.com/invite/xhfNqQv), [R…
-
-
From my understanding the target network updates are implemented wrong in the notebook Double-Dueling-DQN.ipynb.
As it updates the same step as the main network (every 4th). In this simple environmen…
-
当使用demo_DQN_Dueling_Double_DQN 训练结束的的pt文件无法作为测试时的权重文件 ,是否需要将保存pt文件
由torch.save(actor, actor_path)
更改为torch.save(actor.state_dict(), actor_path)
-
Thank you so much that you program the Double DQN algorithm. However when I run this algorithm I faced a memory increase consistantly when trainning. Do you have any idea where the memory leak could h…
-
```
IndexError Traceback (most recent call last)
in
43 try:
44 clear_output(True)
---> 45 plot_all_data(log_dir, env_i…
-
I haven't been able to reproduce the results of the Breakout benchmark with Double DQN when using similar hyperparameter values than the ones presented in the original paper. After more than 20M obser…
-
I changed the parameter in examples/dqn.py to this and I get an error:
```
def main():
env_name = 'CartPole-v1'
# env_name = 'PongNoFrameskip-v4'
use_prioritization = True
use_…
jt70 updated
3 months ago