-
# Asynchronous Methods for Deep Reinforcement Learning #
- Author: Volodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy P. Lillicrap, Tim Harley, David Silver, Koray Kavukcuo…
-
I am new to pyTorch, just cloned your codes and ran them, but got an error. I hope you to point me to the right direction to fix this issue.
More specifics:
1. used conda env with python 3.6
…
-
Hi @rarilurelo,
can I ask if you have been able to modify your code to work with continuous actions - eg pendulum or mountain car? I tired to modify @ikostrikov 's implementation, see here
https…
-
This is quite nice!
There are several A3C PyTorch implementations for Atari.
Is it possible to do the same with this Truck environment?
Thank you.
-
Dear Hongzi,
sorry to bother you, but I went through a few problems about critic gradient when reproducing Pensieve with PyTorch.
In ```/sim/a3c.py```,you used the mean square error of **R_batc…
-
where is multi-step loss function and entropy loss function?
-
Hi, I am wondering why there is no break statement in the local_test in this script [https://github.com/uvipen/Super-mario-bros-A3C-pytorch/blob/master/src/process.py]. It seems like the testing loop …
-
Looking at a handful A3C implementations and results on Seaquest, they appear to score around 50K:
- https://gym.openai.com/evaluations/eval_pjjgc9POQJK4IuVw8nXlBw (ConvNet)
- https://gym.openai.com…
beniz updated
7 years ago
-
I can't see that the local gradients are ever reset. The values are overwritten by the global weights, but the optimizer `opt` is assigned to the global parameters, so won't this accumulate gradients …
-
I run the Cart_Pole.py with A3C&A2C on windows and got the error.
Traceback (most recent call last):
File "D:/学习/Deep-Reinforcement-Learning-Algorithms-with-PyTorch-master/results/Cart_Pole.py",…