-
## Motivation
Add A3C implementation as presented [here](https://arxiv.org/abs/1602.01783) so that the experiments of [Rainbow](https://arxiv.org/pdf/1710.02298.pdf) can be added to the examples in…
-
Dear
I wonder if the **A3C-LSTM** actually implements **DRQN** but on A3C instead of DQN ( https://arxiv.org/pdf/1507.06527.pdf )
If not, then could you please share the details/reference?
Finally…
-
Hi,
I really like your work, and want to ask for some clarifications on your new observation on training a flat A3C agent without the meta-controller. In this case are the sub-goals randomly genera…
-
Hi Arthur
Instead of basic scenario, I used **health_gathering.cfg** [scenario](https://github.com/mwydmuch/ViZDoom/blob/master/scenarios/health_gathering.cfg)
![doom_health](https://cloud.githu…
-
Hi,
Sorry for stupid question
I've done training process. Now I'm looking for execute the agent to play the Doom game but didn't see that piece of code. Could you please show me how to make it run…
-
Hello,
as a part of my master thesis, I have been trying to reproduce the results from the DeepMind paper. I have implemented the A3C algorithm and am currently testing it on the MoveToBeacon minigam…
-
Hello
Thank you for sharing file
I run
python inference.py
Traceback (most recent call last):
File "/home/osboxes/Desktop/Traffic-Signal-Control/inference.py", line 13, in
from ray.rlli…
-
Hi there,
I noticed that even though policy net and value net share some parameters (in a3c/estimators.py), their gradient were [clipped](https://github.com/dennybritz/reinforcement-learning/blob/m…
-
Hi
I wanted to conduct some extra pre-processing for the state (games image). And see how this may speed up the training or enhance the agent performance.
For example, in Break-out, I thought th…
-
Hi there,
Thanks for the implementation!
I tried running the code to solve the breakout game(v0). However, after 3 days of training the results were quite far from one a DQN would obtain - the a…