Zeta36 / Asynchronous-Methods-for-Deep-Reinforcement-Learning

Using a paper from Google DeepMind I've developed a new version of the DQN using threads exploration instead of memory replay as explain in here: http://arxiv.org/pdf/1602.01783v1.pdf I used the one-step-Q-learning pseudocode, and now we can train the Pong game in less than 20 hours and without any GPU or network distribution.
84 stars 31 forks source link

It is GREAT! #2

Open MingyanZhao opened 8 years ago

MingyanZhao commented 8 years ago

Hi zeta:

I saw your reply and came here to have a look.

This is so cool! I really need a method that can avoid memory replay since the memory space is a big problem and time consuming.

Thank you. I will look into it further. Mingyan

Zeta36 commented 8 years ago

You are welcome, friend. ;)