-
https://datawhalechina.github.io/easy-rl/#/chapter7/chapter7
Description
-
Hi Denny,
Thanks for this wonderful resource. It's been hugely helpful. Can you say what your results are when training the DQN solution? I've been unable to reproduce the results of the DeepMind p…
-
## 🐛 Bug
On Linux, when using the CPU (whether a CPU device in a CUDA-capable PyTorch, or using a CPU-only PyTorch distribution), merely _instantiating_ modules with an LSTM in them claims memory t…
-
Hi @dgriff777 . Thank you for your repo. It's great that it can achieve such a high score. But I met a problem when I try to apply it to MsPacman-v0.
I simply used this command `python main.py --e…
-
I got this error:
Traceback (most recent call last):
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/multiprocessing/process.py", line 315, in _bootstrap
self.run()
…
-
### What happened + What you expected to happen
I am currently running trials with the A3C algorithm in an episodic environment. Since the horizon option in the 2.3.0 build has been removed, it perio…
-
Can I ask what this move reward used for? It seems the move reward will dominate the total reward.
-
Hi @EndingCredits,
this is really cool that you got the `NEC` working :+1:
Have you tried to run your code on the Atari environments, in Open AI gym?
I tried to train on `Pong`, but I got th…
-
I've looked into the available documentation and examples, but haven't been able to figure out if it is possible to use the ML.NET in its current state for (non-deep) reinforcement learning. If it is …
-
Make an agent that does [model free episodic control](https://arxiv.org/pdf/1606.04460.pdf) because that is nice and easy and uses the dictionary.