-
### Search before asking
- [X] I have searched the MuZero [issues](https://github.com/werner-duvaud/muzero-general/issues) and found no similar feature requests.
### Description
Perfect ideas and…
-
hi,
i tried out this project and it is one of the few that actually works off the shelf, thank you for your work.
Is there a way to enable self play when training an agent? My usecase is to use Dr…
-
A big difference between the training methods of MuZero and AlphaZero is: Muzero uses K consequential steps samples, and AlphaZero uses only one step samples. Muzero's paper has shown the advantage …
-
Hello,
This issue is closely related to #229, I am trying to reproduce results of the original Muzero paper on Atari (or at least on a small subset of games, for the moment I tried MsPacman and Spa…
-
# MuZero Intuition
Posts and writings by Julian Schrittwieser
[http://www.furidamu.org/blog/2020/12/22/muzero-intuition/](http://www.furidamu.org/blog/2020/12/22/muzero-intuition/)
-
I see the examples have [MuZero for discrete control tasks](https://github.com/deepmind/acme/blob/31528f87711c1c94b3d99b5a21f347424759e29e/examples/baselines/rl_discrete/run_muzero.py).
But this d…
-
I'm trying to fill the pseudocode.py DeepMind attached to its paper about MuZero. for the network part I used a very similar structure to yours in this repository but I'm not able to make the training…
-
Hi,Yuri!how are you there? and I want to know about your recent progress in Muzero project, has your model converged?I built my Muzero to play renju,but after several hundred epochs of training, it s…
-
MuZero: https://arxiv.org/abs/1911.08265
-
Dear Chiamp,
Thank you for sharing your effort with the community. I was trying your muZero code and I fixed some not updated issues relative to the use of the discontinued Monitor. Once I fixed th…