-
hi,
i was wondering if i can apply the baselines algo on the game chess either https://github.com/genyrosk/gym-chess or a custom chess env. based on python-chess.
i was thinking 18x8x8 input and 6…
-
I tried to run the code for Atari Freeway using the following command with the default settings in the code:
```bash
python main.py --env FreewayNoFrameskip-v4 \
--case atari \
--opr train \
--am…
-
Hi @enpasos,
Any idea how to implement single player games?
E.g. Cartpole, Lunar Lander or Breakout?
And print out current max score archived.
Thanks in advance.
-
If I have access to the environment model, is it faster/better to train alphazero instead?
thanks
-
`Agent`s are entities with a `sample_action` and `update` method, in potence.
We exclude from the list exploration strategies and curricula.
_Implement_ means either to produce new code from the pape…
-
### Search before asking
- [X] I have searched the MuZero [issues](https://github.com/werner-duvaud/muzero-general/issues) and found no similar feature requests.
### Description
Hi,
I will appr…
-
可以加入AI训练吗?
我们可以训练AI
我们需要最新的AI棋谱
https://katagotraining.org/ 这是一个非常成功的AI训练例子,你可以参考看看。虽然它是围棋的,它现在已经拥有2千万张棋谱,还在不断的增长。
象棋很多人还用那些古谱来练习象棋,那些已经过时了,象棋比围棋更有优势,象棋可以马上让人进入深度计算思维。
仙女鳕鱼 象棋AI引擎
https://gith…
-
Self play, and generally multi-LM-agent settings are something we are very interested in exploring. What does it take to support this? Does it already work without big overheads?
-
Hi there,
First of all, great work and thank you for opensourcing your code!
I have a question regarding reanalyze: you chose to reanalyze 99% of policy targets and 100% of value targets. I am j…
-
There is no target network which is noted in the article in this implementation?