-
Hi, I am trying to create a good Ai for the backgammon game, as there are no examples for it. I tried pure PPO, but it did not give very good results. I was thinking to apply the AlphaZero algorithm, …
-
Should we add functions `parse_state(env, string)` and `parse_action(env, string)` to help building interactive debugging tools?
It would be tempting to just let the user define `parse_state` by o…
-
你好,我了解到AlphaZero好像是一个on-policy算法,on-policy算法是不适用experience replay的,但是我又在代码里面看到使用了experience replay, 我想知道我的看法是不是对的。
https://github.com/initial-h/AlphaZero_Gomoku_MPI/blob/95867cb7e524ebe9c77a926c82091…
-
请问在大棋盘上的实验效果如何,比如 15 * 15 的棋盘大小
-
It looks like AlphaZero's implementation samples uniformly from a replay buffer for each training step. I wonder why they do this instead of iterating through in batches. I am not sure why you would c…
-
-
Some powerful contributors recently started helping this project. While we thank and kudo them for their resources, some changes are needed for the training pipeline, as the AlphaZero pipeline require…
-
I originally had lsf running with a spool size of 100, and went to adjust it back to its default of 1024. After changing it back and puppet bouncing the service, all lines were being truncated that w…
-
請寫『姓名:到』!
-
japanese chess(Shogi) https://github.com/kobanium/aobazero
chinese chess(Xiangqi) https://github.com/leedavid/leela-chess-to-Chinese-Chess
If you could supports korean chess(janggi), could you ple…
KDKOA updated
4 years ago