-
# To Do (Urgent)
- [x] 3 types of State functions - Code Template
- [x] 3 types of Action functions - Code Template
- [x] 3 types of Reward functions - Code Template
- [x] Finish code template f…
-
**Describe the bug**
When using PER with QMIX, an issue arises with the idx_range returned by the insert function of RecPolicyBuffer:
> line 267, in insert
for idx in range(idx_range[0], idx_…
-
Hi,
I would like to know if you support MultiDiscrete action space or not?
thanks in advance
Andrea
-
Are there any demos that can illustrate the process of using MATD3 / MADDPG to process discrete actions?
-
您好,我没看懂您关于vdn和qmix不能通信的解释“vdn和qmix是直接根据Q值选动作的,没有办法进行通信。coma是AC方法,正好可以把commnet的结构用在actor上”,按说vdn和qmix虽然不是AC方法,但是critic网络同样可以通过通信来共享局部观察
-
Hi, bro
I have a question that I have always wondered. Is it correct that in rode_controller()'s "forward()" function, the role selector you defined chooses new roles for agents no matter in decentra…
-
Hello, thank you for your contribution of rllib version.
When I run run_ppo.py and run_qmix.py, there is a Value Error as below
(RolloutWorker pid=3950203) File "../python3.9/site-packages/ray/rl…
-
Hello, I encoutered following error while i tried to run the code:
Traceback (most recent call last):
File "run_madqn.py", line 150, in
main()
File "run_madqn.py", line 45, in main
t…
-
I was having a look at the `build_td_lambda_targets` function into `utils/rl_utils.py`, and I was wondering if line 8:
`ret[:, -1] = target_qs[:, -1] * (1 - th.sum(terminated, dim=1))`
is really cor…
-
Reading the pseudocode in paper [Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning](https://arxiv.org/abs/2003.08839)
![image](https://user-images.githubusercontent…