issues
search
shariqiqbal2810
/
MAAC
Code for "Actor-Attention-Critic for Multi-Agent Reinforcement Learning" ICML 2019
MIT License
676
stars
173
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Is this code applicable to continuous actions?
#40
Lxsogo
opened
10 months ago
0
add cluster attention
#39
phillipinseoul
closed
2 years ago
0
issue thanks!
#38
djm0539
opened
2 years ago
0
How to visualize during training
#37
AI-Hello
opened
3 years ago
0
The function names of "update_policies" and "update_critic" are reversed
#36
TangYuan96
closed
3 years ago
0
Why does your implementation of MADDPG not work in your fork of MPE?
#35
Henry668
opened
3 years ago
0
Critic function learning
#34
yesiam-png
opened
3 years ago
0
Where is the code to load the model?
#33
vinsheldon
opened
3 years ago
0
How to solve env_id?
#32
lylyjy
opened
3 years ago
2
Memory Leak
#31
yuchen-x
closed
3 years ago
1
Memory usage increases a lot when use the latest version of OpenAI baselines
#30
yuchen-x
closed
3 years ago
0
Bias on value extractors ?
#29
jeanibarz
opened
4 years ago
0
Critic encoders as shared modules ?
#28
jeanibarz
closed
3 years ago
3
When I run "python main.py fullobs_collect_treasure V1" I meet error "ImportError: cannot import name 'Wall'"
#27
leeruibin
closed
4 years ago
0
How does the gradient back-propagate from Q to the action $a_i$?
#26
xihuai18
opened
4 years ago
2
About query, key and value input embedding
#25
TianQi-777
closed
3 years ago
1
How to implement MADDPG+SAC and COMA+SAC
#24
Ben-NLP
opened
4 years ago
2
question about reward
#23
ShuangLI59
closed
4 years ago
10
About SAC implementation
#22
yesiam-png
closed
4 years ago
1
Seeding fails to produce deterministic results
#21
sanjeevanahilan
closed
4 years ago
9
Problem of optimizing policy
#20
zsano1
closed
4 years ago
4
does training advantage soft actor critic based on replay buffer has large bias?
#19
KK666-AI
closed
4 years ago
7
hello , i have a question on your paper
#18
qiufengsly
opened
4 years ago
1
connection.py
#17
agharibr2019
closed
4 years ago
3
cannot handle some scenarios
#16
KatnissY
closed
4 years ago
1
About the reporduction of experiment Cooperative Treasure Collection
#15
zwfightzw
closed
4 years ago
1
hi,my machine memory always overflow when I run your code,but I don't find the reason. Can you help me?
#14
wang88256187
closed
5 years ago
1
when I run"python main.py fullobs_collect_treasure dir_1",there are the errors
#13
Gezx
closed
5 years ago
1
How to evaluate and test the training result?
#12
wanghuimu
closed
5 years ago
3
VecEnv, why closing the Pipe()
#11
GoingMyWay
closed
5 years ago
2
How to visualize the attention weights between agents in the testing phase?
#10
soada
closed
5 years ago
8
How to run your code in other scenarios, i.e., cooperative navigation
#9
soada
closed
5 years ago
4
2 agents modification
#8
emanuelepesce
closed
5 years ago
1
multi-agent particle environments
#7
SHYang1210
closed
5 years ago
8
About environment
#6
YaozuGen
closed
5 years ago
1
State Action Encoding in Critic
#5
HassamSheikh
closed
5 years ago
1
How to visualize the result?
#4
sanshibayuan
closed
5 years ago
1
Update README.md
#3
HassamSheikh
closed
5 years ago
1
About the results
#2
ewanlee
closed
5 years ago
2
About the code
#1
lucasliunju
closed
5 years ago
1