Open tessavdheiden opened 4 years ago
Have you noticed any improvement in average rewards per episode due to this change? I am curious.
Hi!
No, because it will just add a constant negative reward everytime, so the gradient is not affected ;-)
Swarm intelligence on reinforcement learning for more than 50 agents without collision: https://github.com/Edision-liu/Reinforcement-learning-on-MAPE
Hi!
Is it true that in simple_spread.py collisions are checked with the agents themselves:
if agent.collide:
for a in world.agents:
if self.is_collision(a, agent):
line: https://github.com/openai/multiagent-particle-envs/blob/69ee7f85811c77ee651722bc3c332677b2195da1/multiagent/scenarios/simple_spread.py#L78Maybe do something like:
if agent.collide:
for a in world.agents:
if agent == a: continue
if self.is_collision(a, agent):