Farama-Foundation / PettingZoo

An API standard for multi-agent reinforcement learning environments, with popular reference environments and related utilities
https://pettingzoo.farama.org
Other
2.45k stars 400 forks source link

[Question] Inconsistent MPE Observations between Documentation and Code #1185

Closed shuoyang2000 closed 3 months ago

shuoyang2000 commented 4 months ago

Question

Hi,

I am using MPE recently (simple adversary) and thanks for this great environment!

However, when I am trying to understand agents' observations, I find it not consistent with what is introduced in the documentation (https://pettingzoo.farama.org/environments/mpe/simple_adversary/): _Agent observation space: [self_pos, self_vel, goal_rel_position, landmark_rel_position, other_agent_relpositions]

In low-level code implementation, what users actually obtain is from the function here https://github.com/Farama-Foundation/PettingZoo/blob/master/pettingzoo/mpe/simple_adversary/simple_adversary.py#L229 in which all returned observation (for agent) is
[relative_pos_with_goal, relative_pos_with_landmarks, relative_pos_with_other_agents]

Could you please double check the documentation and code since this disagreement might be confusing :) But again, MPE is very great and thanks for your great efforts maintaining it.

Best, Shuo

elliottower commented 3 months ago

Thanks this looks like the docs are just out of date, will fix it now