Harry67Hu / MADPS

Official implementation of the AAMAS 2024 paper Measruing Policy Distance for MARL.
2 stars 3 forks source link

Multi-agent Dynamic Parameter Sharing

Official implementation of the AAMAS 2024 paper Measruing Policy Distance for MARL.

Requirements

To run this repository you need:

i) Install the code's requirements. A virtual environment based on Conda is recommended (We will update the docker approach soon). Install with

conda create --name madps --file requirements.txt

ii) Install the supported MARL-environments, for example:

Running

You can then simply run ac_NF.py using:

python ac_NF.py with env_name='pettingzoo:pz-mpe-large-spread-v1' time_limit=50

This command runs our multi-agent actor-critic training framework, with pettingzoo:pz-mpe-large-spread-v1 as the training scenario and the maximum episode step length as 50. The scenario versions, from large-spread-v1 to v6, correspond to the 15a_3c, \ 30a_3c, \ 30a_5c, \ 30a_5c_super, \ 15a_3c_shuffle, and 30a_3c_shuffle scenarios in our paper.

Structure of MADPS

The MADPS code is structured as follows:

1. Actor-Critic Training and Execution Framework (ac_NF.py)

ac_NF.py includes:

More details can be found in the comments of main function in ac_NF.py.

2. Multi-agent Policy Distance Computing and Multi-agent Dynamic Parameter Sharing (MADPS_NF.py)

MADPS_NF.py includes:

3. Multi-agent Nerual Network Models (model_NF.py)

model_NF.py includes:

Note: More details are coming soon.

Cite:

Before the AAMAS 2024 conference, please cite the arxiv version.

@inproceedings{tianyihu-MAPD,
   title={Measruing Policy Distance for Multi-agent Reinforcement Learning},
   author={Tianyi Hu et.al},
   booktitle={International Conference on Autonomous Agents and Multiagent Systems (AAMAS)},
   year={2024}
}