neardws / Game-Theoretic-Deep-Reinforcement-Learning

Code of Paper "Joint Task Offloading and Resource Optimization in NOMA-based Vehicular Edge Computing: A Game-Theoretic DRL Approach", JSA 2022.
https://www.sciencedirect.com/science/article/abs/pii/S138376212200265X
GNU General Public License v3.0
237 stars 49 forks source link
deep-reinforcement-learning noma potential-game resource-allocation task-offloading vehicular-networks

Game-Theoretic-Deep-Reinforcement-Learning

This is the code of paper, named "Joint Task Offloading and Resource Optimization in NOMA-based Vehicular Edge Computing: A Game-Theoretic DRL Approach", and the proposed solution and comparison algorithms are implemented.

Environment

The conda environment file is located in environment.yml.
It can be used to create the environment by:

conda env create -f environment.yml

File Structure

Main Function

The main() function of the repo is located in Experiment/experiment.py.

Algorithms

Didi Dataset

The vehicular trajectories for November 16, 2016, generated in Chengdu and extracted from the Didi GAIA Open Data Set, can be found on neardws/Vehicular-Trajectories-Processing-for-Didi-Open-Data.

Citing this paper

@article{xu2022joint,
  title={Joint task offloading and resource optimization in NOMA-based vehicular edge computing: A game-theoretic DRL approach},
  author={Xu, Xincao and Liu, Kai and Dai, Penglin and Jin, Feiyu and Ren, Hualing and Zhan, Choujun and Guo, Songtao},
  journal={Journal of Systems Architecture},
  pages={102780},
  year={2022},
  issn = {1383-7621},
  doi = {https://doi.org/10.1016/j.sysarc.2022.102780},
  url = {https://www.sciencedirect.com/science/article/pii/S138376212200265X},
  publisher={Elsevier}
}