ShunLu91 / PA-DA

[CVPR '23] PA&DA: Jointly Sampling PAth and DAta for Consistent NAS
MIT License
34 stars 4 forks source link
neural-architecture-search one-shot-nas

PA&DA: Jointly Sampling PAth and DAta for Consistent NAS (CVPR 2023)

license python pytorch

This repository contains the code for our paper "PA&DA: Jointly Sampling PAth and DAta for Consistent NAS", where we propose to explicitly minimize the gradient variance of the supernet training by jointly optimizing the sampling distributions of PAth and DAta (PA&DA).

pa-da_framework

Prerequisites and Dependencies

To run our code, please see the prerequisites below:

  1. Download the datasets of NAS-Bench-201 and NAS-Bench-1Shot1, and pre-trained checkpoints from the [Google Drive].
  2. Install the necessary packages below. We adopt Python 3.7 in our experiments.

NAS-Bench-201 Experiments

We first construct the supernet of NAS-Bench-201 based on the Awesome AutoDL. According to previous papers, we then provide our implementations for baseline methods such as SPOS, FairNAS, and SUMNAS. The implementation of PA&DA are provided subsequently and adopts the same training configuration as baseline methods. If you have any questions or concerns, please feel free to raise an issue and discuss with us.

Acknowledgments

During our implementations, we referred the following code and we sincerely appreciate their valuable contributions:

Citation

If you find this work helpful in your research, please consider citing our paper:

@inproceedings{lu2023pa-da,
  title     = {PA&DA: Jointly Sampling PAth and DAta for Consistent NAS},
  author    = {Lu, Shun and Hu, Yu and Yang, Longxing and Sun, Zihao and Mei, Jilin and Tan, Jianchao and Song, Chengru},
  booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year      = {2023}
}