Here we propose a method to extremely accelerate NAS, without reinforcement learning or gradient, just by sampling architectures from a distribution and comparing these architectures, estimating their relative performance rather than absolute performance, iteratively updating parameters of the distribution while training.
Our paper is available at 《Multinomial Distribution Learning for Effective Neural Architecture Search》
Here we provide our test codes and pretrained models, our code is based on DARTS and ProxylessNAS, pretrained models can be downloaded here
You need to modified your path to dataset in data_providers/cifar10.py
and data_providers/imagenet.py
,config.sh
is used to prepare your environment, you should write this file by yourself and here we use it to prepare dataset and packages
To evaluate the model in DARTS setting, just run
for cifar10
:
chmod +x run_darts_cifar.sh
./run_darts_cifar.sh
for imagenet
:
chmod +x run_darts_imagenet.sh
./run_darts_imagenet.sh
To evaluate the model in Mobile setting, just run
for gpu
:
chmod +x run_gpu_imagenet.sh
./run_gpu_imagenet.sh
for cpu
:
chmod +x run_cpu_imagenet.sh
./run_cpu_imagenet.sh