AKG(Auto Kernel Generator) is an optimizer for operators in Deep Learning Networks. It provides the ability to automatically fuse ops with specific patterns. AKG works with MindSpore-GraphKernel to improve the performance of networks running on different hardware backends.
AKG composes with three basic optimization module, normalization, auto schedule and backend optimization.
backend optimization. The backend optimization module mainly consists of TensorCore acceleration, double buffer optimization, storage flatten optimization and inject sync optimization.
At present, Ascend910
, NVIDIA V100/A100
and CPU
are supported. More Backends are on the list.
See MindSpore README.md for details.
We suggest you build and run akg together with MindSpore. And we also provide a way to run case in standalone mode for convenience sake. Refer to MindSpore Installation for more information about compilation dependencies.
Build on Ascend910
git-lfs needs to be installed before cloning the source codes.
git clone https://gitee.com/mindspore/akg.git
cd akg
bash build.sh -e ascend -j8
Build on GPU
git clone https://gitee.com/mindspore/akg.git
cd akg
bash build.sh -e gpu -j8
Build on CPU
git clone https://gitee.com/mindspore/akg.git
cd akg
bash build.sh -e cpu -j8
Ascend910
cd tests
source ./test_env.sh
NVIDIA V100/A100
cd tests
source ./test_env.sh gpu
CPU
cd tests
source ./test_env.sh cpu
cd tests/st
python run.py -e gpu -o add -l level0 # run add operator on GPU
Detailed instructions see:python run.py -h
Use specific case:
Ascend910
cd tests/st/ops/
pytest -s test_abs.py -m "level0 and platform_x86_ascend_training" # run level0 testcases on Ascend
NVIDIA V100/A100
cd tests/st/ops/
pytest -s test_abs.py -m "level0 and platform_x86_gpu_training" # run level0 testcases on GPU
CPU
cd tests/st/ops/
pytest -s test_abs.py -m "level0 and platform_x86_cpu" # run level0 testcases on CPU
See Wiki.
Welcome contributions. See MindSpore Contributor Wiki for more details.
The release notes, see our RELEASE.