This project is an implementation of the Cavs system presented in the paper: Cavs: An Efficient Runtime System for Dynamic Neural Networks, ATC'18, sponsored by Petuum Inc.
Recent deep learning (DL) models are moving more and more to dynamic neural network (NN) architectures, where the NN structure changes for every data sample. However, existing DL programming models are inefficient in handling dynamic network architectures because of:
In this paper, we present Cavs, a runtime system that overcomes these bottlenecks and achieves efficient training and inference of dynamic NNs. Cavs represents a dynamic NN as a static vertex function and a dynamic instance-specific graph . It avoids the overhead of repeated graph construction by only declaring and constructing once, and allows for the use of static graph optimization techniques on pre-defined operations in . Cavs performs training and inference by scheduling the execution of following the dependencies in , hence naturally exposing batched execution opportunities over different samples.
Experiments comparing Cavs to state-of-the-art frameworks for dynamic NNs (TensorFlow Fold, PyTorch and DyNet) demonstrate the efficacy of our approach: Cavs achieves a near one order of magnitude speedup on training of dynamic NN architectures, and ablations verify the effectiveness of our proposed design and optimizations.
@inproceedings{xu2018cavs,
title={Cavs: An efficient runtime system for dynamic neural networks},
author={Xu, Shizhen and Zhang, Hao and Neubig, Graham and Dai, Wei and Kim, Jin Kyu and Deng, Zhijie and Ho, Qirong and Yang, Guangwen and Xing, Eric P},
booktitle={2018 $\{$USENIX$\}$ Annual Technical Conference ($\{$USENIX$\}$$\{$ATC$\}$ 18)},
pages={937--950},
year={2018}
}