Open Ishitori opened 5 years ago
@mxnet-label-bot Add [Feature Request]
@Ishitori do you have an application using SRU now?
We have a CPU implementation in house. If your case works on CPU, we can consider to upstream the initial version. @ciyongch @TaoLv
+1 on this.
These bar charts looks nice, but SRU need more layers to catch the accuracy of LSTM. In total, the speed would be nearly equalization because LSTM is rapidly developing in CUDNN.
Description
Simple Recurrent Unit (SRU) is a new recurrent netowkr architecture that provides better parallelization compare to regular LSTM/GRU cells and in some cases better performance compare to CNN used for sequential data. The original code written in PyTorch + CUDA is open and available at https://github.com/taolei87/sru
It would be great to port this layer into MXNet as it seems to be a basic building block for other models.
@szha