Closed miccio-dk closed 5 years ago
Indeed, this doc wasn't very helpful! you can find a better explanation of LSTMs in part 2.4.3 (Recurrent Neural Networks and LSTMs) of the overview [1] that Lars and Ulrik sent us. I really like this document, it is well explained in general :) [1]: Single-Channel Speech Enhancement using Deep Learning, Hulser, 2018
For the concept of LSTM, these 2 documents [18,19] seem to be references (I haven't looked into them though, but they were cited in [2]): [2] Discriminatively Trained Recurrent Neural Networks for Single-Channel Speech Separation [18]S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997. [19] F. Gers, J. Schmidhuber, and F. Cummins, “Learning to forget: Continual prediction with LSTM,” Neural Computation, vol. 12, no. 10, pp. 2451– 2471, 2000.
And to complete the picture, this one: [20]. Doesn't bring much explanation, but uses LSTM for speech separation. And cites [2]. [20]: PHASE-SENSITIVE AND RECOGNITION-BOOSTED SPEECH SEPARATION USING DEEP RECURRENT NEURAL NETWORKS, 2015, Erdogan
I hope it helps
A tuto that looks good: https://adventuresinmachinelearning.com/keras-lstm-tutorial/
thanks for the resources, i'll give it another try today :)
good source for lstm NOT applied to text: https://github.com/keras-team/keras/blob/master/examples/conv_lstm.py
lol nvm
not much literature on denoising autoencoders with lstm layers:
as opposed to those using conv layers:
should we just downscale temporarely? :thinking: