Christie22 / SingleChannelDenoising

3 stars 2 forks source link

lstm model not so straightforward to implement #2

Closed miccio-dk closed 5 years ago

miccio-dk commented 5 years ago

not much literature on denoising autoencoders with lstm layers:

as opposed to those using conv layers:

should we just downscale temporarely? :thinking:

Christie22 commented 5 years ago

Indeed, this doc wasn't very helpful! you can find a better explanation of LSTMs in part 2.4.3 (Recurrent Neural Networks and LSTMs) of the overview [1] that Lars and Ulrik sent us. I really like this document, it is well explained in general :) [1]: Single-Channel Speech Enhancement using Deep Learning, Hulser, 2018

For the concept of LSTM, these 2 documents [18,19] seem to be references (I haven't looked into them though, but they were cited in [2]): [2] Discriminatively Trained Recurrent Neural Networks for Single-Channel Speech Separation [18]S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997. [19] F. Gers, J. Schmidhuber, and F. Cummins, “Learning to forget: Continual prediction with LSTM,” Neural Computation, vol. 12, no. 10, pp. 2451– 2471, 2000.

And to complete the picture, this one: [20]. Doesn't bring much explanation, but uses LSTM for speech separation. And cites [2]. [20]: PHASE-SENSITIVE AND RECOGNITION-BOOSTED SPEECH SEPARATION USING DEEP RECURRENT NEURAL NETWORKS, 2015, Erdogan

I hope it helps

Christie22 commented 5 years ago

A tuto that looks good: https://adventuresinmachinelearning.com/keras-lstm-tutorial/

miccio-dk commented 5 years ago

thanks for the resources, i'll give it another try today :)

miccio-dk commented 5 years ago

good source for lstm NOT applied to text: https://github.com/keras-team/keras/blob/master/examples/conv_lstm.py

miccio-dk commented 5 years ago

lol nvm