madhavmk / Noise2Noise-audio_denoising_without_clean_training_data

Source code for the paper titled "Speech Denoising without Clean Training Data: a Noise2Noise Approach". Paper accepted at the INTERSPEECH 2021 conference. This paper tackles the problem of the heavy dependence of clean speech data required by deep learning based audio denoising methods by showing that it is possible to train deep speech denoising networks using only noisy speech samples.
MIT License
172 stars 42 forks source link

I would like to ask how long it takes for such an experiment to run a type of noise, and the minimum equipment requirements #1

Closed guandachen closed 3 years ago

guandachen commented 3 years ago

I would like to ask how long it takes for such an experiment to run a type of noise, and the minimum equipment requirements,Thanks!!

anujstam commented 3 years ago

Hi. It took us about 48 hours on an Nvidia K80 per noise type + training method - this is the same GPU available on Colab or on Azure data science VMs. For our exact network and dataset, you will need 12 GB of GPU memory. If you're looking to test it faster, you could use the smaller DCUnet10 model but we did not test with those, or alternatively you could use a different dataset.

guandachen commented 3 years ago

你好。 在Nvidia K80上,每種噪聲類型+訓練方法花了我們大約48小時-這與Colab或Azure數據科學VM上可用的GPU相同。對於我們確切的網絡和數據集,您將需要12 GB的GPU內存。 如果您想更快地測試它,可以使用較小的DCUnet10模型,但我們沒有對此進行測試,或者可以使用其他數據集。

ok,thanks,i will try on the colab