-
Hello,
I **think** I've found a bug, and if this is a mess up on my part, I apologize in advance.
**Describe the bug**
It appears that the mel-scale spectrogram y-axis is displayed incorrectl…
-
Hi,
I am trying to use the NISQA Corpus dataset published by you. I found that the speech samples in the NISQA Corpus are 48k sampling and their true frequency distribution is only 0-8k ( some …
-
I trained Tacotron2 with my own dataset in Swedish. Alignment looks terrible but it produces very good pronunciation. The issue I'm having is that the infer result is very low pitched. The source is a…
-
I have gone through the code for plotting of weights in GitHub. Issue #5573
But when I tried to implement same thing in my code ,I'm not able to plot it. Please help me in this .I want to extract th…
-
Most of librosa only supports monophonic audio. But for many of the analyses we'd like to do, stereo or multi-channel support would be very useful and not all that difficult at this point.
This is…
-
This is not an issue and is more of a discussion. I read the WaveGrad paper today (which may be found [here](https://arxiv.org/pdf/2009.00713.pdf)) and listened to the samples [here](https://wavegrad.…
-
Any chance you have PyTorch model files saved, in addition to the Keras models present in this repo?
-
Hi,how to add "emobase2010.conf" to the project, is there a tutorial?
-
![图片](https://user-images.githubusercontent.com/26831425/99934861-33128f80-2d9a-11eb-8d0d-bddce32e0f79.png)
hi, this a 48k wav trained 120k steps, you can see the noise above 16kHz. How can I rem…
-
Question :
Trying to run the Nemo for the audio recorded with the phone /zoom meeting , but the text output is very bad, the same audio which has been recorded with good microphone ,the text output…