-
Does it make sense to use the Wavenet vocoder as it is for speech to speech? For example, Can I record my voice, generate a melspectrogram, then use a pre-trained model on LJSpeech dataset to respeak …
-
@jaron Can you please explain how number of frames in each window is set in 3-us8k-cnn-extract-train.ipynb script. To be exact, I am referring to **extract_features** method. I understand that the scr…
-
Hi.
I'm interesting to deep voice3 so I think I want to make this model myself.
Now, I read the report of this model but I can't understand input of decoder.
The report say, it's mel-scale log magn…
-
Reasons
- I don't come close to synthesizing comprehensible wave files. And I don't know where to begin fixing it.
- There is no manual for hparams.py for beginners
- Hardly any support in this f…
-
thanks for this lib !
i want to know how can i get mfccs if i had a matrix and every row of this matrix is a voice frame after VAD test, i mean i don't have a time series , because before i get mfccs…
-
Hi,
I'm not quite understand the mel spectrum normalization code:
```python
def _normalize(S):
return np.clip((S - hparams.min_level_db) / -hparams.min_level_db, 0, 1)
```
why do it like …
-
Hi,
I am new to pytorch and following the example of jsut here. And I encountered the following assertion error which is hard for me to look in further. Could anyone help me out?
```
[kwon@ssi-…
-
Hey Keith! Thanks for your work.
It looks like the normalization of the Spectrogram is different from the normalization of the Mel-Spectrogram, i.e. there's a difference in offset by ref_level_db. …
-
I'm a bit stumped on this.
TLDR; im getting weird behavior using kapre as a replacement for spectrogram feats
I have a model and traditionally i have precomputed 64 mel x 128 frame specs and fed t…
-
Can You Please Help:
Traceback (most recent call last):
File "easy_feature_extraction.py", line 105, in
main(txt_path, out_path, n_jobs)
File "easy_feature_extraction.py", line 73, in mai…