-
Hello,
I would like to connect a Tacotron2 model to LPCNet.
Is there a way to convert the 80-mel coefficients (output of Taco2) into the 18 Bark scale + 2 pitch parameters (input of LPCNet) ?
…
-
### Describe the bug
I met an error when I run the train_tacotron_ddc.py in TTS/recipes/ljspeech/tacotron2-DDC with the default config. The error and the config are as follows:
ERROR
![6208dece60…
lyjgo updated
2 weeks ago
-
Thank you for such an exciting project.
"To start, double click the xVASynth.exe file, and make sure to click Allow"
Where is xVASynth.exe file?
-
I want to deploy this model in the tensorflow serving, but saved-model is needed. It seems complex and difficult for me to figure the raw inputs and final outputs of the model, I think in the tensorfl…
-
Hey everyone,
I am working with **Hindi** Language. My dataset is pretty small right now(about 700 clips) and the speaker is a male. I wanted to check if a reasonable output can be produced using …
-
It's really amazing that Glow-TTS does such a good job. I have some confusion about the decoder framework:
There is no post-net at the end of the decoder. I understand that the invertible flows requ…
-
Whenever I preprocess the custom dataset, this is the output:
`C:\Users\Graham\Desktop\Lip2Wav-master>python preprocess.py --speaker_root Dataset/larry --speaker larry
C:\Users\Graham\AppData\Loca…
-
Hi,
I trained tacotron on a custom dataset, about 3700 samples, about 3.8 seconds each for 676k steps.
The loss has came down to ~0.4706
The model can output words that kind of sound like wor…
-
What is the relationship between the alignment and the synthesis quality? In particular, my alignment looks bad, but the synthesized utterances sound quite good. My training set is relatively small --…
-
I've trained WaveRNN on LJSpeech dataset with mel as condition. When generating waves, there are some bad cases occasionally shown in the following pictures.(They are the same sentence generated at di…