-
https://github.com/NVIDIA/waveglow/issues/54
In this issue, they were talking about lower some parameters to maximize inference speed. But I dont know how to do it properly, what can be reduced and w…
-
Hi, @dathudeptrai, thanks for great repo about Text to speech.
I have question about preprocess ljspeech. In .csv data has format
12_wav | | hai cái đầu sẽ nghĩ ra những cái mà một cái đầu không ng…
-
### Describe the bug
I am training fast_speech on the ljspeech en_US set using this [recipe](https://github.com/coqui-ai/TTS/blob/dev/recipes/ljspeech/fast_speech/train_fast_speech.py) (on dev bran…
-
No idea what could be causing it, but it seems to rely on speaker count: an 8-speaker model of mine had the inference IDs match the training ones, whereas a 20-speaker one is jumbled up. `dhama the ll…
Sobsz updated
2 years ago
-
Not sure how to share this, but feel free to add.
https://drive.google.com/drive/folders/1jrMWhkK3KrOKsbPMYtCCcPczPmf8V-eQ?usp=sharing
It uses the default English encoder. To adapt tacotron text p…
-
Hey guys,
first of all: thanks for the great work!
Then: I was training Tacotron2 with the provided /examples/tts/tacotron2.py on my own dataset.
The first ~13k steps everything seemed alright b…
-
Hi i couldn't understand how can i use pretrained model in hifigan. For example in tacotron i was running script like that
`python train.py --output_directory=outdir --log_directory=logdir -c pre-…
-
I'm trying to perform inference without a file for copying style from, just using this repo like a multi-speaker Tacotron2 without GSTs.
What code do I need for that?
In Tacotron2 you only need th…
-
Tacotron team has released Parallel Tacotron 2, with novel duration modelling technique.
```
This section introduces the proposed Parallel Tacotron 2 model,
specifically differentiable duration mod…
-
Hi Thorsten.
I have a question about tokenization based on phonemes ! I know this is not related to this repo but I would appreciate any help.
I have used your dataset and produced German text to …