-
First of all, thank you for great implementation of Tacotron 2. I'm trying to training Vietnamese language with my own dataset. The dataset is 22.05kHz, about 6 hours, already trimmed all silence. But…
-
Hi @r9y9, I'm training on German audio. I have added the german characters (Ä, Ö, Ü, ß, ä, ö, ü) to the symbolset and am using basic_cleaners.
The problem is the alignment on test-audio. Look at so…
-
Traceback (most recent call last):
File "synthesis.py", line 181, in
waveform = wavegen(model, length, c=c, g=speaker_id, initial_value=initial_value, fast=True)
File "synthesis.py", line …
-
It seems the decreasing of loss during WaveNet training unsteady. Is it all right for the results or should I wait more steps? The predicted wavs under `logs-WaveNet/wavs` sound OK but the ones under …
-
First of all, I'm really appreciate all of your job here, make this thing become wonderful repo. I'm a student and trying to use this implementation to synthesize Vietnamese for pet project. Because V…
-
Hi again,
I trained single Korean speaker successfully and moving to multiple Korean speaker. Again, I encountered such Assertion error as shown below. I tracked down and looks like self.encoder
…
-
I get
```
E:\Python\Projekte\tacotron(keithito)>E:\Python\Anaconda\python.exe preprocess.py --dataset ljspeech
20%|███████████████ | 2…
-
I'm using the default training dataset(LJSpeech-1.1) and got no GPU on my machine .The CPU training speed is about 7.182 sec/step, for fast envaluation ,how many steps do you think it would be OK to s…
-
(Questions in bold font)
I am trying to synthesize my own voice files.
As I did not find a documentation apart from `You can use other datasets if you convert them to the right format. See ljspeec…