-
@bigpon Hi
I'm trying to reproduce the denoising code.
https://github.com/facebookresearch/AudioDec?tab=readme-ov-file#bonus-track-denoising
You mentioned following the requirements in `submit_den…
-
Hi, I want to know what config does you use when you train the hifigan model of DataBaker to get the samples in the webset https://light1726.github.io/vaenar-tts/.
With these parameters clearifie…
-
Dear,
Thank you for your great works.
I followed the procedure to train our TTS by the fs2 + hifigan joint training approach using the corresponding pretrained models with the config provided in …
-
Hi
## The issue
I trained a model based on [Matcha TTS](https://github.com/shivammehta25/Matcha-TTS), and I tried to use Vocos with it. Unfortunately, vocoding using a checkpoint trained with th…
-
import logging
import math
import random
from dataclasses import dataclass, field, fields
import dataclasses
from math import pi
from typing import Sequence, Tuple, Union, Optional
import tor…
-
Hi,
I recently discovered in experiments that the audio generated using fluentspeech is also different from the original audio in the non-modified parts. Here are the waveforms of the unmodified port…
-
Can you provide me with the code to convert wav to wav to do it the indirect way?
-
Hello!
I'm trying to FineTuning HiFi with GlowTTS npy
i generate npy with this code:
```
def TTS(tst_stn, path):
if getattr(hps.data, "add_blank", False):
text_norm = text_to_sequen…
-
hello, guys, I have a question here:
If my data **sample_rate **is **16k**, and I want use this **16k** data to train model . How can I modify parameter ? and the model of **Hifi-Gan**, how can get…
-
I have tried to inference mel-spectrogram with my own pre-trained model. But I don't understand why with the same text the mel-spectrogram inferred by tacotron have different size
For example: I ha…