-
If you have, for example, a Korean road name, speech synthesis using `AVSpeechSynthesizer` skips straight over this.
I assume that this is fixable with better annotation of the voice instructions tex…
-
**Describe the bug**
A call to `SpeechSynthesizer.StopSpeakingAsync()` does not stop synthesis for a very long time, up to 30 seconds. The log file is here: [speech.log](https://github.com/Azure-Sa…
-
Hello p0p4k,
I am deeply grateful for the code you have provided.
I have a question while adapting it to a Korean version. I am preparing to use a speech-to-text dataset with approximately 2000 …
-
hello @steveash!
I'm very interested in your project, so I checked your datasets of Korean/Japanese.
(I'm native speaker of Korean, and I'm doing research about Japanese transliteration in master c…
-
**Describe the bug**
from cosyvoice.cli.cosyvoice import CosyVoice
from cosyvoice.utils.file_utils import load_wav
import torchaudio
cosyvoice = CosyVoice('pretrained_models/CosyVoice-300M')
…
-
Can someone share a voice sample he created with this repository based on a given and/or a custom set of voice files
-
Thank you for creating a great repository.
I wonder why there is no bert when converting a pytorch model of MeloTTS to an Onnx model.
https://github.com/k2-fsa/sherpa-onnx/blob/963aaba82b01a425ae8…
-
Hello p0p4k,
I'm reaching out to you again with a question.
Thanks to your great help, I've successfully trained and inferred the Korean pflow model. During the inference process, I observed a f…
-
I would like to know how to train a glow-tts 2 model for another language, using another dataset which have the same structure as LJ Speech dataset?
Could you give some hints about how to train it or…
-
_Last updated: 3rd October, 2023_
A number of the APIs have voices available that we either don't know the IDs for (if closed source like TikTok for example) or we do know them but for some reason …