-
Another small issue. Sometimes the transcriptions are all shifted significantly forward in time. So that the transcription occurs seconds before the speech. This usually adjusts itself later in the tr…
-
Hi,
I am trying to run en model
`python -m TTS.bin.synthesize --text "Hi, how are you?" --model_path en/fastpitch/best_model.pth --config_path en/fastpitch/config.json --vocoder_path en/hifiga…
-
Additionally, I want to ask why some losses increase during training. Should I choose the last checkpoint or the best checkpoint?
-
Udało mi się odtworzyć dźwięk, natomiast jest on tylko z pierwszej pary głośników 2.1
0dBFS level = 32768.0
--Csound version 6.10 (double samples) Dec 21 2017
[commit: 2b6f2d970da3d1abf09b70bd02…
-
参考 https://github.com/PaddlePaddle/PaddleSpeech/tree/develop/demos/text_to_speech 的 multi-speaker 命令,但是遇到以下错误
**Describe the bug**
(venv) λ paddlespeech tts --am fastspeech2_mix --voc hifigan_csms…
-
### The problem
I have a large song library on my NAS and i'm trying to play music to my sonos speaker using the media_player integration. Some of the songs when played with home assistant dont conta…
-
Hi.
I am exploring about speed of training and inference different multi speaker TTS models on single CPU or on singe GPU.
Thanks for any explanation in this case for current model or any other mode…
-
Hello!
I have used xTTS v2 for a while and made great voices.
I sih to know one thing:
every voice made, when it "speaks" has the same cadence and pronounciation (clearly from a trained model).
Ho…
-
Hi,
I was going through your repository. I could not find results of LSTM and BLSTM on the 2 speaker .wav (audio files) generated by you. Can you please add them?
Also, have you tried this algorit…
-
Just a suggestion - it would be great if stable-ts did speaker diarization too so that we have accurate timestamps for multi-speaker audio.
Thanks!