-
First of all, thank you for your excellent work,The data set I currently have contains 8 speaker, each of which is expected for about 20 minutes. Train according to your method and the results of mode…
-
I read you can't publish your multi-speaker corpus.
so I want to train multi-speaker model with my corpus.
how to setup dataset in multi-speaker?
put one traing_data?
and one more question…
-
Hello,
I have a question regarding the current pitch models, specifically the differences between Reflow and DDPM. With the latest update, it seems like Reflow has become the new default and recomm…
-
### Describe the bug
I was trying to record screen with only systems audio using quicktime player and blackhole audio driver.
Created a multi-output Device as primary audio output device to send a…
-
As discussed in Issue #4 speakers are not able to create multi-speaker events?
manno updated
12 years ago
-
Hi ,
I am using conformer transformer AED model for my custom ASR. I have used conformer encoder with 256 dim, 8 attention heads and 12 encoder layers. Decoder with 6 layers with 8 attention hea…
-
你好,关于multi speaker的tts是基于迁移学习的吗
-
# Task Name
Speaker Diarization with ASR
[Description]: To do multi-speaker ASR where each speeches may have overlap.
## Task Objective
Most of the time, we do ASR on audio with only one main sp…
-
Hi,
We are trying to train a multi-speaker model starting from the LibriTTS data and using the latest FastPitch commit. We selected the 50 speakers which have the most utterances in the dataset, an…
-
Hi @KdaiP nice work, just like to know is this architecture is intended to support zero-shot TTS or normal multi-speaker kind of TTS,