-
The text for the transcript and the audio transcript can be found below :
I used both mfa validate and mfa align but they don't give any errors. Why is this happening ?
example 1 : change in wor…
-
Hi
I am using stt_en_citrinet_1024 model and able to get good transcript , I am using the recorded audios with microphone and WER is varying from 3.5% to 15%. This has names of person and place, ho…
-
Hi @archiki,
I appreciate this work very much, and thanks for providing the implementation. Could you please tell me that how long does the training cost? BTW, did you take the model checkpoint trai…
-
-
see discussion here: https://github.com/facebookresearch/hydra/issues/1352
the basic issue is that the best parameters recommended by ax (as exposed by hydra integration) are often, in fact, not be…
-
I want to use wav2letter to adapt it to my training dataset. I am following the colab notebook FineTuneCTC.ipynb which uses
./flashlight/build/bin/asr/fl_asr_tutorial_finetune_ctc model.bin
I ha…
-
-
Hi,
I'm trying to train a multi-speaker Tacotron model from scratch using VCTK + LibriTTS databases. The model trains fine until about 50K global steps but after that I start running into "CUDA out…
-
Please check whether this paper is about 'Voice Conversion' or not.
## article info.
- title: **Defense for Black-box Attacks on Anti-spoofing Models by Self-Supervised
Learning**
- summary: High-…
-
Thank you for your code. I have a question about function mvdr.
https://github.com/snsun/cgmm_mvdr/blob/cfc88aec4e26a6536eb4642566db9701ee9d6369/test.m#L37
According to your description of functio…