-
Awesome work!
I have read your paper and noticed that your training data is extracted from VGG2 and Voxceleb, and substantial efforts have been put into it. Could you please let me know if there are …
-
Hello! I would like to use WhisperX and Pyannote to combine automatic transcription and diarization. I can do it on Colab using the Huggingface (HF) token, but I would like to avoid entering the HF to…
-
Thanks for your excellent work.
I have a question about the experiment in section 4.3. As far as I know, Voxceleb is a monocular dataset. How did you evaluate Novel View Synthesis quantitatively in t…
-
Hi there, I am following your work. It impresses me very much.
I used your scripts to download taichi and voxceleb dataset, and met 2 problems.
- 1. some links broken, I have downloaded 476 tra…
-
- The ECAPA-TDNN model trained on "development part of the `VoxCeleb2`".
- The datasets contains 5994 speakers.
- Some speakers have many more files than other speakers (`imbalanced classes)`.
- …
-
Having an explicit script to do this would help people downloading this dataset convert the files ready to be used in pyannote audio
-
WWW
Authors in [Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation](https://arxiv.org/abs/2202.09082) use the UASpeech dataset to contruct their experi…
-
In the file `kaldi/egs/chime6/s5_track2/local/train_diarizer.sh` the comments suggest to use `local/make_voxceleb1_v2.pl` when voxceleb was downloaded recently.
https://github.com/kaldi-asr/kaldi/…
-
I tried to run [voxceleb recipe](/kaldi-asr/kaldi/tree/master/egs/voxceleb/v2) but I encountered a problem with the [VoxCeleb1 dataset](http://www.robots.ox.ac.uk/~vgg/data/voxceleb/vox1.html). The fo…
kukas updated
4 years ago
-
You said "Since Voxceleb2 official videos are of low resolution, we redownload the original YouTube videos and re-crop the videos" in the paper, but where is the youtube link of these videos?