-
It is said in your paper that
> We train our network using Adam optimizer with learning rate 2e-4 and batch size 20. The model converges in
> approximately 2 days using 2 TitanX gpus for Tai-Chi-…
-
Hello,
I am wondering if it is possible to fine-tune the pre-trained model using videos of two persons in much smaller quantities than VoxCeleb. Also, does it make sense to fine-tune on a single spea…
-
Hello,
I am pretty new to ESPnet and I am attempting to perform inference using the `vctk_tts_train_xvector_transformer_raw_phn_tacotron_g2p_en_no_space_train.loss.ave` pretrained model.
Steps …
-
I am trying the repo for the first time. While preparing the data I find that we need the text annotations of the voxceleb files. But I find the [dataset](https://www.robots.ox.ac.uk/~vgg/data/voxcele…
-
Hi, thanks for sharing this amazing repository.
When using `aamsoftmax`, I am getting this error for all of the models.
```
Save path: exps/nist_old_bang_youtube_aamsoft_rawnet
Embedding size is…
-
In DatasetHelper.md
I have downloaded the VoxCeleb dataset
and I successfully run the 3. Extract facial landmarks from videos.
But when I run the 4. Extract coefficients for videos
I met an error:…
qxzha updated
6 months ago
-
If I wanted to use one of the larger wespeaker models - say 293 - would I just download the .pt file and point to it in the config.yaml?
-
### Describe the bug
For a real time project, I need to convert [Spkrec-Ecapa-voxceleb](https://huggingface.co/speechbrain/spkrec-ecapa-voxceleb) to ONNX in order to use it in cpp with [OnnxRuntime…
-
```
requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/pyannote/wespeaker-voxceleb-resnet34-LM/resolve/main/speaker-embedding.onnx
```
The above exception …
-
Hello author, I would like to ask you what the three files in the open-source VQVAE code represent? ([2022-xx-xx/xx-xx])