-
Hi, is it possible to extract what time (or where) the speech of each speaker start and end?
I want to extract speech of each speaker so it needs to know when the speech matched to the speakers and e…
-
- Changing `--n_mels` from 40 to 64 leads to a small increase in performance.
- Using `--log_input` also leads to a small increase in performance.
- Combining two loss functions (e.g. `angleproto` a…
-
We should start creating example recipes for some data sets and tasks. I'll post an initial list here, and we can modify or extend it based on discussions. I'll sort it by the level of implementation …
-
I'd like to thank all the contributors for their efforts. We know that DDP should be faster than DP, but how many times faster is DDP than DP in speechbrain? I mean that I use 8 RTX 2080 Ti GPUs in a…
-
In the tutorial, the AMI dataset is used to train speech activity and change detection. However, the voxceleb data set is used to train speaker embedding. Does the speaker embedding model necessarily …
-
The README mentions the argument `--config finetuning-base` in the fine-tuning step and a training configuration `configs/default.yaml` in the training step. I suppose the `config` directory was not c…
-
For this year's GSOC, TFDS is going forward with community datasets. However, I had some queries regarding it as I would like to contribute to TFDS as part of this year's GSOC.
1. The [community da…
-
**Describe the bug**
Running
```
pyannote-audio emb validate --subset=development --to=1 --parallel=8 /media/ben/datadrive/Software/pyannote-audio/data/ami/voxceleb_finetuneexp1/train/myexpt.Spea…
-
* Name of dataset: VoxCeleb
* URL of dataset: http://www.openslr.org/49/
* License of dataset: Not copyrighted
* Short description of dataset and use case(s):
This resource contains files for t…
-
**Describe the bug**
I can run validation on AMI as per tutorial, but when I run on my own fine-tuned training folder and protocol I get TypeError: get_protocol() got an unexpected keyword argument '…