-
Would it be possible to publish trained models for other languages...like Japanese and mandarin.
Would such models be useful in general speaker isolation for unknown speakers sequence not in traini…
-
## Describe the question
when I tested my own datasets which haven't been trained ,the result is really bad,can you give me some suggestion?
## My background
Have I read the `README.md` file?
…
-
Sir,
I've trained the d-vectors by your code, but it takes much more time than UIS-RNN's own test sequence which have the same sequence length and dims during each iteration. The phenomenon is ju…
-
## Describe the question
A clear and concise description of what the question is.
**Summary of work:**
Audio signal is transformed into frames (log-mel-filterbank energies features) with frame widt…
-
@wq2012
The overlap rate seems strongly influenced the number of speakers.
Since when overlap size is larger, the speaker embedding will change more smoothly, and the change points will hard to det…
-
Hi,
I am working on training a uis-rnn model with dataset `voxceleb2: http://www.robots.ox.ac.uk/~vgg/data/voxceleb/vox2.html`.
- Step 1: use my embedding model to generate npz files.
- Step 2: …
-
## Describe the bug
I have tested with my audio file for speaker Diarization which is not accurate. i have [attached audio file](https://drive.google.com/open?id=1Hw4y4NhFh0MW4EazfF_CxZvAOBxTqKw_)(sp…
-
sir i have one doubt.
is this d-vector embedding PyTorch implementation of "Generalized End-to-End Loss for Speaker Verification" is moreover common as UIS-RNN or not?
accuracy for d-vector emb…
-
In your code, librosa.effects.split(utter, top_db=30) for d-vector and webrtcvad for uis-rnn.
-
Sorry if this sounds like a dumb question. I am not an expert in eighter python or speaker diarization. After I have trained the model, how can I use it to determine how is speaking from a wave file…