-
I am having a lot of trouble with speaker diarization across lots of different platforms and models.
```toml
[target.'cfg(any(windows, target_os = "linux"))'.dependencies]
sherpa-rs = { version =…
-
Hi, thank you for your very nice work! I have rerun this project, and it has run 90K steps. the loss_id_psnt is around 0.07. And I tried to feed into a in-domain speaker's melspec and his speaker emb…
-
import logging
import math
import random
from dataclasses import dataclass, field, fields
import dataclasses
from math import pi
from typing import Sequence, Tuple, Union, Optional
import tor…
-
Hi,
I wanna use **wavlm** model to extract speaker embedding for speaker verification task. In [the paper](https://arxiv.org/pdf/2110.13900.pdf) it is mentioned that for the task of speaker verificat…
-
Hi,
I am trying out Open Voice (v1), and it mechanically worked, but the cloned voice is far from its reference speaker. Sometimes, I gave a male reference speaker mp3, and got back a female voice…
-
```python
from pyannote.audio.pipelines.utils.hook import ProgressHook
pipeline = Pipeline.from_pretrained("pyannote/speaker-diarization")
```
On my Mac, default `pipeline.embedding_batch_size…
-
Theoretically, the original speaker embedding information has already been contained in the spectrogram. The network will automatically squeeze the original speaker embedding information out after c…
nkcdy updated
3 years ago
-
Dear Developers!
In the demo_part1.ipynb, it is written that
```
Obtain Tone Color Embedding
The source_se is the tone color embedding of the base speaker. It is an average of multiple senten…
-
Hi.
I trained the flowtron on two speakers, for a total of 50 hours, 25 for each. After that, I wanted to train the model for 10 speakers for 20-30 minutes each using the basic checkpoint of the mode…
-
hello,
the provided vocoder checkpoint using mHubert does not support multi-speaker. Do you have a multi-speaker checkpoint?
`mhubert_vp_en_es_fr_it3_400k_layer11_km1000_lj`