-
你好
感谢你的这个工程,为什么你的推理结果会有三个人?如果我的使用情况只有2个人,应该怎么做?是导出onnx 模型需要配置参数吗?
-
Hello, your job is very interesting. Regarding the issue of emotional leakage, may I ask if you labeled the issue of emotional leakage with accents or speakers?
-
The extract_features method of the ContentVec model (not the legacy one) needs speaker embedding as a parameter.
So do I have to get the speaker embedding (with Resemblyzer) for every audio sample I …
-
Hello,
I am using **diart.stream microphone** from command line for inference but it is **not detecting for more than 2 speakers** even if there are.
For ex. if I play the recording of 3 people …
-
**Describe the bug**
The implementation of Global Style Tokens (GSTs) in FastPitch introduced in #6417 does not follow the prescription of the original [paper](https://arxiv.org/abs/1803.09017). In…
-
I can get 23 seconds out of it but you'd think that it would be possible.
Also are the emotional stresses automatic?
-
Hello,
I'm currently working with two systems, each equipped with different graphic card drivers supporting CUDA 12.2 and 12.3, respectively, utilizing Tesla M40 and RTX 2060 Super. However, it see…
-
Since its introduction in `pyannote.audio` 3.x, the ONNX dependency seems to cause lots of problem to `pyannote` users: #1526 #1523 #1517 #1510 #1508 #1481 #1478 #1477 #1475
WeSpeaker does provi…
-
### Describe the bug
When using the xtts-1 model on windows (python 3.11.6), every time I run the `tts_to_file` function, it gives the error `TypeError: Invalid file: None`
### To Reproduce
On wind…
-
In was facing an issue of vad progress being stuck in diarize_parallel.py [Discussion link](https://github.com/MahmoudAshraf97/whisper-diarization/discussions/36#discussioncomment-7292384)
Which did …