Closed leviethung2103 closed 8 months ago
Thank you for your issue.You might want to check the FAQ if you haven't done so already.
Feel free to close this issue if you found an answer in the FAQ.
If your issue is a feature request, please read this first and update your request accordingly, if needed.
If your issue is a bug report, please provide a minimum reproducible example as a link to a self-contained Google Colab notebook containing everthing needed to reproduce the bug:
Providing an MRE will increase your chance of getting an answer from the community (either maintainers or other power users).
Companies relying on pyannote.audio
in production may contact me via email regarding:
This is an automated reply, generated by FAQtory
Hello,
The pyannote/voice-activity-detection
uses pyannote/segmentation
(2.1), as you can see in the config.yml on huggingface :
pipeline:
name: pyannote.audio.pipelines.VoiceActivityDetection
params:
segmentation: pyannote/segmentation@Interspeech2021
So, this is not the same segmentation model as in pyannote/segmentation-3.0
, which explains why you have different results.
Have a nice day!
Thank you for your help.
Hello,
I am wondering what is the difference between the
pyannote/voice-actitivy-detection
andpyannote/segmentation-3.0
pyannote/voice-activity-detection: https://huggingface.co/pyannote/voice-activity-detection pyannote/segmentation-3.0: https://huggingface.co/pyannote/segmentation-3.0
In the segmentation-3.0, there is voice activity detection part.
In the pyanote/voice-activity-detection
I've tested these two methods and got different results.
Could you please tell the reason why? Are they using the same vad models ?
Thank you