-
Hi I have a audio data like this:
`audio = np.frombuffer(audio_msg.data, dtype=np.int16)`
```
self._oww = Model(
wakeword_models=[self._OWW_MODEL_PATH],
infe…
-
I tested some videos
if the silence duration is long , then enable vad_filter will be effective
but if video is as normal, then enable vad_filter may cause more timestamp mismatch
is there …
-
I'm trying to debug an issue with mobile use of an STT pipeline. Basically VAD creates audio arrays, and Whisper running via Transformers.js then transcribes that audio.
This works great on desktop…
-
First of all, thanks for this project, it's very easy to set up and run locally.
Transcribing on this webui, the large-v2 model skips the first three sentences in a file I tested, just like what ha…
-
## 🐛 Bug
qwen-audio + vad 运行报错
### To Reproduce
1. Run cmd `python qwen_demo.py`
2. See error
```
2024-05-14 11:09:35,110 - modelscope - INFO - PyTorch version 2.3.0 Found.
2024-05-14 11:…
-
Would it be possible to add a similar option to stable-ts that Whisper-WebUI (https://gitlab.com/aadnk/whisper-webui) uses to split up the audio into segments of perceived audio using VAD?
The main…
-
I am trying to access the content of ProcessHeaps without success, I am using volatility3 Framework 2.5.2, and to test and be sure that all the code works and that only the ProcessHeaps part fails me,…
-
Nice work! Any chance of also porting functions from [here](https://github.com/ggerganov/whisper.cpp/blob/c23588cc4b2c4be4ba3634a21c7f2b302c5f370e/examples/common.cpp#L126) specifically vad_simple? Th…
-
-
# Project Status Overview
### Project Objective:
- Incorporate color tokens into our components as well as allow color tokens to be used standalone by the app. Inherit color tokens from VADS where…