-
**Is your feature request related to a problem? Please describe.**
It can be useful to modify audio passed to STT plugins to remove silence and normalize audio levels for better accuracy. There are a…
-
-
Hi,
Tried executing your project ! Able to run all the python files without any errors, but how can I create a model for the same ? And no Model folder is generated inside the current working direc…
-
I have many audio files with human speech.
I want to group it by speaker.
For the test I get one long file (about 18 minutes) and get embedings for it (about 80 vectors). It meas each vector has abo…
-
1. text
- paraphrase
- q/a
- translation
- summarization
2. image
3. audio
-
https://github.com/alphacep/vosk-api/
This project has a libre, offline-capable speech recognition engine that's over 90% accurate - perfect for autoedit!
Vosk is a speech recognition toolkit. T…
-
I have tried
`python speakerlab/bin/infer_sv.py --model_id $model_id --wavs input.wav`
This exports the numpy array file. How can I get the inference info from trained model that this object is…
-
Hi,
I'm not so much into the details of whisper or whisper.cpp and I don't know if it is currently even possible with the foundation, but it would be nice if speakers could be marked or speaker-cha…
-
Hello,
I have a speech recognition project (written in tensorflow), and I would like to know how could I be able to use your trained model inside my project. My initial idea is to use the layer be…
-
Dear all,
CoreNLP 3.7.0 with the cleanxml annotator apparently fails to remove tags in certain conditions. The problems I encountered so far resulted in the following tokens:
```
=_<
~_<
-_<
x…