-
Dataloader name: `cmu_wilderness_multilingual_speech_dataset/cmu_wilderness_multilingual_speech_dataset.py`
DataCatalogue: http://seacrowd.github.io/seacrowd-catalogue/card.html?cmu_wilderness_multili…
-
Hi,
What type of server config would it need to process/decode 10 concurrent speech recognition? How many cores and ram? Not training only for decoding.
Thanks
-
We repeatedly get questions in the forum about how to decode just one file.
It would be good if we had a documentation page or sub-page about this, that would describe the various steps required to do…
-
Hi alumae,
thank you for sharing this code. i was able to work through your examples.
i am new to kaldi and the gstreamer plugin.
I have a question though. what is the reason behind choosing …
-
-
To be updated.
-
I'm writing a native application (Rust, C/C++) that needs to perform forced alignment. Not training, just the alignment part. So I'm wondering how best to integrate MFA into an application.
My unde…
-
Thank you for the excellent work!
In order to install samia-speech on a RPi4 with Raspbian Buster, the changes below were made to the steps in https://github.com/gooofy/zamia-speech#raspbian-9-stre…
-
I have been using `egs/wsj/s5/steps/segmentation/detect_speech_activity.sh` and a pre-trained model to do SAD on recordings. One thing I noticed is that the resulting segments may be consecutive. As a…
-
I am using the standard python web socket server implementation (vosk-server/websockeet/asr_server.py). I am running with the (large) English model available from http://alphacephei.com/kaldi/vosk-mo…