-
Need to at least expose a spectrogram audio feature extraction. There is code for mel spectrogram in emlearn, so that would be the primary candidate https://emlearn.readthedocs.io/en/stable/eml_audio.…
-
Hi,
I am a newbie, and I want to know how can I use this library to extract the visual facial embedding of videos so that I can use it for a real-time speech separation model so that it have the tem…
-
-
### First Steps Update
- [x] **Project Initialization**: Set up the repository and created the initial README file.
- [ ] **Data Organization**: Collected and organized Telugu music into four catego…
-
When I am using the pipeline, I meet an error: KeyError: "Unknown task depth-estimation, available tasks are ['audio-classification', 'automatic-speech-recognition', 'conversational', 'feature-extrac…
-
**Description:** I intend to work on implementing the MFCC for feature extraction from audio signals. The MFCC is a popular technique for extracting features from audio signals such as voice or music,…
-
Using a distributed database (https://github.com/orbitdb/orbit-db) it should be possible for every user to contribute to a cache of audio data.
This would start with caching and sharing the wavefor…
-
If I'd like to use CMG on my own dataset (for video and audio), how should I prepare the data? I've got video-audio pairs, whether should I extract their features? If yes, what feature extraction mode…
-
Are there any plans to support torchaudio, such as StreamReader and StreamWriter classes
-
The feature extract used 'MFCC',however, It appears that TFLM (TensorFlow Lite for Microcontrollers) does not support the MFCC (Mel-Frequency Cepstral Coefficients) operator.How can I use it on TFLM?