-
### Description
Hi, most of the time when driving I listen to music. When osmand has something to say, the music volume decreases during that, so audio ducking works. However, even when the music i…
-
model = stable_whisper.load_model('small')
result = model.transcribe(file)
result.to_srt_vtt('audio.vtt', False, True)
for caption in webvtt.read('audio.vtt'):
print(caption.start +" "+caption…
-
-
Here is the document:
> Bark can generate all types of audio, and, in principle, doesn't see a difference between speech and music. Sometimes Bark chooses to generate text as music, but you can hel…
-
## Description
i'd like to ask about a new 2nd 2-stems pretrained model , - spliting for speech-voice and music! It's highly wanted for audiobooks, - very often needed to split audiobook's stream!…
DI555 updated
4 years ago
-
**Which OS are you using?**
- OS: [Ubuntu 24.04]
I am trying to transcribe videos meetings recorded in serveral formats, mostly mkv and mp4. Some of them are transcribed good, some excellent, de…
-
https://github.com/balisujohn/tortoise.cpp is what we're about to use.
Is there anything better? Report as comment in this issue.
-
We have lost synchronisation between Principles and Technics examples for Additional Accessibility Information.
### Current examples in Principles :
- Descriptives
- Content is enhanced wit…
-
**Help !!**
Wtf is the plot ,, need TLDR and cohesive **_"Autism to coherent speech"_** Text .
Thanks !
- Burger from Music composer department
xrt1 updated
5 months ago
-
Thanks for your contributions to the opensource community. There is some confusion about the training code. In `anygpt/src/stage1_pretrain.py`, I only find that the image/speech/music data is loaded, …