-
See here...start thinking about true batching.. 😉
https://github.com/shashikg/WhisperS2T/issues/33
-
Hey all, after a nice conversation with @MahmoudAshraf97 on a different repo I wanted to share some of my benchmark data. This was created using an RTX 4090 on Windows, no flash attention, with 5 be…
-
Hey @shashikg great repo and cheers to the insane efforts in building this repo.
I have a finetuned whisper model (both in original openai and HF formats ) which I want to use in TensorRT backend …
-
Hi @sanchit-gandhi !
I'm in the process of integrating multiple whisper backends into a unified package that includes VAD-based chunking. During testing, I observed significantly higher inference t…
-
**Describe the solution you'd like**
With the recent [TensorRT-LLM support for Whipser](https://github.com/NVIDIA/TensorRT-LLM/tree/main/examples/whisper), and now that PyTriton supports TensorRT-LLM…
-
In CTranslate2 Whisper model, batch generate is not faster than looping one by one. I tried the same thing on Translator model and it shows batching is far superior (a lot faster). I used Whisper smal…
-
When my script batch processes a bunch of audio files using the approach you gave me to use a list of files and their settings when processing, if a single file fails for any reason, it prevents the t…
-
I noticed that the latest benchmark from whisper.cpp is from February 2, 2023 and presumably the others are equally outdated. Can someone please do updated benchmarks for the three backends tested? …
-
Thanks for putting so much work into this, its so polished already!
Just want to understand if speaker diarization is something planned in the future?
Thanks!
-
Hi WhisperX Team,
I was wondering if you consider support https://github.com/NVIDIA/TensorRT-LLM/tree/main/examples/whisper the tensorrt-llm backend of whisper. I have done several benchmark test us…