Port of OpenAI's whisper model.
Clone the repository and instantiate it.
LocalPreferences.toml
file (either AMDGPU
or CUDA
) if using GPU for inference.
julia> using AMDGPU # If using AMDGPU for inference.
julia> using CUDA # If using CUDA for inference.
julia> using Whisper, Flux
julia> Whisper.transcribe( "./input.flac", "./output.srt"; model_name="tiny.en", dev=gpu, precision=f16)
julia> Whisper.transcribe( "./input.flac", "./output.srt"; model_name="tiny.en", dev=cpu, precision=f32)
**Multilingual support**
To perform transcribtion from non-English language,
specify `language` argument (optional) and drop `.en` from the model name.
```julia
julia> Whisper.transcribe(
"ukrainian-sample.flac", "./output.srt";
model_name="medium", language="ukrainian", dev=cpu, precision=f32)
To see what languages are supported, execute:
julia> values(Whisper.LANGUAGES)
.flac
with 1 channel and 16k sample rate.ffmpeg
which must be installed on your system and accessible from PATH.