Open Theking1313 opened 9 months ago
Try smaller models and if they work then this is probably a WhsiperX
issue, go to their repo and see if they support large-v3
.
For Whisper-timestamped
, you should see the cuda
option if you have a cuda device.
If there is no cuda option despite having a cuda device then you should check if Pytorch GPU is installed correctly and can see your cuda device by running
import torch
print(torch.cuda.is_available())
Turns out I was missing cuda, so managed to fix that,
Still working on the other issue.
I'm getting this error with large-v2:
TypeError: TranscriptionOptions.__new__() missing 6 required positional arguments: 'repetition_penalty', 'no_repeat_ngram_size', 'prompt_reset_on_temperature', 'max_new_tokens', 'clip_timestamps', and 'hallucination_silence_threshold'
I was trying to use "m-bain/whisperX" on large-V3 model with speaker labels. but get these errors:
Other models work well without issue such as faster-whisper or whisper.cpp.
Whisper-timestamped seems to have no cuda option despite mention it can run on gpu in git page.