Closed jerome83136 closed 7 months ago
Thanks for opening your first issue here! Be sure to follow the relevant issue templates, or risk having this issue marked as invalid.
The following is the upstream repo I mentioned on Discord: https://github.com/SYSTRAN/faster-whisper
The following is the upstream repo I mentioned on Discord: https://github.com/SYSTRAN/faster-whisper
Oh, sorry. I will then move my issue in this other repo. Thank you
Is there an existing issue for this?
Current Behavior
Hello all,
I'm running faster-whisper this way:
docker run -d --gpus all --runtime=nvidia --name=faster-whisper --privileged=true -e WHISPER_BEAM=10 -e WHISPER_LANG=fr -e WHISPER_MODEL=medium-int8 -e NVIDIA_DRIVER_CAPABILITIES=all -e NVIDIA_VISIBLE_DEVICES=all -p 10300:10300/tcp -v /mnt/docker/data/faster-whisper/:/config:rw ghcr.io/linuxserver/lspipepr-faster-whisper:gpu-version-1.0.1
Hardware:
It works fine and the container PID is affected to the GPU (nvidia-smi)
But after ~1h of inactivity I get the following Out Of Memory error: https://pastebin.com/raw/c3s4wYAm If I check nvidia-smi; I still see the container PID using ~1,3GB of memory (so: quite less than the 6GB available on the GPU)
Is there someone that could be so kind to point me to some fix ? When searching, I spotted this config: max_split_size_mb but I don't know if it could help and I reallydon't know how to apply it Am I using a too big model for my GPU ? Or should I reduce beams number ?
Thank you very much for your help Best regards
Expected Behavior
faster-whisper stays available on long term; without OOM error
NB: when problem occurs; nvidia-smi shows the container PID using 1,3GB of GPU memory (so there is still ~4.7GB available memory)
Steps To Reproduce
Environment
CPU architecture
x86-64
Docker creation
Container logs