Open edisonzf2020 opened 1 month ago
docker build self image=core , The q5_0 model is ok.
8:29AM INF Loading model 'ggml-large-v3-q5_0.bin' with backend whisper
8:30AM INF Success ip=127.0.0.1 latency="27.248µs" method=GET status=200 url=/readyz
8:30AM INF Success ip=192.168.1.28 latency=48.973942329s method=POST status=200 url=/v1/audio/transcriptions
Hello, I have the same issue with the aio docker image v2.16.0-aio-gpu-nvidia-cuda-12. This model doesn't work : https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-medium-q5_0.bin But this one works: https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-medium.bin
**LocalAI version: 2.15.0
quay.io/go-skynet/local-ai:master-cublas-cuda12-ffmpeg **Environment, CPU architecture, OS, and Version: ubuntu 22.04Describe the bug
To Reproduce
curl http://192.168.1.19:8090/v1/audio/transcriptions -H "Content-Type: multipart/form-data" -F file="@$PWD/gb1.ogg" -F model="whisper-1"
Expected behavior
Logs
Additional context