Closed eddieoz closed 3 months ago
You need to create a similar function to that which checks that the vLLM server is running. I am not too sure what this endpoint would be looking at the Deepgram docs but if they had some thing that returned "ready" that would be ideal. But maybe you can hit the listen or speak endpoint and see if the response is a 200? Not the best way I know
My workaround was something like this:
stt = DeepgramSTTService(
name="STT",
api_key='<key>',
url='https://api.deepgram.com/v1/listen'
)
tts = ClearableDeepgramTTSService(
name="Voice",
aiohttp_session=session,
api_key='<key>',
voice=deepgram_voice,
base_url="https://api.deepgram.com/v1/speak"
)
I tried to bring the key with get_secret
from my dashboard, but it was not working. Than I hardcoded the key and somehow it worked.
It is working fine by consuming on their API.
If you are using the local Deepgram model, the API key should be set to "None".
It sounds like you are maybe calling the deepgram models while the models are still connecting and then causes it to error. If you set min replicas to 1 and then call it for the first time does it work?
Closing this issue as it is a month old
After sucessfully building the project, I'm getting an error when running it:
Live Replica View log:
Run log
I believe both errors are connected because on the code it adresses:
main.py
How can I check if the deepgram server is running ok locally, or debugging it?