Open BriceGUILLAUME opened 8 months ago
@BriceGUILLAUME I believe the clip model we're using expects png bytes only, is webp supported by the OpenAI api?
@abetlen, thank you for your answer and for the awesome projet as well.
I have managed to get answers from the Azure OpenAI with webp so I guess that this is supported. However I will make a try with png data to see if the issue happens as well, and I will post the results here maybe tomorrow.
@BriceGUILLAUME looks like you're right https://platform.openai.com/docs/guides/vision/what-type-of-files-can-i-upload
The fix would be to use something like Pillow to convert the image in the chat handler which is do-able but I'll keep it as an optional dependency for the vision chat handlers.
Hey has anyone being assigned to fix this issue ?
Expected Behavior
I was trying to integrate llama-cpp-python with LibreChat, has both use OpenAI API to communicate. The API implementation authorizes the sending of images with
image_url
that can be embedded in the request body. I was expecting the image to be analyzed by llava but I have an decoding error instead.Current Behavior
The request is correctly received by llama-cpp-python but I have the following error in logs:
Environment and Context
Linux *** 6.2.0-1019-azure #19~22.04.1-Ubuntu SMP Wed Jan 10 22:57:03 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
Steps to Reproduce
I run a docker instance of llama-cpp-python behind a reverse proxy (for SSL handling) and I tried to send a completion request from LibreChat. Here is the request intercepted by my nginx reverse proxy:
As you can see the
image_url
is send in base 64 but somehow there is an issue when it is being decoded.I have tried to see in the code where this action is happening but I would be glad if someone could help me find where this is comming from.
Thank you for your support!