BerriAI / litellm

Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
https://docs.litellm.ai/docs/
Other
13.62k stars 1.59k forks source link

[Feature]: Phi-3-vision and other VLM Support on Nvidia NIM (ai.api.nvidia.com Endpoints) #4780

Open BillJones-SectorFlow opened 3 months ago

BillJones-SectorFlow commented 3 months ago

The Feature

Support was added for all Nvidia NIM LLM models (integrate.api.nvidia.com addresses), but so far, I don't believe support was added for the VLM models (ai.api.nvidia.com addresses).

Examples (each endpoint listed below links to the associated Nvidia build):

https://ai.api.nvidia.com/v1/vlm/microsoft/phi-3-vision-128k-instruct: https://ai.api.nvidia.com/v1/vlm/google/paligemma https://ai.api.nvidia.com/v1/vlm/nvidia/neva-22b

It would be great if support for these could be added.

Thanks!

Motivation, pitch

Since support for regular Nvidia LLMs has been added, it would be good to also have support for their set of VLMs.

Twitter / LinkedIn details

No response

BillJones-SectorFlow commented 3 months ago

Does anyone have a workaround for this, possibly?