Open koreanssam opened 3 weeks ago
@koreanssam, Gemini models are already supported (refer #69), what you see are warnings so it should be fine, could you print the output of zerox api to check if you are getting some sensible output. Also I am assuming that you are setting correct api keys are per example:
os.environ["GEMINI_API_KEY"] = "your-api-key"
`model = "gemini/gemini-1.5-flash-002"``
2024-10-25 23:04:56,975 - INFO - HTTP Request: POST https://generativelanguage.googleapis.com/v1beta/models/gemini-1.5-flash-002:generateContent?key=secret^^ "HTTP/1.1 200 OK" 2024-10-25 23:05:16,321 - INFO -
LiteLLM completion() model= gemini-1.5-flash-002; provider = gemini [92m23:05:16 - LiteLLM:WARNING[0m: vertex_ai_non_gemini.py:198 - No text in user content. Adding a blank text to user content, to ensure Gemini doesn't fail the request. Relevant Issue - https://github.com/BerriAI/litellm/issues/5515 2024-10-25 23:05:16,332 - WARNING - No text in user content. Adding a blank text to user content, to ensure Gemini doesn't fail the request. Relevant Issue - https://github.com/BerriAI/litellm/issues/5515
How can I use Gemini?