getomni-ai / zerox

PDF to Markdown with vision models
https://getomni.ai/ocr-demo
MIT License
6.58k stars 358 forks source link

supporting Gemini #74

Open koreanssam opened 3 weeks ago

koreanssam commented 3 weeks ago

`model = "gemini/gemini-1.5-flash-002"``


2024-10-25 23:04:56,975 - INFO - HTTP Request: POST https://generativelanguage.googleapis.com/v1beta/models/gemini-1.5-flash-002:generateContent?key=secret^^ "HTTP/1.1 200 OK" 2024-10-25 23:05:16,321 - INFO -

LiteLLM completion() model= gemini-1.5-flash-002; provider = gemini 23:05:16 - LiteLLM:WARNING: vertex_ai_non_gemini.py:198 - No text in user content. Adding a blank text to user content, to ensure Gemini doesn't fail the request. Relevant Issue - https://github.com/BerriAI/litellm/issues/5515 2024-10-25 23:05:16,332 - WARNING - No text in user content. Adding a blank text to user content, to ensure Gemini doesn't fail the request. Relevant Issue - https://github.com/BerriAI/litellm/issues/5515

How can I use Gemini?

pradhyumna85 commented 2 weeks ago

@koreanssam, Gemini models are already supported (refer #69), what you see are warnings so it should be fine, could you print the output of zerox api to check if you are getting some sensible output. Also I am assuming that you are setting correct api keys are per example:

os.environ["GEMINI_API_KEY"] = "your-api-key"