assafelovic / gpt-researcher

LLM based autonomous agent that conducts local and web research on any topic and generates a comprehensive report with citations.
https://gptr.dev
Apache License 2.0
15k stars 2.01k forks source link

LLM RATE LIMIT WITH GROQ MODELS #985

Closed Bless50 closed 2 days ago

Bless50 commented 1 week ago

Describe the bug warnings.warn( Error in generate_report: Error code: 413 - {'error': {'message': 'Request too large for model llama-3.1-70b-versatile in organization org_01hw8f0aa0ekmrn70ffm2zkdv1 on tokens per minute (TPM): Limit 6000, Requested 14368, please reduce your message size and try again. Visit https://console.groq.com/docs/rate-limits for more information.', 'type': 'tokens', 'code': 'rate_limit_exceeded'}}

To Reproduce Steps to reproduce the behavior:

  1. Go to '...'
  2. Click on '....'
  3. Scroll down to '....'
  4. See error

Expected behavior A clear and concise description of what you expected to happen.

Screenshots If applicable, add screenshots to help explain your problem.

Desktop (please complete the following information):

Smartphone (please complete the following information):

Additional context Add any other context about the problem here.

ElishaKay commented 1 week ago

Welcome @Bless50,

Try changing your LLM config variables:

FAST_LLM="ollama:qwen2:1.5b"
SMART_LLM="ollama:qwen2:1.5b"
EMBEDDING="ollama:all-minilm:22m"

OR: llama3.1

Looks from the error message that the rate limit on the API is shared across anyone else in your organization using the same model