farizrahman4u / loopgpt

Modular Auto-GPT Framework
MIT License
1.43k stars 131 forks source link

browser.py call to summarizer exceeds token limits #38

Open jtac opened 1 year ago

jtac commented 1 year ago

Please check that this issue hasn't been reported before.

Expected Behavior

even if a scraped page exceeds token limits, it should be properly chunked?

Current behaviour

NEXT_COMMAND: browser, Args: {'url': 'https://www.soundguys.com/sony-wf-1000xm4-review-31815/', 'question': 'Extract specs, prices, and reviews for Sony WF-1000XM4.'}

SYSTEM: Executing command: browser Summarizing text...: 0%| | 0/2 [00:00<?, ?it/s] Summarizing text...: 50%|█████ | 1/2 [00:00<00:00, 2.66it/s] Summarizing text...: 50%|█████ | 1/2 [00:00<00:00, 1.45it/s] SYSTEM: browser output: An error occurred while scraping the website: This model's maximum context length is 8192 tokens. However, your messages resulted in 15534 tokens. Please reduce the length of the messages.. Make sure the URL is valid.

Steps to reproduce

run the browser command with arguments to repeat

Possible solution

check chunking logic, maybe getting bypassed after soup handles the content <> links splitting

Which Operating Systems are you using?

Python Version

LoopGPT Version

feature/azure_openai 0.0.13

Acknowledgements