RateLimitError: Error code: 429 - {'error': {'message': 'Rate limit reached for model llama3-70b-8192 in organization org_01hrx3emwtett8bq1cyh7w230q on tokens per minute (TPM): Limit 6000, Used 0, Requested 6194. Please try again in 1.94s. Visit https://console.groq.com/docs/rate-limits for more information.', 'type': 'tokens', 'code': 'rate_limit_exceeded'}}
Traceback:
File "/usr/local/lib/python3.11/site-packages/streamlit/runtime/scriptrunner/script_runner.py", line 600, in _run_script
exec(code, module.dict)File "/app/app.py", line 52, in <module>
result = melody_crew.run()
^^^^^^^^^^^^^^^^^File "/app/crew.py", line 59, in run
return crew.kickoff()
^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/crew.py", line 252, in kickoff
result = self._run_sequential_process()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/crew.py", line 293, in _run_sequential_process
output = task.execute(context=task_output)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/task.py", line 173, in execute
result = self._execute(
^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/task.py", line 182, in _execute
result = agent.execute_task(
^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/agent.py", line 221, in execute_task
result = self.agent_executor.invoke(
^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain/chains/base.py", line 163, in invoke
raise eFile "/usr/local/lib/python3.11/site-packages/langchain/chains/base.py", line 153, in invoke
self._call(inputs, run_manager=run_manager)File "/usr/local/lib/python3.11/site-packages/crewai/agents/executor.py", line 124, in _call
next_step_output = self._take_next_step(
^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain/agents/agent.py", line 1138, in _take_next_step
[File "/usr/local/lib/python3.11/site-packages/langchain/agents/agent.py", line 1138, in <listcomp>
[File "/usr/local/lib/python3.11/site-packages/crewai/agents/executor.py", line 186, in _iter_next_step
output = self.agent.plan(
^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain/agents/agent.py", line 397, in plan
for chunk in http://self.runnable.stream(inputs, config={"callbacks": callbacks}):File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 2875, in stream
yield from self.transform(iter([input]), config, kwargs)File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 2862, in transform
yield from self._transform_stream_with_config(File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 1880, in _transform_stream_with_config
chunk: Output = http://context.run(next, iterator) # type: ignore
^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 2826, in _transform
for output in final_pipeline:File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 1283, in transform
for chunk in input:File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 4728, in transform
yield from self.bound.transform(File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 1300, in transform
yield from http://self.stream(final, config, kwargs)File "/usr/local/lib/python3.11/site-packages/langchain_core/language_models/chat_models.py", line 249, in stream
raise eFile "/usr/local/lib/python3.11/site-packages/langchain_core/language_models/chat_models.py", line 229, in stream
for chunk in self._stream(messages, stop=stop, kwargs):File "/usr/local/lib/python3.11/site-packages/langchain_groq/chat
_models.py", line 321, in _stream
for chunk in self.client.create(messages=message_dicts, params):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/resources/chat/completions.py", line 289, in create
return self._post(
^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1225, in post
return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 920, in request
return self._request(
^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1003, in _request
return self._retry_request(
^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1051, in _retry_request
return self._request(
^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1003, in _request
return self._retry_request(
^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1051, in _retry_request
return self._request(
^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1018, in _request
raise self._make_status_error_from_response(err.response) from None
AI Thought Bubble - Next Action:
Thought: I need to read more content from other relevant websites and articles to gather more information about the topic.
Action: Read website content
Action Input: {"website_url": "https://www.forbes.com/sites/roberthart/2024/05/28/elon-musk-is-feuding-with-ai-godfather-yann-lecun-again-heres-why/"}
RateLimitError: Error code: 429 - {'error': {'message': 'Rate limit reached for model
llama3-70b-8192
in organizationorg_01hrx3emwtett8bq1cyh7w230q
on tokens per minute (TPM): Limit 6000, Used 0, Requested 6194. Please try again in 1.94s. Visit https://console.groq.com/docs/rate-limits for more information.', 'type': 'tokens', 'code': 'rate_limit_exceeded'}} Traceback: File "/usr/local/lib/python3.11/site-packages/streamlit/runtime/scriptrunner/script_runner.py", line 600, in _run_script exec(code, module.dict)File "/app/app.py", line 52, in <module> result = melody_crew.run() ^^^^^^^^^^^^^^^^^File "/app/crew.py", line 59, in run return crew.kickoff() ^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/crew.py", line 252, in kickoff result = self._run_sequential_process() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/crew.py", line 293, in _run_sequential_process output = task.execute(context=task_output) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/task.py", line 173, in execute result = self._execute( ^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/task.py", line 182, in _execute result = agent.execute_task( ^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/crewai/agent.py", line 221, in execute_task result = self.agent_executor.invoke( ^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain/chains/base.py", line 163, in invoke raise eFile "/usr/local/lib/python3.11/site-packages/langchain/chains/base.py", line 153, in invoke self._call(inputs, run_manager=run_manager)File "/usr/local/lib/python3.11/site-packages/crewai/agents/executor.py", line 124, in _call next_step_output = self._take_next_step( ^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain/agents/agent.py", line 1138, in _take_next_step [File "/usr/local/lib/python3.11/site-packages/langchain/agents/agent.py", line 1138, in <listcomp> [File "/usr/local/lib/python3.11/site-packages/crewai/agents/executor.py", line 186, in _iter_next_step output = self.agent.plan( ^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain/agents/agent.py", line 397, in plan for chunk in http://self.runnable.stream(inputs, config={"callbacks": callbacks}):File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 2875, in stream yield from self.transform(iter([input]), config, kwargs)File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 2862, in transform yield from self._transform_stream_with_config(File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 1880, in _transform_stream_with_config chunk: Output = http://context.run(next, iterator) # type: ignore ^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 2826, in _transform for output in final_pipeline:File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 1283, in transform for chunk in input:File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 4728, in transform yield from self.bound.transform(File "/usr/local/lib/python3.11/site-packages/langchain_core/runnables/base.py", line 1300, in transform yield from http://self.stream(final, config, kwargs)File "/usr/local/lib/python3.11/site-packages/langchain_core/language_models/chat_models.py", line 249, in stream raise eFile "/usr/local/lib/python3.11/site-packages/langchain_core/language_models/chat_models.py", line 229, in stream for chunk in self._stream(messages, stop=stop, kwargs):File "/usr/local/lib/python3.11/site-packages/langchain_groq/chat _models.py", line 321, in _stream for chunk in self.client.create(messages=message_dicts, params): ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/resources/chat/completions.py", line 289, in create return self._post( ^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1225, in post return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 920, in request return self._request( ^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1003, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1051, in _retry_request return self._request( ^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1003, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1051, in _retry_request return self._request( ^^^^^^^^^^^^^^File "/usr/local/lib/python3.11/site-packages/groq/_base_client.py", line 1018, in _request raise self._make_status_error_from_response(err.response) from None