Closed gileneusz closed 1 year ago
AFAIU we are limited by Open AI's limits.
I would suggest you use the 16k model gpt-4-0613
I think it is, but you might find the same issue if your application is too big.
Thank you for your prompt response and suggestion.
I understand that we are working within the constraints set by OpenAI. However, at this moment, I don't have access to the 16k model of GPT-4. I attempted to run the process using the gpt-3.5-16k model, and unfortunately, the same error occurred.
Given this context, I was wondering if introducing some form of token limitation per minute in the application might provide a solution. This could potentially keep us within the context window limit and prevent the chunked encoding error we're experiencing.
Hey @gileneusz,
Happy to add this to reliableGPT
(wrapper for the openai endpoint). How would you envision the token limitation working?
I encountered an issue while using the GPT-4 model for processing a quite extensive and intricate prompt related to a Laravel application setup.
Issue:
When I used the model to generate output based on the following prompt:
We are writing an application in Laravel for the backend, inertia, and Vue for the frontend. This application is for testing doctors' final exam with ABCDE single answer tests. Those tests will be split into categories - such as specialties. All files such as models, controllers, migrations, views etc. are split into separate files or folders according to Laravel folders structure.
The application returned an error after a few minutes of processing, specifically the requests.exceptions.ChunkedEncodingError was thrown. This suggests an issue related to the handling of chunked encoding in the HTTP response.
Based on the traceback, the error seems to originate from urllib3, where it's trying to read a chunk of the HTTP response, but encounters an invalid chunk length. This error may be a potential indication that the application has reached the context window limit of 8192 tokens set for the GPT-4 model.
Another possible cause could be related to network limitations or rate restrictions associated with GPT-4, which might be causing the issue when handling larger inputs or running longer processes.
Reproduction:
The issue can be reproduced by running the aforementioned prompt in your application with the GPT-4 model.
Please investigate this issue, as it is causing some difficulties in handling larger prompts and processing them for extended periods in the application. If more information is required, feel free to ask.
Thank you for your attention to this matter.
Here's error log: