Closed ghost closed 1 year ago
The reported results with the Davinci model are for the CodeX Davinci model which is accessible through the Microsoft Azure. The CodeX Davinci model has a larger context window of 8001 token. You can see the context window sizes of models from this link: https://platform.openai.com/docs/models/codex
Hi Mohammadreza,
I wanted to run the script with GPT-3.5 as I don't have access to GPT-4 (on the waitlist). After changing the
model
variable inGPT4_generation
togpt-3.5-turbo
and running the script, I'm getting the following error:My understanding is that we need to reduce the prompt size as GPT-3.5 doesn't support the same number of tokens as GPT-4 (8096). The interesting thing is that Davinci has around the same context size as GPT-3.5 (4096), so the question boils down to figuring out how to run the script with Davinci, which was one of the LLMs used to test DIN-SQL in the paper.
Thanks for your help.