Closed adieyal closed 10 months ago
Ah, this is for sure a context window overflow then, because the prompt looks to have 10K chars in it so I'm guessing with that + the context injection it is overflowing. Will need to handle that aspect. Did you sent a big chat through the input? If so this issue is related to #62
I just realized are you using chat
mode? Chat is currently non-stable since it has spotty support. Only the pinecone vector db supports it right now but there certainly more work to be done to support it
I didn't explicitly set it to chat as far as I recall, but that was what I expected. Out of interest, why don't you offload all that logic to langchain? I haven't worked with langchain js but langchain python can handle chat with memory natively.
any update for this issue, after some query, i also got problem with 400 bad request
@kiendt7 Set the workspace history setting to a lower value in the workspace setting. You likely are using large prompts or something to the effect
thanks @timothycarambat
i try to setting chat history with lower value (from 20 -> 5) but sometimes still got 400 bad request, do you have an idea for limit size of prompt
This has been solved with the infinite prompt window
feature that is live! Marking as closed
I'm getting this error after two dozen or so queries (trimmed for brevity and privacy):
the workspace history seems pretty long
According to the error, the actual payload is probably in the 15000 character range. In any case, it works again if I delete the chats from the workspace.