Locally deployed docker version. When using it, I found that when doing multiple rounds of Q&A in the same session window, the answers become not based on the content in the knowledge base, but on the original knowledge of the qwen2:7b, and the citations are messed up (citing snippets that were cited in other historical sessions and are not relevant to the current question). However, a new session is opened to re-ask the question (same assistant and same knowledge base) and the answer is again faithful to the knowledge base and the citation is correct.
Relevant snippets from the knowledge base
Answers during multiple rounds of questions and answers
First Q&A in a new session window
Such questions occur frequently, and answers in a new session are significantly better than multiple Q&A sessions in a single session window.
All the chat history will be transfered to LLM.
I guess LLM thought the first answer was not good enough so came the same quesion again, and it's also confused how to answer that.
Describe your problem
Locally deployed docker version. When using it, I found that when doing multiple rounds of Q&A in the same session window, the answers become not based on the content in the knowledge base, but on the original knowledge of the qwen2:7b, and the citations are messed up (citing snippets that were cited in other historical sessions and are not relevant to the current question). However, a new session is opened to re-ask the question (same assistant and same knowledge base) and the answer is again faithful to the knowledge base and the citation is correct.
Relevant snippets from the knowledge base
Answers during multiple rounds of questions and answers
First Q&A in a new session window
Such questions occur frequently, and answers in a new session are significantly better than multiple Q&A sessions in a single session window.