Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
[X] I confirm that I am using English to submit this report (我已阅读并同意 Language Policy).
[X] [FOR CHINESE USERS] 请务必使用英文提交 Issue,否则会被关闭。谢谢!:)
[X] Please do not modify this template :) and fill in all the required fields.
1. Is this request related to a challenge you're experiencing? Tell me about your story.
Currently, the retrieval model and the re-ranking stage share the score_threshold and top_k parameters, which currently have the following problems:
The retrieval model and the re-ranking model have different score distributions and need to use different thresholds, which cannot be supported under the current solution;
In order to ensure the integrity of the information, as much text as possible is generally recalled during the retrieval stage, and compressed to a relatively small number after the re-ranking is completed to match the input context of the large model. Currently, when top_k is shared, if top_k is set too small, the retrieval stage cannot recall enough content, affecting the recall rate, or if top_k is set too large, it exceeds the context of the large model;
2. Additional context or comments
No response
3. Can you help us with this feature?
[ ] I am interested in contributing to this feature.
Self Checks
1. Is this request related to a challenge you're experiencing? Tell me about your story.
Currently, the retrieval model and the re-ranking stage share the score_threshold and top_k parameters, which currently have the following problems:
2. Additional context or comments
No response
3. Can you help us with this feature?