Hi,
first I want to say that you guys are building a useful tool!
Great job!
We’re exploring ways to enhance our workflows with RAGFlow and have a few questions about integrating it with tools like Agenta and improving observability:
Automating Testing of Retrieval and Model Replies
We aim to automate testing for both retrieval results and model-generated replies. Specifically:
We can write custom code using RAGFLow API to validate retrieval results and responses.
Is there a recommended or built-in approach in RAGFLow to perform such automated tests?
Iterating on Prompts and Parameters
We’re interested in integrating prompt management and evaluation tools, such as Agenta, to improve iteration and experimentation.
Does RAGFLow support or provide guidance for integrating with tools like Agenta for:
Creating and versioning prompts?
Running evaluations on prompts and analyzing results?
Connecting prompt outcomes to observability for performance monitoring?
Monitoring and Logging Improvements
To enhance observability and debugging, have you considered adopting OpenTelemetry standards for logging OpenTelemetry’s documentation.
Does RAGFLow currently support or have plans to support OpenTelemetry for structured logging?
What options are available for tracking logs and metrics across retrieval, generation, and overall system performance?
We are happy to contribute on this if helpful.
Your input is much appreciated!
Looking forward to your insights and suggestions!
Describe your problem
Hi, first I want to say that you guys are building a useful tool! Great job! We’re exploring ways to enhance our workflows with RAGFlow and have a few questions about integrating it with tools like Agenta and improving observability:
Automating Testing of Retrieval and Model Replies We aim to automate testing for both retrieval results and model-generated replies. Specifically:
We can write custom code using RAGFLow API to validate retrieval results and responses. Is there a recommended or built-in approach in RAGFLow to perform such automated tests?
Iterating on Prompts and Parameters
We’re interested in integrating prompt management and evaluation tools, such as Agenta, to improve iteration and experimentation.
To enhance observability and debugging, have you considered adopting OpenTelemetry standards for logging OpenTelemetry’s documentation.
We are happy to contribute on this if helpful. Your input is much appreciated! Looking forward to your insights and suggestions!