Closed vicenteherrera closed 1 month ago
To address the concern around the non-determinism of LLMs, one approach is to develop a continuous testing module that regularly checks the correctness of the model's output by establishing a set of benchmark question-answer pairs. Here's how you could approach this:
Contact Details
vicente.herrera@control-plane.io
What is the idea
This is a request for direct feedback and answers to the following question regarding a threat described in the governance framework.
TR-6, Non-deterministic behaviour Section: governance-framework/_threats/tr-6.md File (private repo link): https://github.com/finos/ai-readiness-private/blob/10e31ea7ccf3893983404de3484f7c57f9934d57/governance-framework/_threats/tr-6.md Diff (private repo link): https://github.com/finos/ai-readiness-private/pull/5/files#diff-87d41e3ec96687ff49345546307c878da1b45b272916021f5446d007c681abbe
Title: Non-deterministic behaviour Type: Integrity External references:
Description: Given the immaturity of the products, the vector store may not have capabilities expected of enterprise software (access control, encryption at rest, audit logging etc). Misconfiguration may allow unauthorized access to data. Internal user accesses data and leaks/tampers it.
Question:
Why this is important
Discussion surrounding this threat has been going on live, on today's meeting (2024-09-03).
We have opened this issue to better capture's everyone's feedback, and not being constraint by the time alloted to the meeting.
Code of Conduct