Closed smcgregor closed 9 months ago
Let's test run any taxonomies produced on the following incidents, which I selected to produce a variety of tasks, technologies, and factors. All are heavily reported, but not necessarily at a depth that would allow for greatly reducing the uncertainty.
https://incidentdatabase.ai/cite/1
Example potential classifications
https://incidentdatabase.ai/cite/72
https://incidentdatabase.ai/cite/112
I have begun work on developing the three taxonomies described above. An indicative snapshot of the current (malleable) status is:
Indicative next steps include:
I will share additional info (documentation, tools, repositories) shortly.
The full report on this: https://arxiv.org/abs/2211.07280
Closing this historical tracking issue and linking the improvements that we will prioritize per risk checklisting.
(This is an issue intended to engender discussion)
One of the projects we would like to pursue is the development of a technical failure taxonomy that indicates the technical factors at play in producing an AI incident. Technical factors are challenging to determine since most incidents give only high-level information about what happened. Details about the technology and why the incident happened is typically lacking. Still, some incidents do provide more information, and those lacking concrete technical information are often amenable to technical analysis derived from first principles on how such systems are typically implemented and can fail. Thus if we establish a hierarchical taxonomy, in which the classifications are marked as speculative or grounded in evidence, then we can produce a system whereby we state (1; see below) with factual grounding for most incidents, while (2) and (3) would be increasingly speculative,
The nice thing about this is that there is a strong probabilistic relationship between each of these hierarchical levels.
--> The set of potential technologies involved in an incident is limited to those technologies that may be applied to a task.
-->Similarly, technologies can fail in specific ways when applied to a task.
Thus, while it is proving difficult to develop a taxonomy of AI failures on the basis of just news reports (we have tried...), we can likely annotate (1) and (2), then leverage this context to generate a set of potential technical failures. We won't know what definitely happened, but we can develop the set of potential technical factors and share this context across incidents having the same applications and technologies.
I am proposing this project to be developed in the following segments, to be tracked across a set of new incidents to be created and linked below,