FabienCharmet / XAISurvey

0 stars 0 forks source link

XAI desiderata #7

Open cloudgravity opened 2 years ago

cloudgravity commented 2 years ago

While reading one of the papers, I came upon the XAI desiderata from Hansen, L. K., & Rieger, L. (2019). Interpretability in intelligent systems–a new concept?. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning (pp. 41-49). Springer, Cham.

It lists the following "requirements" :

D1. Fidelity: the explanation must be a reasonable representation of what the system actually does. D2. Understandability: Involves multiple usability factors including terminology, user competencies, levels of abstraction and interactivity. D3. Sufficiency: Should be able to explain function and terminology and be detailed enough to justify decision. D4. Low Construction overhead: The explanation should not dominate the cost of designing AI. D5. Efficiency: The explanation system should not slow down the AI significantly.

It could have been great to evaluate for each work in the survey if it complies with these requirements.

cloudgravity commented 2 years ago

The below article has an interesting explanation evaluation section (3.2) Antwarg, L., Miller, R. M., Shapira, B., & Rokach, L. (2021). Explaining anomalies detected by autoencoders using Shapley Additive Explanations. Expert Systems with Applications, 186, 115736

Notable methods/metrics include:

Other methods seem to compute explanations of explanations...

[1] Doshi-Velez, F., & Kim, B. (2017). A roadmap for a rigorous science of interpretability. Stat, 1050, 28. [2] Gunning, D. (2017). Explainable artificial intelligence (xai). Defense Advanced Research Projects Agency (DARPA), Nd Web. [3] Hoffman, R. R., Mueller, S. T., Klein, G., & Litman, J. (2018). Metrics for explainable AI: Challenges and prospects. ArXiv preprint arXiv:1812.04608. [4] Melis, D. A., & Jaakkola, T. (2018). Towards robust interpretability with self-explaining neural networks. In Advances in neural information processing systems (pp.7786–7795).