hon9g / papers__Today-I-Read

0 stars 0 forks source link

"Why Should I Trust you?" Explaining the Prediction of Any Classifier #3

Open hon9g opened 5 years ago

hon9g commented 5 years ago

Abstract

Introduction

It is important to differentiate between two different (but related) definitions of trust:

  1. trusting a prediction,
    • i.e. whether a user trusts an individual prediction sufficiently to take some action based on it
  2. trusting a model,
    • i.e. whether the user trusts a model to behave inreasonable ways if deployed.
    • Both are directly impacted by how much the human understands a model’s behaviour, as opposed to seeing it as a black box.

    • Determining trust in individual predictions is an important problem when the model is used for decision making.
    • When using machine learning for medical diagnosis or terrorism detection, for example, predictions cannot be acted upon on blind faith, as the consequences may be catastrophic.

Conclusion

hon9g commented 5 years ago

Details:

The Case for Explanation

1. Model or its evaluation can go wrong.

image

2. Machine learning practitioners often have to select a model from a number of alternatives, requiring them to assess the relative trust between two or more models.

Desired Characteristics for Explainers:

Local Interpretable Model-Agnostic explanations

Interpretable Data Representation

Fidelity-Interpretability Trade-off

Sampling for Local Exploration

Sparse Linear Explanations

Example

  1. Text classification with SVMs
  2. Deep networks for images image

Submodular pick for Explaining model

hon9g commented 5 years ago

User Study:

Simulated User Experiments

Are explanations faithful to the model?

Should I trust this prediction?

Can I trust this model?

Evaluation with Human subjects

Can users select the best classifier?

Can non-experts improve a classifier?

Do explanations lead to insights?