Interpretability and Explainability: A Machine Learning Zoo Mini-tour


Loading...

Date

2020-12-03

Publication Type

Working Paper

ETH Bibliography

yes

Citations

Altmetric

Data

Abstract

In this review, we examine the problem of designing interpretable and explainable machine learning models. Interpretability and explainability lie at the core of many machine learning and statistical applications in medicine, economics, law, and natural sciences. Although interpretability and explainability have escaped a clear universal definition, many techniques motivated by these properties have been developed over the recent 30 years with the focus currently shifting towards deep learning methods. In this review, we emphasise the divide between interpretability and explainability and illustrate these two different research directions with concrete examples of the state-of-the-art. The review is intended for a general machine learning audience with interest in exploring the problems of interpretation and explanation beyond logistic regression or random forest variable importance. This work is not an exhaustive literature survey, but rather a primer focusing selectively on certain lines of research which the authors found interesting or informative.

Publication status

published

External links

Editor

Book title

Journal / series

Volume

Pages / Article No.

Publisher

ETH Zurich, Department of Computer Science, Institute for Machine Learning

Event

Edition / version

Methods

Software

Geographic location

Date collected

Date created

Subject

Machine Learning; Interpretability; explainability

Organisational unit

09670 - Vogt, Julia / Vogt, Julia check_circle

Notes

Funding

Related publications and datasets

Is previous version of: