Interpretable and explainable machine learning: A methods-centric overview with concrete examples
OPEN ACCESS
Loading...
Author / Producer
Date
2023-02-28
Publication Type
Review Article
ETH Bibliography
yes
Citations
Altmetric
OPEN ACCESS
Data
Rights / License
Abstract
Interpretability and explainability are crucial for machine learning (ML) and statistical applications in medicine, economics, law, and natural sciences and form an essential principle for ML model design and development. Although interpretability and explainability have escaped a precise and universal definition, many models and techniques motivated by these properties have been developed over the last 30 years, with the focus currently shifting toward deep learning. We will consider concrete examples of state-of-the-art, including specially tailored rule-based, sparse, and additive classification models, interpretable representation learning, and methods for explaining black-box models post hoc. The discussion will emphasize the need for and relevance of interpretability and explainability, the divide between them, and the inductive biases behind the presented “zoo” of interpretable models and explanation methods.
Permanent link
Publication status
published
External links
Editor
Book title
Journal / series
Volume
13 (3)
Pages / Article No.
Publisher
Wiley
Event
Edition / version
Methods
Software
Geographic location
Date collected
Date created
Subject
Explainable AI; Machine Learning; Interpretability
Organisational unit
09670 - Vogt, Julia / Vogt, Julia
Notes
Funding
Related publications and datasets
Cites: