arXiv Analytics

Sign in

arXiv:2006.01272 [cs.LG]AbstractReferencesReviewsResources

Shapley-based explainability on the data manifold

Christopher Frye, Damien de Mijolla, Laurence Cowton, Megan Stanley, Ilya Feige

Published 2020-06-01Version 1

Explainability in machine learning is crucial for iterative model development, compliance with regulation, and providing operational nuance to model predictions. Shapley values provide a general framework for explainability by attributing a model's output prediction to its input features in a mathematically principled and model-agnostic way. However, practical implementations of the Shapley framework make an untenable assumption: that the model's input features are uncorrelated. In this work, we articulate the dangers of this assumption and introduce two solutions for computing Shapley explanations that respect the data manifold. One solution, based on generative modelling, provides flexible access to on-manifold data imputations, while the other directly learns the Shapley value function in a supervised way, providing performance and stability at the cost of flexibility. While the commonly used ``off-manifold'' Shapley values can (i) break symmetries in the data, (ii) give rise to misleading wrong-sign explanations, and (iii) lead to uninterpretable explanations in high-dimensional data, our approach to on-manifold explainability demonstrably overcomes each of these problems.

Related articles: Most relevant | Search more
arXiv:2308.13792 [cs.LG] (Published 2023-08-26)
Out-of-distribution detection using normalizing flows on the data manifold
arXiv:2210.07100 [cs.LG] (Published 2022-10-13)
Dissipative residual layers for unsupervised implicit parameterization of data manifolds
arXiv:2204.08624 [cs.LG] (Published 2022-04-19)
Topology and geometry of data manifold in deep learning