arXiv Analytics

Sign in

arXiv:2212.03131 [cs.LG]AbstractReferencesReviewsResources

Explainability as statistical inference

Hugo Henri Joseph Senetaire, Damien Garreau, Jes Frellsen, Pierre-Alexandre Mattei

Published 2022-12-06Version 1

A wide variety of model explanation approaches have been proposed in recent years, all guided by very different rationales and heuristics. In this paper, we take a new route and cast interpretability as a statistical inference problem. We propose a general deep probabilistic model designed to produce interpretable predictions. The model parameters can be learned via maximum likelihood, and the method can be adapted to any predictor network architecture and any type of prediction problem. Our method is a case of amortized interpretability models, where a neural network is used as a selector to allow for fast interpretation at inference time. Several popular interpretability methods are shown to be particular cases of regularised maximum likelihood for our general model. We propose new datasets with ground truth selection which allow for the evaluation of the features importance map. Using these datasets, we show experimentally that using multiple imputation provides more reasonable interpretations.

Comments: 10 pages, 22 figures, submitted at ICLR 2023
Categories: cs.LG, cs.AI, stat.ME
Related articles: Most relevant | Search more
arXiv:2403.06425 [cs.LG] (Published 2024-03-11)
A Differential Geometric View and Explainability of GNN on Evolving Graphs
arXiv:2208.11782 [cs.LG] (Published 2022-08-24)
Maximum Likelihood on the Joint (Data, Condition) Distribution for Solving Ill-Posed Problems with Conditional Flow Models
arXiv:2201.10528 [cs.LG] (Published 2022-01-25)
Explainability in Music Recommender Systems