arXiv Analytics

Sign in

arXiv:2106.11905 [cs.LG]AbstractReferencesReviewsResources

Dangers of Bayesian Model Averaging under Covariate Shift

Pavel Izmailov, Patrick Nicholson, Sanae Lotfi, Andrew Gordon Wilson

Published 2021-06-22Version 1

Approximate Bayesian inference for neural networks is considered a robust alternative to standard training, often providing good performance on out-of-distribution data. However, Bayesian neural networks (BNNs) with high-fidelity approximate inference via full-batch Hamiltonian Monte Carlo achieve poor generalization under covariate shift, even underperforming classical estimation. We explain this surprising result, showing how a Bayesian model average can in fact be problematic under covariate shift, particularly in cases where linear dependencies in the input features cause a lack of posterior contraction. We additionally show why the same issue does not affect many approximate inference procedures, or classical maximum a-posteriori (MAP) training. Finally, we propose novel priors that improve the robustness of BNNs to many sources of covariate shift.

Related articles: Most relevant | Search more
arXiv:2003.00343 [cs.LG] (Published 2020-02-29)
Calibrated Prediction with Covariate Shift via Unsupervised Domain Adaptation
arXiv:1608.00250 [cs.LG] (Published 2016-07-31)
On Regularization Parameter Estimation under Covariate Shift
arXiv:2106.09848 [cs.LG] (Published 2021-06-17)
PAC Prediction Sets Under Covariate Shift