Predictive variational inference: Learn the predictively optimal posterior distribution
Vanilla variational inference finds an optimal approximation to the Bayesian posterior distribution, but even the exact Bayesian posterior is often not meaningful under model misspecification. We propose predictive variational inference (PVI): a general inference framework that seeks and samples fro...
Saved in:
Main Authors | , |
---|---|
Format | Journal Article |
Language | English |
Published |
18.10.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Vanilla variational inference finds an optimal approximation to the Bayesian
posterior distribution, but even the exact Bayesian posterior is often not
meaningful under model misspecification. We propose predictive variational
inference (PVI): a general inference framework that seeks and samples from an
optimal posterior density such that the resulting posterior predictive
distribution is as close to the true data generating process as possible, while
this this closeness is measured by multiple scoring rules. By optimizing the
objective, the predictive variational inference is generally not the same as,
or even attempting to approximate, the Bayesian posterior, even asymptotically.
Rather, we interpret it as implicit hierarchical expansion. Further, the
learned posterior uncertainty detects heterogeneity of parameters among the
population, enabling automatic model diagnosis. This framework applies to both
likelihood-exact and likelihood-free models. We demonstrate its application in
real data examples. |
---|---|
DOI: | 10.48550/arxiv.2410.14843 |