In this paper we leverage on probability over Riemannian manifolds to rethink the interpretation of priors and posteriors in Bayesian inference. The main mindshift is to move away from the idea that “a prior distribution establishes a probability distribution over the parameters of our model” to the idea that “a prior distribution establishes a probability distribution over probability distributions”. To do that we assume that our probabilistic model is a Riemannian manifold with the Fisher metric. Under this mindset, any distribution over probability distributions should be “intrinsic”, that is, invariant to the specific parametrization which is selected for the manifold. We exemplify our ideas through a simple analysis of distributions over the manifold of Bernoulli distributions. One of the major shortcomings of maximum a posteriori estimates is that they depend on the parametrization. Based on the understanding developed here, we can define the maximum a posteriori estimate which is independent of the parametrization.
[1]
X. Pennec.
Probabilities and Statistics on Riemannian Manifolds : A Geometric approach
,
2004
.
[2]
Christian P. Robert,et al.
The Bayesian choice : from decision-theoretic foundations to computational implementation
,
2007
.
[3]
L. Campbell.
An extended Čencov characterization of the information metric
,
1986
.
[4]
Kevin P. Murphy,et al.
Machine learning - a probabilistic perspective
,
2012,
Adaptive computation and machine learning series.
[5]
L. Wasserman,et al.
The Selection of Prior Distributions by Formal Rules
,
1996
.
[6]
J. Jost.
Riemannian geometry and geometric analysis
,
1995
.
[7]
L. M. M.-T..
Theory of Probability
,
1929,
Nature.
[8]
F. Opitz.
Information geometry and its applications
,
2012,
2012 9th European Radar Conference.
[9]
N. Čencov.
Statistical Decision Rules and Optimal Inference
,
2000
.