edit
The Structure of Neural Network Posteriors
, 2001.
Abstract
Exact inference in Bayesian neural networks is non analytic to compute and as a result approximate approaches such as the evidence procedure, Monte-Carlo sampling and variational inference have been proposed. In this paper we explore the structure of the posterior distributions in such a model through a new approximating distribution based on mixtures of Gaussian distributions and show how it may be implemented.