Next: Gaussian regression
Up: Mixtures of Gaussian process
Previous: Introduction
  Contents
The Bayesian model
Let us consider
the following random variables:
- 1.
- , representing (a vector of)
independent, visible variables (`measurement situations'),
- 2.
- , being (a vector of)
dependent, visible variables (`measurement results'),
and
- 3.
- , being the hidden variables (`possible states of Nature').
A Bayesian approach is based on two model inputs
[1,11,4,12]:
- 1.
- A likelihood model ,
describing the density of observing given and .
Regarded as function of , for fixed and ,
the density
is also known as the (-conditional) likelihood of .
- 2.
- A prior model , specifying
the a priori density of
given some a priori information denoted by
(but before training data have been taken into account).
Furthermore,
to decompose a possibly complicated prior density
into simpler components,
we introduce
continuous hyperparameters
and discrete hyperparameters
(extending the set of hidden variables to = ),
|
(1) |
In the following, the summation
over will be treated exactly,
while the -integral will
be approximated.
A Bayesian approach aims in calculating the predictive density
for outcomes in test situations
|
(2) |
given data =
consisting of
a priori data
and
i.i.d. training data
=
.
The vector of all ()
will be denoted .
Fig.1 shows a graphical representation
of the considered probabilistic model.
In saddle point approximation
(maximum a posteriori approximation)
the -integral becomes
|
(3) |
|
(4) |
assuming
to be slowly varying at the stationary point.
The posterior density is related to
(-conditional) likelihood and prior
according to Bayes' theorem
|
(5) |
where the -independent denominator (evidence)
can be skipped when maximising with respect to .
Treating the -integral within
also in saddle point approximation
the posterior must be maximised
with respect to and simultaneously .
Figure 1:
Graphical representation of
the considered probabilistic model,
factorising according to
=
.
(The variable is introduced
in Section 3.)
Circles indicate visible variables.
|
Next: Gaussian regression
Up: Mixtures of Gaussian process
Previous: Introduction
  Contents
Joerg_Lemm
1999-12-21