In the setting of empirical learning available training data are used to obtain information about new test situations. Clearly, this generalization from training to test data requires knowledge about their dependencies. In this paper, such knowledge concerning dependencies between training and test data, in some contexts also known as rules or axioms, will be called prior knowledge.
To be specific,
we consider a typical function approximation problem.
Assume a given a set of training data =
sampled i.i.d. from an unknown but fixed
``true state of Nature''.
The aim is to obtain an approximation function
to predict unknown outcomes
for
test situations
by
.
Relying on the fact that the
generalization ability of any learning system
is crucially based on the dependencies it implements,
our goal has to be a strict empirical measurement and control
of the prior or ``dependency'' data
which represent our prior knowledge.
It is interesting to note
that for the common situation with
an infinite number of potential test situations
also the number of dependencies
to be controlled empirically
becomes infinite.
Empirical measurement of an infinite number of data, however,
seems at first glance impossible.
On the other hand,
for an infinite set
of test situations
any learning system has to
use an infinite number of data,
either explicitly or implicitly.
To discuss this empirical measurement problem
let us have a closer look at two examples:
Thus, one can say: Infinite a-priori information can be empirically measured by a-posteriori control at the time of testing. From this point of view, related to that of constructivism, (also infinite) a-priori information can (and should) be explicitly related to empirical control of the application situation.