Consider a Markov network defined by the potentials
each dependent on a set of nodes
.
The probability to find the network in a certain state
is
proportional to
| (1) |
![]() |
Let us define the set of separator nodes,
, to be the nodes
that separate
from the rest of the network. One can take,
for instance, the union of the nodes in all potentials that contain at
least one marginal node, while excluding the marginal nodes itself in that
union (which is the Markov blanket of
):
| (2) |
a-c. It is not necessary,
however, that
d).
A sufficient condition for the rest of the theory is that the size of
the total state space,
Suppose that we know a particular setting of the separator nodes,
, then computing the conditional distribution of
given
is easy:
| (3) |
Unfortunately, in general we do not know the distribution
and therefore we cannot compute
. We can, however,
still derive certain properties of the marginal distribution, namely
upper and lower bounds for each state of
. For instance,
an upper bound on
can easily be found by computing
as a simple linear programming
(LP) problem with
Upto now we did not include any information about
we might have into the equation above. But suppose that we
already computed upper and lower bounds on the marginals of other
nodes than
. How can we make use of this knowledge?
We investigate our table of all sets of marginal nodes we bounded
earlier and take out those that are subsets of
. Obviously,
whenever we find already computed bounds on
where
1,
we can be sure that
.
There can be several
In this way, information collected by bounds computed earlier, can
propagate to
, thereby finding tighter bounds for these
marginals. This process can be repeated for all sets of marginal
nodes we have selected until convergence is reached.