In this subsection we show how the original Expectation-Maximization algorithm [13] can be extended to a Quasi-Bayesian Expectation-Maximization (QEM) algorithm with the same convergence properties. We must maximize the posterior log-likelihood L(Theta) defined previously. The algorithm begins by assuming that the transparent variables are actual random quantities with distributions specified by Theta. An initial estimate Theta0 is assumed for Theta.
Suppose we had i sets of complete data for the transformed network, i.e.,
we had observed i trials for all variables in the network, including the
transparent variables. The log-likelihood for this complete data would be
L(Theta) = sumijk li(j,k)
The first step of the QEM algorithm is to obtain the expected value of the log-likelihood given the evidence and assuming Theta0 is correct [10]:
The second step of the QEM algorithm is to maximize Q(Theta|Thetak) for Theta. Only a few terms in the expression for Q(Theta|Thetak) will be free, since only the thetaij for z'i are estimated. Collecting these terms we obtain:
sumij p(z'i = j | xq = a, e)
Fri May 30 15:55:18 EDT 1997