7.3 Factor analysis and principal-components analysis
Retaining the Gaussian emissions from the GMM but exchanging the categorical latent variable for a standard normal variate yields “factor analysis” (see Section 2.1.2).
We also restrict the emission covariance to be diagonal in order to remove a degree of freedom that is not (as we shall see) identifiable from the data.
The model is fully described by Eq. 2.20.
However, we depart slightly from that formulation by augmenting the latent variable vector
The learning problem starts once again with minimization of the joint cross entropy:
The M step.
The model prior distribution does not depend on any parameters, so only the model emission is differentiated.
Starting with the emission matrix
the normal equations.
Thus, in a fully observed model, finding
The emission covariance also takes on a familiar form:
The final line can be simplified using our newly acquired formula for
Now, we require
The E step.
In Section 2.1.2, we derived the posterior distribution for factor analysis:
(7.3) |
In the E step, then, the expected sufficient statistics for
and
7.3.1 Principal-components analysis
We saw that in the limit of equal and infinite emission precisions, EM for the GMM reduces to
From Eq. 7.3, we see that the posterior covariance goes to zero as
The final expression is the Moore-Penrose pseudo-inverse of
[……]
[Tipping1999]
Iterative Principal-Components Analysis
|
|
|