5.2 Unsupervised learning
So far we have considered discriminative models in which the “outputs,” , have been observed. Indeed, it might seem rather quixotic to try to learn a discriminative model purely from its inputs, . Nevertheless, let us consider the intuitive objective of maximizing information transmission [1], and see where it leads us.
5.2.1 “InfoMax” in deterministic, invertible models
To motivate the objective, consider the deterministic input-output relationship
(5.25) |
Here the “final” output is computed by passing each element of the vector through some sort of invertible, element-wise “squashing” function (cf. a layer of a neural network, Eq. 5.17), i.e. a monotonic function from the reals to some bounded interval on the real line:
We will justify the need for such functions below. Now, when there are fewer outputs than inputs, , changing so as to maximize mutual information between inputs and outputs will yield a function that recodes its inputs more efficiently. However, we begin with a seemingly more näive, but mathematically tractable, alternative: Let the number of outputs equal the number inputs, , and furthermore let be invertible (in its first argument), so that . This may seem somewhat perverse, since invertible transformations are automatically information preserving. However, recall that mutual information nevertheless depends on :
That is, although the conditional entropy is zero for all values of (under the assumption that doesn’t lose its invertibility), the output entropy still depends on . In fine, for invertible , maximizing mutual information amounts to maximizing output entropy.88 8 If there were noise in the map from to , this would add an extra term to the gradient, and the goal of maximizing output entropy would have to be balanced against noise-proofing the transmission. Since a vector random variable is maximally entropic only if its components are statistically independent, this suggests that the “InfoMax” criterion can underwrite a form of independent-components analysis††margin: independent-components analysis . That is, maximizing input-output mutual information or output entropy will “unmix” the inputs into their independent components.
Now we can say why the squashing function is necessary. If the input-output map were unbounded, the output entropy could be driven arbitrarily large. For concreteness, and without loss of generality, we can let be an increasing function, with its range the interval . Now let us try to re-express the objective in Eq. 5.27. More precisely, we consider its negation, , for consistency with the standard objectives of this book, which are losses. First we note that, although we have not specified a distribution for , is inherited directly from the data distribution via the deterministic relationship in Eq. 5.25. In particular, since the relationship is (by assumption) invertible, the two distributions are related by the standard change-of-variables formula, which we apply on the third line:
The last line follows because the squashing functions are assumed to act element-wise, so the determinant is just the product of the diagonal elements. Evidently, maximizing mutual information through an invertible transformation is identical to minimizing our standard loss, the relative entropy between data and model distributions, with the latter equal to the Jacobian determinant of that transformation:
The assimilation to density estimation can be completed by reinterpreting Eq. 5.25 as defining (via its inverse) the emission density of a generative model [3]. We defer this reinterpretion until Chapter 6, when we take up learning in generative models in earnest.
InfoMax independent-components analysis.
One interesting special case of the unsupervised, discriminative learning problem just described is so-called InfoMax ICA [1]. Here, the discriminative map in Eq. 5.25 is defined to be a linear transformation by a full-rank square matrix:
(5.30) |
We leave the output nonlinearity undefined for now, except to insist (as we have been) that it be invertible and bounded. The gradient, whether of the mutual information, output entropy, relative input entropy, or input cross entropy, is then
(recalling the derivative of the log determinant). Here is the row of . For example, when is the logistic function,
the gradient inside the expectation becomes
Setting Eq. 5.31 to zero yields an iterpretable, albeit implicit, equation for the optimal solution:
where the function is applied element-wise. Thus when (for all ), and is approximately an identity function, , proportional to the whitening transformation.
“Semi-supervised” clustering.
[[XXX]]