
Other related research includes the charting
algorithm of Brand [6], which computes a pseudo-
invertible mapping of the data as well as the intrinsic
dimensionality of the manifold. The dimensionality is
estimated by examining the rate of growth of the num-
ber of points contained in hyper-spheres as a function of
the radius. Linear patches, areas of curvature and noise
can be correctly classified using the proposed measure.
At a subsequent stage, a global coordinate system for
the embedding is defined. This produces a mapping
between the input space and the embedding space.
Weinberger and Saul [17] developed Semidefinite
Embedding (SDE) which addresses manifold learning
by enforcing local isometry . The lengths of the sides of
triangles formed by neighboring points are preserved
during the embedding. These constraints can be
expressed in terms of pairwise distances and the optimal
embedding can be found by semidefinite programming.
The method is computationally demanding, but can reli-
ably estimate the underlying dimensionality of the inputs
by locating the largest gap between the eigen values of the
Gram matrix of the outputs. As in the case of the authors’
approach , this estimate does not require a threshold.
The method for manifold learning described [8]by
Mordohai and Medioni [8] is based on inferring the
geometric properties of the manifold locally via tensor
voting. An estimate of the local tangent space allows one
to traverse the manifold estimating geodesic distances
between points and generating novel observations on the
manifold. In this method it is not necessary that the
manifold is differentiable, or even connected. It can
process data from intersecting manifolds with different
dimensionality and is very robust to outliers. Unlike
most of the other approaches, the authors did not per-
form dimensionality reduction, but conducted all opera-
tions in the ambient space instead. If dimensionality
reduction is desired for visualization or memory saving,
any technique can be applied after this.
Applications
There are two main areas of application of manifold
learning techniques in biometrics: estimation of the
degrees of freedom of the data and visualization.
Given labeled data, the degrees of freedom can be
separated into those that are related to the identity of
the subject and those that are due to other factors, such
as pose. Visualization is enabled by reducing the
dimensionality of the data to two or three to generate
datasets suitable for display. This can be achieved by
selecting the most relevant dimensions of the manifold
and mapping them to a linear 2-D or 3-D space.
An example of both visualization and estimation of
the important mode s of variability of face images has
been discussed by [4]. The input is a set of images of the
face of a single person undergoing expression and view-
point changes. The images are vectorized, that is the
pixels of each 28*20 image are stacked to form a 560-D
vector, and used as observations. LLE is able to deter-
mine the two most dominant degrees of freedom which
are related to head pose and expression variations.
Embedding the manifold from the 560-D ambient
space to a 2-D space provides a visualization in which
similar images appear close to each other. Similar experi-
ments have been described in Tenenbaum et al. [5].
Prince and Elder [18] addressed the issue of face
recognition from a manifold learning perspective by
creating invariance to degrees of freedom that do not
depend on identity. They labeled these degrees of free-
dom, namely, pose and illumination, ‘‘nuisance para-
meters’’ and were able to isolate their effects using a
training dataset in which the value of the nuisance para-
meters is known and each individual has at least two
different values of each nuisance parameter. The images
are converted to 32 32 and subsequently to 1024-D
vectors. Varying a nuisance parameter generates a mani-
fold, which has little value for recognition. Therefore,
once these manifolds are learned, their observations are
mapped to a single point, which corresponds to the
identity of the imaged person, in a new space.
Liao and Medioni [19] studied face tracking and
expression inference from video sequences using tensor
voting to learn manifolds that corr espond to basic expres-
sions, such as smile and surprise. During training, land-
mark points are tracked in the video sequence and their
3-D positions are obtained using a 3-D model of the head.
Facial deformation manifolds are learned from labeled
sequences of the basic expressions. A parameter that cor-
responds to the magnitude of the expression is estimated
for each frame. During testing, the observation vector is
the position of the landmarks and the goal is to jointly
estimate head pose and the magnitude of each expression.
This is accomplished by computing the probability that
the observation was generated by each manifold. The
posterior probability is inferred using a combination
model of all manifolds. Some results of deformable track-
ing and expression inference are presented in Fig. 1.
956
M
Manifold Learning