
62 4. Unsupervised Learning Neural Networks
In the above, d
k,p
is the Euclidean distance as defined in equation (4.20), I is the total
number of input units, and g
k
(0) = 0. Thus, b
k
(0) =
1
I
, which initially gives each
output unit an equal chance to be the winner; b
k
(t) is the conscience factor defined
for each output unit. The more an output unit wins, the larger the value of g
k
(t)
becomes, and b
k
(t) becomes larger negative. Consequently, a factor |b
k
(t)| is added to
the distance d
k,p
. Usually, for normalized inputs, β =0.0001 and γ = 10.
4.5 Self-Organizing Feature Maps
Kohonen developed the self-organizing feature map (SOM) [474, 475, 476], as moti-
vated by the self-organization characteristics of the human cerebral cortex. Studies of
the cerebral cortex showed that the motor cortex, somatosensory cortex, visual cortex
and auditory cortex are represented by topologically ordered maps. These topological
maps form to represent the structures sensed in the sensory input signals.
The self-organizing feature map is a multidimensional scaling method to project an
I-dimensional input space to a discrete output space, effectively performing a com-
pression of input space onto a set of codebook vectors. The output space is usually a
two-dimensional grid. The SOM uses the grid to approximate the probability density
function of the input space, while still maintaining the topological structure of input
space. That is, if two vectors are close to one another in input space, so is the case
for the map representation.
The SOM closely resembles the learning vector quantizer discussed in the previous
section. The difference between the two unsupervised algorithms is that neurons are
usually organized on a rectangular grid for SOM, and neighbors are updated to also
perform an ordering of the neurons. In the process, SOMs effectively cluster the
input vectors through a competitive learning process, while maintaining the topological
structure of the input space.
Section 4.5.1 explains the standard stochastic SOM training rule, while a batch version
is discussed in Section 4.5.2. A growing approach to SOM is given in Section 4.5.3.
Different approaches to speed up the training of SOMs are overviewed in Section 4.5.4.
Section 4.5.5 explains the formation of clusters for visualization purposes. Section 4.5.6
discusses in brief different ways how the SOM can be used after training.
4.5.1 Stochastic Training Rule
SOM training is based on a competitive learning strategy. Assume I-dimensional
input vectors z
p
, where the subscript p denotes a single training pattern. The first
step of the training process is to define a map structure, usually a two-dimensional
grid (refer to Figure 4.3). The map is usually square, but can be of any rectangular
shape. The number of elements (neurons) in the map is less than the number of
training patterns. Ideally, the number of neurons should be equal to the number of
independent training patterns.