
a training set beforehand, incremental learning shows
several advantages: (1) It does not require a sufficient
training set before learning; (2) It can continuously
learn to improve when the system is running; (3) It
can adapt to changes of the target concept; (4) It
requires less computation and storage resources than
learning from scratch; (5) It naturally matches the
applications depending on time series. Nevertheless,
incremental learning is not suitable for many non-
incremental learning tasks due to the fact that it is
inherently ‘‘myopic’’ and tends to ignore the global
information in the entire training set.
Related Entries
▶ Machine-Learning
References
1. Giraud-Carrier, C.G.: A note on the utility of incremental
learning. AI Commun. 13(4), 215–224 (2000)
2. Ourston, D., Mooney, R.J.: Theory refinement combining ana-
lytical and empirical methods. Artif. Intell. 66(2), 273–309
(1994)
3. Elman, J.L.: Learning and development in neural net-
works: The importance of starting small. Cognition 46(1),
71–99 (1993)
4. Cheng, L., Vishwanathan, S.V.N., Schuurmans, D., Wang, S.,
Caelli, T.: Implicit online learning with kernels. In: Advances
in Neural Information Processing Systems 19, pp. 249–256.
Vancouver, Canada (2006)
5. Huo, Q., Lee, C.H.: On-line adaptive learning of the continuous
density hidden markov model based on approximate recursive
bayes estimate. IEEE Trans. Speech Audio Process. 5(2), 161–172
(1997)
6. Pan, S.J., Kwok, J.T., Yang, Q.: Transfer learning via dimension-
ality reduction. In: Proceedings of the AAAI Conference on
Artificial Intelligence, pp. 677–682. Chicago, IL (2008)
7. Cover, T.M., Hart, P.E.: Nearest neighbour pattern classification.
Trans. Inf. Theory 13, 21–27 (1967)
8. Schlimmer, J.C., Fisher, D.H.: A case study of incremental con-
cept induction. In: Proceedings of the National Conference on
Artifical Intelligence, pp. 496–501. San Mateo, CA (1986)
9. Aha, D.W., Kibler, D.F., Albert, M.K.: Instance-based learning
algorithms. Mach. Learn. 6, 37–66 (1991)
10. Syed, N.A., Liu, H., Sung, K.K.: Handling concept drifts in incre-
mental learning with support vector machines. In: Proceedings of
ACM International Conference on Knowledge Discovery and Data
Mining, pp. 317–321. San Diego, CA (1999)
11. Ross, D.A., Lim, J., Lin, R.S., Yang, M.H.: Incremental learning
for robust visual tracking. Int. J. Comput. Vis. 77(1-3), 125–141
(2008)
12. Schlimmer, J.C., Granger, R.H.: Incremental learning from noisy
data. Mach. Learn. 1(3), 317–354 (1986)
13. Zhou, Z.H., Chen, Z.: Hybrid decision tree. Knowl. Based Syst.
15(8), 515–528 (2002)
Independent Component Analysis
SEUNGJIN CHOI
Department of Computer Science, Pohang University
of Science and Technology, Korea
Synonyms
Blind source separation; Independent factor analysis
Definition
Independent component analysis (ICA) is a statistical
method, the goal of which is to decompose multivari-
ate data into a linear sum of non-orthogonal basis
vectors with coefficients (encoding variables, latent
variables, hidden variables) being statistically indepen-
dent. ICA generalizes a widely-used subspace analysis
method such as principal component analysis (PCA)
and factor analysis, allowing latent variables to be non-
Gaussian and basis vectors to be non-orthogonal in
general. Thus, ICA is a density estimation method
where a linear model is learnt such that the probability
distribution of the observed data is best captured,
while factor analysis aims at best modeling the covari-
ance structure of the observed data.
Introduction
Linear latent variable model assumes that m-dimensional
observed data x
t
2 R
m
is generated by
x
t
¼ a
1
s
1;t
þ a
2
s
2;t
þa
n
s
n;t
þ E
t
; ð1Þ
where a
i
2 R
m
are basis vectors and s
i,t
are latent
variables (hidden variables, coefficients, encoding
variables) which are introduced for parsimonious
representation (n m). Modeling uncertainty or
noise is absorbed in E
t
2 R
m
. Neglecting the uncertainty
E
t
in (1), the linear latent varia ble model is nothing but
Independent Component Analysis
I
735
I