1 results
11 - On-line Learning of Prototypes and Principal Components
-
- By Michael Biehl, Institut für Theoretische Physik Universität Würzburg Am Hubland D-97074 Würzburg, Germany, Ansgar Freking, Matthias Hölzer, Georg Reents, Enno Schlösser, Institut für Theoretische Physik Universität Würzburg Am Hubland D-97074 Würzburg, Germany
- Edited by David Saad, Aston University
-
- Book:
- On-Line Learning in Neural Networks
- Published online:
- 28 January 2010
- Print publication:
- 28 January 1999, pp 231-250
-
- Chapter
- Export citation
-
Summary
Abstract
We review our recent investigation of on–line unsupervised learning from high–dimensional structured data. First, on–line competitive learning is studied as a method for the identification of prototype vectors from overlapping clusters of examples. Specifically, we analyse the dynamics of the well–known winner–takes–all or K–means algorithm. As a second standard learning technique, the application of Sanger's rule for principal component analysis is investigated. In both scenarios the necessary process of student specialization may be delayed significantly due to underlying symmetries.
Introduction
Methods from statistical physics have been applied to the theory of adaptive systems with great success in recent years. Perhaps the most prominent example is the analysis of feedforward neural networks which can learn from example data. The statistical mechanics approach allows to investigate typical properties of very large systems on average over the randomness contained in the data. It complements results from computational learning theory and other disciplines.
Most of the investigations concern the supervised learning of a rule. For reviews of the field see for instance (Watkin et al. 1993; Opper and Kinzel, 1996). A particularly successful line of research was initiated in (Kinzel and Rujan, 1990; Kinouchi and Caticha, 1992) and aims at the analysis of the physics of on–line learning schemes (Amari, 1967 and 1993; Hertz et al., 1991). On–line learning is attractive from a practical point of view because it uses only the latest in the sequence of examples. Obviously storage needs and computational effort are reduced in comparison with batch– or off–line learning (Hertz et al., 1991).