Abstract
In order to create spatially ordered and organized representations of input occurrences in a neural network, the most essential principle seems to be to confine the learning corrections to subsets of network units that lie in the topological neighborhood of the best-matching unit. There seems to exist an indefinite number of ways to define the matching of an input occurrence with the internal representations, and even the neighborhood of a unit can be defined in many ways. It is neither necessary to define the corrections as gradient steps in the parameter space: improvements in matching may be achieved by batch computation or evolutionary changes. Consequently, all such cases will henceforth be regarded to belong to the broader category of the Self-Organizing Map (SOM) algorithms. This category may also be thought to contain both supervised and unsupervised learning methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1997 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Kohonen, T. (1997). Variants of SOM. In: Self-Organizing Maps. Springer Series in Information Sciences, vol 30. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-97966-8_5
Download citation
DOI: https://doi.org/10.1007/978-3-642-97966-8_5
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-62017-4
Online ISBN: 978-3-642-97966-8
eBook Packages: Springer Book Archive