Abstract

Kohonen's self-organizing map (SOM), which performs topology-preserving transformation from a high dimensional data vector space to a low-dimensional map space, provides a powerful tool for data analysis, classification and visualization in many application fields. Despite its power, SOM can only deal with vectorized data, although many expansions have been proposed for various data-type cases. This study aims to develop a novel generalization of SOM called modular network SOM (mnSOM), which enables users to deal with general data classes in a consistent manner. mnSOM has an array structure consisting of function modules that are trainable neural networks, e.g. multi-layer perceptrons (MLPs), instead of the vector units of the conventional SOM family. In the case of MLP-modules, mnSOM learns a group of systems or functions in terms of the input-output relationships, and at the same time, mnSOM generates a feature map that shows distances between the learned systems. Thus, mnSOM with MLP modules is an SOM in function space rather than in vector space. From this point of view, the conventional SOM of Kohonen's can be regarded as a special case of mnSOM, the modules consisting of fixed-value bias units. In this paper, mnSOM with MLP modules is described along with some application examples.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call