222826308.Pdf

Total Page:16

File Type:pdf, Size:1020Kb

Load more

View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Aston Publications Explorer Published in: Neural Computation 18, 446-469 (2006) Magnification Control in Self-Organizing Maps and Neural Gas Thomas Villmann Clinic for Psychotherapy, University of Leipzig, 04107 Leipzig, Germany Jens Christian Claussen Institute of Theoretical Physics and Astrophysics, Christian-Albrecht University Kiel, 24098 Kiel, Germany We consider different ways to control the magnification in self-organizing maps (SOM) and neural gas (NG). Starting from early approaches of magnification control in vector quantization, we then concentrate on different approaches for SOM and NG. We show that three structurally similar approaches can be applied to both algorithms: localized learning, concave-convex learning, and winner relaxing learning. Thereby, the approach of concave-convex learning in SOM is extended to a more general description, whereas the concave-convex learning for NG is new. In general, the control mechanisms generate only slightly different behavior comparing both neural algorithms. However, we emphasize that the NG results are valid for any data dimension, whereas in the SOM case the results hold only for the one-dimensional case. 1 Introduction Vector quantization is an important task in data processing, pattern recog- nition and control (Fritzke, 1993; Haykin, 1994; Linde, Buzo, & Gray, 1980; Ripley, 1996). A large number of different types have been dis- cussed, (for an overview, refer to Haykin, 1994; Kohonen, 1995; Duda & Hart, 1973). Neural maps are a popular type of neural vector quan- tizers that are commonly used in, for example, data visualization, feature extraction, principle component analysis, image processing, classification tasks, and acceleration of common vector quantization (de Bodt, Cot- trell, Letremy, & Verleysen, 2004). Well known approaches are the Self- arXiv:cond-mat/0609517v1 [cond-mat.dis-nn] 20 Sep 2006 Organizing Map (SOM) (Kohonen, 1995), the neural gas (NG) (Martinetz, Berkovich, & Schulten, 1993), elastic net (EN) (Durbin & Willshaw, 1987) and generative topographic mapping (GTM) (Bishop, Svens´en, & Williams 1998). In vector quantization, data vectors v ∈ Rd are represented by a few codebooks or weight vectors wi, where i is an arbitrary index. Several cri- teria exist to evaluate the quality of a vector quantizer. The most common 1 one is the squared reconstruction error. However, other quality criteria are also known, for instance topographic quality for neighborhood preserving mapping approaches (Bauer & Pawelzik, 1992; Bauer, Der, & Villmann, 1999), optimization of mutual information (Linsker, 1989) and other cri- teria (for an overview, see Haykin, 1994). Generally, a faithful represen- tation of the data space by the codebooks is desired. This property is closely related to the so-called magnification, which describes the relation between data and weight vector density for a given model. The knowledge of magnification of a map is essential for correct interpretation of its output (Hammer & Villmann, 2003). In addition, explicit magnification control is a desirable property of learning algorithms, if depending on the respective application, only sparsely covered regions of the data space have to be em- phasized or, conversely, suppressed. The magnification can be explicitly expressed for several vector quantization models. Usually, for these ap- proaches the magnification can be expressed by a power law between the codebook vector density ρ and the data density P . The respective exponent is called magnification exponent or magnification factor. As explained in more detail below, the magnification is also related to other properties of the map, for example, reconstruction error as well as mutual information. Hence, control of magnification is influencing these properties too. In biologically motivated approaches, magnification can also be seen in the context of information representation in brains, for instance, in the senso-motoric cortex (Ritter, Martinetz, & Schulten, 1992). Magnification and its control can be related to biological phenomena like the perceptual magnet effect, which refers to the fact that rarely occurring stimuli are dif- ferentiated with high precision whereas frequent stimuli are distinguished only in a rough manner (Kuhl, 1991; Kuhl, Williams, Lacerda, Stevens, & Lindblom, 1992). It is a kind of attention-based learning with inverted magnification, that is, rarely occurring input samples are emphasized by an increased learning gain (Der & Herrmann, 1992; Herrmann, Bauer, & Der, 1994). This effect is also beneficial in technical systems. In remote- sensing image analysis, for instance, seldomly found ground cover classes should be detected, whereas usual (frequent) classes with broad variance should be suppressed (Mer´enyi & Jain, 2004; Villmann, Mer´enyi & Ham- mer, 2003). Another technical environment for magnification control is robotics for accurate description of dangerous navigation states (Villmann & Heinze, 2000). In this article we concentrate on a general framework for magnification control in SOM and NG. In this context, we briefly review the most im- portant approaches. One approach for SOM is generalized, and afterward, it is transferred to NG. For this purpose, we first give the basic notations, followed in section 3 by a more detailed description of magnification and early approaches related to the topic of magnification control, including a unified approach for controlling strategies. The magnification control approaches of SOM are described according to the unified framework in 2 section 4, whereby one of them is significantly extended. The same pro- cedure is applied to NG in section 5. Again, one of the control approaches presented in this section is new. A short discussion concludes the article. 2 Basic Concepts and Notations in SOM and NG In general, neural maps project data vectors v from a (possibly high- dimensional) data manifold D ⊆Rd onto a set A of neurons i, which is formally written as ΨD→A : D → A. Each neuron i is associated with a w Rd W w pointer i ∈ , all of which establish the set = { i}i∈A. The map- ping description is a winner-take-all rule, that is, a stimulus vector v ∈D is mapped onto that neuron s ∈ A with the pointer ws being closest to the actual presented stimulus vector v, ΨD→A : v 7→ s (v) = argmin kv − wik . (2.1) i∈A The neuron s is called winner neuron. The set Ri = {v ∈ D|ΨD→A (v)= i} is called the (masked) receptive field of the neuron i. The weight vectors are adapted during the learning process such that the data distribution is represented. For further investigations, we describe SOM and NG as our focused neural maps in more detail. During the adaptation process a sequence of data points v ∈D is presented to the map with respect to the data distribu- tion P (D). Then the most proximate neuron s according to equation (2.1) is determined, and the pointer ws, as well as all pointers wi of neurons in the neighborhood of s, are shifted towards v, according to △wi = ǫh (i, v, W)(v − wi) . (2.2) The property of “being in the neighborhood of s” is represented by a neigh- borhood function h (i, v, W). The neighborhood function is defined as k (v, W) h (i, v, W) = exp − i (2.3) λ λ for the NG, where ki (v, W) yields the number of pointers wj for which the relation kv − wjk < kv − wik is valid (Martinetz et al., 1993); espe- cially, we have hλ (s, v, W)=1.0. In case of SOM the set A of neurons has a topological structure usually chosen as a hypercube or hexagonal lat- tice. Each neuron i has a fixed position r (i). The neighborhood function has the form kr (i) − r (s (v))k h (i, v, W) = exp − A . (2.4) σ 2σ2 3 In contrast to the NG, the neighborhood function of SOM is evaluated in the output space A according to its topological structure. This difference causes the significantly different properties of both algorithms. For the SOM there does not exist any energy function such that the adaptation rule follows the gradient descent (Erwin, Obermayer, & Schulten, 1992). Moreover, the convergence proofs are only valid for the one-dimensional setting (Cottrell, Fort & Pages, 1998, Ritter et al., 1992). The introduction of an energy function leads to different dynamics as in the EN (Durbin & Willshaw, 1987) or new winner determination rule (Heskes 1999). The advantage of the SOM is the ordered topological structure of neurons in A. In contrast, in the original NG, such an order is not given. One can extend the NG to the topology representing network (TRN) such that topological relations between neurons are installed during learning, although generally they do not achieve the simple structure as in SOM lattices (Martinetz & Schulten, 1994). Finally, the important advantage of the NG is that the adaptation dynamic of the weight vectors follows a potential minimizing dynamics (Martinetz et al., 1993). 3 Magnification and Magnification Control in Vector Quantization 3.1 Magnification in Vector Quantization. Usually vector quantization aims to minimize the reconstruction error RE = kv − w k2 P (v) dv. However, other quality criteria are i Ri i also known,P R for instance, topographic quality (Bauer & Pawelzik, 1992; Bauer et al., 1999). More generally, one can consider the generalized dis- tortion error, γ Eγ = kws − vk P (v) dv. (3.1) ZD This error is closely related to other properties of the (neural) vector quan- tizer. One important property is the achieved weight vector density ρ (w) after learning in relation to the data density P (D). Generally, for vector quantizers one finds the relation P (w) ∝ ρ (w)α (3.2) after the converged learning process (Zador 1982). The exponent α is called magnification exponent or magnification factor. The magnification is coupled with the generalized distortion error (3.1) by d α = (3.3) d + γ 4 Table 1: Magnification of Different Neural Maps and Vector Quantization Ap- proaches.
Recommended publications
  • Growing Neural Gas with Correntropy Induced Metric

    Growing Neural Gas with Correntropy Induced Metric

    Growing Neural Gas with Correntropy Induced Metric Naoki Masuyama Chu Kiong Loo Faculty of Computer Science and Information Technology, Faculty of Computer Science and Information Technology, University of Malaya, University of Malaya, 50603 Kuala Lumpur, MALAYSIA 50603 Kuala Lumpur, MALAYSIA Email: [email protected] Email: [email protected] Abstract—This paper discusses the Correntropy Induced Met- of arbitrary shapes without any assumptions on the number ric (CIM) based Growing Neural Gas (GNG) architecture. CIM is of clusters. Boonmee et al. [13] introduced a hybrid GNG a kernel method based similarity measurement from the informa- by considering not only the distance between the clusters, tion theoretic learning perspective, which quantifies the similarity between probability distributions of input and reference vectors. but also the centroids of the each cluster to obtain more We apply CIM to find a maximum error region and node insert practical topology structure. Mohebi and Bagirov [14] applied criterion, instead of euclidean distance based function in original an algorithm based on split and merge of clusters to initialize GNG. Furthermore, we introduce the two types of Gaussian neurons. The initialization algorithm speeds up the learning kernel bandwidth adaptation methods for CIM. The simulation process in large high-dimensional data sets. Palomo et al. experiments in terms of the affect of kernel bandwidth σ in CIM, the self-organizing ability, and the quantitative comparison show [15] combined GNG and tree structure to represent a network that proposed model has the superior abilities than original GNG. topology, which is called Growing Neural Forest (GNF). Index Terms—Growing Neural Gas; Correntropy Induced GNF learns a set of trees so that each tree represents a Metric; Kernel Method; Clustering; connected cluster of data.
  • Probabilistic Graphical Model Based on Growing Neural Gas for Long Time Series Classification

    Probabilistic Graphical Model Based on Growing Neural Gas for Long Time Series Classification

    Modern Applied Science; Vol. 9, No. 2; 2015 ISSN 1913-1844 E-ISSN 1913-1852 Published by Canadian Center of Science and Education Probabilistic Graphical Model Based on Growing Neural Gas for Long Time Series Classification Irina N. Palamar1 & Sergey S. Yulin1 1 Soloviev Rybinsk State Aviation Technical University, Rybinsk, Russia Correspondence: Sergey S. Yulin, Soloviev Rybinsk State Aviation Technical University, Rybinsk, Russia. Received: September 10, 2014 Accepted: October 16, 2014 Online Published: December 11, 2014 doi:10.5539/mas.v9n2p108 URL: http://dx.doi.org/10.5539/mas.v9n2p108 Abstract This article proposes a generative probabilistic graphical model with hidden states (Neural Gas Graphical Model (NGGM)) based on data approximation with a grid of "neural gas" nodes aimed at solving the problem of long time series classification. Such time series include information about changes in economic, weather and health values, as well as information about changes in values of operation sensors of technical objects during a quite long period. The most difficult task of classification of such long time series using probabilistic graphical models with hidden states is the selection of the optimum number of hidden states. This work proposes a method for automatic selection of the optimum number of hidden states of the model in the course of the model learning. The model proposed in the article and the methods of its learning are based on a combination of elements used in the metric and Bayesian approaches to classification. The basic NGGM purpose is to match hidden states of a graphical model and nodes (neurons) of the approximating grid.
  • Neural Network Classification [Roya Asadi*, Haitham Sabah Hasan, Sameem Abdul Kareem]

    Neural Network Classification [Roya Asadi*, Haitham Sabah Hasan, Sameem Abdul Kareem]

    Proc. of the Intl. Conf. on Advances in Computer Science and Electronics Engineering -- CSEE 2014 Copyright © Institute of Research Engineers and Doctors. All rights reserved. ISBN: 978-1-63248-000-2 doi: 10.15224/ 978-1-63248-000-2-36 Review of current Online Dynamic Unsupervised Feed Forward Neural Network classification [Roya Asadi*, Haitham Sabah Hasan, Sameem Abdul Kareem] Abstract— Online Dynamic Unsupervised Feed Forward Neural networks are able to dynamically learn the types of Neural Network (ODUFFNN) classification is suitable to be input values based on their weights and properties. A feed applied in different research areas and environments such as forward neural network is a popular tool for statistical decision email logs, networks, credit card transactions, astronomy and making and is a software version of the brain. The neural satellite communications. Currently, there are a few strong network is flexible algorithm that allows us to encode methods as ODUFFNN classification, although they have general nonlinear relationships between input and desirable outputs [5- problems. The goal of this research is an investigation of the 7]. The online dynamic unsupervised classification must solve critical problems and comparison of current ODUFFNN several problems such as a huge data with high dimensions classification. For experimental results, Evolving Self-Organizing which causes huge memory usage, high processing time and Map (ESOM) and Dynamic Self-Organizing Map (DSOM) as low accuracy [2, 8, 9]; losing data details because of dividing strong related methods are considered; and also they applied some difficult datasets for clustering from the UCI Dataset input values in a few clusters [8, 10, 11]; defining the Repository.
  • Batch and Median Neural Gas Marie Cottrell, Barbara Hammer, Alexander Hasenfuss, Thomas Villmann

    Batch and Median Neural Gas Marie Cottrell, Barbara Hammer, Alexander Hasenfuss, Thomas Villmann

    Batch and median neural gas Marie Cottrell, Barbara Hammer, Alexander Hasenfuss, Thomas Villmann To cite this version: Marie Cottrell, Barbara Hammer, Alexander Hasenfuss, Thomas Villmann. Batch and median neural gas. Neural Networks, Elsevier, 2006, 19, pp.762-771. 10.1016/j.neunet.2006.05.018. hal-00107462 HAL Id: hal-00107462 https://hal.archives-ouvertes.fr/hal-00107462 Submitted on 18 Oct 2006 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Batch and median neural gas Marie Cottrell SAMOS-MATISSE, Universite´ Paris I, Paris, France Barbara Hammer 1 Institute of Computer Science, Clausthal University of Technology, Clausthal-Zellerfeld, Germany Alexander Hasenfuß Institute of Computer Science, Clausthal University of Technology, Clausthal-Zellerfeld, Germany Thomas Villmann Clinic for Psychotherapy, Universitat¨ Leipzig, Leipzig, Germany ccsd-00107462, version 1 - 18 Oct 2006 1 corresponding author: Barbara Hammer, Institute of Computer Science, Clausthal University of Technology, Julius-Albert-Str. 4, D-38678 Clausthal-Zellerfeld, Germany, [email protected] Preprint submitted to Elsevier Science 18 October 2006 Abstract Neural Gas (NG) constitutes a very robust clustering algorithm given euclidian data which does not suffer from the problem of local minima like simple vector quantization, or topo- logical restrictions like the self-organizing map.
  • Arxiv:1509.07093V1 [Cs.LG] 23 Sep 2015

    Arxiv:1509.07093V1 [Cs.LG] 23 Sep 2015

    Noname manuscript No. (will be inserted by the editor) A Review of Learning Vector Quantization Classifiers David Nova · Pablo A. Estévez the date of receipt and acceptance should be inserted later Abstract In this work we present a review of the state of the art of Learning Vector Quantization (LVQ) classifiers. A taxonomy is proposed which inte- grates the most relevant LVQ approaches to date. The main concepts associ- ated with modern LVQ approaches are defined. A comparison is made among eleven LVQ classifiers using one real-world and two artificial datasets. Keywords Learning Vector Quantization · Supervised Learning · Neural Networks · Margin Maximization · Likelihood Ratio Maximization 1 Introduction Learning Vector Quantization (LVQ) is a family of algorithms for statistical pattern classification, which aims at learning prototypes (codebook vectors) representing class regions. The class regions are defined by hyperplanes be- tween prototypes, yielding Voronoi partitions. In the late 80’s Teuvo Kohonen introduced the algorithm LVQ1 [36, 38], and over the years produced several variants. Since their inception LVQ algorithms have been researched by a small but active community. A search on the ISI Web of Science in November, 2013, found 665 journal articles with the keywords “Learning Vector Quantization” or “LVQ” in their titles or abstracts. This paper is a review of the progress made in the field during the last 25 years. D. Nova Department of Electrical Engineering, Faculty of Physical and Mathematical Sciences, Uni- versity of Chile, Santiago, Chile E-mail: [email protected] P. A. Estévez arXiv:1509.07093v1 [cs.LG] 23 Sep 2015 Department of Electrical Engineering and Advanced Mining Technology Center, Faculty of Physical and Mathematical Sciences, University of Chile, Santiago, Chile E-mail: [email protected] 2 David Nova, Pablo A.
  • A New Self-Organizing Neural Gas Model Based on Bregman Divergences

    A New Self-Organizing Neural Gas Model Based on Bregman Divergences

    A New Self-Organizing Neural Gas Model based on Bregman Divergences Esteban J. Palomo∗, Miguel A. Molina-Cabello∗, Ezequiel Lopez-Rubio´ ∗ and Rafael Marcos Luque-Baena∗ ∗Department of Computer Languages and Computer Science University of Malaga,´ Bulevar Louis Pasteur, 35, 29071 Malaga,´ Spain Emails: fejpalomo,miguelangel,ezeqlr,[email protected] Abstract—In this paper, a new self-organizing neural gas by using Bregman divergences the most suitable divergence model that we call Growing Hierarchical Bregman Neural according to input data can be specified. In this paper, a new Gas (GHBNG) has been proposed. Our proposal is based self-organizing neural network called the Growing Hierar- on the Growing Hierarchical Neural Gas (GHNG) in which chical Bregman Neural Gas (GHBNG) is proposed, which Bregman divergences are incorporated in order to compute is grounded in the GHNG model and in which Bregman the winning neuron. This model has been applied to anomaly divergences have been considered. detection in video sequences together with a Faster R-CNN On the other hand, the proliferation in recent years of as an object detector module. Experimental results not only a huge amount of visual information in the form of data confirm the effectiveness of the GHBNG for the detection sequences has led to a growth in the field of intelligent video of anomalous object in video sequences but also its self- surveillance. In particular, one of the most important tasks organization capabilities. to consider is to automatically detect moving objects that are not very frequent in a scene and can be considered as anomalies. In recent years, the appearance of deep learning 1.
  • 1 Introduction

    1 Introduction

    Controversial Empirical Results on Batch versus One Pass Online Algorithms CONTROVERSIAL EMPIRICAL RESULTS ON BATCH VERSUS ONE PASS ONLINE ALGORITHMS A. I. González, M. Graña1 Grupo de Inteligencia Computacional Facultad de Informática, UPV/EHU Apdo. 649, 20080 San Sebastián, España [email protected] Abstract – We present controversial empirical results about the relative convergence of batch and online neural network vector quantization (VQ) learning algorithms. It is the commonplace belief that online algorithms have a very slow convergence, while batch versions are faster and give better results. However we propose an online realization of the Self-Organizing Map (SOM) and the Neural Gas algorithm, with learning parameters tuned to perform the training in a one-pass over the sample, that are faster than their batch counterparts and that improve them on some benchmark data sets. Although these results are too limited to claim any kind of superiority of the One Pass algorithms over their batch counterparts, they nevertheless must be taken into account as existence proofs contradicting the claim that the batch realizations are faster and more accurate than online algorithms. Keywords – SOM, Neural Gas, Batch, One pass online 1 Introduction Vector Quantization [9, 10, 12] is a technique that maps a set of input vectors into a finite collection of predetermined codevectors. The set of all codevectors is called the codebook. In designing a vector quantizer, the goal is to construct a codebook for which the expected distortion of approximating any input vector by a codevector is minimized. This task has been tackled by with Competitive Neural Networks (CNN) [1].
  • Self-Organizing Neural Networks for Sequence Processing

    Self-Organizing Neural Networks for Sequence Processing

    University of OsnabruckÄ Institute of Computer Science Department of Mathematics and Computer Science Self-Organizing Neural Networks for Sequence Processing Thesis by Marc Strickert In partial ful¯llment of the requirements for the degree of Doctor of Computer Science (Dr. rer. nat.) Osnabruck,Ä Germany June 7, 2004 Referees: Prof. Dr. Barbara Hammer, Research Group `Learning with Neural Methods on Structured Data' (LNM), Technical University of Clausthal, Germany. Prof. Dr. Helge Ritter, Faculty of Technology, Neuroinformatics Group, University of Bielefeld, Germany. Self-Organizing Neural Networks for Sequence Processing Arbeitsgruppe LNM Institut furÄ Informatik am Fachbereich furÄ Mathematik, Informatik und Angewandte Systemwissenschaft UniversitÄat OsnabruckÄ D-49069 Osnabruck,Ä Germany Arbeitsgruppenleiterin: Dr. Barbara Hammer Institutsdirektor: Prof. Dr. Oliver Vornberger Autor: Dipl.-Systemwiss. Marc Strickert Osnabruck,Ä Juni 2004 Abstract This work investigates the self-organizing representation of temporal data in prototype- based neural networks. Extensions of the supervised learning vector quantization (LVQ) and the unsupervised self-organizing map (SOM) are considered in detail. The principle of Hebbian learning through prototypes yields compact data models that can be easily interpreted by similarity reasoning. In order to obtain a robust prototype dynamic, LVQ is extended by neighborhood cooperation between neurons to prevent a strong dependence on the initial prototype locations. Additionally, implementations of more general, adap- tive metrics are studied with a particular focus on the built-in detection of data attributes involved for a given classi¯cation task. For unsupervised sequence processing, two modi¯- cations of SOM are pursued: the SOM for structured data (SOMSD) realizing an e±cient back-reference to the previous best matching neuron in a triangular low-dimensional neural lattice, and the merge SOM (MSOM) expressing the temporal context as a fractal combi- nation of the previously most active neuron and its context.