A general framework for unsupervised processing of structured data

被引:51
作者
Hammer, B
Micheli, A
Sperduti, A
Strickert, M
机构
[1] Univ Osnabruck, Dept Math & Comp Sci, Res Grp, LNM, D-49069 Osnabruck, Germany
[2] Univ Pisa, Dipartimento Informat, Pisa, Italy
[3] Univ Padua, Dipartimento Matemat Pura & Applicata, Padua, Italy
关键词
self-organizing map; Kohonen map; recurrent networks; SOM for structured data;
D O I
10.1016/j.neucom.2004.01.008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-organization constitutes an,important paradigm in machine learning with successful applications e.g. in data- and web-mining. Most approaches, however, have been proposed for processing data contained in a fixed and finite dimensional vector space. In this article, we will focus on extensions to more general data structures like sequences and tree structures. Various modifications of the standard self-organizing map (SOM) to sequences or tree structures have been proposed in the literature some of which are the temporal Kohonen map, the recursive SOM, and SOM for structured data. These methods enhance the standard SOM by utilizing recursive connections. We define a general recursive dynamic in this article which provides recursive processing of complex data structures by recursive computation of internal representations for the given context. The above mentioned mechanisms of SOMs for structures are special cases of the proposed general dynamic. Furthermore, the dynamic covers the supervised case of recurrent and recursive networks. The general framework offers an uniform notation for training mechanisms such as Hebbian learning. Moreover, the transfer of computational alternatives such as vector quantization or the neural gas algorithm to structure processing networks can be easily achieved. One can formulate general cost functions corresponding to vector quantization, neural gas, and a modification of SOM. The cost functions can be compared to Hebbian learning which can be interpreted as an approximation of a stochastic gradient descent. For comparison, we derive the exact gradients for general cost functions. (C) 2004 Elsevier B.V. All rights reserved.
引用
收藏
页码:3 / 35
页数:33
相关论文
共 52 条
  • [1] [Anonymous], 2001, P 3 IAPR TC 15 WORKS
  • [2] [Anonymous], WORKSH SELF ORG MAPS
  • [3] [Anonymous], ESANN
  • [4] Structural basis for triplet repeat disorders: a computational analysis
    Baldi, P
    Brunak, S
    Chauvin, Y
    Pedersen, AG
    [J]. BIOINFORMATICS, 1999, 15 (11) : 918 - 929
  • [5] BARRETO GA, 2001, INT J COMPUT RES, V10, P139
  • [6] LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT
    BENGIO, Y
    SIMARD, P
    FRASCONI, P
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02): : 157 - 166
  • [7] Application of cascade correlation networks for structures to chemistry
    Bianucci, AM
    Micheli, A
    Sperduti, A
    Starita, A
    [J]. APPLIED INTELLIGENCE, 2000, 12 (1-2) : 117 - 146
  • [8] THE TEMPORAL KOHONEN MAP
    CHAPPELL, GJ
    TAYLOR, JG
    [J]. NEURAL NETWORKS, 1993, 6 (03) : 441 - 445
  • [9] COSTA F, IN PRESS APPL INTELL
  • [10] Theoretical aspects of the SOM algorithm
    Cottrell, M
    Fort, JC
    Pagès, G
    [J]. NEUROCOMPUTING, 1998, 21 (1-3) : 119 - 138