Distributed associative memory network with memory refreshing loss

被引:4
作者
Park, Taewon [1 ]
Choi, Inchul [2 ]
Lee, Minho [1 ,2 ]
机构
[1] Kyungpook Natl Univ, Dept Artificial Intelligence, 80 Daehak Ro, Daegu, South Korea
[2] NEOALI, 80 Daehak Ro, Daegu, South Korea
关键词
Memory augmented neural network; Relational reasoning; Distributed representation; Auxiliary loss; Machine learning; NEURAL-NETWORKS; MAINTENANCE REHEARSAL; ITEM;
D O I
10.1016/j.neunet.2021.07.030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite recent progress in memory augmented neural network (MANN) research, associative memory networks with a single external memory still show limited performance on complex relational reasoning tasks. Especially the content-based addressable memory networks often fail to encode input data into rich enough representation for relational reasoning and this limits the relation modeling performance of MANN for long temporal sequence data. To address these problems, here we introduce a novel Distributed Associative Memory architecture (DAM) with Memory Refreshing Loss (MRL) which enhances the relation reasoning performance of MANN. Inspired by how the human brain works, our framework encodes data with distributed representation across multiple memory blocks and repeatedly refreshes the contents for enhanced memorization similar to the rehearsal process of the brain. For this procedure, we replace a single external memory with a set of multiple smaller associative memory blocks and update these sub-memory blocks simultaneously and independently for the distributed representation of input data. Moreover, we propose MRL which assists a task's target objective while learning relational information existing in data. MRL enables MANN to reinforce an association between input data and task objective by reproducing stochastically sampled input data from stored memory contents. With this procedure, MANN further enriches the stored representations with relational information. In experiments, we apply our approaches to Differential Neural Computer (DNC), which is one of the representative content-based addressing memory models and achieves the state-of-the-art performance on both memorization and relational reasoning tasks. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页码:33 / 48
页数:16
相关论文
共 66 条
[1]  
ACKLEY DH, 1985, COGNITIVE SCI, V9, P147
[2]  
Alonso HM, 2017, 15TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2017), VOL 1: LONG PAPERS, P44
[3]  
[Anonymous], 1994, THESIS
[4]   Long short-term memory [J].
Hochreiter, S ;
Schmidhuber, J .
NEURAL COMPUTATION, 1997, 9 (08) :1735-1780
[5]  
[Anonymous], 2017, P 55 ANN M ASS COMP, DOI DOI 10.18653/v1/P17-1194
[6]   Distributed associative memories for high-speed symbolic reasoning [J].
Austin, J .
FUZZY SETS AND SYSTEMS, 1996, 82 (02) :223-233
[7]   DISTRIBUTED ASSOCIATIVE MEMORY FOR USE IN SCENE ANALYSIS [J].
AUSTIN, J ;
STONHAM, TJ .
IMAGE AND VISION COMPUTING, 1987, 5 (04) :251-260
[8]  
Ba J., 2016, ARXIV160706450, V1050, P21
[9]  
Banino A., 2020, INT C LEARN REPR
[10]   Exploiting task relatedness for multiple task learning [J].
Ben-David, S ;
Schuller, R .
LEARNING THEORY AND KERNEL MACHINES, 2003, 2777 :567-580