Comparing feedforward and recurrent neural network architectures with human behavior in artificial grammar learning

被引:0
作者
Andrea Alamia
Victor Gauducheau
Dimitri Paisios
Rufin VanRullen
机构
[1] CerCo,Laboratoire Cognition, Langues, Langage, Ergonomie
[2] CNRS,undefined
[3] CNRS,undefined
[4] Université Toulouse,undefined
[5] ANITI,undefined
[6] Université de Toulouse,undefined
来源
Scientific Reports | / 10卷
关键词
D O I
暂无
中图分类号
学科分类号
摘要
In recent years artificial neural networks achieved performance close to or better than humans in several domains: tasks that were previously human prerogatives, such as language processing, have witnessed remarkable improvements in state of the art models. One advantage of this technological boost is to facilitate comparison between different neural networks and human performance, in order to deepen our understanding of human cognition. Here, we investigate which neural network architecture (feedforward vs. recurrent) matches human behavior in artificial grammar learning, a crucial aspect of language acquisition. Prior experimental studies proved that artificial grammars can be learnt by human subjects after little exposure and often without explicit knowledge of the underlying rules. We tested four grammars with different complexity levels both in humans and in feedforward and recurrent networks. Our results show that both architectures can “learn” (via error back-propagation) the grammars after the same number of training sequences as humans do, but recurrent networks perform closer to humans than feedforward ones, irrespective of the grammar complexity level. Moreover, similar to visual processing, in which feedforward and recurrent architectures have been related to unconscious and conscious processes, the difference in performance between architectures over ten regular grammars shows that simpler and more explicit grammars are better learnt by recurrent architectures, supporting the hypothesis that explicit learning is best modeled by recurrent networks, whereas feedforward networks supposedly capture the dynamics involved in implicit learning.
引用
收藏
相关论文
共 50 条
  • [21] Learning behavior using multiresolution recurrent neural network
    Morita, S
    COMPUTER ANALYSIS OF IMAGES AND PATTERNS, 1999, 1689 : 157 - 166
  • [22] Neural network Architectures and learning
    Wilamowski, BM
    2003 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY, VOLS 1 AND 2, PROCEEDINGS, 2003, : TU1 - TU12
  • [23] Individual behavior in learning of an artificial grammar
    Zimmerer, Vitor C.
    Cowell, Patricia E.
    Varley, Rosemary A.
    MEMORY & COGNITION, 2011, 39 (03) : 491 - 501
  • [24] Individual behavior in learning of an artificial grammar
    Vitor C. Zimmerer
    Patricia E. Cowell
    Rosemary A. Varley
    Memory & Cognition, 2011, 39 : 491 - 501
  • [25] Characterize and Compare the Performance of Deep Learning Optimizers in Recurrent Neural Network Architectures
    Zaeed, Mohammad
    Islam, Tanzima Z.
    Indic, Vladimir
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 39 - 44
  • [26] COMPARISON OF FEEDFORWARD AND RECURRENT NEURAL NETWORK LANGUAGE MODELS
    Sundermeyer, M.
    Oparin, I.
    Gauvain, J. -L.
    Freiberg, B.
    Schlueter, R.
    Ney, H.
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 8430 - 8434
  • [27] RECURRENT NEURAL-NETWORK TRAINING WITH FEEDFORWARD COMPLEXITY
    OLUROTIMI, O
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02): : 185 - 197
  • [28] Neural Smithing – Supervised Learning in Feedforward Artificial Neural Networks
    Barbara Hammer
    Pattern Analysis & Applications, 2001, 4 (1) : 73 - 74
  • [29] Spatial Grammar-Based Recurrent Neural Network for Design Form and Behavior Optimization
    Stump, Gary M.
    Miller, Simon W.
    Yukish, Michael A.
    Simpson, Timothy W.
    Tucker, Conrad
    JOURNAL OF MECHANICAL DESIGN, 2019, 141 (12)
  • [30] Regularization and feedforward artificial neural network training with noise
    Chandra, P
    Singh, Y
    PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS 2003, VOLS 1-4, 2003, : 2366 - +