LSTM Networks Can Perform Dynamic Counting

被引:0
|
作者
Suzgun, Mirac [1 ]
Gehrmann, Sebastian [1 ]
Belinkov, Yonatan [1 ,2 ]
Shieber, Stuart M. [1 ]
机构
[1] Harvard A John Paulson Sch Engn & Appl Sci, Cambridge, MA 02138 USA
[2] MIT, Comp Sci & Artificial Intelligence Lab, 77 Massachusetts Ave, Cambridge, MA 02139 USA
来源
DEEP LEARNING AND FORMAL LANGUAGES: BUILDING BRIDGES | 2019年
关键词
CONTEXT-FREE; LEARN;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we systematically assess the ability of standard recurrent networks to perform dynamic counting and to encode hierarchical representations. All the neural models in our experiments are designed to be small-sized networks both to prevent them from memorizing the training sets and to visualize and interpret their behaviour at test time. Our results demonstrate that the Long Short-Term Memory (LSTM) networks can learn to recognize the well-balanced parenthesis language (Dyck-1) and the shuffles of multiple Dyck-1 languages, each defined over different parenthesis-pairs, by emulating simple realtime k-counter machines. To the best of our knowledge, this work is the first study to introduce the shuffle languages to analyze the computational power of neural networks. We also show that a single-layer LSTM with only one hidden unit is practically sufficient for recognizing the Dyck-1 language. However, none of our recurrent networks was able to yield a good performance on the Dyck-2 language learning task, which requires a model to have a stack-like mechanism for recognition.
引用
收藏
页码:44 / 54
页数:11
相关论文
共 2 条
  • [1] Use Coupled LSTM Networks to Solve Constrained Optimization Problems
    Chen, Zheyu
    Leung, Kin K.
    Wang, Shiqiang
    Tassiulas, Leandros
    Chan, Kevin
    Towsley, Don
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2023, 9 (02) : 304 - 316
  • [2] Yes we can: Improving performance in dynamic tasks
    Qudrat-Ullah, Hassan
    DECISION SUPPORT SYSTEMS, 2014, 61 : 23 - 33