PhraseTransformer: an incorporation of local context information into sequence-to-sequence semantic parsing

被引:0
|
作者
Phuong Minh Nguyen
Tung Le
Huy Tien Nguyen
Vu Tran
Minh Le Nguyen
机构
[1] Japan Advanced Institute of Science and Technology,
[2] Faculty of Information Technology,undefined
[3] University of Science,undefined
[4] Vietnam National University,undefined
[5] Institute of Statistical Mathematics,undefined
来源
Applied Intelligence | 2023年 / 53卷
关键词
Semantic parsing; Local context; Transformer; Neural machine translation;
D O I
暂无
中图分类号
学科分类号
摘要
Semantic parsing is a challenging task mapping a natural language utterance to machine-understandable information representation. Recently, approaches using neural machine translation (NMT) have achieved many promising results, especially the Transformer. However, the typical drawback of adapting the vanilla Transformer to semantic parsing is that it does not consider the phrase in expressing the information of sentences while phrases play an important role in constructing the sentence meaning. Therefore, we propose an architecture, PhraseTransformer, that is capable of a more detailed meaning representation by learning the phrase dependencies in the sentence. The main idea is to incorporate Long Short-Term Memory into the Self-Attention mechanism of the original Transformer to capture the local context of a word. Experimental results show that our proposed model performs better than the original Transformer in terms of understanding sentences structure as well as logical representation and raises the model local context-awareness without any support from external tree information. Besides, although the recurrent architecture is integrated, the number of sequential operations of the PhraseTransformer is still O(1)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\boldsymbol {\mathcal {O}}\mathbf {(1)}$\end{document} similar to the original Transformer. Our proposed model achieves strong competitive performance on Geo and MSParS datasets, and leads to SOTA performance on the Atis dataset for methods using neural networks. In addition, to prove the generalization of our proposed model, we also conduct extensive experiments on three translation datasets IWLST14 German-English, IWSLT15 Vietnamese-English, WMT14 English-German, and show significant improvement. Our code is available at https://github.com/phuongnm94/PhraseTransformer.git.
引用
收藏
页码:15889 / 15908
页数:19
相关论文
共 50 条
  • [1] PhraseTransformer: an incorporation of local context information into sequence-to-sequence semantic parsing
    Nguyen, Phuong Minh
    Le, Tung
    Nguyen, Huy Tien
    Tran, Vu
    Nguyen, Minh Le
    APPLIED INTELLIGENCE, 2023, 53 (12) : 15889 - 15908
  • [2] Sequence-to-sequence AMR Parsing with Ancestor Information
    Yu, Chen
    Gildea, Daniel
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 571 - 577
  • [3] Improving Sequence-to-Sequence Constituency Parsing
    Liu, Lemao
    Zhu, Muhua
    Shi, Shuming
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 4873 - 4880
  • [4] Deterministic Attention for Sequence-to-Sequence Constituent Parsing
    Ma, Chunpeng
    Liu, Lemao
    Tamura, Akihiro
    Zhao, Tiejun
    Sumita, Eiichiro
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3237 - 3243
  • [5] Neural AMR: Sequence-to-Sequence Models for Parsing and Generation
    Konstas, Ioannis
    Iyer, Srinivasan
    Yatskar, Mark
    Choi, Yejin
    Zettlemoyer, Luke
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 146 - 157
  • [6] Semantic Matching for Sequence-to-Sequence Learning
    Zhang, Ruiyi
    Chen, Changyou
    Zhang, Xinyuan
    Bai, Ke
    Carin, Lawrence
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 212 - 222
  • [7] A Sequence-to-Sequence Model for Semantic Role Labeling
    Daza, Angel
    Frank, Anette
    REPRESENTATION LEARNING FOR NLP, 2018, : 207 - 216
  • [8] Improving AMR Parsing with Sequence-to-Sequence Pre-training
    Xu, Dongqin
    Li, Junhui
    Zhu, Muhua
    Min Zhang
    Zhou, Guodong
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2501 - 2511
  • [9] Unleashing the True Potential of Sequence-to-Sequence Models for Sequence Tagging and Structure Parsing
    He, Han
    Choi, Jinho D.
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 582 - 599
  • [10] Concept Identification with Sequence-to-Sequence Models in Abstract Meaning Representation Parsing
    Batiz, Orsolya Bernadeu
    Helmer, Robert Paul
    Pop, Roxana
    Macicasan, Florin
    Lemnaru, Camelia
    2020 IEEE 16TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTER COMMUNICATION AND PROCESSING (ICCP 2020), 2020, : 83 - 90