Improving question answering performance using knowledge distillation and active learning

被引:5
|
作者
Boreshban, Yasaman [1 ]
Mirbostani, Seyed Morteza [2 ]
Ghassem-Sani, Gholamreza [1 ]
Mirroshandel, Seyed Abolghasem [2 ]
Amiriparian, Shahin [3 ]
机构
[1] Sharif Univ Technol, Comp Engn Dept, Tehran, Iran
[2] Univ Guilan, Dept Comp Engn, Rasht, Iran
[3] Univ Augsburg, Embedded Intelligence Hlth Care & Wellbeing, Augsburg, Germany
关键词
Natural language processing; Question answering; Deep learning; Knowledge distillation; Active learning; Performance;
D O I
10.1016/j.engappai.2023.106137
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Contemporary question answering (QA) systems, including Transformer-based architectures, suffer from increasing computational and model complexity which render them inefficient for real-world applications with limited resources. Furthermore, training or even fine-tuning such models requires a vast amount of labeled data which is often not available for the task at hand. In this manuscript, we conduct a comprehensive analysis of the mentioned challenges and introduce suitable countermeasures. We propose a novel knowledge distillation (KD) approach to reduce the parameter and model complexity of a pre-trained bidirectional encoder representations from transformer (BERT) system and utilize multiple active learning (AL) strategies for immense reduction in annotation efforts. We show the efficacy of our approach by comparing it with four state-of-the-art (SOTA) Transformers-based systems, namely KroneckerBERT, EfficientBERT, TinyBERT, and DistilBERT. Specifically, we outperform KroneckerBERT21 and EfficientBERTTINY by 4.5 and 0.4 percentage points in EM, despite having 75.0% and 86.2% fewer parameters, respectively. Additionally, our approach achieves comparable performance to 6-layer TinyBERT and DistilBERT while using only 2% of their total trainable parameters. Besides, by the integration of our AL approaches into the BERT framework, we show that SOTA results on the QA datasets can be achieved when we only use 40% of the training data. Overall, all results demonstrate the effectiveness of our approaches in achieving SOTA performance, while extremely reducing the number of parameters and labeling efforts. Finally, we make our code publicly available at https://github.com/mirbostani/QA-KD-AL.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Study of a lifelong learning scenario for question answering
    Echegoyen G.
    Rodrigo Á.
    Peñas A.
    Expert Systems with Applications, 2022, 209
  • [42] KNOWLEDGE ENHANCED LATENT RELEVANCE MINING FOR QUESTION ANSWERING
    Wang, Dong
    Shen, Ying
    Zheng, Hai-Tao
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 4282 - 4286
  • [43] Question Formulation and Question Answering for Knowledge Graph Completion
    Khvalchik, Maria
    Blaschke, Christian
    Revenko, Artem
    DATABASE AND EXPERT SYSTEMS APPLICATIONS (DEXA 2019), 2019, 1062 : 166 - 171
  • [44] QUESTION ANSWERING AND LEARNING WITH HYPERTEXT
    ROUET, JF
    LESSONS FROM LEARNING, 1994, 46 : 39 - 52
  • [45] Knowledge-based question answering
    Hermjakob, U
    Hovy, EH
    Lin, CY
    6TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL XVI, PROCEEDINGS: COMPUTER SCIENCE III, 2002, : 66 - 71
  • [46] Question answering as a knowledge discovery technique
    Chali, Yllias
    WEBIST 2008: PROCEEDINGS OF THE FOURTH INTERNATIONAL CONFERENCE ON WEB INFORMATION SYSTEMS AND TECHNOLOGIES, VOL 2, 2008, : 339 - 342
  • [47] Joint Learning of Question Answering and Question Generation
    Sun, Yibo
    Tang, Duyu
    Duan, Nan
    Qin, Tao
    Liu, Shujie
    Yan, Zhao
    Zhou, Ming
    Lv, Yuanhua
    Yin, Wenpeng
    Feng, Xiaocheng
    Qin, Bing
    Liu, Ting
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2020, 32 (05) : 971 - 982
  • [48] Question Answering on Scholarly Knowledge Graphs
    Jaradeh, Mohamad Yaser
    Stocker, Markus
    Auer, Soeren
    DIGITAL LIBRARIES FOR OPEN KNOWLEDGE, TPDL 2020, 2020, 12246 : 19 - 32
  • [49] Query-Driven Knowledge Graph Construction using Question Answering and Multimodal Fusion
    Peng, Yang
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 1119 - 1126
  • [50] Using Semantic Constraints for Question Answering
    Zeng, Qingpeng
    Wu, Shuixiu
    ENGINEERING SOLUTIONS FOR MANUFACTURING PROCESSES, PTS 1-3, 2013, 655-657 : 1750 - +