Improving question answering performance using knowledge distillation and active learning

被引:5
|
作者
Boreshban, Yasaman [1 ]
Mirbostani, Seyed Morteza [2 ]
Ghassem-Sani, Gholamreza [1 ]
Mirroshandel, Seyed Abolghasem [2 ]
Amiriparian, Shahin [3 ]
机构
[1] Sharif Univ Technol, Comp Engn Dept, Tehran, Iran
[2] Univ Guilan, Dept Comp Engn, Rasht, Iran
[3] Univ Augsburg, Embedded Intelligence Hlth Care & Wellbeing, Augsburg, Germany
关键词
Natural language processing; Question answering; Deep learning; Knowledge distillation; Active learning; Performance;
D O I
10.1016/j.engappai.2023.106137
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Contemporary question answering (QA) systems, including Transformer-based architectures, suffer from increasing computational and model complexity which render them inefficient for real-world applications with limited resources. Furthermore, training or even fine-tuning such models requires a vast amount of labeled data which is often not available for the task at hand. In this manuscript, we conduct a comprehensive analysis of the mentioned challenges and introduce suitable countermeasures. We propose a novel knowledge distillation (KD) approach to reduce the parameter and model complexity of a pre-trained bidirectional encoder representations from transformer (BERT) system and utilize multiple active learning (AL) strategies for immense reduction in annotation efforts. We show the efficacy of our approach by comparing it with four state-of-the-art (SOTA) Transformers-based systems, namely KroneckerBERT, EfficientBERT, TinyBERT, and DistilBERT. Specifically, we outperform KroneckerBERT21 and EfficientBERTTINY by 4.5 and 0.4 percentage points in EM, despite having 75.0% and 86.2% fewer parameters, respectively. Additionally, our approach achieves comparable performance to 6-layer TinyBERT and DistilBERT while using only 2% of their total trainable parameters. Besides, by the integration of our AL approaches into the BERT framework, we show that SOTA results on the QA datasets can be achieved when we only use 40% of the training data. Overall, all results demonstrate the effectiveness of our approaches in achieving SOTA performance, while extremely reducing the number of parameters and labeling efforts. Finally, we make our code publicly available at https://github.com/mirbostani/QA-KD-AL.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] KNOWLEDGE DISTILLATION FOR IMPROVED ACCURACY IN SPOKEN QUESTION ANSWERING
    You, Chenyu
    Chen, Nuo
    Zou, Yuexian
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7793 - 7797
  • [2] Improving Question Answering over Knowledge Graphs with a Chunked Learning Network
    Zuo, Zicheng
    Zhu, Zhenfang
    Wu, Wenqing
    Wang, Wenling
    Qi, Jiangtao
    Zhong, Linghui
    ELECTRONICS, 2023, 12 (15)
  • [3] A Video Question Answering Model Based on Knowledge Distillation
    Shao, Zhuang
    Wan, Jiahui
    Zong, Linlin
    INFORMATION, 2023, 14 (06)
  • [4] Video Question Answering Scheme Base on Multimodal Knowledge Active Learning
    Liu M.
    Wang R.
    Zhou F.
    Lin G.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (04): : 889 - 902
  • [5] Improving complex knowledge base question answering via structural information learning
    Zhang, Jinhao
    Zhang, Lizong
    Hui, Bei
    Tian, Ling
    KNOWLEDGE-BASED SYSTEMS, 2022, 242
  • [6] Adversarial Knowledge Distillation Based Biomedical Factoid Question Answering
    Bai, Jun
    Yin, Chuantao
    Zhang, Jianfei
    Wang, Yanmeng
    Dong, Yi
    Rong, Wenge
    Xiong, Zhang
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2023, 20 (01) : 106 - 118
  • [7] Malayalam Question Answering System Using Deep Learning Approaches
    Rahmath, Reji K.
    Raj, P. C. Reghu
    Rafeeque, P. C.
    IETE JOURNAL OF RESEARCH, 2023, 69 (12) : 8889 - 8901
  • [8] Improving the performance of question answering with semantically equivalent answer patterns
    Kosseim, Leila
    Yousefi, Jamileh
    DATA & KNOWLEDGE ENGINEERING, 2008, 66 (01) : 53 - 67
  • [9] A review of deep learning in question answering over knowledge bases
    Zhang, Chen
    Lai, Yuxuan
    Feng, Yansong
    Zhao, Dongyan
    AI OPEN, 2021, 2 : 205 - 215
  • [10] Developing a Vietnamese Tourism Question Answering System Using Knowledge Graph and Deep Learning
    Phuc Do
    Phan, Truong H., V
    Gupta, Brij B.
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (05)