Parameter-efficient feature-based transfer for paraphrase identification

被引:1
|
作者
Liu, Xiaodong [1 ]
Rzepka, Rafal [2 ]
Araki, Kenji [2 ]
机构
[1] Hokkaido Univ, Grad Sch Informat Sci & Technol, Sapporo, Hokkaido, Japan
[2] Hokkaido Univ, Fac Informat Sci & Technol, Sapporo, Hokkaido, Japan
关键词
Parameter-efficient feature-based transfer; Paraphrase identification; Natural language inference; Semantic textual similarity; Continual learning;
D O I
10.1017/S135132492200050X
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are many types of approaches for Paraphrase Identification (PI), an NLP task of determining whether a sentence pair has equivalent semantics. Traditional approaches mainly consist of unsupervised learning and feature engineering, which are computationally inexpensive. However, their task performance is moderate nowadays. To seek a method that can preserve the low computational costs of traditional approaches but yield better task performance, we take an investigation into neural network-based transfer learning approaches. We discover that by improving the usage of parameters efficiently for feature-based transfer, our research goal can be accomplished. Regarding the improvement, we propose a pre-trained task-specific architecture. The fixed parameters of the pre-trained architecture can be shared by multiple classifiers with small additional parameters. As a result, the computational cost left involving parameter update is only generated from classifier-tuning: the features output from the architecture combined with lexical overlap features are fed into a single classifier for tuning. Furthermore, the pre-trained task-specific architecture can be applied to natural language inference and semantic textual similarity tasks as well. Such technical novelty leads to slight consumption of computational and memory resources for each task and is also conducive to power-efficient continual learning. The experimental results show that our proposed method is competitive with adapter-BERT (a parameter-efficient fine-tuning approach) over some tasks while consuming only 16% trainable parameters and saving 69-96% time for parameter update.
引用
收藏
页码:1066 / 1096
页数:31
相关论文
共 50 条
  • [31] Simultaneous feature-based identification and track fusion
    Blasch, E
    Hong, L
    PROCEEDINGS OF THE 37TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-4, 1998, : 239 - 244
  • [32] Feature-based interaction: an identification and classification methodology
    Hounsell, MD
    Case, K
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART B-JOURNAL OF ENGINEERING MANUFACTURE, 1999, 213 (04) : 369 - 380
  • [33] Communication-Efficient and Privacy-Preserving Feature-based Federated Transfer Learning
    Wang, Feng
    Gursoy, M. Cenk
    Velipasalar, Senem
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 3875 - 3880
  • [34] CAN FEATURE-BASED INDUCTIVE TRANSFER LEARNING HELP PERSON RE-IDENTIFICATION?
    Wu, Yang
    Li, Wei
    Minoh, Michihiko
    Mukunoki, Masayuki
    2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 2812 - 2816
  • [35] Parameter-Efficient Cross-lingual Transfer of Vision and Language Models via Translation-based Alignment
    Zhang, Zhen
    Wang, Jialu
    Wang, Xin Eric
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 7258 - 7268
  • [36] VL-ADAPTER: Parameter-Efficient Transfer Learning for Vision-and-Language Tasks
    Sung, Yi-Lin
    Cho, Jaemin
    Bansal, Mohit
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 5217 - 5227
  • [37] Composing Parameter-Efficient Modules with Arithmetic Operations
    Zhang, Jinghan
    Chen, Shiqi
    Liu, Junteng
    He, Junxian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [38] The Power of Scale for Parameter-Efficient Prompt Tuning
    Lester, Brian
    Al-Rfou, Rami
    Constant, Noah
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3045 - 3059
  • [39] PARAMETER-EFFICIENT VISION TRANSFORMER WITH LINEAR ATTENTION
    Zhao, Youpeng
    Tang, Huadong
    Jiang, Yingying
    Yong, A.
    Wu, Qiang
    Wang, Jun
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1275 - 1279
  • [40] Parameter-Efficient Tuning with Special Token Adaptation
    Yang, Xiaocong
    Huang, James Y.
    Zhou, Wenxuan
    Chen, Muhao
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 865 - 872