LiteMuL: A Lightweight On-Device Sequence Tagger using Multi-task Learning

被引:6
|
作者
Kumari, Sonal [1 ]
Agarwal, Vibhav [1 ]
Challa, Bharath [1 ]
Chalamalasetti, Kranti [1 ]
Ghosh, Sourav [1 ]
Harshavardhana, Harshavardhana [1 ]
Raja, Barath Raj Kandur [1 ]
机构
[1] Samsung R&D Inst Bangalore, Bangalore 560037, Karnataka, India
来源
2021 IEEE 15TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC 2021) | 2021年
关键词
Sequence labeling; mobile device; multi-task learning; informal conversation;
D O I
10.1109/ICSC50631.2021.00007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Named entity detection and Parts-of-speech tagging are the key tasks for many NLP applications. Although the current state of the art methods achieved near perfection for long, formal, structured text there are hindrances in deploying these models on memory-constrained devices such as mobile phones. Furthermore, the performance of these models is degraded when they encounter short, informal, and casual conversations. To overcome these difficulties, we present LiteMuL a lightweight on-device sequence tagger that can efficiently process the user conversations using a Multi-Task Learning (MTL) approach. To the best of our knowledge, the proposed model is the first on-device MTL neural model for sequence tagging. Our LiteMuL model is about 2.39 MB in size and achieved an accuracy of 0.9433 (for NER), 0.9090 (for POS) on the CoNLL 2003 dataset. The proposed LiteMuL not only outperforms the current state of the art results but also surpasses the results of our proposed on-device task-specific models, with accuracy gains of up to 11% and model-size reduction by 50%-56%. Our model is competitive with other MTL approaches for NER and POS tasks while outshines them with a low memory footprint. We also evaluated our model on custom-curated user conversations and observed impressive results.
引用
收藏
页码:1 / 8
页数:8
相关论文
共 50 条
  • [1] On-Device Deep Multi-Task Inference via Multi-Task Zipping
    He, Xiaoxi
    Wang, Xu
    Zhou, Zimu
    Wu, Jiahang
    Yang, Zheng
    Thiele, Lothar
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (05) : 2878 - 2891
  • [2] Multi-Task Adapters for On-Device Audio Inference
    Tagliasacchi, Marco
    Quitry, Felix de Chaumont
    Roblek, Dominik
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 630 - 634
  • [3] Hierarchical Tagger with Multi-task Learning for Cross-domain Slot Filling
    Wei, Xiao
    Si, Yuke
    Wang, Shiquan
    Wang, Longbiao
    Dang, Jianwu
    INTERSPEECH 2022, 2022, : 3273 - 3277
  • [4] Meta Multi-Task Learning for Sequence Modeling
    Chen, Junkun
    Qiu, Xipeng
    Liu, Pengfei
    Huang, Xuanjing
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 5070 - 5077
  • [5] Multi-Task Temporal Shift Attention Networks for On-Device Contactless Vitals Measurement
    Liu, Xin
    Fromm, Josh
    Patel, Shwetak
    McDuff, Daniel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Multiple Device Segmentation for Fluoroscopic Imaging Using Multi-task Learning
    Breininger, Katharina
    Wuerfl, Tobias
    Kurzendorfer, Tanja
    Albarqouni, Shadi
    Pfister, Marcus
    Kowarschik, Markus
    Navab, Nassir
    Maier, Andreas
    INTRAVASCULAR IMAGING AND COMPUTER ASSISTED STENTING AND LARGE-SCALE ANNOTATION OF BIOMEDICAL DATA AND EXPERT LABEL SYNTHESIS, 2018, 11043 : 19 - 27
  • [7] Learning Multi-Task Communication with Message Passing for Sequence Learning
    Liu, Pengfei
    Fu, Jie
    Dong, Yue
    Qiu, Xipeng
    Cheung, Jackie Chi Kit
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 4360 - 4367
  • [8] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [9] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [10] Lightweight Multi-Task Learning Method for System Log Anomaly Detection
    Pham, Tuan-Anh
    Lee, Jong-Hoon
    IEEE ACCESS, 2024, 12 : 147739 - 147752