Combining Self-supervised Learning and Active Learning for Disfluency Detection

被引:4
|
作者
Wang, Shaolei [1 ]
Wang, Zhongyuan [1 ]
Che, Wanxiang [1 ]
Zhao, Sendong [1 ]
Liu, Ting [1 ]
机构
[1] Harbin Inst Technol, 2 YiKuang St,Tech & Innovat Bldg,HIT Sci Pk, Harbin 150001, Heilongjiang, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Disfluency detection; self-supervised learning; active learning; pre-training technology;
D O I
10.1145/3487290
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spoken language is fundamentally different from the written language in that it contains frequent disfluencies or parts of an utterance that are corrected by the speaker. Disfluency detection (removing these disfluencies) is desirable to clean the input for use in downstream NLP tasks. Most existing approaches to disfluency detection heavily rely on human-annotated data, which is scarce and expensive to obtain in practice. To tackle the training data bottleneck, in this work, we investigate methods for combining self-supervised learning and active learning for disfluency detection. First, we construct large-scale pseudo training data by randomly adding or deleting words fromunlabeled data and propose two self-supervised pre-training tasks: (i) a tagging task to detect the added noisy words and (ii) sentence classification to distinguish original sentences from grammatically incorrect sentences. We then combine these two tasks to jointly pre-train a neural network. The pre-trained neural network is then fine-tuned using human-annotated disfluency detection training data. The self-supervised learning method can capture task-special knowledge for disfluency detection and achieve better performance when fine-tuning on a small annotated dataset compared to other supervised methods. However, limited in that the pseudo training data are generated based on simple heuristics and cannot fully cover all the disfluency patterns, there is still a performance gap compared to the supervised models trained on the full training dataset. We further explore how to bridge the performance gap by integrating active learning during the fine-tuning process. Active learning strives to reduce annotation costs by choosing the most critical examples to label and can address the weakness of self-supervised learning with a small annotated dataset. We show that by combining self-supervised learning with active learning, our model is able to match state-of-the-art performance with just about 10% of the original training data on both the commonly used English Switchboard test set and a set of in-house annotated Chinese data.
引用
收藏
页数:25
相关论文
共 50 条
  • [21] Contrastive self-supervised learning for diabetic retinopathy early detection
    Jihong Ouyang
    Dong Mao
    Zeqi Guo
    Siguang Liu
    Dong Xu
    Wenting Wang
    Medical & Biological Engineering & Computing, 2023, 61 : 2441 - 2452
  • [22] Repeatable adaptive keypoint detection via self-supervised learning
    Yan, Pei
    Tan, Yihua
    Tai, Yuan
    SCIENCE CHINA-INFORMATION SCIENCES, 2022, 65 (11)
  • [23] Pavement anomaly detection based on transformer and self-supervised learning
    Lin, Zijie
    Wang, Hui
    Li, Shenglin
    AUTOMATION IN CONSTRUCTION, 2022, 143
  • [24] Self-Supervised Learning for Anomaly Detection With Dynamic Local Augmentation
    Yoa, Seungdong
    Lee, Seungjun
    Kim, Chiyoon
    Kim, Hyunwoo J.
    IEEE ACCESS, 2021, 9 : 147201 - 147211
  • [25] Rumor detection with self-supervised learning on texts and social graph
    Yuan GAO
    Xiang WANG
    Xiangnan HE
    Huamin FENG
    Yongdong ZHANG
    Frontiers of Computer Science, 2023, 17 (04) : 153 - 167
  • [26] Domain adaptation and self-supervised learning for surgical margin detection
    Alice M. L. Santilli
    Amoon Jamzad
    Alireza Sedghi
    Martin Kaufmann
    Kathryn Logan
    Julie Wallis
    Kevin Y. M Ren
    Natasja Janssen
    Shaila Merchant
    Jay Engel
    Doug McKay
    Sonal Varma
    Ami Wang
    Gabor Fichtinger
    John F. Rudan
    Parvin Mousavi
    International Journal of Computer Assisted Radiology and Surgery, 2021, 16 : 861 - 869
  • [27] Understanding the limitations of self-supervised learning for tabular anomaly detection
    Mai, Kimberly T.
    Davies, Toby
    Griffin, Lewis D.
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (02)
  • [28] Rumor detection with self-supervised learning on texts and social graph
    Gao, Yuan
    Wang, Xiang
    He, Xiangnan
    Feng, Huamin
    Zhang, Yongdong
    FRONTIERS OF COMPUTER SCIENCE, 2023, 17 (04)
  • [29] Hierarchical Detection of Network Anomalies : A Self-Supervised Learning Approach
    Kye, Hyoseon
    Kim, Miru
    Kwon, Minhae
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1908 - 1912
  • [30] Deep anomaly detection with self-supervised learning and adversarial training
    Zhang, Xianchao
    Mu, Jie
    Zhang, Xiaotong
    Liu, Han
    Zong, Linlin
    Li, Yuangang
    PATTERN RECOGNITION, 2022, 121