Universal Language Model Fine-tuning for Text Classification

被引:0
|
作者
Howard, Jeremy [1 ]
Ruder, Sebastian [2 ,3 ]
机构
[1] Univ San Francisco, Fast Ai, San Francisco, CA 94117 USA
[2] NUI Galway, Insight Ctr, Galway, Ireland
[3] Aylien Ltd, Dublin, Ireland
来源
PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1 | 2018年
基金
爱尔兰科学基金会;
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Inductive transfer learning has greatly impacted computer vision, but existing approaches in NLP still require task-specific modifications and training from scratch. We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP, and introduce techniques that are key for fine-tuning a language model. Our method significantly outperforms the state-of-the-art on six text classification tasks, reducing the error by 18-24% on the majority of datasets. Furthermore, with only 100 labeled examples, it matches the performance of training from scratch on 100x more data. We opensource our pretrained models and code(1).
引用
收藏
页码:328 / 339
页数:12
相关论文
共 50 条
  • [21] On Transferability of Bias Mitigation Effects in Language Model Fine-Tuning
    Jin, Xisen
    Barbieri, Francesco
    Kennedy, Brendan
    Davani, Aida Mostafazadeh
    Neves, Leonardo
    Ren, Xiang
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3770 - 3783
  • [22] Fine-tuning natural language imperatives
    Kaufmann, Magdalena
    JOURNAL OF LOGIC AND COMPUTATION, 2019, 29 (03) : 321 - 348
  • [23] On Surgical Fine-tuning for Language Encoders
    Lodha, Abhilasha
    Belapurkar, Gayatri
    Chalkapurkar, Saloni
    Tao, Yuanming
    Ghosh, Reshmi
    Basu, Samyadeep
    Petrov, Dmitrii
    Srinivasan, Soundararajan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3105 - 3113
  • [24] Fine-Tuning DARTS for Image Classification
    Tanveer, Muhammad Suhaib
    Khan, Muhammad Umar Karim
    Kyung, Chong-Min
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 4789 - 4796
  • [25] Multi-phase Fine-Tuning: A New Fine-Tuning Approach for Sign Language Recognition
    Sarhan, Noha
    Lauri, Mikko
    Frintrop, Simone
    KUNSTLICHE INTELLIGENZ, 2022, 36 (01): : 91 - 98
  • [26] A Novel Machine Learning Approach for Sentiment Analysis on Twitter Incorporating the Universal Language Model Fine-Tuning and SVM
    AlBadani, Barakat
    Shi, Ronghua
    Dong, Jian
    APPLIED SYSTEM INNOVATION, 2022, 5 (01)
  • [27] Multi-phase Fine-Tuning: A New Fine-Tuning Approach for Sign Language Recognition
    Noha Sarhan
    Mikko Lauri
    Simone Frintrop
    KI - Künstliche Intelligenz, 2022, 36 : 91 - 98
  • [28] Meta Fine-Tuning Neural Language Models for Multi-Domain Text Mining
    Wang, Chengyu
    Qiu, Minghui
    Huang, Jun
    He, Xiaofeng
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3094 - 3104
  • [29] Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
    Zhang, Zhen-Ru
    Tan, Chuanqi
    Xu, Haiyang
    Wang, Chengyu
    Huang, Jun
    Huang, Songfang
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1239 - 1248
  • [30] Personalizing Text-to-Image Diffusion Models by Fine-Tuning Classification for AI Applications
    Hidalgo, Rafael
    Salah, Nesreen
    Jetty, Rajiv Chandra
    Jetty, Anupama
    Varde, Aparna S.
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, INTELLISYS 2023, 2024, 822 : 642 - 658