CLMSM: A Multi-Task Learning Framework for Pre-training on Procedural Text

被引:0
|
作者
Nandy, Abhilash [1 ]
Kapadnis, Manav Nitin [1 ]
Goyal, Pawan [1 ]
Ganguly, Niloy [1 ]
机构
[1] Indian Inst Technol Kharagpur, Kharagpur, W Bengal, India
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023) | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose CLMSM, a domainspecific, continual pre-training framework, that learns from a large set of procedural recipes. CLMSM uses a Multi-Task Learning Framework to optimize two objectives - a) Contrastive Learning using hard triplets to learn fine-grained differences across entities in the procedures, and b) a novel Mask-Step Modelling objective to learn step-wise context of a procedure. We test the performance of CLMSM on the downstream tasks of tracking entities and aligning actions between two procedures on three datasets, one of which is an open-domain dataset not conforming with the pre-training dataset. We show that CLMSM not only outperforms baselines on recipes (indomain) but is also able to generalize to opendomain procedural NLP tasks.
引用
收藏
页码:8793 / 8806
页数:14
相关论文
共 50 条
  • [1] Improving AMR-to-text Generation with Multi-task Pre-training
    Xu D.-Q.
    Li J.-H.
    Zhu M.-H.
    Zhou G.-D.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (10): : 3036 - 3050
  • [2] Multi-Task Collaborative Pre-Training and Adaptive Token Selection: A Unified Framework for Brain Representation Learning
    Jiang, Ning
    Wang, Gongshu
    Ye, Chuyang
    Liu, Tiantian
    Yan, Tianyi
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (09) : 5528 - 5539
  • [3] Pre-training Multi-task Contrastive Learning Models for Scientific Literature Understanding
    Zhang, Yu
    Cheng, Hao
    Shen, Zhihong
    Liu, Xiaodong
    Wang, Ye-Yi
    Gao, Jianfeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12259 - 12275
  • [4] Learning general multi-agent decision model through multi-task pre-training
    Wang, Jiawei
    Xu, Lele
    Sun, Changyin
    NEUROCOMPUTING, 2025, 627
  • [5] Multi-task Pre-training with Soft Biometrics for Transfer-learning Palmprint Recognition
    Xu, Huanhuan
    Leng, Lu
    Yang, Ziyuan
    Teoh, Andrew Beng Jin
    Jin, Zhe
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2341 - 2358
  • [6] Multi-task Pre-training with Soft Biometrics for Transfer-learning Palmprint Recognition
    Huanhuan Xu
    Lu Leng
    Ziyuan Yang
    Andrew Beng Jin Teoh
    Zhe Jin
    Neural Processing Letters, 2023, 55 : 2341 - 2358
  • [7] A Multi-Task Semantic Decomposition Framework with Task-specific Pre-training for Few-Shot NER
    Dong, Guanting
    Wang, Zechen
    Zhao, Jinxu
    Zhao, Gang
    Guo, Daichi
    Fu, Dayuan
    Hui, Tingfeng
    Zeng, Chen
    He, Keqing
    Li, Xuefeng
    Wang, Liwen
    Cui, Xinyue
    Xu, Weiran
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 430 - 440
  • [8] A Multi-Task Learning Framework for Abstractive Text Summarization
    Lu, Yao
    Liu, Linqing
    Jiang, Zhile
    Yang, Min
    Goebel, Randy
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9987 - 9988
  • [9] Multi-task Pre-training Language Model for Semantic Network Completion
    Li, Da
    Zhu, Boqing
    Yang, Sen
    Xu, Kele
    Yi, Ming
    He, Yukai
    Wang, Huaimin
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (11)
  • [10] Multi-Task Pre-Training of Deep Neural Networks for Digital Pathology
    Mormont, Romain
    Geurts, Pierre
    Maree, Raphael
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (02) : 412 - 421