Incorporating Scenario Knowledge into A Unified Fine-tuning Architecture for Event Representation

被引:21
|
作者
Zheng, Jianming [1 ]
Cai, Fei [1 ]
Chen, Honghui [1 ]
机构
[1] Natl Univ Def Technol, Sci & Technol Informat Syst Engn Lab, Changsha, Peoples R China
来源
PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20) | 2020年
基金
中国国家自然科学基金;
关键词
event representation; pre-training; fine-tuning; scenario knowledge;
D O I
10.1145/3397271.3401173
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Given an occurred event, human can easily predict the next event or reason the preceding event, yet which is difficult for machine to perform such event reasoning. Event representation bridges the connection and targets to model the process of event reasoning as a machine-readable format, which then can support a wide range of applications in information retrieval, e.g., question answering and information extraction. Existing work mainly resorts to a joint training to integrate all levels of training loss in event chains by a simple loss summation, which is easily trapped into a local optimum. In addition, the scenario knowledge in event chains is not well investigated for event representation. In this paper, we propose a unified fine-tuning architecture, incorporated with scenario knowledge for event representation, i.e., UniFA-S, which mainly consists of a unified fine-timing architecture (UniFA) and a scenario-level variational auto-encoder (S-VAE). In detail, UniFA employs a multi-step fine-tuning to integrate all levels of training and S-VAE applies a stochastic variable to implicitly represent the scenario-level knowledge. We evaluate our proposal from two aspects, i.e., the representation and inference abilities. For the representation ability, our ensemble model UniFA-S can beat state-of-the-art base-lines for two similarity tasks. For the inference ability, UniFA-S can outperform the best baseline, achieving 4.1% 8.2% improvements in terms of accuracy for various inference tasks.
引用
收藏
页码:249 / 258
页数:10
相关论文
共 22 条
  • [1] Event Recognition on Images by Fine-Tuning of Deep Neural Networks
    Yudin, Dmitry
    Zeno, Bassel
    PROCEEDINGS OF THE SECOND INTERNATIONAL SCIENTIFIC CONFERENCE INTELLIGENT INFORMATION TECHNOLOGIES FOR INDUSTRY (IITI'17), VOL 1, 2018, 679 : 479 - 487
  • [2] Low fine-tuning with heavy higgsinos in Yukawa unified SUSY GUTs
    Un, Cem Salih
    TURKISH JOURNAL OF PHYSICS, 2024, 48 (01): : 1 - 27
  • [3] Boosting fine-tuning via Conditional Online Knowledge Transfer
    Liu, Zhiqiang
    Li, Yuhong
    Huang, Chengkai
    Luo, KunTing
    Liu, Yanxia
    NEURAL NETWORKS, 2024, 169 : 325 - 333
  • [4] RAFNet: Interdomain Representation Alignment and Fine-Tuning for Image Series Classification
    Gong, Maoguo
    Qiao, Wenyuan
    Li, Hao
    Qin, A. K.
    Gao, Tianqi
    Luo, Tianshi
    Xing, Lining
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [5] KNOWLEDGE DISTILLATION FROM BERT IN PRE-TRAINING AND FINE-TUNING FOR POLYPHONE DISAMBIGUATION
    Sun, Hao
    Tan, Xu
    Gan, Jun-Wei
    Zhao, Sheng
    Han, Dongxu
    Liu, Hongzhi
    Qin, Tao
    Liu, Tie-Yan
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 168 - 175
  • [6] Unveiling Key Aspects of Fine-Tuning in Sentence Embeddings: A Representation Rank Analysis
    Jung, Euna
    Kim, Jaeill
    Ko, Jungmin
    Park, Jinwoo
    Rhee, Wonjong
    IEEE ACCESS, 2024, 12 : 159877 - 159888
  • [7] AstroMAE: Redshift Prediction Using a Masked Autoencoder with a Novel Fine-Tuning Architecture
    Fathkouhi, Amirreza Dolatpour
    Fox, Geoffrey Charles
    2024 IEEE 20TH INTERNATIONAL CONFERENCE ON E-SCIENCE, E-SCIENCE 2024, 2024,
  • [8] Fine-Tuning Channel-Pruned Deep Model via Knowledge Distillation
    Zhang, Chong
    Wang, Hong-Zhi
    Liu, Hong-Wei
    Chen, Yi-Lin
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2024, 39 (06) : 1238 - 1247
  • [9] Chinese Medical Named Entity Recognition based on Expert Knowledge and Fine-tuning Bert
    Zhang, Bofeng
    Yao, Xiuhong
    Li, Haiyan
    Aini, Mirensha
    2023 IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH, ICKG, 2023, : 84 - 90
  • [10] Efficient Fine-Tuning Large Language Models for Knowledge-Aware Response Planning
    Minh Nguyen
    Kishan, K. C.
    Toan Nguyen
    Chadha, Ankit
    Thuy Vu
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 : 593 - 611