A transformer-based neural network framework for full names prediction with abbreviations and contexts

被引:0
|
作者
Ye, Ziming [1 ,2 ]
Li, Shuangyin [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Shandong Univ, Sch Comp Sci & Technol, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Full name inference; Multi-attention mask; Various grained contexts; Abbreviation; MODEL;
D O I
10.1016/j.datak.2023.102275
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid spread of information, abbreviations are used more and more common because they are convenient. However, the duplication of abbreviations can lead to confusion in many cases, such as information management and information retrieval. The resultant confusion annoys users. Thus, inferring a full name from an abbreviation has practical and significant advantages. The bulk of studies in the literature mainly inferred full names based on rule-based methods, statistical models, the similarity of representation, etc. However, these methods are unable to use various grained contexts properly. In this paper, we propose a flexible framework of Multi-attention mask Abbreviation Context and Full name language model, named MACF to address the problem. With the abbreviation and contexts as the inputs, the MACF can automatically predict a full name by generation, where the contexts can be variously grained. That is, different grained contexts ranging from coarse to fine can be selected to perform such complicated tasks in which contexts include paragraphs, several sentences, or even just a few keywords. A novel multi-attention mask mechanism is also proposed, which allows the model to learn the relationships among abbreviations, contexts, and full names, a process that makes the most of various grained contexts. The three corpora of different languages and fields were analyzed and measured with seven metrics in various aspects to evaluate the proposed framework. According to the experimental results, the MACF yielded more significant and consistent outputs than other baseline methods. Moreover, we discuss the significance and findings, and give the case studies to show the performance in real applications.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Learning continuous dynamic network representation with transformer-based temporal graph neural network
    Li, Yingji
    Wu, Yue
    Sun, Mingchen
    Yang, Bo
    Wang, Ying
    INFORMATION SCIENCES, 2023, 649
  • [32] TF-Predictor: Transformer-Based Prerouting Path Delay Prediction Framework
    Cao, Peng
    He, Guoqing
    Yang, Tai
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (07) : 2227 - 2237
  • [33] Remaining Useful Life Prediction of Lithium-Ion Batteries by Using a Denoising Transformer-Based Neural Network
    Han, Yunlong
    Li, Conghui
    Zheng, Linfeng
    Lei, Gang
    Li, Li
    ENERGIES, 2023, 16 (17)
  • [34] A Transformer-Based Ensemble Framework for the Prediction of Protein-Protein Interaction Sites
    Mou, Minjie
    Pan, Ziqi
    Zhou, Zhimeng
    Zheng, Lingyan
    Zhang, Hanyu
    Shi, Shuiyang
    Li, Fengcheng
    Sun, Xiuna
    Zhu, Feng
    RESEARCH, 2023, 6
  • [35] TXtreme: transformer-based extreme value prediction framework for time series forecasting
    Yadav, Hemant
    Thakkar, Amit
    DISCOVER APPLIED SCIENCES, 2025, 7 (02)
  • [36] Novel Transformer-based deep neural network for the prediction of post-refracturing production from oil wells
    Jia, Jing
    Li, Diquan
    Wang, Lichang
    Fan, Qinghu
    ADVANCES IN GEO-ENERGY RESEARCH, 2024, 13 (02): : 119 - 131
  • [37] VDGCNeT: A novel network-wide Virtual Dynamic Graph Convolution Neural network and Transformer-based traffic prediction model
    Zheng, Ge
    Chai, Wei Koong
    Zhang, Jiankang
    Katos, Vasilis
    KNOWLEDGE-BASED SYSTEMS, 2023, 275
  • [38] RPConvformer: A novel Transformer-based deep neural networks for traffic flow prediction
    Wen, Yanjie
    Xu, Ping
    Li, Zhihong
    Xu, Wangtu
    Wang, Xiaoyu
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 218
  • [39] A transformer-based network for speech recognition
    Tang L.
    International Journal of Speech Technology, 2023, 26 (02) : 531 - 539
  • [40] Carbon emissions forecasting based on temporal graph transformer-based attentional neural network
    Wu, Xingping
    Yuan, Qiheng
    Zhou, Chunlei
    Chen, Xiang
    Xuan, Donghai
    Song, Jinwei
    JOURNAL OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING, 2024, 24 (03) : 1405 - 1421