A transformer-based neural network framework for full names prediction with abbreviations and contexts

被引:0
|
作者
Ye, Ziming [1 ,2 ]
Li, Shuangyin [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Shandong Univ, Sch Comp Sci & Technol, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Full name inference; Multi-attention mask; Various grained contexts; Abbreviation; MODEL;
D O I
10.1016/j.datak.2023.102275
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid spread of information, abbreviations are used more and more common because they are convenient. However, the duplication of abbreviations can lead to confusion in many cases, such as information management and information retrieval. The resultant confusion annoys users. Thus, inferring a full name from an abbreviation has practical and significant advantages. The bulk of studies in the literature mainly inferred full names based on rule-based methods, statistical models, the similarity of representation, etc. However, these methods are unable to use various grained contexts properly. In this paper, we propose a flexible framework of Multi-attention mask Abbreviation Context and Full name language model, named MACF to address the problem. With the abbreviation and contexts as the inputs, the MACF can automatically predict a full name by generation, where the contexts can be variously grained. That is, different grained contexts ranging from coarse to fine can be selected to perform such complicated tasks in which contexts include paragraphs, several sentences, or even just a few keywords. A novel multi-attention mask mechanism is also proposed, which allows the model to learn the relationships among abbreviations, contexts, and full names, a process that makes the most of various grained contexts. The three corpora of different languages and fields were analyzed and measured with seven metrics in various aspects to evaluate the proposed framework. According to the experimental results, the MACF yielded more significant and consistent outputs than other baseline methods. Moreover, we discuss the significance and findings, and give the case studies to show the performance in real applications.
引用
收藏
页数:20
相关论文
共 50 条
  • [41] Empirical assessment of transformer-based neural network architecture in forecasting pollution trends
    Nath, Pritthijit
    Middya, Asif Iqbal
    Roy, Sarbani
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2023,
  • [42] Transformer-Based Wireless Traffic Prediction and Network Optimization in O-RAN
    Habib, Md Arafat
    Rivera, Pedro Enrique Iturria
    Ozcan, Yigit
    Elsayed, Medhat
    Bavand, Majid
    Gaigalas, Raimundus
    Erol-Kantarci, Melike
    2024 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS 2024, 2024, : 1 - 6
  • [43] TR-Net: A Transformer-Based Neural Network for Point Cloud Processing
    Liu, Luyao
    Chen, Enqing
    Ding, Yingqiang
    MACHINES, 2022, 10 (07)
  • [44] Progressive Guidance Categorization Using Transformer-Based Deep Neural Network Architecture
    Aurpa, Tanjim Taharat
    Ahmed, Md Shoaib
    Sadik, Rifat
    Anwar, Sabbir
    Adnan, Md Abdul Mazid
    Anwar, Md Musfique
    HYBRID INTELLIGENT SYSTEMS, HIS 2021, 2022, 420 : 344 - 353
  • [45] A Transformer-Based Framework for Tiny Object Detection
    Liao, Yi-Kai
    Lin, Gong-Si
    Yeh, Mei-Chen
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 373 - 377
  • [46] Generalizability of Convolutional Neural Network and Vision Transformer-Based OCT Segmentation Models
    Pely, Adam
    Wu, Zhichao
    Leng, Theodore
    Gao, Simon S.
    Chen, Hao
    Hejrati, Mohsen
    Zhang, Miao
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2023, 64 (08)
  • [47] A Hybrid Transformer-based Spatial-Temporal Network for Traffic Flow Prediction
    Tian, Guanqun
    Li, Dequan
    2024 IEEE 19TH CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS, ICIEA 2024, 2024,
  • [48] Korean Sign Language Recognition Using Transformer-Based Deep Neural Network
    Shin, Jungpil
    Musa Miah, Abu Saleh
    Hasan, Md. Al Mehedi
    Hirooka, Koki
    Suzuki, Kota
    Lee, Hyoun-Sup
    Jang, Si-Woong
    APPLIED SCIENCES-BASEL, 2023, 13 (05):
  • [49] A transformer-based framework for enterprise sales forecasting
    Sun, Yupeng
    Li, Tian
    PEERJ COMPUTER SCIENCE, 2024, 10 : 1 - 14
  • [50] Time-Aware Transformer-based Network for Clinical Notes Series Prediction
    Zhang, Dongyu
    Thadajarassiri, Jidapa
    Sen, Cansu
    Rundensteiner, Elke
    MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 126, 2020, 126 : 566 - 587