Robust Representation Learning via Sparse Attention Mechanism for Similarity Models

被引:0
|
作者
Ermilova, Alina [1 ]
Baramiia, Nikita [1 ]
Kornilov, Valerii [1 ]
Petrakov, Sergey [1 ]
Zaytsev, Alexey [1 ,2 ]
机构
[1] Skolkovo Inst Sci & Technol, Moscow 121205, Russia
[2] Sber, Risk Management, Moscow 121165, Russia
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Transformers; Oil insulation; Task analysis; Time series analysis; Meteorology; Training; Deep learning; Representation learning; efficient transformer; robust transformer; representation learning; similarity learning; TRANSFORMER;
D O I
10.1109/ACCESS.2024.3418779
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The attention-based models are widely used for time series data. However, due to the quadratic complexity of attention regarding input sequence length, the application of Transformers is limited by high resource demands. Moreover, their modifications for industrial time series need to be robust to missing or noisy values, which complicates the expansion of their application horizon. To cope with these issues, we introduce the class of efficient Transformers named Regularized Transformers (Reguformers). We implement the regularization technique inspired by the dropout ideas to improve robustness and reduce computational expenses without significantly modifying the pipeline. The focus in our experiments is on oil&gas data. For well-interval similarity task, our best Reguformer configuration reaches ROC AUC 0.97, which is comparable to Informer (0.978) and outperforms baselines: the previous LSTM model (0.934), the classical Transformer model (0.967), and three recent most promising modifications of the original Transformer, namely, Performer (0.949), LRformer (0.955), and DropDim (0.777). We also conduct the corresponding experiments on three additional datasets from different domains and obtain superior results. The increase in the quality of the best Reguformer relative to Transformer for different datasets varies from 3.7% to 9.6%, while the increase range relative to Informer is wider: from 1.7% to 18.4%.
引用
收藏
页码:97833 / 97850
页数:18
相关论文
共 50 条
  • [31] Network representation learning: models, methods and applications
    Mohan, Anuraj
    Pramod, K., V
    SN APPLIED SCIENCES, 2019, 1 (09):
  • [32] Towards Understanding the Mechanism of Contrastive Learning via Similarity Structure: A Theoretical Analysis
    Waida, Hiroki
    Wada, Yuichiro
    Andeol, Leo
    Nakagawa, Takumi
    Zhang, Yuhui
    Kanamori, Takafumi
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT IV, 2023, 14172 : 709 - 727
  • [33] Network representation learning: models, methods and applications
    Anuraj Mohan
    K. V. Pramod
    SN Applied Sciences, 2019, 1
  • [34] Representation Learning via Variational Bayesian Networks
    Barkan, Oren
    Caciularu, Avi
    Rejwan, Idan
    Katz, Ori
    Weill, Jonathan
    Malkiel, Itzik
    Koenigstein, Noam
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 78 - 88
  • [35] Learning an SAR Image Despeckling Model Via Weighted Sparse Representation
    Zhang, Junchao
    Chen, Jianlai
    Yu, Hanwen
    Yang, Degui
    Xu, Xiaoqing
    Xing, Mengdao
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 : 7148 - 7158
  • [36] Boosting Representation Learning via Similarity-based Active Data Sampling from the Web
    Ueno, Shiryu
    Kato, Kunihito
    2024 INTERNATIONAL WORKSHOP ON INTELLIGENT SYSTEMS, IWIS 2024, 2024,
  • [37] Chest x-ray diagnosis via spatial-channel high-order attention representation learning
    Gao, Xinyue
    Jiang, Bo
    Wang, Xixi
    Huang, Lili
    Tu, Zhengzheng
    PHYSICS IN MEDICINE AND BIOLOGY, 2024, 69 (04)
  • [38] SURREAL: Subgraph Robust Representation Learning
    Saba A. Al-Sayouri
    Danai Koutra
    Evangelos E. Papalexakis
    Sarah S. Lam
    Applied Network Science, 4
  • [39] SURREAL: Subgraph Robust Representation Learning
    Al-Sayouri, Saba A.
    Koutra, Danai
    Papalexakis, Evangelos E.
    Lam, Sarah S.
    APPLIED NETWORK SCIENCE, 2019, 4 (01)
  • [40] Multi-Noise Representation Learning for Robust Speaker Recognition
    Cho, Sunyoung
    Wee, Kyungchul
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 681 - 685