BERT Probe: A python']python package for probing attention based robustness evaluation of BERT models

被引:2
作者
Khan, Shahrukh [1 ]
Shahid, Mahnoor [1 ]
Singh, Navdeeppal [1 ]
机构
[1] Saarland Univ, Saarbrucken, Germany
关键词
Deep learning; BERT; Transformers; Adversarial machine learning;
D O I
10.1016/j.simpa.2022.100310
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Transformer models based on attention-based architectures have been significantly successful in establishing state-of-the-art results in natural language processing (NLP). However, recent work about adversarial robustness of attention-based models show that their robustness is susceptible to adversarial inputs causing spurious outputs thereby raising questions about trustworthiness of such models. In this paper, we present BERT Probe which is a python-based package for evaluating robustness to attention attribution based on character-level and word-level evasion attacks and empirically quantifying potential vulnerabilities for sequence classification tasks. Additionally, BERT Probe also provides two out-of-the-box defenses against character-level attention attribution-based evasion attacks.
引用
收藏
页数:3
相关论文
共 50 条
[31]   BERT Learns From Electroencephalograms About Parkinson's Disease: Transformer-Based Models for Aid Diagnosis [J].
Nogales, Alberto ;
Garcia-Tejedor, Alvaro J. ;
Maitin, Ana M. ;
Perez-Morales, Antonio ;
Dolores Del Castillo, Maria ;
Pablo Romero, Juan .
IEEE ACCESS, 2022, 10 :101672-101682
[32]   Research on News Text Classification Based on BERT-BiLSTM-TextCNN-Attention [J].
Wang, Jia ;
Li, Zongting ;
Ma, Chenyang .
PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, :295-298
[33]   HBert: A Long Text Processing Method Based on BERT and Hierarchical Attention Mechanisms [J].
Lv, Xueqiang ;
Liu, Zhaonan ;
Zhao, Ying ;
Xu, Ge ;
You, Xindong .
INTERNATIONAL JOURNAL ON SEMANTIC WEB AND INFORMATION SYSTEMS, 2023, 19 (01)
[34]   Assessing the use of attention weights to interpret BERT-based stance classification [J].
Cordova Saenz, Carlos Abel ;
Becker, Karin .
2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT 2021), 2021, :194-201
[35]   BVMHA: Text classification model with variable multihead hybrid attention based on BERT [J].
Peng, Bo ;
Zhang, Tao ;
Han, Kundong ;
Zhang, Zhe ;
Ma, Yuquan ;
Ma, Mengnan .
JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2024, 46 (01) :1443-1454
[36]   Creation of time series models based on deep learning for generation of probabilistic forecasts using GluonTS and Python']Python [J].
Cancino-Villatoro, Karina ;
Solis, Alfredo Castillo ;
Castillo-Estrada, Christian ;
Juarez-Ramirez, Reyes .
2023 11TH INTERNATIONAL CONFERENCE IN SOFTWARE ENGINEERING RESEARCH AND INNOVATION, CONISOFT 2023, 2023, :265-274
[37]   Learning-based short text compression using BERT models [J].
Ozturk, Emir ;
Mesut, Altan .
PEERJ COMPUTER SCIENCE, 2024, 10
[38]   Research frontiers of pre-training mathematical models based on BERT [J].
Li, Guang ;
Wang, Wennan ;
Zhu, Liukai ;
Peng, Jun ;
Li, Xujia ;
Luo, Ruijie .
2022 IEEE INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, BIG DATA AND ALGORITHMS (EEBDA), 2022, :154-158
[39]   INNOVATIVE BERT-BASED RERANKING LANGUAGE MODELS FOR SPEECH RECOGNITION [J].
Chiu, Shih-Hsuan ;
Chen, Berlin .
2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, :266-271
[40]   Learning to rank with BERT-based confidence models in ASR rescoring [J].
Wu, Ting-Wei ;
Chen, I-Fan ;
Gandhe, Ankur .
INTERSPEECH 2022, 2022, :1651-1655