BERT Probe: A python']python package for probing attention based robustness evaluation of BERT models

被引:2
作者
Khan, Shahrukh [1 ]
Shahid, Mahnoor [1 ]
Singh, Navdeeppal [1 ]
机构
[1] Saarland Univ, Saarbrucken, Germany
关键词
Deep learning; BERT; Transformers; Adversarial machine learning;
D O I
10.1016/j.simpa.2022.100310
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Transformer models based on attention-based architectures have been significantly successful in establishing state-of-the-art results in natural language processing (NLP). However, recent work about adversarial robustness of attention-based models show that their robustness is susceptible to adversarial inputs causing spurious outputs thereby raising questions about trustworthiness of such models. In this paper, we present BERT Probe which is a python-based package for evaluating robustness to attention attribution based on character-level and word-level evasion attacks and empirically quantifying potential vulnerabilities for sequence classification tasks. Additionally, BERT Probe also provides two out-of-the-box defenses against character-level attention attribution-based evasion attacks.
引用
收藏
页数:3
相关论文
共 50 条
[41]   ENHANCED BERT-BASED RANKING MODELS FOR SPOKEN DOCUMENT RETRIEVAL [J].
Lin, Hsiao-Yun ;
Lo, Tien-Hong ;
Chen, Berlin .
2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, :601-606
[42]   PyRaDiSe: A Python']Python package for DICOM-RT-based auto-segmentation pipeline construction and DICOM-RT data conversion [J].
Rufenacht, Elias ;
Kamath, Amith ;
Suter, Yannick ;
Poel, Robert ;
Ermis, Ekin ;
Scheib, Stefan ;
Reyes, Mauricio .
COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2023, 231
[43]   Sens-BERT: A BERT-Based Approach for Enabling Transferability and Re-Calibration of Calibration Models for Low-Cost Sensors Under Reference Measurements Scarcity [J].
Narayana, M. V. ;
Rachavarapu, Kranthi Kumar ;
Jalihal, Devendra ;
Nagendra, S. M. Shiva .
IEEE SENSORS JOURNAL, 2024, 24 (07) :11362-11373
[44]   Enhancing predictive modeling for Indian banking stock trends: A fusion of BERT and attention-based BiLSTM approach [J].
Buche, Arti ;
Chandak, M. B. .
JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (05) :8761-8773
[45]   Similarity-Based Resume Matching via Triplet Loss with BERT Models [J].
Ozlu, O. Anil ;
Orman, Gunce Keziban ;
Danis, F. Serhan ;
Turhan, Sultan N. ;
Kara, K. Can ;
Yucel, T. Arda .
INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 3, 2023, 544 :520-532
[46]   ICRM: An intelligent citation recommendation mechanism based on BERT and weighted BoW models [J].
Chang C.-Y. ;
Yang Y.-T. ;
Zhang Q. ;
Lin Y.-T. ;
Roy D.S. .
Journal of Intelligent and Fuzzy Systems, 2024, 46 (04) :10135-10150
[47]   Exploring BERT-Based Pretrained Models for Polarity Analysis of Tweets in Spanish [J].
Gonzalez, Erick Barrios ;
Vidal, Mireya Tovar ;
Reyes-Ortiz, Jose A. ;
Flores, Fernando Zacarias ;
Lopez, Pedro Bello .
INTERNATIONAL JOURNAL OF COMBINATORIAL OPTIMIZATION PROBLEMS AND INFORMATICS, 2023, 14 (01) :27-38
[48]   Enhancing Arabic Word Sense Disambiguation with Ensemble BERT-Based Models [J].
Djaidri, Asma ;
Aliane, Hassina ;
Azzoune, Hamid .
ARABIC LANGUAGE PROCESSING: FROM THEORY TO PRACTICE, ICALP 2023, PT I, 2025, 2339 :261-272
[49]   Fine-Tuning and Efficacy Assessment of BERT-Based Models in Detecting Early Signs of Depression [J].
Kumar, Vaibhav ;
Singh, Vibhav Prakash .
ADVANCED NETWORK TECHNOLOGIES AND INTELLIGENT COMPUTING, ANTIC 2024, PT III, 2025, 2335 :371-383
[50]   Advancements in Text Subjectivity Analysis: From Simple Approaches to BERT-Based Models and Generalization Assessments [J].
Antal, Margit ;
Buza, Krisztian ;
Nemes, Szilard .
ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2024, PART I, 2024, 2165 :245-255