A Practical Black-Box Attack on Source Code Authorship Identification Classifiers

被引:11
|
作者
Liu, Qianjun [1 ]
Ji, Shouling [1 ]
Liu, Changchang [2 ]
Wu, Chunming [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] IBM Thomas J Watson Res Ctr, Dept Distributed AI, Yorktown Hts, NY 10598 USA
基金
中国国家自然科学基金;
关键词
Feature extraction; Tools; Training; Syntactics; Predictive models; Perturbation methods; Transforms; Source code; authorship identification; adversarial stylometry; ROBUSTNESS;
D O I
10.1109/TIFS.2021.3080507
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Existing researches have recently shown that adversarial stylometry of source code can confuse source code authorship identification (SCAI) models, which may threaten the security of related applications such as programmer attribution, software forensics, etc. In this work, we propose source code authorship disguise (SCAD) to automatically hide programmers' identities from authorship identification, which is more practical than the previous work that requires to known the output probabilities or internal details of the target SCAI model. Specifically, SCAD trains a substitute model and develops a set of semantically equivalent transformations, based on which the original code is modified towards a disguised style with small manipulations in lexical features and syntactic features. When evaluated under totally black-box settings, on a real-world dataset consisting of 1,600 programmers, SCAD induces state-of-the-art SCAI models to cause above 30% misclassification rates. The efficiency and utility-preserving properties of SCAD are also demonstrated with multiple metrics. Furthermore, our work can serve as a guideline for developing more robust identification methods in the future.
引用
收藏
页码:3620 / 3633
页数:14
相关论文
共 34 条
  • [31] Black-Box Universal Adversarial Attack for DNN-Based Models of SAR Automatic Target Recognition
    Wan, Xuanshen
    Liu, Wei
    Niu, Chaoyang
    Lu, Wanjie
    Du, Meng
    Li, Yuanli
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 8673 - 8696
  • [32] Complex Cases of Source Code Authorship Identification Using a Hybrid Deep Neural Network
    Kurtukova, Anna
    Romanov, Aleksandr
    Shelupanov, Alexander
    Fedotova, Anastasia
    FUTURE INTERNET, 2022, 14 (10):
  • [33] Adversarial infrared blocks: A multi-view black-box attack to thermal infrared detectors in physical world
    Hu, Chengyin
    Shi, Weiwen
    Jiang, Tingsong
    Yao, Wen
    Tian, Ling
    Chen, Xiaoqian
    Zhou, Jingzhi
    Li, Wen
    NEURAL NETWORKS, 2024, 175
  • [34] A Black-Box Adversarial Attack Method via Nesterov Accelerated Gradient and Rewiring Towards Attacking Graph Neural Networks
    Zhao, Shu
    Wang, Wenyu
    Du, Ziwei
    Chen, Jie
    Duan, Zhen
    IEEE TRANSACTIONS ON BIG DATA, 2023, 9 (06) : 1586 - 1597