Black-box Attacks Against Neural Binary Function Detection

被引:0
作者
Bundt, Joshua [1 ,2 ]
Davinroy, Michael [1 ]
Agadakos, Ioannis [1 ,3 ]
Oprea, Alina [1 ]
Robertson, William [1 ]
机构
[1] Northeastern Univ, Boston, MA 02115 USA
[2] Army Cyber Inst, West Point, NY 10996 USA
[3] Amazon, Seattle, WA USA
来源
PROCEEDINGS OF THE 26TH INTERNATIONAL SYMPOSIUM ON RESEARCH IN ATTACKS, INTRUSIONS AND DEFENSES, RAID 2023 | 2023年
基金
美国国家科学基金会;
关键词
binary analysis; disassembly; deep neural network; function boundary detection; CODE;
D O I
10.1145/3607199.3607200
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Binary analyses based on deep neural networks (DNNs), or neural binary analyses (NBAs), have become a hotly researched topic in recent years. DNNs have been wildly successful at pushing the performance and accuracy envelopes in the natural language and image processing domains. Thus, DNNs are highly promising for solving binary analysis problems that are hard due to a lack of complete information resulting from the lossy compilation process. Despite this promise, it is unclear that the prevailing strategy of repurposing embeddings and model architectures originally developed for other problem domains is sound given the adversarial contexts under which binary analysis often operates. In this paper, we empirically demonstrate that the current state of the art in neural function boundary detection is vulnerable to both inadvertent and deliberate adversarial attacks. We proceed from the insight that current generation NBAs are built upon embeddings and model architectures intended to solve syntactic problems. We devise a simple, reproducible, and scalable black-box methodology for exploring the space of inadvertent attacks - instruction sequences that could be emitted by common compiler toolchains and configurations - that exploits this syntactic design focus. We then show that these inadvertent misclassifications can be exploited by an attacker, serving as the basis for a highly effective black-box adversarial example generation process. We evaluate this methodology against two state-of-the-art neural function boundary detectors: XDA and DeepDi. We conclude with an analysis of the evaluation data and recommendations for how future research might avoid succumbing to similar attacks.
引用
收藏
页码:1 / 16
页数:16
相关论文
共 78 条
  • [41] Marcelli A, 2022, PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, P2099
  • [42] SAFE: Self-Attentive Function Embeddings for Binary Similarity
    Massarelli, Luca
    Di Luna, Giuseppe Antonio
    Petroni, Fabio
    Baldoni, Roberto
    Querzoni, Leonardo
    [J]. DETECTION OF INTRUSIONS AND MALWARE, AND VULNERABILITY ASSESSMENT (DIMVA 2019), 2019, 11543 : 309 - 329
  • [43] Mikolov T, 2013, Arxiv, DOI [arXiv:1301.3781, 10.48550/arXiv.1301.3781]
  • [44] Probabilistic Disassembly
    Miller, Kenneth
    Kwon, Yonghwi
    Sun, Yi
    Zhang, Zhuo
    Zhang, Xiangyu
    Lin, Zhiqiang
    [J]. 2019 IEEE/ACM 41ST INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2019), 2019, : 1187 - 1198
  • [45] DeepFool: a simple and accurate method to fool deep neural networks
    Moosavi-Dezfooli, Seyed-Mohsen
    Fawzi, Alhussein
    Frossard, Pascal
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2574 - 2582
  • [46] NSA, 2019, Ghidra
  • [47] Ott M., 2019, NAACL HLT 2019 DEMON
  • [48] SoK: Security and Privacy in Machine Learning
    Papernot, Nicolas
    McDaniel, Patrick
    Sinha, Arunesh
    Wellman, Michael P.
    [J]. 2018 3RD IEEE EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY (EUROS&P 2018), 2018, : 399 - 414
  • [49] Pei KX, 2020, Arxiv, DOI [arXiv:2010.00770, 10.48550/arXiv.2010.00770, DOI 10.48550/ARXIV.2010.00770]
  • [50] Pei Kexin., 2021, XDA