Pre-trained Encoders in Self-Supervised Learning Improve Secure and Privacy-preserving Supervised Learning

被引:0
作者
Liu, Hongbin [1 ]
Qu, Wenjie [2 ]
Jia, Jinyuan [3 ]
Gong, Neil Zhenqiang [1 ]
机构
[1] Duke Univ, Durham, NC 27706 USA
[2] Natl Univ Singapore, Singapore, Singapore
[3] Penn State Univ, University Pk, PA USA
来源
PROCEEDINGS 45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS, SPW 2024 | 2024年
关键词
ROBUSTNESS; ATTACKS;
D O I
10.1109/SPW63631.2024.00019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classifiers in supervised learning have various security and privacy issues, e.g., 1) data poisoning attacks, backdoor attacks, and adversarial examples on the security side as well as 2) inference attacks to the training data on the privacy side. Various secure and privacy-preserving supervised learning algorithms with formal guarantees have been proposed to address these issues. However, they suffer from various limitations such as accuracy loss, small certified security guarantees, and/or inefficiency. Self-supervised learning pre-trains encoders using unlabeled data. Given a pre-trained encoder as a feature extractor, supervised learning can train a simple yet accurate classifier using a small amount of labeled training data. In this work, we perform the first systematic, principled measurement study to understand whether and when a pre-trained encoder can address the limitations of secure or privacy-preserving supervised learning algorithms. Our key findings are that a pre-trained encoder substantially improves 1) both accuracy under no attacks and certified security guarantees against data poisoning and backdoor attacks of state-of-the-art secure learning algorithms (i.e., bagging and KNN), 2) certified security guarantees of randomized smoothing against adversarial examples without sacrificing its accuracy under no attacks, 3) accuracy of differentially private classifiers.
引用
收藏
页码:144 / 156
页数:13
相关论文
共 58 条
  • [1] Deep Learning with Differential Privacy
    Abadi, Martin
    Chu, Andy
    Goodfellow, Ian
    McMahan, H. Brendan
    Mironov, Ilya
    Talwar, Kunal
    Zhang, Li
    [J]. CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, : 308 - 318
  • [2] AN INTRODUCTION TO KERNEL AND NEAREST-NEIGHBOR NONPARAMETRIC REGRESSION
    ALTMAN, NS
    [J]. AMERICAN STATISTICIAN, 1992, 46 (03) : 175 - 185
  • [3] [Anonymous], 2021, Public pre-trained image encoder of CLIP by OpenAI
  • [4] [Anonymous], 2021, MicroImageNet classification challenge
  • [5] Ateniese Giuseppe, 2015, International Journal of Security and Networks, V10, P137
  • [6] Athalye A, 2018, PR MACH LEARN RES, V80
  • [7] Barreno M, 2006, P 2006 ACM S INF COM, P16, DOI DOI 10.1145/1128817.1128824
  • [8] Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds
    Bassily, Raef
    Smith, Adam
    Thakurta, Abhradeep
    [J]. 2014 55TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS 2014), 2014, : 464 - 473
  • [9] Biggio B., 2012, arXiv preprint arXiv:1206.6389, P1467, DOI 10.48550/arxiv.1206.6389
  • [10] Bishop C. M., 2006, PATTERN RECOGN