On Optimal Learning Under Targeted Data Poisoning

被引:0
作者
Hanneke, Steve [1 ]
Karbasi, Amin [2 ,3 ]
Mahmoody, Mohammad [4 ]
Mehalel, Idan [5 ]
Moran, Shay [3 ,5 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
[2] Yale Univ, New Haven, CT USA
[3] Google Res, Mountain View, CA USA
[4] Univ Virginia, Charlottesville, VA USA
[5] Technion, Haifa, Israel
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022 | 2022年
关键词
STABILITY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Consider the task of learning a hypothesis class H in the presence of an adversary that can replace up to an eta fraction of the examples in the training set with arbitrary adversarial examples. The adversary aims to fail the learner on a particular target test point x which is known to the adversary but not to the learner. In this work we aim to characterize the smallest achievable error epsilon = epsilon(eta) by the learner in the presence of such an adversary in both realizable and agnostic settings. We fully achieve this in the realizable setting, proving that epsilon = Theta(VC(H) center dot eta), where VC(H) is the VC dimension of H. Remarkably, we show that the upper bound can be attained by a deterministic learner. In the agnostic setting we reveal a more elaborate landscape: we devise a deterministic learner with a multiplicative regret guarantee of epsilon <= C center dot OPT + O(VC(H) center dot eta), where C > 1 is a universal numerical constant. We complement this by showing that for any deterministic learner there is an attack which worsens its error to at least 2 center dot OPT. This implies that a multiplicative deterioration in the regret is unavoidable in this case. Finally, the algorithms we develop for achieving the optimal rates are inherently improper. Nevertheless, we show that for a variety of natural concept classes, such as linear classifiers, it is possible to retain the dependence epsilon = Theta(H)(eta) by a proper algorithm in the realizable setting. Here Theta(H) conceals a polynomial dependence on VC(H).
引用
收藏
页数:13
相关论文
共 40 条
[1]  
[Anonymous], 2018, arXiv
[2]  
[Anonymous], 2019, P INT C MACHINE LEAR
[3]   The Power of Localization for Efficiently Learning Linear Separators with Noise [J].
Awasthi, Pranjal ;
Balcan, Maria Florina ;
Long, Philip M. .
STOC'14: PROCEEDINGS OF THE 46TH ANNUAL 2014 ACM SYMPOSIUM ON THEORY OF COMPUTING, 2014, :449-458
[4]  
Balcan Maria-Florina, 2022, ARXIV220304160
[5]  
Barreno M., 2006, P 2006 ACM S INF COM, P16
[6]  
Blum Avrim, 2021, C LEARN THEOR
[7]   Stability and generalization [J].
Bousquet, O ;
Elisseeff, A .
JOURNAL OF MACHINE LEARNING RESEARCH, 2002, 2 (03) :499-526
[8]  
Bousquet Olivier, 2020, P MACHINE LEARNING R, V125
[9]  
Braverman Mark, 2019, ARXIV190903547
[10]   PAC learning with nasty noise [J].
Bshouty, NH ;
Eiron, N ;
Kushilevitz, E .
THEORETICAL COMPUTER SCIENCE, 2002, 288 (02) :255-275