Remote State Estimation in the Presence of an Active Eavesdropper

被引:49
作者
Ding, Kemi [1 ]
Ren, Xiaoqiang [2 ]
Leong, Alex S. [3 ]
Quevedo, Daniel E. [4 ]
Shi, Ling [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Kowloon, Hong Kong, Peoples R China
[2] Shanghai Univ, Sch Mech Engn & Automat, Shanghai 200444, Peoples R China
[3] Paderborn Univ, Dept Elect Engn & Informat Technol, D-33098 Paderborn, Germany
[4] Queensland Univ Technol, Sch Elect Engn & Robot, Brisbane, Qld 4001, Australia
基金
国家重点研发计划;
关键词
Active eavesdropper; cyber-physical system (CPS) privacy; Markov decision process (MDP); state estimation; MARKOV DECISION-PROCESSES; POLICIES; SYSTEMS;
D O I
10.1109/TAC.2020.2980730
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We consider remote state estimation in the presence of an active eavesdropper. A sensor forward local state estimates to a remote estimator over a network, which may be eavesdropped by an intelligent adversary. Aiming at improving the eavesdropping performance efficiently, the adversary may adaptively alternate between an eavesdropping and an active mode. In contrast to eavesdropping, the active attack enables the adversary to sabotage the data transfer to the estimator, and improve the data reception to itself at the same time. However, launching active attacks may increase the risk of being detected. As a result, a tradeoff between eavesdropping performance and stealthiness arises. We present a generalized framework for active eavesdropping and propose a criterion based on the packet reception rate at the estimator to evaluate the stealthiness of the eavesdropper. Moreover, the tradeoff is formulated as a constrained Markov decision process. After deriving a sufficient condition under which at least one stationary policy satisfies the stealthiness constraint and also bounds the eavesdropping performance, we develop an optimal attack policy for the eavesdropper and focus on the structural analysis of the optimal policy. Furthermore, numerical examples are provided to illustrate the developed results.
引用
收藏
页码:229 / 244
页数:16
相关论文
共 40 条
[1]  
Altman E, 1999, STOCH MODEL SER
[2]  
Anderson B. D., 2005, Optimal filtering
[3]  
[Anonymous], Q-learning decision transformer: Leveraging dynamic pro, DOI DOI 10.1002/9780470316887
[4]  
[Anonymous], 2011, SUPERMODULARITY COMP, DOI DOI 10.1515/9781400822539
[5]  
[Anonymous], 2013, Applied Probability Models with Optimization Applications
[6]   OPTIMAL POLICIES FOR CONTROLLED MARKOV-CHAINS WITH A CONSTRAINT [J].
BEUTLER, FJ ;
ROSS, KW .
JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 1985, 112 (01) :236-252
[7]  
Bloch M., 2011, PHYS LAYER SECURITY, DOI DOI 10.1017/CBO9780511977985
[8]   An actor-critic algorithm for constrained Markov decision processes [J].
Borkar, VS .
SYSTEMS & CONTROL LETTERS, 2005, 54 (03) :207-213
[9]  
Cardenas Alvaro A., 2008, 2008 28th International Conference on Distributed Computing Systems Workshops (ICDCS Workshops), P495, DOI 10.1109/ICDCS.Workshops.2008.40
[10]   COMPARING RECENT ASSUMPTIONS FOR THE EXISTENCE OF AVERAGE OPTIMAL STATIONARY POLICIES [J].
CAVAZOSCADENA, R ;
SENNOTT, LI .
OPERATIONS RESEARCH LETTERS, 1992, 11 (01) :33-37