Resource-Efficient Sensor Data Management for Autonomous Systems Using Deep Reinforcement Learning

被引:5
作者
Jeong, Seunghwan [1 ]
Yoo, Gwangpyo [2 ]
Yoo, Minjong [2 ]
Yeom, Ikjun [1 ]
Woo, Honguk [1 ]
机构
[1] Sungkyunkwan Univ, Dept Software, Suwon 16419, South Korea
[2] Sungkyunkwan Univ, Dept Math, Suwon 16419, South Korea
基金
新加坡国家研究基金会;
关键词
autonomous system; sensor network; real-time data; digital twin; reinforcement learning; action embedding;
D O I
10.3390/s19204410
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Hyperconnectivity via modern Internet of Things (IoT) technologies has recently driven us to envision "digital twin", in which physical attributes are all embedded, and their latest updates are synchronized on digital spaces in a timely fashion. From the point of view of cyberphysical system (CPS) architectures, the goals of digital twin include providing common programming abstraction on the same level of databases, thereby facilitating seamless integration of real-world physical objects and digital assets at several different system layers. However, the inherent limitations of sampling and observing physical attributes often pose issues related to data uncertainty in practice. In this paper, we propose a learning-based data management scheme where the implementation is layered between sensors attached to physical attributes and domain-specific applications, thereby mitigating the data uncertainty between them. To do so, we present a sensor data management framework, namely D2WIN, which adopts reinforcement learning (RL) techniques to manage the data quality for CPS applications and autonomous systems. To deal with the scale issue incurred by many physical attributes and sensor streams when adopting RL, we propose an action embedding strategy that exploits their distance-based similarity in the physical space coordination. We introduce two embedding methods, i.e., a user-defined function and a generative model, for different conditions. Through experiments, we demonstrate that the D2WIN framework with the action embedding outperforms several known heuristics in terms of achievable data quality under certain resource restrictions. We also test the framework with an autonomous driving simulator, clearly showing its benefit. For example, with only 30% of updates selectively applied by the learned policy, the driving agent maintains its performance about 96.2%, as compared to the ideal condition with full updates.
引用
收藏
页数:22
相关论文
共 35 条
[1]  
[Anonymous], 2018, P INT C MACH LEARN
[2]  
[Anonymous], 2018, ARXIV180803196
[3]  
[Anonymous], 2018, ARXIV181106776
[4]  
[Anonymous], 2017, PREPRINT
[5]  
[Anonymous], 2016, INT C MACH LEARN
[6]  
[Anonymous], 2014, Neural Information Processing Systems
[7]  
Bao YX, 2019, IEEE INFOCOM SER, P505, DOI [10.1109/INFOCOM.2019.8737460, 10.1109/infocom.2019.8737460]
[8]  
BARTLE R. G., 1995, ELEMENTS INTEGRATION
[9]   Optimization and Control of Cyber-Physical Vehicle Systems [J].
Bradley, Justin M. ;
Atkins, Ella M. .
SENSORS, 2015, 15 (09) :23020-23049
[10]  
CHINCHALI S, 2018, P 32 AAAI C ART INT