Connected and Autonomous Vehicles in Web3: An Intelligence-Based Reinforcement Learning Approach

被引:12
作者
Ren, Yuzheng [1 ,2 ]
Xie, Renchao [3 ]
Yu, Fei Richard [4 ]
Zhang, Ran [3 ]
Wang, Yuhang [5 ]
He, Ying [6 ]
Huang, Tao [3 ]
机构
[1] Univ Sci & Technol Beijing, Sch Comp & Commun Engn, Beijing 100083, Peoples R China
[2] Univ Sci & Technol Beijing, Beijing Engn & Technol Res Ctr Convergence Networ, Beijing 100083, Peoples R China
[3] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[4] Shenzhen Univ, Shenzhen Key Lab Digital & Intelligent Technol &, Shenzhen 518107, Peoples R China
[5] Hong Kong Univ Sci & Technol, Informat Hub, Guangzhou 511466, Peoples R China
[6] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen 518052, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Computational modeling; Internet; Data models; Reinforcement learning; Nonfungible tokens; Cognition; Smart contracts; Intelligence; Web3; connected and autonomous vehicles (CAVs);
D O I
10.1109/TITS.2024.3355179
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
"Read-write-own" based Web3 has been proposed as a promising user-centric Internet to open the new generation of the World Wide Web, where Web3 users can independently manage data and derive value from creating content without relying on intermediaries. Connected and autonomous vehicles (CAVs) in Web3 can trade models in a self-controlled and decentralized credible way, which is a fundamentally and principally innovation based on novel architecture. Effectively implementing such paradigms involves proper model trading strategies. However, reinforcement learning (RL)-based strategies face challenges of poor generalization ability, low feasibility, and the exploration-exploitation dilemma. It is also difficult to define an explicit and appropriate reward function. Therefore, in this paper, we propose an intelligence-based reinforcement learning (IRL) approach for CAVs in Web3. We present a framework to enable model transactions between CAVs. Also, we provide a decentralized identifier (DID)-based identity management system for resource description and data verification to access Web3, followed by the mechanism and supporting smart contracts. Furthermore, we formulate the model trading issue as an active inference to form higher-level cognition about the environment without rewards. Then we use IRL to solve it. And we use "intelligence," a high-level indicator, to quantify the efficiency of such cognition. It can evaluate the difference between the predicted state and the real state in policy exploration. The proposed scheme shows good generalization and can auto-balance exploration and exploitation, simultaneously achieving outperforming performance on the model trading issue with no rewards. In simulations, the performance of the proposed scheme is compared with existing methods.
引用
收藏
页码:9863 / 9877
页数:15
相关论文
共 43 条
[1]  
Avellaneda O., 2019, IEEE Communications Standards Magazine, V3, P10, DOI [10.1109/MCOMSTD.2019.9031542, DOI 10.1109/MCOMSTD.2019.9031542]
[2]  
Brys T, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P3352
[3]   On Latency and Reliability of Road Hazard Warnings Over the Cellular V2X Sidelink Interface [J].
Campolo, Claudia ;
Molinaro, Antonella ;
Berthet, Antoine O. ;
Vinel, Alexey .
IEEE COMMUNICATIONS LETTERS, 2019, 23 (11) :2135-2138
[4]   When Digital Economy Meets Web3.0: Applications and Challenges [J].
Chen, Chuan ;
Zhang, Lei ;
Li, Yihao ;
Liao, Tianchi ;
Zhao, Siran ;
Zheng, Zibin ;
Huang, Huawei ;
Wu, Jiajing .
IEEE OPEN JOURNAL OF THE COMPUTER SOCIETY, 2022, 3 :233-245
[5]   DeSci Based on Web3 and DAO: A Comprehensive Overview and Reference Model [J].
Ding, Wenwen ;
Hou, Jiachen ;
Li, Juanjuan ;
Guo, Chao ;
Qin, Jirong ;
Kozma, Robert ;
Wang, Fei-Yue .
IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2022, 9 (05) :1563-1573
[6]   Quantum reinforcement learning [J].
Dong, Daoyi ;
Chen, Chunlin ;
Li, Hanxiong ;
Tarn, Tzyh-Jong .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2008, 38 (05) :1207-1220
[7]  
Friston K, 2019, Arxiv, DOI arXiv:1906.10184
[8]   What Is Optimal about Motor Control? [J].
Friston, Karl .
NEURON, 2011, 72 (03) :488-498
[9]   Active inference and agency: optimal control without cost functions [J].
Friston, Karl J. ;
Samothrakis, Spyridon ;
Montague, Read .
BIOLOGICAL CYBERNETICS, 2012, 106 (8-9) :523-541
[10]   The free-energy principle: a unified brain theory? [J].
Friston, Karl J. .
NATURE REVIEWS NEUROSCIENCE, 2010, 11 (02) :127-138