DDQP: A Double Deep Q-Learning Approach to Online Fault-Tolerant SFC Placement

被引:37
作者
Wang, Lei [1 ,2 ]
Mao, Weixi [1 ,2 ]
Zhao, Jin [1 ,2 ]
Xu, Yuedong [3 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200438, Peoples R China
[2] Shanghai Key Lab Intelligent Informat Proc, Shanghai 200438, Peoples R China
[3] Fudan Univ, Sch Informat Sci & Technol, Shanghai 200438, Peoples R China
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2021年 / 18卷 / 01期
基金
中国国家自然科学基金;
关键词
Fault tolerant systems; Fault tolerance; Software; Routing; Real-time systems; Software reliability; Reinforcement learning; Deep reinforcement learning; service function chain; network function virtualization; fault tolerance; NETWORK; ALLOCATION; GAME; GO;
D O I
10.1109/TNSM.2021.3049298
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Since Network Function Virtualization (NFV) decouples network functions (NFs) from the underlying dedicated hardware and realizes them in the form of software called Virtual Network Functions (VNFs), they are enabled to run in any resource-sufficient virtual machines. A service function chain (SFC) is composed of a sequential set of VNFs. As VNFs are vulnerable to various faults such as software failures, we consider how to deploy both active and standby SFC instances. Given the complexity and unpredictability of the network state, we propose a double deep Q-networks based online SFC placement scheme DDQP. Specifically, DDQP uses deep neural networks to deal with large continuous network state space. In the case of stateful VNFs, we offer constant generated state updates from active instances to standby instances to guarantee seamless redirection after failures. With the goal of balancing the waste of resources and ensuring service reliability, we introduce five progressive schemes of resource reservations to meet different customer needs. Our experimental results demonstrate that DDQP responds rapidly to arriving requests and reaches near-optimal performance. Specifically, DDQP outweighs the state-of-the-art method by 16.30% and 38.51% higher acceptance ratio under different schemes with 82x speedup on average. In order to enhance the integrity of the SFC state transition, we further proposed DDQP+, which extends DDQP by adding the delayed placement mechanism. Compared with DDQP, the design of the DDQP+ algorithm is more reasonable and comprehensive. The experiment results also show that DDQP+ achieved further improvement in multiple performance indicators.
引用
收藏
页码:118 / 132
页数:15
相关论文
共 50 条
  • [31] Intelligent fault diagnosis method based on unsupervised feature representation and deep Q-learning
    Wu W.
    Chen J.
    Liu S.
    Zhou Z.
    Zhongnan Daxue Xuebao (Ziran Kexue Ban)/Journal of Central South University (Science and Technology), 2022, 53 (05): : 1750 - 1759
  • [32] A learning-based approach to fault detection and fault-tolerant control of permanent magnet DC motors
    Sardashti A.
    Nazari J.
    Journal of Engineering and Applied Science, 2023, 70 (01):
  • [33] A Deep Q-Learning Bisection Approach for Power Allocation in Downlink NOMA Systems
    Youssef, Marie-Josepha
    Nour, Charbel Abdel
    Lagrange, Xavier
    Douillard, Catherine
    IEEE COMMUNICATIONS LETTERS, 2022, 26 (02) : 316 - 320
  • [34] A Distributed Double Deep Q-Learning Method for Object Redundancy Mitigation in Vehicular Networks
    Ghnaya, Imed
    Aniss, Hasnaa
    Ahmed, Toufik
    Mosbah, Mohamed
    2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [35] Multi-Agent Double Deep Q-Learning for Beamforming in mmWave MIMO Networks
    Wang, Xueyuan
    Gursoy, M. Cenk
    2020 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (IEEE PIMRC), 2020,
  • [36] Dueling double Q-learning based reinforcement learning approach for the flow shop scheduling problem
    Kim S.J.
    Kim B.W.
    Transactions of the Korean Institute of Electrical Engineers, 2021, 70 (10) : 1497 - 1508
  • [37] Deep Q-Learning for Self-Organizing Networks Fault Management and Radio Performance Improvement
    Mismar, Faris B.
    Evans, Brian L.
    2018 CONFERENCE RECORD OF 52ND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2018, : 1457 - 1461
  • [38] Human-like Autonomous Vehicle Speed Control by Deep Reinforcement Learning with Double Q-Learning
    Zhang, Yi
    Sun, Ping
    Yin, Yuhan
    Lin, Lin
    Wang, Xuesong
    2018 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2018, : 1251 - 1256
  • [39] Energy management based on reinforcement learning with double deep Q-learning for a hybrid electric tracked vehicle
    Han, Xuefeng
    He, Hongwen
    Wu, Jingda
    Peng, Jiankun
    Li, Yuecheng
    APPLIED ENERGY, 2019, 254
  • [40] Blockchain-Based Computing Resource Trading in Autonomous Multi-Access Edge Network Slicing: A Dueling Double Deep Q-Learning Approach
    Kwantwi, Thomas
    Sun, Guolin
    Kuadey, Noble Arden Elorm
    Maale, Gerald Tietaa
    Liu, Guisong
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2023, 20 (03): : 2912 - 2928