DDQP: A Double Deep Q-Learning Approach to Online Fault-Tolerant SFC Placement

被引:37
作者
Wang, Lei [1 ,2 ]
Mao, Weixi [1 ,2 ]
Zhao, Jin [1 ,2 ]
Xu, Yuedong [3 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200438, Peoples R China
[2] Shanghai Key Lab Intelligent Informat Proc, Shanghai 200438, Peoples R China
[3] Fudan Univ, Sch Informat Sci & Technol, Shanghai 200438, Peoples R China
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2021年 / 18卷 / 01期
基金
中国国家自然科学基金;
关键词
Fault tolerant systems; Fault tolerance; Software; Routing; Real-time systems; Software reliability; Reinforcement learning; Deep reinforcement learning; service function chain; network function virtualization; fault tolerance; NETWORK; ALLOCATION; GAME; GO;
D O I
10.1109/TNSM.2021.3049298
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Since Network Function Virtualization (NFV) decouples network functions (NFs) from the underlying dedicated hardware and realizes them in the form of software called Virtual Network Functions (VNFs), they are enabled to run in any resource-sufficient virtual machines. A service function chain (SFC) is composed of a sequential set of VNFs. As VNFs are vulnerable to various faults such as software failures, we consider how to deploy both active and standby SFC instances. Given the complexity and unpredictability of the network state, we propose a double deep Q-networks based online SFC placement scheme DDQP. Specifically, DDQP uses deep neural networks to deal with large continuous network state space. In the case of stateful VNFs, we offer constant generated state updates from active instances to standby instances to guarantee seamless redirection after failures. With the goal of balancing the waste of resources and ensuring service reliability, we introduce five progressive schemes of resource reservations to meet different customer needs. Our experimental results demonstrate that DDQP responds rapidly to arriving requests and reaches near-optimal performance. Specifically, DDQP outweighs the state-of-the-art method by 16.30% and 38.51% higher acceptance ratio under different schemes with 82x speedup on average. In order to enhance the integrity of the SFC state transition, we further proposed DDQP+, which extends DDQP by adding the delayed placement mechanism. Compared with DDQP, the design of the DDQP+ algorithm is more reasonable and comprehensive. The experiment results also show that DDQP+ achieved further improvement in multiple performance indicators.
引用
收藏
页码:118 / 132
页数:15
相关论文
共 50 条
  • [41] Efficient Replay Deep Meta-Reinforcement Learning for Active Fault-Tolerant Control of Solid Oxide Fuel Cell Systems Considering Multivariable Coordination
    Li, Jiawen
    Zhou, Tao
    [J]. IEEE TRANSACTIONS ON TRANSPORTATION ELECTRIFICATION, 2025, 11 (01): : 4803 - 4817
  • [42] Brain-Inspired Deep Meta-Reinforcement Learning for Active Coordinated Fault-Tolerant Load Frequency Control of Multi-Area Grids
    Li, Jiawen
    Zhou, Tao
    Cui, Haoyang
    [J]. IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (03) : 2518 - 2530
  • [43] Dynamic Courier Capacity Acquisition in Rapid Delivery Systems: A Deep Q-Learning Approach
    Auad, Ramon
    Erera, Alan
    Savelsbergh, Martin
    [J]. TRANSPORTATION SCIENCE, 2024, 58 (01) : 67 - 93
  • [44] Enhancing Intersection Signal Control: Distributional Double Dueling Deep Q-learning Network with Priority Experience Replay and NoisyNet Approach
    He, Yue
    Mu, Chen
    Sun, Yu
    [J]. 2023 19TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING, MSN 2023, 2023, : 794 - 799
  • [45] DHQN: a Stable Approach to Remove Target Network from Deep Q-learning Network
    Yang, Guang
    Li, Yang
    Fei, Di'an
    Huang, Tian
    Li, Qingyun
    Chen, Xingguo
    [J]. 2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 1474 - 1479
  • [46] Double Deep Q-Learning based Backhaul Spectrum Allocation in Integrated Access and Backhaul Network
    Park, Jeonghun
    Jin, Heetae
    Joo, Jaehan
    Choi, Geonho
    Kim, Suk Chan
    [J]. 2023 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION, ICAIIC, 2023, : 706 - 708
  • [47] Optimization of NB-IoT Uplink Resource Allocation via Double Deep Q-Learning
    Zhong, Han
    Zhang, Runzhou
    Jin, Fan
    Ning, Lei
    [J]. COMMUNICATIONS, SIGNAL PROCESSING, AND SYSTEMS, VOL. 1, 2022, 878 : 775 - 781
  • [48] Intelligent Handover Algorithm for Vehicle-to-Network Communications With Double-Deep Q-Learning
    Tan, Kang
    Bremner, Duncan
    Le Kernec, Julien
    Sambo, Yusuf
    Zhang, Lei
    Imran, Muhammad Ali
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (07) : 7848 - 7862
  • [49] Fault detection and isolation for a small CMG-based satellite: A fuzzy Q-learning approach
    Choi, Young-Cheol
    Son, Ji-Hwan
    Ahn, Hyo-Sung
    [J]. AEROSPACE SCIENCE AND TECHNOLOGY, 2015, 47 : 340 - 355
  • [50] A Double Deep Q Network Guided Online Learning Differential Evolution Algorithm
    Zhao, Fuqing
    Yang, Mingxiang
    [J]. ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT I, ICIC 2024, 2024, 14862 : 196 - 208