DDQP: A Double Deep Q-Learning Approach to Online Fault-Tolerant SFC Placement

被引:37
|
作者
Wang, Lei [1 ,2 ]
Mao, Weixi [1 ,2 ]
Zhao, Jin [1 ,2 ]
Xu, Yuedong [3 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200438, Peoples R China
[2] Shanghai Key Lab Intelligent Informat Proc, Shanghai 200438, Peoples R China
[3] Fudan Univ, Sch Informat Sci & Technol, Shanghai 200438, Peoples R China
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2021年 / 18卷 / 01期
基金
中国国家自然科学基金;
关键词
Fault tolerant systems; Fault tolerance; Software; Routing; Real-time systems; Software reliability; Reinforcement learning; Deep reinforcement learning; service function chain; network function virtualization; fault tolerance; NETWORK; ALLOCATION; GAME; GO;
D O I
10.1109/TNSM.2021.3049298
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Since Network Function Virtualization (NFV) decouples network functions (NFs) from the underlying dedicated hardware and realizes them in the form of software called Virtual Network Functions (VNFs), they are enabled to run in any resource-sufficient virtual machines. A service function chain (SFC) is composed of a sequential set of VNFs. As VNFs are vulnerable to various faults such as software failures, we consider how to deploy both active and standby SFC instances. Given the complexity and unpredictability of the network state, we propose a double deep Q-networks based online SFC placement scheme DDQP. Specifically, DDQP uses deep neural networks to deal with large continuous network state space. In the case of stateful VNFs, we offer constant generated state updates from active instances to standby instances to guarantee seamless redirection after failures. With the goal of balancing the waste of resources and ensuring service reliability, we introduce five progressive schemes of resource reservations to meet different customer needs. Our experimental results demonstrate that DDQP responds rapidly to arriving requests and reaches near-optimal performance. Specifically, DDQP outweighs the state-of-the-art method by 16.30% and 38.51% higher acceptance ratio under different schemes with 82x speedup on average. In order to enhance the integrity of the SFC state transition, we further proposed DDQP+, which extends DDQP by adding the delayed placement mechanism. Compared with DDQP, the design of the DDQP+ algorithm is more reasonable and comprehensive. The experiment results also show that DDQP+ achieved further improvement in multiple performance indicators.
引用
收藏
页码:118 / 132
页数:15
相关论文
共 50 条
  • [1] Online Fault-tolerant VNF Chain Placement: A Deep Reinforcement Learning Approach
    Mao, Weixi
    Wang, Lei
    Zhao, Jin
    Xu, Yuedong
    2020 IFIP NETWORKING CONFERENCE AND WORKSHOPS (NETWORKING), 2020, : 163 - 171
  • [2] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100
  • [3] Active fault-tolerant attitude control based on Q-learning for rigid spacecraft with actuator faults
    Rafiee, Sajad
    Kankashvar, Mohammadrasoul
    Mohammadi, Parisa
    Bolandi, Hossein
    ADVANCES IN SPACE RESEARCH, 2024, 74 (03) : 1261 - 1275
  • [4] Fault-Tolerant Deep Learning Using Regularization
    Joardar, Biresh Kumar
    Arka, Aqeeb Iqbal
    Doppa, Janardhan Rao
    Pande, Partha Pratim
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [5] A Double Q-Learning Routing in Delay Tolerant Networks
    Yuan, Fan
    Wu, Jaogao
    Zhou, Hongyu
    Liu, Linfeng
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [6] An Online Home Energy Management System using Q-Learning and Deep Q-Learning
    Izmitligil, Hasan
    Karamancioglu, Abdurrahman
    SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2024, 43
  • [7] Output feedback fault-tolerant Q-learning for discrete-time linear systems with actuator faults
    Rafiee, Sajad
    Kankashvar, Mohammadrasoul
    Bolandi, Hossein
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 138
  • [8] A New Method for Fault Tolerant Control through Q-Learning
    Hua, Changsheng
    Ding, Steven X.
    Shardt, Yuri A. W.
    IFAC PAPERSONLINE, 2018, 51 (24): : 38 - 45
  • [9] Q-learning based fault estimation and fault tolerant iterative learning control for MIMO systems
    Wang, Rui
    Zhuang, Zhihe
    Tao, Hongfeng
    Paszke, Wojciech
    Stojanovic, Vladimir
    ISA TRANSACTIONS, 2023, 142 : 123 - 135
  • [10] Fault-Tolerant Pole-Placement in Double-Integrator Networks
    Locatelli, Arturo
    Schiavoni, Nicola
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2012, 57 (11) : 2912 - 2917