Dynamic QoS Prediction With Intelligent Route Estimation Via Inverse Reinforcement Learning

被引:2
|
作者
Li, Jiahui [1 ]
Wu, Hao [1 ]
He, Qiang [2 ,3 ]
Zhao, Yiji [4 ]
Wang, Xin [5 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650091, Yunnan, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
[3] Huazhong Univ Sci & Technol, Natl Engn Res Ctr Big Data Technol & Syst, Sch Comp Sci & Technol, Serv Comp Technol & Syst Lab,Cluster & Grid Comp, Wuhan 430074, Peoples R China
[4] Beijing Jiaotong Univ, Sch Comp Sci, Beijing 100044, Peoples R China
[5] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Quality of service; Reinforcement learning; Estimation; Predictive models; Peer-to-peer computing; Network topology; Heuristic algorithms; Deep learning; inverse reinforcement learning; QoS prediction; reinforcement learning; route estimation; LOCATION; RECOMMENDATION; FACTORIZATION; ALGORITHMS; MODEL;
D O I
10.1109/TSC.2023.3342481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dynamic quality of service (QoS) measurement is crucial for discovering services and developing online service systems. Collaborative filtering-based approaches perform dynamic QoS prediction by incorporating temporal information only but never consider the dynamic network environment and suffer from poor performance. Considering different service invocation routes directly reflect the dynamic environment and further lead to QoS fluctuations, we coin the problem of Dynamic QoS Prediction (DQP) with Intelligent Route Estimation (IRE) and propose a novel framework named IRE4DQP. Under the IRE4DQP framework, the dynamic environment is captured by Network Status Representation, and the IRE is modeled as a Markov decision process and implemented by a deep learning agent. After that, the DQP is achieved by a specific neural model with the estimated route as input. Through collaborative training with reinforcement and inverse reinforcement learning, eventually, based on the updated representations of the network status, IRE learns an optimal route policy that matches well with observed QoS values, and DQP achieves accurate predictions. Experimental results demonstrate that IRE4DQP outperforms SOTA methods on the accuracy of response-time prediction by 5.79-31.34% in MAE, by 1.29-20.18% in RMSE, and by 4.43-27.73% in NMAE and with a success rate of nearly 45% on finding routes.
引用
收藏
页码:509 / 523
页数:15
相关论文
共 50 条
  • [41] Dynamic route planning method based on deep reinforcement learning and velocity obstacle
    Lou Mengmeng
    Yang Xiaofei
    Xiang Zhengrong
    Wang Qi
    Hu Jiabao
    2023 IEEE 12TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE, DDCLS, 2023, : 627 - 632
  • [42] Long-Term Feature Extraction via Frequency Prediction for Efficient Reinforcement Learning
    Wang, Jie
    Ye, Mingxuan
    Kuang, Yufei
    Yang, Rui
    Zhou, Wengang
    Li, Houqiang
    Wu, Feng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) : 3094 - 3110
  • [43] Reinforcement learning for QoS-guaranteed intelligent routing in Wireless Mesh Networks with heavy traffic load
    Duong, Thuy-Van T.
    Le Huu Binh
    Ngo, Vuong M.
    ICT EXPRESS, 2022, 8 (01): : 18 - 24
  • [44] Dependable Reinforcement Learning via Timed Differential Dynamic Logic
    Wang, Runhao
    Zhang, Yuhong
    Sun, Haiying
    Liu, Jing
    26TH IEEE SYMPOSIUM ON COMPUTERS AND COMMUNICATIONS (IEEE ISCC 2021), 2021,
  • [45] Portfolio Allocation with Dynamic Risk Preferences via Reinforcement Learning
    Chen, Ting-Fu
    Kuang, Xian-Ji
    Liao, Szu-Lang
    Lin, Shih-Kuei
    COMPUTATIONAL ECONOMICS, 2024, 64 (04) : 2033 - 2052
  • [46] Dynamic Data Publishing with Differential Privacy via Reinforcement Learning
    Gao, Ruichao
    Ma, Xuebin
    2019 IEEE 43RD ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE (COMPSAC), VOL 1, 2019, : 746 - 752
  • [47] Pilot Performance Modeling via Observer-Based Inverse Reinforcement Learning
    Town, Jared
    Morrison, Zachary
    Kamalapurkar, Rushikesh
    IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2024, 32 (06) : 2444 - 2451
  • [48] Learning Adaptive Graph Protection Strategy on Dynamic Networks via Reinforcement Learning
    Wijayanto, Arie Wahyu
    Murata, Tsuyoshi
    2018 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE (WI 2018), 2018, : 534 - 539
  • [49] Intelligent Traffic Light via Policy-based Deep Reinforcement Learning
    Yue Zhu
    Mingyu Cai
    Chris W. Schwarz
    Junchao Li
    Shaoping Xiao
    International Journal of Intelligent Transportation Systems Research, 2022, 20 : 734 - 744
  • [50] Intelligent Traffic Light via Policy-based Deep Reinforcement Learning
    Zhu, Yue
    Cai, Mingyu
    Schwarz, Chris W.
    Li, Junchao
    Xiao, Shaoping
    INTERNATIONAL JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS RESEARCH, 2022, 20 (03) : 734 - 744