Dynamic QoS Prediction With Intelligent Route Estimation Via Inverse Reinforcement Learning

被引:2
|
作者
Li, Jiahui [1 ]
Wu, Hao [1 ]
He, Qiang [2 ,3 ]
Zhao, Yiji [4 ]
Wang, Xin [5 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650091, Yunnan, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
[3] Huazhong Univ Sci & Technol, Natl Engn Res Ctr Big Data Technol & Syst, Sch Comp Sci & Technol, Serv Comp Technol & Syst Lab,Cluster & Grid Comp, Wuhan 430074, Peoples R China
[4] Beijing Jiaotong Univ, Sch Comp Sci, Beijing 100044, Peoples R China
[5] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Quality of service; Reinforcement learning; Estimation; Predictive models; Peer-to-peer computing; Network topology; Heuristic algorithms; Deep learning; inverse reinforcement learning; QoS prediction; reinforcement learning; route estimation; LOCATION; RECOMMENDATION; FACTORIZATION; ALGORITHMS; MODEL;
D O I
10.1109/TSC.2023.3342481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dynamic quality of service (QoS) measurement is crucial for discovering services and developing online service systems. Collaborative filtering-based approaches perform dynamic QoS prediction by incorporating temporal information only but never consider the dynamic network environment and suffer from poor performance. Considering different service invocation routes directly reflect the dynamic environment and further lead to QoS fluctuations, we coin the problem of Dynamic QoS Prediction (DQP) with Intelligent Route Estimation (IRE) and propose a novel framework named IRE4DQP. Under the IRE4DQP framework, the dynamic environment is captured by Network Status Representation, and the IRE is modeled as a Markov decision process and implemented by a deep learning agent. After that, the DQP is achieved by a specific neural model with the estimated route as input. Through collaborative training with reinforcement and inverse reinforcement learning, eventually, based on the updated representations of the network status, IRE learns an optimal route policy that matches well with observed QoS values, and DQP achieves accurate predictions. Experimental results demonstrate that IRE4DQP outperforms SOTA methods on the accuracy of response-time prediction by 5.79-31.34% in MAE, by 1.29-20.18% in RMSE, and by 4.43-27.73% in NMAE and with a success rate of nearly 45% on finding routes.
引用
收藏
页码:509 / 523
页数:15
相关论文
共 50 条
  • [1] Objective Weight Interval Estimation Using Adversarial Inverse Reinforcement Learning
    Takayama, Naoya
    Arai, Sachiyo
    IEEE ACCESS, 2023, 11 : 58532 - 58538
  • [2] Methodologies for Imitation Learning via Inverse Reinforcement Learning: A Review
    Zhang K.
    Yu Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2019, 56 (02): : 254 - 261
  • [3] Accurate Prediction of Required Virtual Resources via Deep Reinforcement Learning
    Huang, Haojun
    Li, Zhaoxi
    Tian, Jialin
    Min, Geyong
    Miao, Wang
    Wu, Dapeng Oliver
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2023, 31 (02) : 920 - 933
  • [4] A dynamic route guidance arithmetic based on reinforcement learning
    Zhang, Z
    Xu, JM
    PROCEEDINGS OF 2005 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-9, 2005, : 3607 - 3611
  • [5] Route Optimization via Environment-Aware Deep Network and Reinforcement Learning
    Guo, Pengzhan
    Xiao, Keli
    Ye, Zeyang
    Zhu, Wei
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (06)
  • [6] Future Trajectory Prediction via RNN and Maximum Margin Inverse Reinforcement Learning
    Choi, Dooseop
    An, Taeg-Hyun
    Ahn, Kyounghwan
    Choi, Jeongdan
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 125 - 130
  • [7] Inverse Reinforcement Learning for Legibility Automation in Intelligent Agents
    Zeng, Buxin
    Zeng, Yifeng
    Pan, Yinghui
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 741 - 746
  • [8] TEMPORAL LINK PREDICTION VIA REINFORCEMENT LEARNING
    Tao, Ye
    Li, Ying
    Wu, Zhonghai
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3470 - 3474
  • [9] Speech Enhancement Using Dynamic Learning in Knowledge Distillation via Reinforcement Learning
    Chu, Shih-Chuan
    Wu, Chung-Hsien
    Su, Tsai-Wei
    IEEE ACCESS, 2023, 11 : 144421 - 144434
  • [10] Connected vehicles' dynamic route planning based on reinforcement learning
    Ma, Kanghua
    Liao, Shubing
    Niu, Yunyun
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 153 : 375 - 390