Exploration Degree Bias: The Hidden Influence of Node Degree in Graph Neural Network-Based Reinforcement Learning

被引:0
作者
Tarabek, Peter [1 ]
Matis, David [1 ]
机构
[1] Univ Zilina, Fac Management Sci & Informat, Zilina 01026, Slovakia
关键词
Convolution; Correlation; Aggregates; Training; Limiting; Hands; Graph convolutional networks; Testing; Taxonomy; Exploration degree bias; graph neural networks; message-passing mechanism; node degree bias; reinforcement learning;
D O I
10.1109/ACCESS.2025.3528878
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Graph Neural Networks (GNNs) have demonstrated remarkable performance in tasks involving graph-structured data, but they also exhibit biases linked to node degrees. This paper explores a specific manifestation of such bias, termed Exploration Degree Bias (EDB), in the context of Reinforcement Learning (RL). We show that EDB arises from the inherent design of GNNs, where nodes with high or low degrees disproportionately influence output logits used for decision-making. This phenomenon impacts exploration in RL, skewing it away from mid-degree nodes, potentially hindering the discovery of optimal policies. We provide a systematic investigation of EDB across widely used GNN architectures-GCN, GraphSAGE, GAT, and GIN-by quantifying correlations between node degrees and logits. Our findings reveal that EDB varies by architecture and graph configuration, with GCN and GIN exhibiting the strongest biases. Moreover, analysis of DQN and PPO RL agents illustrates how EDB can distort exploration patterns, with DQN exhibiting EDB under low exploration rates and PPO showing a partial ability to counteract these effects through its probabilistic sampling mechanism. Our contributions include defining and quantifying EDB, providing experimental insights into its existence and variability, and analyzing its implications for RL. These findings underscore the need to address degree-related biases in GNNs to enhance RL performance on graph-based tasks.
引用
收藏
页码:10746 / 10757
页数:12
相关论文
共 46 条
[1]   Statistical mechanics of complex networks [J].
Albert, R ;
Barabási, AL .
REVIEWS OF MODERN PHYSICS, 2002, 74 (01) :47-97
[2]   Deep reinforcement learning meets graph neural networks: Exploring a routing optimization use case [J].
Almasan, Paul ;
Suarez-Varela, Jose ;
Rusek, Krzysztof ;
Barlet-Ros, Pere ;
Cabellos-Aparicio, Albert .
COMPUTER COMMUNICATIONS, 2022, 196 :184-194
[3]  
Bello I., 2017, NEURAL COMBINATORIAL, DOI DOI 10.48550/ARXIV.1611.09940
[4]   Scale-free networks are rare [J].
Broido, Anna D. ;
Clauset, Aaron .
NATURE COMMUNICATIONS, 2019, 10 (1)
[5]   An overview on vehicle scheduling models [J].
Bunte S. ;
Kliewer N. .
Public Transp., 2009, 4 (299-317) :299-317
[6]  
Cappart Q, 2021, AAAI CONF ARTIF INTE, V35, P3677
[7]   BA-GNN: On Learning Bias-Aware Graph Neural Network [J].
Chen, Zhengyu ;
Xiao, Teng ;
Kuang, Kun .
2022 IEEE 38TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2022), 2022, :3012-3024
[8]   Power-Law Distributions in Empirical Data [J].
Clauset, Aaron ;
Shalizi, Cosma Rohilla ;
Newman, M. E. J. .
SIAM REVIEW, 2009, 51 (04) :661-703
[9]  
Dai HJ, 2017, ADV NEUR IN, V30
[10]   Magnetic control of tokamak plasmas through deep reinforcement learning [J].
Degrave, Jonas ;
Felici, Federico ;
Buchli, Jonas ;
Neunert, Michael ;
Tracey, Brendan ;
Carpanese, Francesco ;
Ewalds, Timo ;
Hafner, Roland ;
Abdolmaleki, Abbas ;
de las Casas, Diego ;
Donner, Craig ;
Fritz, Leslie ;
Galperti, Cristian ;
Huber, Andrea ;
Keeling, James ;
Tsimpoukelli, Maria ;
Kay, Jackie ;
Merle, Antoine ;
Moret, Jean-Marc ;
Noury, Seb ;
Pesamosca, Federico ;
Pfau, David ;
Sauter, Olivier ;
Sommariva, Cristian ;
Coda, Stefano ;
Duval, Basil ;
Fasoli, Ambrogio ;
Kohli, Pushmeet ;
Kavukcuoglu, Koray ;
Hassabis, Demis ;
Riedmiller, Martin .
NATURE, 2022, 602 (7897) :414-+