Accelerating bio-inspired optimizer with transfer reinforcement learning for reactive power optimization

被引:41
作者
Zhang, Xiaoshun [1 ]
Yu, Tao [1 ]
Yang, Bo [2 ]
Cheng, Lefeng [1 ]
机构
[1] South China Univ Technol, Coll Elect Power, Guangzhou 510640, Guangdong, Peoples R China
[2] Kunming Univ Sci & Technol, Fac Elect Power Engn, Kunming 650504, Peoples R China
基金
中国国家自然科学基金;
关键词
Accelerating bio-inspired optimizer; Transfer reinforcement learning; Memory matrix; Cooperating multi-bion; WOLF-PHC; Reactive power optimization; ECONOMIC-DISPATCH; ALGORITHM; FLOW; KNOWLEDGE; DISCRETE; SYSTEMS; MODELS;
D O I
10.1016/j.knosys.2016.10.024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a novel accelerating bio-inspired optimizer (ABO) associated with transfer reinforcement learning (TRL) to solve the reactive power optimization (RPO) in large-scale power systems. A memory matrix is employed to represent the memory of different state-action pairs, which is used for knowledge learning, storage, and transfer among different optimization tasks. Then an associative memory is introduced to significantly reduce the dimension of memory matrix, in which more than one element can be simultaneously updated by the cooperating multi-bion. The win or learn fast policy hill-climbing (WOLF-PHC) is also used to accelerate the convergence. Thus, ABO can rapidly seek the closest solution to the exact global optimum by exploiting the prior knowledge of the source tasks according to their similarities. The performance of ABO has been evaluated for RPO on IEEE 118-bus system and IEEE 300 bus system, respectively. Simulation results verify that ABO outperforms the existing artificial intelligence algorithms in terms of global convergence ability and stability, which can raise one order of magnitude of the convergence rate than that of others. (C) 2016 Elsevier B.V. All rights reserved.
引用
收藏
页码:26 / 38
页数:13
相关论文
共 43 条
  • [1] Optimal economic designing of grid-connected photovoltaic systems with multiple inverters using linear and nonlinear module models based on Genetic Algorithm
    Bakhshi, Reza
    Sadeh, Javad
    Mosaddegh, Hamid-Reza
    [J]. RENEWABLE ENERGY, 2014, 72 : 386 - 394
  • [2] Transferring knowledge as heuristics in reinforcement learning: A case-based approach
    Bianchi, Reinaldo A. C.
    Celiberto, Luiz A., Jr.
    Santos, Paulo E.
    Matsuura, Jackson P.
    Lopez de Mantaras, Ramon
    [J]. ARTIFICIAL INTELLIGENCE, 2015, 226 : 102 - 121
  • [3] Hierarchical reinforcement learning and decision making
    Botvinick, Matthew Michael
    [J]. CURRENT OPINION IN NEUROBIOLOGY, 2012, 22 (06) : 956 - 962
  • [4] Multiagent learning using a variable learning rate
    Bowling, M
    Veloso, M
    [J]. ARTIFICIAL INTELLIGENCE, 2002, 136 (02) : 215 - 250
  • [5] Particle swarm optimization-based optimal power management of plug-in hybrid electric vehicles considering uncertain driving conditions
    Chen, Zeyu
    Xiong, Rui
    Cao, Jiayi
    [J]. ENERGY, 2016, 96 : 197 - 208
  • [6] Seeker Optimization Algorithm for Optimal Reactive Power Dispatch
    Dai, Chaohua
    Chen, Weirong
    Zhu, Yunfang
    Zhang, Xuexia
    [J]. IEEE TRANSACTIONS ON POWER SYSTEMS, 2009, 24 (03) : 1218 - 1231
  • [7] Dorigo M., 1999, P 4 INT C PAR PROBL, P656
  • [8] Adaptive barrier filter-line-search interior point method for optimal power flow with FACTS devices
    Duan, Chao
    Fang, Wanliang
    Jiang, Lin
    Liu, Jun
    [J]. IET GENERATION TRANSMISSION & DISTRIBUTION, 2015, 9 (16) : 2792 - 2798
  • [9] Group counseling optimization
    Eita, M. A.
    Fahmy, M. M.
    [J]. APPLIED SOFT COMPUTING, 2014, 22 : 585 - 604
  • [10] A new optimization method: Big Bang Big Crunch
    Erol, OK
    Eksin, I
    [J]. ADVANCES IN ENGINEERING SOFTWARE, 2006, 37 (02) : 106 - 111