Real-time power optimization based on Q-learning algorithm for direct methanol fuel cell system

被引:1
|
作者
Chi, Xuncheng [1 ]
Chen, Fengxiang [1 ]
Zhai, Shuang [2 ]
Hu, Zhe [2 ]
Zhou, Su [3 ]
Wei, Wei [4 ]
机构
[1] Tongji Univ, Sch Automot Studies, Shanghai, Peoples R China
[2] Shanghai Refire Technol Co Ltd, Shanghai, Peoples R China
[3] Shanghai Zhongqiao Vocat & Tech Univ, Shanghai, Peoples R China
[4] CAS &M Zhangjiagang New Energy Technol Co Ltd, Zhangjiagang, Peoples R China
基金
中国国家自然科学基金;
关键词
Direct methanol fuel cell (DMFC) system; Real-time power optimization; Methanol supply control; Reinforcement learning; Q -learning algorithm; MASS-TRANSPORT MODEL; NUMERICAL-MODEL; PERFORMANCE; DMFC;
D O I
10.1016/j.ijhydene.2024.09.084
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
Efficient real-time power optimization of direct methanol fuel cell (DMFC) system is crucial for enhancing its performance and reliability. The power of DMFC system is mainly affected by stack temperature and circulating methanol concentration. However, the methanol concentration cannot be directly measured using reliable sensors, which poses a challenge for the real-time power optimization. To address this issue, this paper investigates the operating mechanism of DMFC system and establishes a system power model. Based on the established model, reinforcement learning using Q-learning algorithm is proposed to control methanol supply to optimize DMFC system power under varying operating conditions. This algorithm is simple, easy to implement, and does not rely on methanol concentration measurements. To validate the effectiveness of the proposed algorithm, simulation comparisons between the proposed method and the traditional perturbation and observation (P&O) algorithm are implemented under different operating conditions. The results show that proposed power optimization based on Q-learning algorithm improves net power by 1% and eliminates the fluctuation of methanol supply caused by P&O. For practical implementation considerations and real-time requirements of the algorithm, hardware-in-the-loop (HIL) experiments are conducted. The experiment results demonstrate that the proposed methods optimize net power under different operating conditions. Additionally, in terms of model accuracy, the experimental results are well matched with the simulation. Moreover, under varying load condition, compared with P&O, proposed power optimization based on Q-learning algorithm reduces root mean square error (RMSE) from 7.271% to 2.996% and mean absolute error (MAE) from 5.036% to 0.331%.
引用
收藏
页码:1241 / 1253
页数:13
相关论文
共 50 条
  • [11] Indoor Emergency Path Planning Based on the Q-Learning Optimization Algorithm
    Xu, Shenghua
    Gu, Yang
    Li, Xiaoyan
    Chen, Cai
    Hu, Yingyi
    Sang, Yu
    Jiang, Wenxing
    ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2022, 11 (01)
  • [12] Optimization of Shunting Operation Plan for Detaching and Attaching Trains Based on Q-Learning Algorithm
    Shi J.
    Chen L.
    Lin B.
    Meng G.
    Xia S.
    Zhongguo Tiedao Kexue/China Railway Science, 2022, 43 (01): : 163 - 170
  • [13] Dynamic Obstacle Avoidance of Mobile Robots Using Real-Time Q-learning
    Kim, HoWon
    Lee, WonChang
    2022 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2022,
  • [14] Optimization of Electrical System Topology for Offshore Wind Farm Based on Q-learning Particle Swarm Optimization Algorithm
    Qi Y.
    Hou P.
    Jin R.
    Dianli Xitong Zidonghua/Automation of Electric Power Systems, 2021, 45 (21): : 66 - 75
  • [15] Optimization and simulation of distribution system in a supply chain based on Q-learning
    Li Suicheng
    Lin Jun
    Yin Hongying
    2006 INTERNATIONAL CONFERENCE ON SERVICE SYSTEMS AND SERVICE MANAGEMENT, VOLS 1 AND 2, PROCEEDINGS, 2006, : 1445 - 1449
  • [16] Learning-Based Modeling and Optimization for Real-Time System Availability
    Li, Liying
    Zhou, Junlong
    Wei, Tongquan
    Chen, Mingsong
    Hu, Xiaobo Sharon
    IEEE TRANSACTIONS ON COMPUTERS, 2021, 70 (04) : 581 - 594
  • [17] Real-Time Path Planning Through Q-learning's Exploration Strategy Adjustment
    Kim, Howon
    Lee, WonChang
    2021 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2021,
  • [18] Real-time optimization of an experimental solid-oxide fuel-cell system
    Ferreira, T. de Avila
    Wuillemin, Z.
    Marchetti, A. G.
    Salzmann, C.
    Van Herle, J.
    Bonvin, D.
    JOURNAL OF POWER SOURCES, 2019, 429 : 168 - 179
  • [19] Learning Based Power Management for Periodic Real-Time Tasks
    ul Islam, Fakhruddin Muhammad Mahbub
    Lin, Man
    2014 IEEE INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING AND COMMUNICATIONS, 2014 IEEE 6TH INTL SYMP ON CYBERSPACE SAFETY AND SECURITY, 2014 IEEE 11TH INTL CONF ON EMBEDDED SOFTWARE AND SYST (HPCC,CSS,ICESS), 2014, : 534 - 541
  • [20] A real-time simulating non-isothermal mathematical model for the passive feed direct methanol fuel cell
    Shrivastava, Naveen K.
    Thombre, Shashikant B.
    Wasewar, Kailas L.
    INTERNATIONAL JOURNAL OF GREEN ENERGY, 2016, 13 (02) : 213 - 228