Real-time power optimization based on Q-learning algorithm for direct methanol fuel cell system

被引:1
|
作者
Chi, Xuncheng [1 ]
Chen, Fengxiang [1 ]
Zhai, Shuang [2 ]
Hu, Zhe [2 ]
Zhou, Su [3 ]
Wei, Wei [4 ]
机构
[1] Tongji Univ, Sch Automot Studies, Shanghai, Peoples R China
[2] Shanghai Refire Technol Co Ltd, Shanghai, Peoples R China
[3] Shanghai Zhongqiao Vocat & Tech Univ, Shanghai, Peoples R China
[4] CAS &M Zhangjiagang New Energy Technol Co Ltd, Zhangjiagang, Peoples R China
基金
中国国家自然科学基金;
关键词
Direct methanol fuel cell (DMFC) system; Real-time power optimization; Methanol supply control; Reinforcement learning; Q -learning algorithm; MASS-TRANSPORT MODEL; NUMERICAL-MODEL; PERFORMANCE; DMFC;
D O I
10.1016/j.ijhydene.2024.09.084
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
Efficient real-time power optimization of direct methanol fuel cell (DMFC) system is crucial for enhancing its performance and reliability. The power of DMFC system is mainly affected by stack temperature and circulating methanol concentration. However, the methanol concentration cannot be directly measured using reliable sensors, which poses a challenge for the real-time power optimization. To address this issue, this paper investigates the operating mechanism of DMFC system and establishes a system power model. Based on the established model, reinforcement learning using Q-learning algorithm is proposed to control methanol supply to optimize DMFC system power under varying operating conditions. This algorithm is simple, easy to implement, and does not rely on methanol concentration measurements. To validate the effectiveness of the proposed algorithm, simulation comparisons between the proposed method and the traditional perturbation and observation (P&O) algorithm are implemented under different operating conditions. The results show that proposed power optimization based on Q-learning algorithm improves net power by 1% and eliminates the fluctuation of methanol supply caused by P&O. For practical implementation considerations and real-time requirements of the algorithm, hardware-in-the-loop (HIL) experiments are conducted. The experiment results demonstrate that the proposed methods optimize net power under different operating conditions. Additionally, in terms of model accuracy, the experimental results are well matched with the simulation. Moreover, under varying load condition, compared with P&O, proposed power optimization based on Q-learning algorithm reduces root mean square error (RMSE) from 7.271% to 2.996% and mean absolute error (MAE) from 5.036% to 0.331%.
引用
收藏
页码:1241 / 1253
页数:13
相关论文
共 50 条
  • [31] A Real-time Demand-side Management System Considering User Behavior Using Deep Q-Learning in Home Area Network
    Tai, Chia-Shing
    Hong, Jheng-Huang
    Fu, Li-Chen
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 4050 - 4055
  • [32] A closed-loop data-driven optimization framework for the unit commitment problem: A Q-learning approach under real-time operation
    Jimenez, Diego
    Angulo, Alejandro
    Street, Alexandre
    Mancilla-David, Fernando
    APPLIED ENERGY, 2023, 330
  • [33] Towards Real-Time Video Caching at Edge Servers: A Cost-Aware Deep Q-Learning Solution
    Cui, Laizhong
    Ni, Erchao
    Zhou, Yipeng
    Wang, Zhi
    Zhang, Lei
    Liu, Jiangchuan
    Xu, Yuedong
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 302 - 314
  • [34] An advanced real-time dispatching strategy for a distributed energy system based on the reinforcement learning algorithm
    Meng, Fanyi
    Bai, Yang
    Jin, Jingliang
    RENEWABLE ENERGY, 2021, 178 : 13 - 24
  • [35] Adaptive Power Optimization and Control for Automobile PEMFC Air Management System Based on Fuzzy Q-Learning and Prescribed Tracking Performance
    Wang, Yunlong
    Wang, Yongfu
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2025, 26 (02) : 2609 - 2620
  • [36] Optimum design parameters and operating condition for maximum power of a direct methanol fuel cell using analytical model and genetic algorithm
    Tafaoli-Masoule, M.
    Bahrami, A.
    Elsayed, E. M.
    ENERGY, 2014, 70 : 643 - 652
  • [37] Deep Reinforcement Learning Algorithm Based on Fusion Optimization for Fuel Cell Gas Supply System Control
    Yuan, Hongyan
    Sun, Zhendong
    Wang, Yujie
    Chen, Zonghai
    WORLD ELECTRIC VEHICLE JOURNAL, 2023, 14 (02):
  • [38] Improving hybrid tracked vehicles' fuel efficiency through the use of real-time based power distribution system
    Zhang, Yunhe
    Song, Dafeng
    Yang, Chunhua
    Zeng, Xiaohua
    Xu, Zehua
    JOURNAL OF POWER SOURCES, 2025, 641
  • [39] Adaptive Immune System reinforcement Learning-Based algorithm for real-time Cascading Failures prevention
    Babalola, Adeniyi Abdulrasheed
    Belkacemi, Rabie
    Zarrabian, Sina
    Craven, Robert
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2017, 57 : 118 - 133
  • [40] Dispatching Algorithm Design for Elevator Group Control System with Q-Learning based on a Recurrent Neural Network
    Liu, Weipeng
    Liu, Ning
    Sun, Hexu
    Xing, Guansheng
    Dong, Yan
    Chen, Haiyong
    2013 25TH CHINESE CONTROL AND DECISION CONFERENCE (CCDC), 2013, : 3397 - 3402