DRL-dEWMA: a composite framework for run-to-run control in the semiconductor manufacturing process

被引:2
作者
Ma, Zhu [1 ]
Pan, Tianhong [2 ]
机构
[1] Anhui Univ, Sch Comp Sci & Technol, Sch Elect Engn & Automat, Hefei 230601, Peoples R China
[2] Anhui Univ, Sch Elect Engn & Automat, Hefei 230601, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Semiconductor manufacturing; Run-to-run control; Double exponentially weighted moving average; Parameter optimization;
D O I
10.1007/s00521-023-09112-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study aims to develop a weight-adjustment scheme for a double exponentially weighted moving average (dEWMA) controller using deep reinforcement learning (DRL) techniques. Under the run-to-run control framework, the weight adjustment of the dEWMA is formulated as a Markovian decision process in which the candidate weights are viewed as the DRL agent's decision action. Accordingly, a composite control strategy integrating DRL and dEWMA is proposed. Specifically, a well-trained DRL agent serves as an auxiliary controller that produces the preferred weights of the dEWMA. The optimized dEWMA serves as a master controller to provide a suitable recipe for the manufacturing process. Furthermore, two classical deterministic policy-gradient algorithms are leveraged for automatic weight tuning. The simulation results show that the proposed scheme outperforms existing RtR controllers in terms of disturbance rejection and target tracking. The proposed scheme has significant practical application prospects in smart semiconductor manufacturing.
引用
收藏
页码:1429 / 1447
页数:19
相关论文
共 37 条
[1]   Learning Anticipation via Spiking Networks: Application to Navigation Control [J].
Arena, Paolo ;
Fortuna, Luigi ;
Frasca, Mattia ;
Patane, Luca .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (02) :202-216
[2]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[3]   A general tool-based multi-product model for high-mixed production in semiconductor manufacturing [J].
Chen, Liang ;
Chu, Liuxing ;
Ge, Cuicui ;
Zhang, Yueyuan .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2023, 61 (23) :8062-8079
[4]  
Del Castillo E, 1999, IIE TRANS, V31, P1157
[5]   Reinforcement learning for industrial process control: A case study in flatness control in steel industry [J].
Deng, Jifei ;
Sierla, Seppo ;
Sun, Jie ;
Vyatkin, Valeriy .
COMPUTERS IN INDUSTRY, 2022, 143
[6]   A survey and comparative evaluation of actor-critic methods in process control [J].
Dutta, Debaprasad ;
Upreti, Simant R. .
CANADIAN JOURNAL OF CHEMICAL ENGINEERING, 2022, 100 (09) :2028-2056
[7]   A Review of Data Mining Applications in Semiconductor Manufacturing [J].
Espadinha-Cruz, Pedro ;
Godina, Radu ;
Rodrigues, Eduardo M. G. .
PROCESSES, 2021, 9 (02) :1-38
[8]   A New Double Exponentially Weighted Moving Average Run-to-Run Control Using a Disturbance-Accumulating Strategy for Mixed-Product Mode [J].
Fan, Shu-Kai S. ;
Jen, Chih-Hung ;
Hsu, Chia-Yu ;
Liao, Yu-Ling .
IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2021, 18 (04) :1846-1860
[9]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[10]   An Intelligent Non-Integer PID Controller-Based Deep Reinforcement Learning: Implementation and Experimental Results [J].
Gheisarnejad, Meysam ;
Khooban, Mohammad Hassan .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2021, 68 (04) :3609-3618