Self-learning Processes in Smart Factories: Deep Reinforcement Learning for Process Control of Robot Brine Injection

被引:16
作者
Andersen, Rasmus E. [1 ]
Madsen, Steffen [1 ]
Barlo, Alexander B. K. [1 ]
Johansen, Sebastian B. [1 ]
Nor, Morten [1 ]
Andersen, Rasmus S. [1 ,2 ]
Bogh, Simon [1 ,2 ]
机构
[1] Aalborg Univ, Dept Mat & Prod, Fibigerstraede 16, DK-9220 Aalborg, Denmark
[2] Aalborg Univ, Dept Mat & Prod, Robot & Automat Grp, Fibigerstr 16, DK-9220 Aalborg, Denmark
来源
29TH INTERNATIONAL CONFERENCE ON FLEXIBLE AUTOMATION AND INTELLIGENT MANUFACTURING (FAIM 2019): BEYOND INDUSTRY 4.0: INDUSTRIAL ADVANCES, ENGINEERING EDUCATION AND INTELLIGENT MANUFACTURING | 2019年 / 38卷
关键词
Self-learning Smart Factories; Deep Reinforcement Learning; Process Control;
D O I
10.1016/j.promfg.2020.01.023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The goal of this paper is to investigate the application of adaptive learning algorithms, which enables industrial robots to cope with natural variations exhibited in a brine injection process related to the production of bacon. Due to the variations in bacon meat, the traditional needle-based brine injection process is not capable of injecting the correct amount of brine, leading to either ruined or unflavored bacon. In the presented work a Deep Deterministic Policy Gradient (DDPG) reinforcement learning algorithm is introduced in the injection process to improve process control. To accelerate training of the reinforcement learning algorithm, a simulation environment of the brine absorption is generated based on 64 conducted experiments. The simulation environment estimates the amount of absorbed brine given injection pressure and injection time. Tests are run in the simulation where the starting mass is generated from a normal distribution with mean 80.5g, and a standard deviation of 4.8 g and 20.0 g respectively. With a target of 15 % mass increase, the agent can produce an average mass increase of 14.9 % for the first test and 14.6 % for the second test. This indicates that the model can successfully adapt to a high variety input, thereby showing potential for process control in brine injection, coping with natural variation in meat structure. (C) 2019 The Authors. Published by Elsevier B.V.
引用
收藏
页码:171 / 177
页数:7
相关论文
共 11 条
[1]   Reinforcement learning in feedback control Challenges and benchmarks from technical process control [J].
Hafner, Roland ;
Riedmiller, Martin .
MACHINE LEARNING, 2011, 84 (1-2) :137-169
[2]   Preparing ballistic gelatine - review and proposal for a standard method [J].
Jussila, J .
FORENSIC SCIENCE INTERNATIONAL, 2004, 141 (2-3) :91-98
[3]  
Koprinkova-Hristova P., 2014, Information Technologies and Control, V11, P21, DOI 10.2478/itc-2013-0004
[4]  
Lillicrap T. P., 2016, CoRR, abs/1509.02971, P1
[5]   Motion Planning for Industrial Robots using Reinforcement Learning [J].
Meyes, Richard ;
Tercan, Hasan ;
Roggendorf, Simon ;
Thiele, Thomas ;
Buescher, Christian ;
Obdenbusch, Markus ;
Brecher, Christian ;
Jeschke, Sabina ;
Meisen, Tobias .
MANUFACTURING SYSTEMS 4.0, 2017, 63 :107-112
[6]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[7]  
Mnih Volodymyr, 2013, CORR
[8]   Policy gradient methods for robotics [J].
Peters, Jan ;
Schaal, Stefan .
2006 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-12, 2006, :2219-2225
[9]  
Philipsen M., 2017, 6 AALBORG ROBOTICS W, P16
[10]  
Sutton R. S., 2018, INTRO REINFORCEMENT, V2nd