Strategic Workforce Planning with Deep Reinforcement Learning

被引:0
作者
Smit, Yannick [1 ]
Den Hengst, Floris [2 ,3 ]
Bhulai, Sandjai [2 ]
Mehdad, Ehsan [3 ]
机构
[1] Univ Amsterdam, Amsterdam, Netherlands
[2] Vrije Univ Amsterdam, Amsterdam, Netherlands
[3] ING Bank NV, Amsterdam, Netherlands
来源
MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2022, PT II | 2023年 / 13811卷
关键词
Deep reinforcement learning; Optimization; Simulation; Strategic workforce planning; MODEL;
D O I
10.1007/978-3-031-25891-6_9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a simulation-optimization approach to strategic workforce planning based on deep reinforcement learning. A domain expert expresses the organization's high-level, strategic workforce goals over the workforce composition. A policy that optimizes these goals is then learned in a simulation-optimization loop. Any suitable simulator can be used, and we describe how a simulator can be derived from historical data. The optimizer is driven by deep reinforcement learning and directly optimizes for the high-level strategic goals as a result. We compare the proposed approach with a linear programming-based approach on two types of workforce goals. The first type of goal, consisting of a target workforce, is relatively easy to optimize for but hard to specify in practice and is called operational in this work. The second, strategic, type of goal is a possibly non-linear combination of high-level workforce metrics. These goals can easily be specified by domain experts but may be hard to optimize for with existing approaches. The proposed approach performs significantly better on the strategic goal while performing comparably on the operational goal for both a synthetic and a real-world organization. Our novel approach based on deep reinforcement learning and simulation-optimization has a large potential for impact in the workforce planning domain. It directly optimizes for an organization's workforce goals that may be non-linear in the workforce composition and composed of arbitrary workforce composition metrics.
引用
收藏
页码:108 / 122
页数:15
相关论文
共 20 条
[1]  
April J., 2013, ENSURING WORKF UNPUB
[2]   Markov-Chain Simulation-Based Analysis of Human Resource Structure: How Staff Deployment and Staffing Affect Sustainable Human Resource Strategy [J].
Banyai, Tamas ;
Landschuetzer, Christian ;
Banyai, Agota .
SUSTAINABILITY, 2018, 10 (10)
[3]   Simple methods for shift scheduling in multiskill call centers [J].
Bhulai, Sandjai ;
Koole, Ger ;
Pot, Auke .
M&SOM-MANUFACTURING & SERVICE OPERATIONS MANAGEMENT, 2008, 10 (03) :411-420
[4]   The state of the art of nurse rostering [J].
Burke, EK ;
De Causmaecker, P ;
Vanden Berghe, G ;
Van Landeghem, H .
JOURNAL OF SCHEDULING, 2004, 7 (06) :441-499
[5]  
Cotten A., 2007, 7 STEPS EFF WORKF PL
[6]   Stochastic Optimization Models for Workforce Planning, Operations, and Risk Management [J].
Davis, Michael J. ;
Lu, Yingdong ;
Sharma, Mayank ;
Squillante, Mark S. ;
Zhang, Bo .
SERVICE SCIENCE, 2018, 10 (01) :40-57
[7]   Optimizing cost-effectiveness in a stochastic Markov manpower planning system under control by recruitment [J].
De Feyter, Tim ;
Guerry, Marie-Anne ;
Komarudin .
ANNALS OF OPERATIONS RESEARCH, 2017, 253 (01) :117-131
[8]   Planning for potential: efficient safe reinforcement learning [J].
den Hengst, Floris ;
Francois-Lavet, Vincent ;
Hoogendoorn, Mark ;
van Harmelen, Frank .
MACHINE LEARNING, 2022, 111 (06) :2255-2274
[9]   A DISTRIBUTED PARAMETER COHORT PERSONNEL PLANNING-MODEL THAT USES CROSS-SECTIONAL DATA [J].
GAIMON, C ;
THOMPSON, GL .
MANAGEMENT SCIENCE, 1984, 30 (06) :750-764
[10]   OPTIMAL CONTROL OF A GRADED MANPOWER SYSTEM [J].
GRINOLD, RC ;
STANFORD, RE .
MANAGEMENT SCIENCE SERIES B-APPLICATION, 1974, 20 (08) :1201-1216