Self-adapting WIP parameter setting using deep reinforcement learning

•Using real-time information can significantly improve WIP parameter setting.•Frequently re-setting WIP limits leads to better performance than fixed WIP limits.•Deep Reinforcement Learning Agents are better at prescribing WIP limits than Statistical throughput control.•Deep Reinforcement Learning m...

Full description

Saved in:
Bibliographic Details
Published inComputers & operations research Vol. 144; p. 105854
Main Authors Tomé De Andrade e Silva, Manuel, Azevedo, Américo
Format Journal Article
LanguageEnglish
Published New York Elsevier Ltd 01.08.2022
Pergamon Press Inc
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:•Using real-time information can significantly improve WIP parameter setting.•Frequently re-setting WIP limits leads to better performance than fixed WIP limits.•Deep Reinforcement Learning Agents are better at prescribing WIP limits than Statistical throughput control.•Deep Reinforcement Learning models can be run in constant time at decision time. This study investigates the potential of dynamically adjusting WIP cap levels to maximize the throughput (TH) performance and minimize work in process (WIP), according to real-time system state arising from process variability associated with low volume and high-variety production systems. Using an innovative approach based on state-of-the-art deep reinforcement learning (proximal policy optimization algorithm), we attain WIP reductions of up to 50% and 30%, with practically no losses in throughput, against pure-push systems and the statistical throughput control method (STC), respectively. An exploratory study based on simulation experiments was performed to provide support to our research. The reinforcement learning agent’s performance was shown to be robust to variability changes within the production systems.
ISSN:0305-0548
1873-765X
0305-0548
DOI:10.1016/j.cor.2022.105854