Production Scheduling Using Deep Reinforcement Learning and Discrete Event Simulation |
| |
Authors: | Dr-Ing Stefan Hubert Jonas Meintschel Dominik Bleidorn Yak Ortmanns Roderich Wallrath |
| |
Affiliation: | 1. Bayer AG, Engineering & Technology, Kaiser-Wilhelm-Allee 1, 51373 Leverkusen, Germany;2. Technische Universität Dresden (TUD), Group Process Systems Engineering (SVT), Helmholtzstraße 14 (MER/12A), 01062 Dresden, Germany;3. INOSIM Software GmbH, Joseph-von-Fraunhofer-Straße 20, 44227 Dortmund, Germany |
| |
Abstract: | Scheduling in the process industry is a highly demanding task. Having access to optimal production schedules at short notice, for instance, after spontaneous changes, offers numerous advantages in terms of robustness, economics, and ultimately customer satisfaction, as delays are minimized. In this work, we describe our initial efforts to apply and evaluate deep reinforcement learning (DRL) for optimized scheduling in a typical fill-and-finish batch production plant in the chemical industry. Our pilot study demonstrates how DRL can be implemented using an approach based on discrete event simulation. We discuss the results and benefits of DRL, compare it to mathematical programming approaches, and outline a potential path forward. Our study suggests that the application of DRL in the chemical industry is a promising research direction and that DRL can complement established methods such as process simulation and mathematical programming. |
| |
Keywords: | Discrete event simulation Reinforcement learning Scheduling |
|
|