Abstract
An existing concept for sequence planning in production planning and control was extended by a global decision instance based on neural networks. Therefore, information regarding the state of the production and available orders were normalized and analyzed by one agent. In contrast to a partially observable Markow Decision Problem one single agent was allowed and used to process all available information. Feasibility and problems were examined and compared with a concept for decentralized decisions. The implementation consists of two parts, which continuously interact with each other. One part is a simulation of a job shop, including multiple machines. The other parts tackle the Markow Decision Problem with the use of double Q reinforcement learning in order to estimate the best sequence at any given time. Later, problems due to scaling and comparisons to the usage of multiple agents are given.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Claus, T., Herrmann, F., Manitz, M. (eds.): Produktionsplanung und –steuerung. Springer, Heidelberg (2015). https://doi.org/10.1007/978-3-662-43542-7
Jaehn, F., Pesch, E.: Ablaufplanung. Einführung in Scheduling, Springer, Heidelberg (2014). https://doi.org/10.1007/978-3-642-54439-2
Wuendahl, H.-P.: Betriebsorganisation für Ingenieure. Hanser, München (2010)
Swamidass, P.M.: Encyclopedia of Production and Manufacturing Management. Kluwer Academic Publishers, Boston (2000)
Lödding, H.: Verfahren der Fertigungssteuerung. Springer, Heidelberg (2016). https://doi.org/10.1007/978-3-662-48459-3
LNCS Homepage. http://www.springer.com/lncs. Accessed 21 Nov 2016
Lillicrap, T.: Continuous control with deep reinforcement learning, on ICLR (2016)
Thomas, A.: https://adventuresinmachinelearning.com/double-q-reinforcement-learning-in-tensorflow-2/. Accessed 29 Mar 2020
Zhou, H.: Concept for optimization of scheduling in production planning and control using machine learning, TU Darmstadt (2019)
Huang, D.-S., Li, K., Irwin, G.W. (eds.): ICIC 2006. LNCS (LNAI), vol. 4114. Springer, Heidelberg (2006). https://doi.org/10.1007/11816171
Sutton, R.S., Barto, A.G.: Reinforcement Learning. MIT Press, Cambridge (2018)
Rosenblatt, F.: The perceptron: a probabilistic model for information storage and organization in the brain. Psychol. Rev. 65(6), 386–408 (1958)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 IFIP International Federation for Information Processing
About this paper
Cite this paper
Dasbach, T., Olbort, J., Wenk, F., Ander, R. (2022). Sequencing Through a Global Decision Instance Based on a Neural Network. In: Canciglieri Junior, O., Noël, F., Rivest, L., Bouras, A. (eds) Product Lifecycle Management. Green and Blue Technologies to Support Smart and Sustainable Organizations. PLM 2021. IFIP Advances in Information and Communication Technology, vol 639. Springer, Cham. https://doi.org/10.1007/978-3-030-94335-6_24
Download citation
DOI: https://doi.org/10.1007/978-3-030-94335-6_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-94334-9
Online ISBN: 978-3-030-94335-6
eBook Packages: Computer ScienceComputer Science (R0)