12,206
Views
20
CrossRef citations to date
0
Altmetric
Reviews

Reinforcement learning applied to production planning and control

ORCID Icon, ORCID Icon, ORCID Icon & ORCID Icon
Pages 5772-5789 | Received 29 Jul 2021, Accepted 11 Jul 2022, Published online: 06 Aug 2022

References

  • Abadi, Martín, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S. Corrado, et al. 2016. “TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems.” https://www.tensorflow.org/.
  • Alves, Júlio César, and Geraldo Robson Mateus. 2020. “Deep Reinforcement Learning and Optimization Approach for Multi-Echelon Supply Chain with Uncertain Demands.” Lecture Notes in Computer Science, 584–599. doi:10.1007/978-3-030-59747-4_38.
  • Barat, Souvik, Harshad Khadilkar, Hardik Meisheri, Vinay Kulkarni, Vinita Baniwal, Prashant Kumar, and Monika Gajrani. 2019. “Actor Based Simulation for Closed Loop Control of Supply Chain Using Reinforcement Learning.” Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 3, 1802–1804.
  • Bellman, Richard. 1957. Dynamic Programming. Princeton, New Jersey: Princeton University Press.
  • Briegel, Hans J, and Gemma las Cuevas. 2012. “Projective Simulation for Artificial Intelligence.” Scientific Reports 2 (1): 1–16.
  • Brockman, Greg, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. 2016. “OpenAI Gym.” http://arxiv.org/abs/1606.01540.
  • Bueno, Adauto, Moacir Godinho Filho, and Alejandro G. Frank. 2020. “Smart Production Planning and Control in the Industry 4.0 Context: A Systematic Literature Review.” Computers & Industrial Engineering 149 (November): 106774. doi:10.1016/j.cie.2020.106774.
  • Cañas, Héctor, Josefa Mula, and Francisco Campuzano-Bolarín. 2020. “A General Outline of a Sustainable Supply Chain 4.0.” Sustainability 12 (19): 7978. doi:10.3390/su12197978.
  • Cañas, Héctor, Josefa Mula, Manuel Díaz-Madroñero, and Francisco Campuzano-Bolarín. 2021. “Implementing Industry 4.0 Principles.” Computers & Industrial Engineering 158 (August): 107379. doi:10.1016/j.cie.2021.107379.
  • Canese, Lorenzo, Gian Carlo Cardarilli, Luca Di Nunzio, Rocco Fazzolari, Daniele Giardino, Marco Re, and Sergio Spanò. 2021. “Multi-Agent Reinforcement Learning: A Review of Challenges and Applications.” Applied Sciences 11 (11): 4948.
  • Caspi, Itai, Gal Leibovich, Gal Novik, and Shadi Endrawis. 2017. “Reinforcement Learning Coach.” doi:10.5281/zenodo.1134899.
  • Castro, Pablo Samuel, Subhodeep Moitra, Carles Gelada, Saurabh Kumar, and Marc G Bellemare. 2018. “Dopamine: A Research Framework for Deep Reinforcement Learning.” http://arxiv.org/abs/1812.06110.
  • Chien, Chen Fu, Yun Siang Lin, and Sheng Kai Lin. 2020. “Deep Reinforcement Learning for Selecting Demand Forecast Models to Empower Industry 3.5 and an Empirical Study for a Semiconductor Component Distributor.” International Journal of Production Research 58 (9): 2784–2804. doi:10.1080/00207543.2020.1733125.
  • Cunha, Bruno, Ana M. Madureira, Benjamim Fonseca, and Duarte Coelho. 2020. “Deep Reinforcement Learning as a Job Shop Scheduling Solver: A Literature Review.” Intelligent Decision Support Systems – A Journey to Smarter Healthcare, 350–359. doi:10.1007/978-3-030-14347-3_34.
  • Deisenroth, Marc Peter, Gerhard Neumann, and Jan Peters. 2013. A Survey on Policy Search for Robotics. Now Publishers.
  • Elkan, Charles. 2012. “Reinforcement Learning with a Bilinear Q Function.” Lecture Notes in Computer Science, 78–88. doi:10.1007/978-3-642-29946-9_11.
  • Farazi, Nahid Parvez, Tanvir Ahamed, Limon Barua, and Bo Zou. 2020. “Deep Reinforcement Learning and Transportation Research: A Comprehensive Review.” ArXiv Preprint ArXiv:2010.06187.
  • Gabel, Thomas, and Martin Riedmiller. 2012. “Distributed Policy Search Reinforcement Learning for Job-Shop Scheduling Tasks.” International Journal of Production Research 50 (1): 41–61. doi:10.1080/00207543.2011.571443.
  • Garnier, Paul, Jonathan Viquerat, Jean Rabault, Aurélien Larcher, Alexander Kuhnle, and Elie Hachem. 2021. “A Review on Deep Reinforcement Learning for Fluid Mechanics.” Computers & Fluids 225: 104973. doi:10.1016/j.compfluid.2021.104973.
  • Gronauer, Sven, and Klaus Diepold. 2021. “Multi-Agent Deep Reinforcement Learning: A Survey.” Artificial Intelligence Review 55 (2): 895–943.
  • Grondman, Ivo, Lucian Busoniu, Gabriel A D Lopes, and Robert Babuska. 2012. “A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients.” IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 42 (6): 1291–1307.
  • Guadarrama, Sergio, Anoop Korattikara, Oscar Ramirez, Pablo Castro, Ethan Holly, Sam Fishman, Ke Wang, et al. 2018. “TF-Agents: A Library for Reinforcement Learning in TensorFlow.” https://github.com/tensorflow/agents.
  • Haarnoja, Tuomas, Aurick Zhou, Pieter Abbeel, and Sergey Levine. 2018. “Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor.” ArXiv Preprint ArXiv:1801.01290.
  • Han, Miyoung. 2018. “Reinforcement Learning Approaches in Dynamic Environments.”
  • Hessel, Matteo, Joseph Modayil, Hado Van Hasselt, Tom Schaul, Georg Ostrovski, Will Dabney, Dan Horgan, Bilal Piot, Mohammad Azar, and David Silver. 2018. “Rainbow: Combining Improvements in Deep Reinforcement Learning.” In Proceedings of the AAAI Conference on Artificial Intelligence. 32.
  • Hill, Ashley, Antonin Raffin, Maximilian Ernestus, Adam Gleave, Anssi Kanervisto, Rene Traore, Prafulla Dhariwal, et al. 2018. “Stable Baselines.” GitHub Repository. GitHub.
  • Hoffman, Matt, Bobak Shahriari, John Aslanides, Gabriel Barth-Maron, Feryal Behbahani, Tamara Norman, Abbas Abdolmaleki, et al. 2020. “Acme: A Research Framework for Distributed Reinforcement Learning.” https://arxiv.org/abs/2006.00979.
  • Huang, Jing, Qing Chang, and Nilanjan Chakraborty. 2019. “Machine Preventive Replacement Policy for Serial Production Lines Based on Reinforcement Learning.” In 2019 IEEE 15th International Conference on Automation Science and Engineering (CASE), 523–528. IEEE. doi:10.1109/COASE.2019.8843338.
  • Hubbs, Christian D., Can Li, Nikolaos V. Sahinidis, Ignacio E. Grossmann, and John M. Wassick. 2020a. “A Deep Reinforcement Learning Approach for Chemical Production Scheduling.” Computers and Chemical Engineering 141: 106982. doi:10.1016/j.compchemeng.2020.106982.
  • Hubbs, Christian D, Hector D Perez, Owais Sarwar, Nikolaos V Sahinidis, Ignacio E Grossmann, and John M Wassick. 2020b. “OR-Gym: A Reinforcement Learning Library for Operations Research Problems.”
  • Ivanov, Dmitry, Christopher S. Tang, Alexandre Dolgui, Daria Battini, and Ajay Das. 2021. “Researchers’ Perspectives on Industry 4.0: Multi-Disciplinary Analysis and Opportunities for Operations Management.” International Journal of Production Research 59 (7): 2055–2078. doi:10.1080/00207543.2020.1798035.
  • Jeon, Su Min, and Gitae Kim. 2016. “A Survey of Simulation Modeling Techniques in Production Planning and Control (PPC).” Production Planning & Control 27 (5): 360–377. doi:10.1080/09537287.2015.1128010.
  • Jiang, Chengzhi, and Zhaohan Sheng. 2009. “Case-Based Reinforcement Learning for Dynamic Inventory Control in a Multi-Agent Supply-Chain System.” Expert Systems with Applications 36: 6520–6526. doi:10.1016/j.eswa.2008.07.036.
  • Kapturowski, Steven, Georg Ostrovski, Will Dabney, John Quan, and Remi Munos. 2019. “Recurrent Experience Replay in Distributed Reinforcement Learning.” In International Conference on Learning Representations. https://openreview.net/forum?id=r1lyTjAqYX.
  • Karimi-Majd, Amir Mohsen, Masoud Mahootchi, and Amir Zakery. 2017. “A Reinforcement Learning Methodology for a Human Resource Planning Problem Considering Knowledge-Based Promotion.” Simulation Modelling Practice and Theory 79: 87–99. doi:10.1016/j.simpat.2015.07.004.
  • Kayhan, Behice Meltem, and Gokalp Yildiz. 2021. “Reinforcement Learning Applications to Machine Scheduling Problems: A Comprehensive Literature Review.” Journal of Intelligent Manufacturing, doi:10.1007/s10845-021-01847-3.
  • Kim, Byeongseop, Yongkuk Jeong, and Jong Gye Shin. 2020. “Spatial Arrangement Using Deep Reinforcement Learning to Minimise Rearrangement in Ship Block Stockyards.” International Journal of Production Research 58 (16): 5062–5076. doi:10.1080/00207543.2020.1748247.
  • Konda, Vijay R, and John N Tsitsiklis. 2000. “Actor-Critic Algorithms.” Advances in Neural Information Processing Systems 12: 1008–1014.
  • Kuhnle, Andreas, Johannes Jakubik, and Gisela Lanza. 2019. “Reinforcement learning for opportunistic maintenance optimization.” Production Engineering 13 (1): 33–41. https://doi.org/10.1007/s11740-018-0855-7.
  • Kuhnle, Andreas, Jan-Philipp Kaiser, Felix Theiß, Nicole Stricker, and Gisela Lanza. 2021. “Designing an adaptive production control system using reinforcement learning.” Journal of Intelligent Manufacturing 32 (3): 855–876. https://doi.org/10.1007/s10845-020-01612-y.
  • Kuhnle, Andreas, Marvin Carl May, Louis Schäfer, and Gisela Lanza. 2021. “Explainable reinforcement learning in production control of job shop manufacturing system.” International Journal of Production Research: 1–23. https://doi.org/10.1080/00207543.2021.1972179.
  • Kuhnle, Alexander, Michael Schaarschmidt, and Kai Fricke. 2017. “Tensorforce: A TensorFlow Library for Applied Reinforcement Learning.” https://github.com/tensorforce/tensorforce.
  • Lambert, Douglas M, and Martha C Cooper. 2000. “Issues in Supply Chain Management.” Industrial Marketing Management 29 (1): 65–83. doi:10.1016/S0019-8501(99)00113-3.
  • Lang, Sebastian, Fabian Behrendt, Nico Lanzerath, Tobias Reggelin, and Marcel Muller. 2020. “Integration of Deep Reinforcement Learning and Discrete-Event Simulation for Real-Time Scheduling of a Flexible Job Shop Production.” In 2020 Winter Simulation Conference (WSC), 3057–3068. IEEE. doi:10.1109/WSC48552.2020.9383997.
  • Li, Yuxi. 2018. “Deep Reinforcement Learning.” ArXiv Preprint ArXiv:1810.06339.
  • Liang, Eric, Richard Liaw, Robert Nishihara, Philipp Moritz, Roy Fox, Joseph Gonzalez, Ken Goldberg, and I. Stoica. 2017. “Ray RLLib: A Composable and Scalable Reinforcement Learning Library.” ArXiv abs/1712.0.
  • Lillicrap, Timothy P, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. 2015. “Continuous Control with Deep Reinforcement Learning.” ArXiv Preprint ArXiv:1509.02971.
  • Mehra, A. 1995. Hierarchical Production Planning for Job Shops. College Park: University of Maryland, Harvard University and Industry.
  • Mnih, Volodymyr, Adria Puigdomenech Badia, Mehdi Mirza, Alex Graves, Timothy Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. 2016. “Asynchronous Methods for Deep Reinforcement Learning.” In International Conference on Machine Learning, 1928–1937.
  • Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare, Alex Graves, et al. 2015. “Human-Level Control Through Deep Reinforcement Learning.” Nature 518 (7540): 529–533.
  • Mula, Josefa, and Marija Bogataj. 2021. “OR in the Industrial Engineering of Industry 4.0: Experiences from the Iberian Peninsula Mirrored in CJOR.” Central European Journal of Operations Research. doi:10.1007/s10100-021-00740-x.
  • Panzer, Marcel, and Benedict Bender. 2022. “Deep reinforcement learning in production systems: a systematic literature review.” International Journal of Production Research 60 (13): 4316–4341. https://doi.org/10.1080/00207543.2021.1973138.
  • Park, Junyoung, Jaehyeong Chun, Sang Hun Kim, Youngkook Kim, and Jinkyoo Park. 2021. “Learning to Schedule Job-Shop Problems: Representation and Policy Learning Using Graph Neural Network and Reinforcement Learning.” International Journal of Production Research 59 (11): 3360–3377. doi:10.1080/00207543.2020.1870013.
  • Park, In Beom, Jaeseok Huh, Joongkyun Kim, and Jonghun Park. 2020. “A Reinforcement Learning Approach to Robust Scheduling of Semiconductor Manufacturing Facilities.” IEEE Transactions on Automation Science and Engineering 17 (3): 1420–1431. doi:10.1109/TASE.2019.2956762.
  • Paszke, Adam, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. 2017. “Automatic Differentiation in PyTorch.” In 31st Conference on Neural Information Processing Systems (NIPS 2017). Long Beach, USA.
  • Paszke, Adam, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, et al. 2019. “PyTorch: An Imperative Style, High-Performance Deep Learning Library.” In Advances in Neural Information Processing Systems 32, edited by H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. Fox, and R. Garnett, 8024–8035. Curran Associates. http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf.
  • Plappert, Matthias. 2016. “Keras-RL.” GitHub Repository. GitHub.
  • Pontrandolfo, P., A. Gosavi, O. G. Okogbaa, and T. K. Das. 2002. “Global Supply Chain Management: A Reinforcement Learning Approach.” International Journal of Production Research 40 (6): 1299–1317. doi:10.1080/00207540110118640.
  • Qiao, B., and J. Zhu. 2000. Agent-Based Intelligent Manufacturing System for the 21st Century. Nanjing: Mechatronic Engineering Institute, Nanjing University of Aeronautics and Astronautics.
  • Qu, Shuhui, Jie Wang, and Juergen Jasperneite. 2018. “Dynamic Scheduling in Large-Scale Stochastic Processing Networks for Demand-Driven Manufacturing Using Distributed Reinforcement Learning.” IEEE International Conference on Emerging Technologies and Factory Automation, ETFA 2018-Sept (1). IEEE, 433–440. doi:10.1109/ETFA.2018.8502508.
  • Rabe, Markus, and Felix Dross. 2015. “A Reinforcement Learning Approach for a Decision Support System for Logistics Networks.” In 2015 Winter Simulation Conference (WSC), 2020–2032. IEEE. doi:10.1109/WSC.2015.7408317.
  • Raffin, Antonin, Ashley Hill, Adam Gleave, Anssi Kanervisto, Maximilian Ernestus, and Noah Dormann. 2021. “Stable-Baselines3: Reliable Reinforcement Learning Implementations.” Journal of Machine Learning Research 22 (268): 1–8.
  • Rummery, Gavin A, and Mahesan Niranjan. 1994. On-Line Q-Learning Using Connectionist Systems. Cambridge: University of Cambridge, Department of Engineering Cambridge.
  • Rummukainen, Hannu, and Jukka K. Nurminen. 2019. “Practical Reinforcement Learning – Experiences in Lot Scheduling Application.” IFAC-PapersOnLine 52 (13): 1415–1420. doi:10.1016/j.ifacol.2019.11.397.
  • Russell, Stuart J, and Peter Norvig. 2003. Artificial Intelligence: A Modern Approach. New Jersey: Pearson Education.
  • Schaul, Tom, John Quan, Ioannis Antonoglou, and David Silver. 2016. “Prioritized Experience Replay.” ArXiv Preprint ArXiv:1511.05952.
  • Schneckenreither, Manuel, and Stefan Haeussler. 2019. “Reinforcement Learning Methods for Operations Research Applications: The Order Release Problem.” Lecture Notes in Computer Science 11331: 545–559. doi:10.1007/978-3-030-13709-0_46.
  • Schulman, John, Sergey Levine, Pieter Abbeel, Michael Jordan, and Philipp Moritz. 2015. “Trust Region Policy Optimization.” In International Conference on Machine Learning, 1889–1897.
  • Schulman, John, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. “Proximal Policy Optimization Algorithms.” ArXiv Preprint ArXiv:1707.06347.
  • Serrano-Ruiz, J. C., J. Mula, D. Peidro, and M. Díaz-Madroñero. 2021. “A Metamodel for the Supply Chain 4.0.” Journal of Industrial Integration Information. Under Review.
  • Shiue, Yeou Ren, Ken Chuan Lee, and Chao Ton Su. 2018. “Real-Time Scheduling for a Smart Factory Using a Reinforcement Learning Approach.” Computers and Industrial Engineering 125 (101): 604–614. doi:10.1016/j.cie.2018.03.039.
  • Stone, Peter, and Manuela Veloso. 2000. “Multiagent Systems: A Survey from a Machine Learning Perspective.” Autonomous Robots 8 (3): 345–383.
  • Sutton, Richard S, and Andrew G Barto. 2018. Reinforcement Learning: An Introduction. Cambridge: MIT Press.
  • Sze, Vivienne, Yu-Hsin Chen, Tien-Ju Yang, and Joel S Emer. 2017. “Efficient Processing of Deep Neural Networks: A Tutorial and Survey.” Proceedings of the IEEE 105 (12): 2295–2329.
  • Szepesvári, Csaba. 2010. “Algorithms for Reinforcement Learning.” Synthesis Lectures on Artificial Intelligence and Machine Learning 4 (1): 1–103. doi:10.2200/S00268ED1V01Y201005AIM009.
  • Torres, Jordi. 2020. “Deep Reinforcement Learning Explained.” https://torres.ai/deep-reinforcement-learning-explained- series/.
  • Tuncel, Emre, Abe Zeid, and Sagar Kamarthi. 2014. “Solving Large Scale Disassembly Line Balancing Problem with Uncertainty Using Reinforcement Learning.” Journal of Intelligent Manufacturing 25 (4): 647–659. doi:10.1007/s10845-012-0711-0.
  • Usuga, Cadavid, Juan Pablo, Samir Lamouri, Bernard Grabot, Robert Pellerin, and Arnaud Fortin. 2020. “Machine Learning Applied in Production Planning and Control: A State-of-the-Art in the Era of Industry 4.0.” Journal of Intelligent Manufacturing 31 (6): 1531–1558. doi:10.1007/s10845-019-01531-7.
  • Valluri, Annapurna, Michael J. North, and Charles M. MacAl. 2009. “Reinforcement Learning in Supply Chains.” International Journal of Neural Systems 19 (5): 331–344. doi:10.1142/S0129065709002063.
  • Van Hasselt, Hado, Arthur Guez, and David Silver. 2016. “Deep Reinforcement Learning with Double Q-Learning.” In Proceedings of the AAAI Conference on Artificial Intelligence. 30.
  • Vanvuchelen, Nathalie, Joren Gijsbrechts, and Robert Boute. 2020. “Use of Proximal Policy Optimization for the Joint Replenishment Problem.” Computers in Industry 119: 103239. doi:10.1016/j.compind.2020.103239.
  • Vasilev, Ivan, Daniel Slater, Gianmario Spacagna, Peter Roelants, and Valentino Zocca. 2019. Python Deep Learning: Exploring Deep Learning Techniques and Neural Network Architectures with Pytorch, Keras, and TensorFlow. Birmingham: Packt Publishing Ltd.
  • Vollmann, T. E., W. L. Berry, D. C. Whybark, and F. R. Jacobs. 2005. Manufacturing Planning and Control for Supply Chain Management. New York: McGraw Hill.
  • Wan, Xing, Xingquan Zuo, Xiaodong Li, and Xinchao Zhao. 2020. “A Hybrid Multiobjective GRASP for a Multi-Row Facility Layout Problem with Extra Clearances.” International Journal of Production Research 60 (3): 1–20. doi:10.1080/00207543.2020.1847342.
  • Wang, Ziyu, Tom Schaul, Matteo Hessel, Hado Hasselt, Marc Lanctot, and Nando Freitas. 2016. “Dueling Network Architectures for Deep Reinforcement Learning.” In International Conference on Machine Learning, 1995–2003.
  • Watkins, Christopher J C H, and Peter Dayan. 1992. “Q-Learning.” Machine Learning 8 (3–4): 279–292.
  • Weiß, Gerhard. 1995. “Distributed Reinforcement Learning.” In The Biology and Technology of Intelligent Autonomous Agents, edited by L. Steels, 415–428. Berlin, Heidelberg: Springer.
  • Weiss, Gerhard. 1999. Multiagent Systems: A Modern Approach to Distributed Artificial Intelligence. Cambridge: MIT Press.
  • Wijmans, Erik, Abhishek Kadian, Ari Morcos, Stefan Lee, Irfan Essa, Devi Parikh, Manolis Savva, and Dhruv Batra. 2020. “DD-PPO: Learning Near-Perfect PointGoal Navigators from 2.5 Billion Frames.”
  • Williams, Ronald J. 1992. “Simple Statistical Gradient- Following Algorithms for Connectionist Reinforcement Learning.” Machine Learning 8 (3–4): 229–256.
  • Winder, Phil. 2020. Reinforcement Learning: Industrial Applications of Intelligent Agents.
  • Yang, Shengluo, and Zhigang Xu. 2021. “Intelligent scheduling and reconfiguration via deep reinforcement learning in smart manufacturing.” International Journal of Production Research 89: 1–18. https://doi.org/10.1080/00207543.2021.1943037.
  • Yu, Chao, Jiming Liu, and Shamim Nemati. 2020. “Reinforcement Learning in Healthcare: A Survey.” ArXiv Preprint ArXiv:1908.08796.
  • Zhang, Cong, Wen Song, Zhiguang Cao, Jie Zhang, Puay Siew Tan, and Chi Xu. 2020. “Learning to Dispatch for Job Shop Scheduling via Deep Reinforcement Learning,” http://arxiv.org/abs/2010.12367.
  • Zheng, Shuai, Chetan Gupta, and Susumu Serita. 2020. “Manufacturing Dispatching Using Reinforcement and Transfer Learning.” The European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, October. http://arxiv.org/abs/1910.02035.
  • Zheng, Wei, Yong Lei, and Qing Chang. 2017a. “Comparison Study of Two Reinforcement Learning Based Real-Time Control Policies for Two-Machine-One-Buffer Production System.” IEEE International Conference on Automation Science and Engineering, 1163–1168. doi:10.1109/COASE.2017.8256260.
  • Zheng, Wei, Yong Lei, and Qing Chang. 2017b. “Reinforcement Learning Based Real-Time Control Policy for Two-Machine-One-Buffer Production System.” In Volume 3: Manufacturing Equipment and Systems. American Society of Mechanical Engineers. doi:10.1115/MSEC2017-2771.