References
- Bengio, Y. 2009. Learning Deep Architectures for AI. Now Publishers.
- Berkey, J. O., and P. Y. Wang. 1987. “Two-dimensional Finite Bin-packing Algorithms.” Journal of the Operational Research Society 38 (5): 423–429. doi:https://doi.org/10.1057/jors.1987.70.
- Bradtke, S. J., and M. O. Duff. 1995. “Reinforcement Learning Methods for Continuous-time Markov Decision Problems.” In Advances in Neural Information Processing Systems, 393–400.
- Burke, E. K., M. Gendreau, M. Hyde, G. Kendall, G. Ochoa, E. Özcan, and R. Qu. 2013. “Hyper-heuristics: A Survey of the State of the Art.” Journal of the Operational Research Society 64 (12): 1695–1724. doi:https://doi.org/10.1057/jors.2013.71.
- Chen, F., P. Deng, J. Wan, D. Zhang, A. V. Vasilakos, and X. Rong. 2015. “Data Mining for the Internet of Things: Literature Review and Challenges.” International Journal of Distributed Sensor Networks 11 (8): 431047. doi:https://doi.org/10.1155/2015/431047.
- Diuk, C., A. Cohen, and M. L. Littman (2008, July). An Object-oriented Representation for Efficient Reinforcement Learning. In Proceedings of the 25th international conference on Machine learning, 240–247
- Holloway, C. A., and R. T. Nelson. 1974. “Job Shop Scheduling with Due Dates and Variable Processing Times.” Management Science 20 (9): 1264–1275. doi:https://doi.org/10.1287/mnsc.20.9.1264.
- Hovestadt, M., O. Kao, A. Keller, and A. Streit. June 2003. “Scheduling in HPC Resource Management Systems: Queuing Vs. Planning.” In Workshop on Job Scheduling Strategies for Parallel Processing, 1–20. Berlin, Heidelberg: Springer.
- Krizhevsky, A., I. Sutskever, and G. E. Hinton. 2012. “Imagenet Classification with Deep Convolutional Neural Networks.” In Advances in Neural Information Processing Systems, 1097–1105.
- Lee, J., B. Bagheri, and H. A. Kao. 2015. “A Cyber-physical Systems Architecture for Industry 4.0-based Manufacturing Systems.” Manufacturing Letters 3: 18–23. doi:https://doi.org/10.1016/j.mfglet.2014.12.001.
- Leitão, P. 2009. “Agent-based Distributed Manufacturing Control: A State-of-the-art Survey.” Engineering Applications of Artificial Intelligence 22 (7): 979–991. doi:https://doi.org/10.1016/j.engappai.2008.09.005.
- Lin, J., L. Zhu, and K. Gao. 2020. “A Genetic Programming Hyper-heuristic Approach for the Multi-skill Resource Constrained Project Scheduling Problem.” Expert Systems with Applications 140: 112915. doi:https://doi.org/10.1016/j.eswa.2019.112915.
- Mahadevan, S., and G. Theocharous. 1998. “Optimizing Production Manufacturing Using Reinforcement Learning.” FLAIRS Conference 372 (May): 377.
- Mnih, V., K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller. 2013. “Playing Atari with Deep Reinforcement Learning.” arXiv Preprint arXiv:1312.5602.
- Mnih, V., K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, … S. Petersen. 2015. “Human-level Control through Deep Reinforcement Learning.” nature 518 (7540): 529–533. doi:https://doi.org/10.1038/nature14236.
- Nair, V., and G. E. Hinton. January 2010. Rectified Linear Units Improve Restricted Boltzmann Machines. ICML.
- Riedmiller, M., T. Gabel, R. Hafner, and S. Lange. 2009. “Reinforcement Learning for Robot Soccer.” Autonomous Robots 27 (1): 55–73. doi:https://doi.org/10.1007/s10514-009-9120-4.
- Shahrabi, J., M. A. Adibi, and M. Mahootchi. 2017. “A Reinforcement Learning Approach to Parameter Estimation in Dynamic Job Shop Scheduling.” Computers & Industrial Engineering 110: 75–82. doi:https://doi.org/10.1016/j.cie.2017.05.026.
- Singh, N., P. V. Sarngadharan, and P. K. Pal. 2011. “AGV Scheduling for Automated Material Distribution: A Case Study.” Journal of Intelligent Manufacturing 22 (2): 219–228. doi:https://doi.org/10.1007/s10845-009-0283-9.
- Suttan, R., and A. Barto. 1998. Reinforcement Learning: An Introduction. MIT Press.
- Tesauro, G. 1995. “Temporal Difference Learning and TD-Gammon.” Communications of the ACM 38 (3): 58–68. doi:https://doi.org/10.1145/203330.203343.
- Toth, P., and D. Vigo, Eds. 2002. “The Vehicle Routing Problem.” Society for Industrial and Applied Mathematics.
- Wang, S., J. Wan, D. Zhang, D. Li, and C. Zhang. 2016. “Towards Smart Factory for Industry 4.0: A Self-organized Multi-agent System with Big Data Based Feedback and Coordination.” Computer Networks 101: 158–168. doi:https://doi.org/10.1016/j.comnet.2015.12.017.
- Wang, Y. C., and J. M. Usher. 2005. “Application of Reinforcement Learning for Agent-based Production Scheduling.” Engineering Applications of Artificial Intelligence 18 (1): 73–82. doi:https://doi.org/10.1016/j.engappai.2004.08.018.
- Watkins, C. J., and P. Dayan. 1992. “Q-learning.” Machine Learning 8 (3–4): 279–292. doi:https://doi.org/10.1007/BF00992698.
- Zhang, W., and T. G. Dietterich. 1995. “A Reinforcement Learning Approach to Job-shop Scheduling.” IJCAI 95 (August): 1114–1120.