References
- Albus, J. (1981). Brains, behavior and robotics. New York, NY: McGraw-Hill, Inc.
- Bazzan, A. L. C., & Klügl, F. (2013). A review on agent-based technology for traffic and transportation. The Knowledge Engineering Review. FirstView. 10.1017/S0269888913000118. Retrieved from http://journals.cambridge.org/action/displayAbstract?fromPage=online&aid=8907641.
- Busoniu, L., Babuska, R., & De Schutter, B. (2008). A comprehensive survey of multiagent reinforcement learning. IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews, 38(2), 156–172. doi: 10.1109/TSMCC.2007.913919
- Chang, Y.-H., Ho, T., & Kaelbling, L. P. (2004). All learning is local: Multi-agent learning in global reward games. In S. Thrun, L. Saul, & B. Schoelkopf (Eds.), Advances in neural information processing systems 16. Cambridge, MA: MIT Press.
- Dresner, K., & Stone, P. (2008). A multiagent approach to autonomous intersection management. Journal of Artificial Intelligence Research, 31, 591–656.
- El-Tantawy, S., Abdulhai, B., & Abdelgawad, H. (2013). Multiagent reinforcement learning for integrated network of adaptive traffic signal controllers (MARLIN-ATSC): Methodology and large-scale application on downtown Toronto. IEEE Transactions on Early Access Intelligent Transportation Systems, 14(3), 1140–1150. doi: 10.1109/TITS.2013.2255286
- Junges, R., & Bazzan, A. (2008). Evaluating the performance of dcop algorithms in a real world, dynamic problem. Proceedings of the 7th international joint conference on autonomous agents and multiagent systems, International Foundation for Autonomous Agents and Multiagent Systems, Richland, SC, pp. 599–606.
- Klopf, A. H. (1972). Brain function and adaptive systems: A heterostatic theory (Technical Report AFCRL-72-0164). Bedford, MA: Air Force Cambridge Research Laboratories.
- Kuyer, L., Whiteson, S., Bakker, B., & Vlassis, N. (2008). Multiagent reinforcement learning for urban traffic control using coordination graphs. Machine Learning and Knowledge Discovery in Databases, 5211, 656–671. doi: 10.1007/978-3-540-87479-9_61
- Liu, Z. (2007). A survey of intelligence methods in urban traffic signal control. IJCSNS International Journal of Computer Science and Network Security, 7(7), 105–112.
- Minnesota Department of Transportation. (2011). Traffic signal timing and coordination manual. Minnesota Department of Transportation.
- Mirchandani, P., & Wang, F.-Y. (2005). Rhodes to intelligent transportation systems. IEEE Intelligent Systems, 20(1), 10–15. doi: 10.1109/MIS.2005.15
- National Transportation Operations Coalition. (2012). National traffic signal report card, executive summary. Retrieved from http://www.ite.org/reportcard/ExecSummary.pdf.
- Pham, T. T., Tawfik, A., & Taylor, M. E. (2013). A simple, naive agent-based model for the optimization of a system of traffic lights: Insights from an exploratory experiment. Proceedings of the conference on agent-based modeling in transportation planning and operations, Blacksburg, VA.
- Robertson, D., & Bretherton, R. (1991). Optimizing networks of traffic signals in real time-the scoot method. IEEE Transactions on Vehicular Technology, 40(1), 11–15. doi: 10.1109/25.69966
- Rummery, G., & Niranjan, M. (1994). On-line Q-learning using connectionist systems. Cambridge: Department of Engineering, University of Cambridge.
- Shoufeng, M., Ying, L., & Bao, L. (2002). Agent-based learning control method for urban traffic signal of single intersection. Journal of Systems Engineering, 17(6), 526–530.
- Singh, S. P., Jaakkola, T., Littman, M. L., & Szepesvári, C. (2000). Convergence results for single-step on-policy reinforcement-learning algorithms. Machine Learning, 38(3), 287–308. doi: 10.1023/A:1007678930559
- Singh, S. P., & Sutton, R. S. (1996). Reinforcement learning with replacing eligibility traces. Machine Learning, 22(1–3), 123–158. doi: 10.1007/BF00114726
- Sutton, R., & Barto, A. (1998). Reinforcement learning: An introduction. Cambridge: Cambridge University Press.
- Taylor, M. E., Jain, M., Tandon, P., Yokoo, M., & Tambe, M. (2011). Distributed on-line multi-agent optimization under uncertainty: Balancing exploration and exploitation. Advances in Complex Systems (ACS), 14(03), 471–528. doi: 10.1142/S0219525911003104
- Texas Transportation Institute. (2011). Passer (tm) v-09. Retrieved from http://ttisoftware.tamu.edu/
- Thorpe, T., & Andersson, C. (1997). Vehicle traffic light control using SARSA (Master's thesis). Department of Computer Science, Colorado State University, Fort Collins.
- Van Moffaert, K., Drugan, M. M., & Nowé, A. (2013). Scalarized multi-objective reinforcement learning: Novel design techniques. Proceedings of the IEEE symposium on adaptive dynamic programming and reinforcement learning, Singapore, pp. 94–103.