References
- Abu-Khalaf, M., & Lewis, F.L. (2005). Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach. Automatica, 41, 779–791.
- Apostol, T.M. (1974). Mathematical analysis (2nd ed.). Massachusetts: Addison-Wesley.
- Beard, R., Saridis, G., & Wen, J. (1997). Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation. Automatica, 33, 2159–2177.
- Bellman, R.E. (1957). Dynamic programming. New Jersey: Princeton University Press.
- Bertsekas, D.P., & Tsitsiklis, J.N. (1996). Neuro-dynamic programming. Massachusetts: Athena Scientific.
- Bhasin, S., Kamalapurkar, R., Johnson, M., Vamvoudakis, K.G., Lewis, F.L., & Dixon, W.E. (2013). A novel actor-critic-identifier architecture for approximate optimal control of uncertain nonlinear systems. Automatica, 49, 82–92.
- Dierks, T., Thumati, B.T., & Jagannathan, S. (2009). Optimal control of unknown affine nonlinear discrete-time systems using offline-trained neural networks with proof of convergence. Neural Networks, 22, 851–860.
- Hayakawa, T., Haddad, W.M., & Hovakimyan, N. (2008). Neural network adaptive control for nonlinear uncertain dynamical systems with asymptotic stability guarantees. IEEE Transactions on Neural Networks, 19, 80–89.
- Hornic, K., & Stinchombe, M. (1989). Multilayer feedforward neural networks are universal approximators. Neural Networks, 2, 359–366.
- Ioannou, P.A., & Sun, J. (1996). Robust adaptive control. New Jersey: Prentice-Hall.
- Kek, S.L., Teo, K.L., & Ismail, A.A. (2010). An integrated optimal control algorithm for discrete-time nonlinear stochastic system. International Journal of Control, 83, 2536–2545.
- Koshkouei, A.J., Farahi, M.H., & Burnham, K.J. (2012). An almost optimal control design method for nonlinear time-delay systems. International Journal of Control, 85, 147–158.
- Lewis, F.L., Jagannathan, S., & Yesildirek, A. (1999). Neural network control of robot manipulators and nonlinear systems. London: Taylor & Francis.
- Lewis, F.L., & Liu, D. (2013). Reinforcement learning and approximate dynamic programming for feedback control. New York: John Wiley & Sons, Inc.
- Lewis, F.L., & Syrmos, V.L. (1995). Optimal control. New York: John Wiley & Sons, Inc.
- Lewis, F.L., & Vrabie, D. (2009). Reinforcement learning and adaptive dynamic programming for feedback control. IEEE Circuits and Systems Magazine, 9, 32–50.
- Lin, W. (2011). Optimality and convergence of adaptive optimal control by reinforcement synthesis. Automatica, 47, 1047–1052.
- Lin, W., & Zheng, C. (2012). Constrained adaptive optimal control using a reinforcement learning agent. Automatica, 48, 2614–2619.
- Liu, D., Li, H., & Wang, D. (2013). Neural-network-based zero-sum game for discrete-time nonlinear systems via iterative adaptive dynamic programming algorithm. Neurocomputing, 110, 92–100.
- Liu, D., Yang, X., & Li, H. (2012). Adaptive optimal control for a class of continuous-time affine nonlinear systems with unknown internal dynamics. Neural Computing and Applications, doi: 10.1007/s00521-012-1249-y.
- Liu, D., Wang, D., & Yang, X. (2013). An iterative adaptive dynamic programming algorithm for optimal control of unknown discrete-time nonlinear systems with constrained inputs. Information Sciences, 220, 331–342.
- Liu, D., & Wei, Q. (2013). Finite-approximation-error-based optimal control approach for discrete-time nonlinear systems. IEEE Transactions on Cybernetics, 43, 779–789.
- Lyshevski, S.E. (1998). Optimal control of nonlinear continuous-time systems: Design of bounded controllers via generalized nonquadratic functionals. In Proceedings of American Control Conference (pp. 205–209). Philadelphia, PA.
- Modares, H., Lewis, F.L., & Sistani, M. (2012). Online solution of nonquadratic two-player zero-sum games arising in the H∞ control of constrained input systems. International Journal of Adaptive Control and Signal Processing, doi: 10.1002/acs.2348.
- Murray, J.J., Cox, C.J., Lendaris, G.G., & Saeks, R. (2002). Adaptive dynamic programming. IEEE Transactions on Systems, Man and Cybernetics, Part C: Applications and Reviews, 32, 140–153.
- Padhi, R., Unnikrishnan, N., Wang, X., & Balakrishnan, S.N. (2006). A single network adaptive critic (SNAC) architecture for optimal control synthesis for a class of nonlinear systems. Neural Networks, 19, 1648–1660.
- Powell, W.B. (2011). Approximate dynamic programming: Solving the curses of dimensionality (2nd ed. ). Hoboken, NJ: Wiley.
- Prokhorov, D.V., & Wunsch, D.C. (1997). Adaptive critic designs. IEEE Transactions on Neural Networks, 8, 997–1007.
- Sahnoun, M., Andrieu, V., & Nadri, M. (2012). Nonlinear and locally optimal controllers design for input affine locally controllable systems. International Journal of Control, 85, 159–170.
- Si, J., & Wang, Y.T. (2001). On-line learning control by association and reinforcement. IEEE Transactions on Neural Networks, 12, 264–276.
- Sutton, R.S., & Barto, A.G. (1998). Reinforcement learning–an introduction. Massachusetts: MIT Press.
- Vamvoudakis, K.G., & Lewis, F.L. (2010). Online actor-critic algorithm to solve the continuous-time infinite horizon optimal control problem. Automatica, 46, 878–888.
- Wang, F.Y., Jin, N., Liu, D., & Wei, Q. (2011). Adaptive dynamic programming for finite horizon optimal control of discrete-time nonlinear systems with ε-error bound. IEEE Transactions on Neural Networks, 22, 24–36.
- Wang, D., Liu, D., Wei, Q., Zhao, D., & Jin, N. (2012). Optimal control of unknown nonaffine nonlinear discrete-time systems based on adaptive dynamic programming. Automatica, 48, 1825–1832.
- Wang, F.Y., Zhang, H., & Liu, D. (2009). Adaptive dynamic programming: An introduction. IEEE Computational Intelligence Magazine, 4, 39–47.
- Wei, Q., & Liu, D. (2012). An iterative ε-optimal control scheme for a class of discrete-time nonlinear systems with unfixed initial state. Neural Networks, 32, 236–244.
- Werbos, P.J. (1974). Beyond regression: New tools for prediction and analysis in the behavioral sciences ( Ph.D. thesis). Harvard University, Massachusetts.
- Werbos, P.J. (1977). Advanced forecasting methods for global crisis warning and models of intelligence. General Systems Yearbook, 22, 25–38.
- Werbos, P.J. (1992). Approximate dynamic programming for real-time control and neural modeling. In D.A. White & D.A. Sofge (Eds.), Handbook of intelligent control: Neural, fuzzy, and adaptive approaches. New York: Van Nostrand Reinhold.
- Wu, Q.H. (1995). Reinforcement learning control using interconnected learning automata. International Journal of Control, 62, 1–16.
- Xin, M., & Pan, H. (2010). Integrated nonlinear optimal control of spacecraft in proximity operations. International Journal of Control, 83, 347–363.
- Yu, W. (2009). Recent advances in intelligent control systems. London: Springer-Verlag.
- Yu, W., & Li, X. (2001). Some new results on system identification with dynamic neural network. IEEE Transactions on Neural Networks, 2, 412–417.
- Zhang, H., Cui, L., Zhang, X., & Luo, Y. (2011). Data-driven robust approximate optimal tracking control for unknown general nonlinear systems using adaptive dynamic programming method. IEEE Transactions on Neural Networks, 22, 2226–2236.
- Zhang, H., Liu, D., Luo, Y., & Wang D,. (2013). Adaptive dynamic programming for control: Algorithms and stability. London: Springer.
- Zhang, H., Wei, Q., & Liu, D. (2011). An iterative adaptive dynamic programming method for solving a class of nonlinear zero-sum differential games. Automatica, 47, 207–214.