References
- Abu-Khalaf, M., & Lewis, F. L. (2005). Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach. Automatica, 41(5), 779–791. https://doi.org/https://doi.org/10.1016/j.automatica.2004.11.034
- Abu-Khalaf, M., Lewis, F. L., & Huang, J. (2008). Neurodynamic programming and zero-sum games for constrained control systems. IEEE Transactions on Neural Networks, 19(7), 1243–1252. https://doi.org/https://doi.org/10.1109/TNN.2008.2000204
- Bhasin, S., Kamalapurkar, R., Johnson, M., Vamvoudakis, K. G., Lewis, F. L., & W. E. Dixon (2013). A novel actor–critic–identifier architecture for approximate optimal control of uncertain nonlinear systems. Automatica, 49(1), 82–92. https://doi.org/https://doi.org/10.1016/j.automatica.2012.09.019
- Dierks, T., & Jagannathan, S. (2010). Optimal control of affine nonlinear continuous-time systems. In Proceedings of the 2010 American control conference (pp. 1568–1573). IEEE.
- Finlayson, B. A. (2013). The method of weighted residuals and variational principles (Vol. 73). SIAM.
- Hou, D., Na, J., Lv, Y., Gao, G., & Wu, X. (2017). Adaptive optimal tracking control for continuous-time systems using identifier-critic based dynamic programming. In 2017 36th Chinese control conference (CCC) (pp. 2583–2588). IEEE.
- Kiumarsi, B., Lewis, F. L., Modares, H., Karimpour, A., & Naghibi-Sistani, M. B. (2014). Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics. Automatica, 50(4), 1167–1175. https://doi.org/https://doi.org/10.1016/j.automatica.2014.02.015
- Liu, C., Zhang, H., Ren, H., & Liang, Y. (2019). An analysis of IRL-based optimal tracking control of unknown nonlinear systems with constrained input. Neural Processing Letters, 50, 2681–2700. https://doi.org/https://doi.org/10.1007/s11063-019-10029-5
- Liu, D., Yang, X., Wang, D., & Wei, Q. (2015). Reinforcement-learning-based robust controller design for continuous-time uncertain nonlinear systems subject to input constraints. IEEE Transactions on Cybernetics, 45(7), 1372–1385. https://doi.org/https://doi.org/10.1109/TCYB.2015.2417170
- Lv, Y., Na, J., & Ren, X. (2019). Online H∞ control for completely unknown nonlinear systems via an identifier–critic-based adp structure. International Journal of Control, 92(1), 100–111. https://doi.org/https://doi.org/10.1080/00207179.2017.1381763
- Lv, Y., Na, J., Yang, Q., Wu, X., & Guo, Y. (2016). Online adaptive optimal control for continuous-time nonlinear systems with completely unknown dynamics. International Journal of Control, 89(1), 99–112. https://doi.org/https://doi.org/10.1080/00207179.2015.1060362
- Lyashevskiy, S. (1996). Constrained optimization and control of nonlinear systems: New results in optimal control. In Proceedings of 35th IEEE conference on decision and control (Vol. 1, pp. 541–546). IEEE.
- Mishra, A., & Ghosh, S. (2019a). Critic-only integral reinforcement learning driven by variable gain gradient descent for optimal tracking control. Preprint, arXiv:1911.04153
- Mishra, A., & Ghosh, S. (2019b). Variable gain gradient descent-based robust reinforcement learning for optimal tracking control of unknown nonlinear system with input-constraints. Preprint, arXiv:1911.04157
- Mishra, A., & Ghosh, S. (2020). H∞ tracking control via variable gain gradient descent-based integral reinforcement learning for unknown continuous time nonlinear system. Preprint, arXiv:2001.07355
- Modares, H., & Lewis, F. L. (2014). Optimal tracking control of nonlinear partially-unknown constrained-input systems using integral reinforcement learning. Automatica, 50(7), 1780–1792. https://doi.org/https://doi.org/10.1016/j.automatica.2014.05.011
- Modares, H., Lewis, F. L., & Jiang, Z. P. (2015). H∞ tracking control of completely unknown continuous-time systems via off-policy reinforcement learning. IEEE Transactions on Neural Networks and Learning Systems, 26(10), 2550–2562. https://doi.org/https://doi.org/10.1109/TNNLS.2015.2441749
- Modares, H., Lewis, F. L., & Naghibi-Sistani, M. B. (2013). Adaptive optimal control of unknown constrained-input systems using policy iteration and neural networks. IEEE Transactions on Neural Networks and Learning Systems, 24(10), 1513–1525. https://doi.org/https://doi.org/10.1109/TNNLS.2013.2276571
- Modares, H., Lewis, F. L., & Naghibi-Sistani, M. B. (2014). Integral reinforcement learning and experience replay for adaptive optimal control of partially-unknown constrained-input continuous-time systems. Automatica, 50(1), 193–202. https://doi.org/https://doi.org/10.1016/j.automatica.2013.09.043
- Na, J., & Herrmann, G. (2014). Online adaptive approximate optimal tracking control with simplified dual approximation structure for continuous-time unknown nonlinear systems. IEEE/CAA Journal of Automatica Sinica, 1(4), 412–422. https://doi.org/https://doi.org/10.1109/JAS.2014.7004668
- Na, J., Lv, Y., Wu, X., Guo, Y., & Chen, Q. (2014). Approximate optimal tracking control for continuous-time unknown nonlinear systems. In Proceedings of the 33rd Chinese control conference (pp. 8990–8995). IEEE.
- Ren, X., Lewis, F. L., & Zhang, J. (2009). Neural network compensation control for mechanical systems with disturbances. Automatica, 45(5), 1221–1226. https://doi.org/https://doi.org/10.1016/j.automatica.2008.12.009
- Vamvoudakis, K. G., & Lewis, F. L. (2010). Online actor–critic algorithm to solve the continuous-time infinite horizon optimal control problem. Automatica, 46(5), 878–888. https://doi.org/https://doi.org/10.1016/j.automatica.2010.02.018
- Vamvoudakis, K. G., Vrabie, D., & Lewis, F. L. (2014). Online adaptive algorithm for optimal control with integral reinforcement learning. International Journal of Robust and Nonlinear Control, 24(17), 2686–2710. https://doi.org/https://doi.org/10.1002/rnc.3018
- Vrabie, D., & Lewis, F. (2009). Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems. Neural Networks, 22(3), 237–246. https://doi.org/https://doi.org/10.1016/j.neunet.2009.03.008
- Yang, X., Liu, D., & Wang, D. (2014). Reinforcement learning for adaptive optimal control of unknown continuous-time nonlinear systems with input constraints. International Journal of Control, 87(3), 553–566. https://doi.org/https://doi.org/10.1080/00207179.2013.848292
- Yang, X., Liu, D., & Wei, Q. (2015). Robust tracking control of uncertain nonlinear systems using adaptive dynamic programming. In International conference on neural information processing (pp. 9–16). Springer.
- Zhang, H., Cui, L., Zhang, X., & Luo, Y. (2011). Data-driven robust approximate optimal tracking control for unknown general nonlinear systems using adaptive dynamic programming method. IEEE Transactions on Neural Networks, 22(12), 2226–2236. https://doi.org/https://doi.org/10.1109/TNN.2011.2168538
- Zhang, H., Cui, X., Luo, Y., & Jiang, H. (2017). Finite-horizon H∞ tracking control for unknown nonlinear systems with saturating actuators. IEEE Transactions on Neural Networks and Learning Systems, 29(4), 1200–1212. https://doi.org/https://doi.org/10.1109/TNNLS.2017.2669099
- Zhu, Y., Zhao, D., & Li, X. (2016). Using reinforcement learning techniques to solve continuous-time non-linear optimal tracking problem without system dynamics. IET Control Theory & Applications, 10(12), 1339–1347. https://doi.org/https://doi.org/10.1049/cth2.v10.12