References
- Asiain, E., Clempner, J. B., & Poznyak, A. S. (2018). A reinforcement learning approach for solving the mean variance customer portfolio for partially observable models. International Journal on Artificial Intelligence Tools, 27(8), 1850034.
- Bernstein, D. S., Givan, R., Immerman, N., & Zilberstein, S. (2002). The complexity of decentralized control of markov decision processes. Mathematics of Operations Research, 27(4), 819–840.
- Bopardikar, S. D., & Hespanha, J. P. (2011). Randomized solutions to partial information dynamic zero-sum games. In American control conference (acc) (p. 4039–4044)San Francisco, CA, USA.
- Clempner, J. B. (2021). A proximal/gradient approach for computing the nash equilibrium in controllable markov games. Journal of Optimization Theory and Applications, 188(3), 847–862.
- Clempner, J. B. (2022). A dynamic mechanism design for controllable and ergodic markov games. Computational Economics. (To be Published). https://doi.org/10.1007/s10614-022-10240-y
- Clempner, J. B., & Poznyak, A. S. (2019). Observer and control design in partially observable finite markov chains. Automatica, 110, 108587.
- Filar, J., & Vrieze, K. (1997). Competitive markov decision processes. Springer.
- Ghosh, M. K., McDonald, D., & Sinha, S. (2004). Zero-sum stochastic games with partial information. Journal of Optimization Theory and Applications, 121(1), 99–118.
- Hespanha, J. P., & Prandini, M. (2001). Nash equilibria in partial-information games on markov chains. IEEE Conference on Decision and Control (p. 2102–2107).
- McEneaney, W. M. (2004). Some classes of imperfect information nite statespace stochastic games with nite-dimensional solutions. Applied Mathematics and Optimization, 50(2), 87–118.
- Poznyak, A. S., Najim, K., & Gomez-Ramirez, E. (2000). Self-learning control of finite markov chains. Marcel Dekker, Inc.
- Raghavan, T. E. S., & Filar, J. A. (1991). Algorithms for stochastic games - a survey. Methods and Models of Operations Research.
- Sragovich, V. G. (2006). Mathematical theory of adaptive control. World Scientific Publishing Company.
- Trejo, K. K., Clempner, J. B., & Poznyak, A. S. (2015). Computing the stackelberg/nash equilibria using the extraproximal method: Convergence analysis and implementation details for markov chains games. International Journal of Applied Mathematics and Computer Science, 25(2), 337–351.
- Trejo, K. K., Clempner, J. B., & Poznyak, A. S. (2016). An optimal strong equilibirum solution for cooperative multi-leader-follower stackelberg markov chains games. Kybernetika, 52(2), 258–279.
- Trejo, K. K., Clempner, J. B., & Poznyak, A. S. (2017). Computing the lp-strong nash equilibrium for markov chains games. Applied Mathematical Modelling, 41, 399–418.
- Trejo, K. K., Juarez, R., Clempner, J. B., & Poznyak, A. S. (2020). Non-cooperative bargaining with unsophisticated agents. Computational Economics, (To be published). https://doi.org/10.1007/s10614-020-10003-7