73
Views
2
CrossRef citations to date
0
Altmetric
Research Articles

Revealing perceived individuals’ self-interest

Pages 336-345 | Received 15 Nov 2021, Accepted 16 Mar 2023, Published online: 05 Apr 2023

References

  • Asiain, E., Clempner, J. B., & Poznyak, A. S. (2018). A reinforcement learning approach for solving the mean variance customer portfolio for partially observable models. International Journal on Artificial Intelligence Tools, 27(8), 1850034.
  • Bernstein, D. S., Givan, R., Immerman, N., & Zilberstein, S. (2002). The complexity of decentralized control of markov decision processes. Mathematics of Operations Research, 27(4), 819–840.
  • Bopardikar, S. D., & Hespanha, J. P. (2011). Randomized solutions to partial information dynamic zero-sum games. In American control conference (acc) (p. 4039–4044)San Francisco, CA, USA.
  • Clempner, J. B. (2021). A proximal/gradient approach for computing the nash equilibrium in controllable markov games. Journal of Optimization Theory and Applications, 188(3), 847–862.
  • Clempner, J. B. (2022). A dynamic mechanism design for controllable and ergodic markov games. Computational Economics. (To be Published). https://doi.org/10.1007/s10614-022-10240-y
  • Clempner, J. B., & Poznyak, A. S. (2019). Observer and control design in partially observable finite markov chains. Automatica, 110, 108587.
  • Filar, J., & Vrieze, K. (1997). Competitive markov decision processes. Springer.
  • Ghosh, M. K., McDonald, D., & Sinha, S. (2004). Zero-sum stochastic games with partial information. Journal of Optimization Theory and Applications, 121(1), 99–118.
  • Hespanha, J. P., & Prandini, M. (2001). Nash equilibria in partial-information games on markov chains. IEEE Conference on Decision and Control (p. 2102–2107).
  • McEneaney, W. M. (2004). Some classes of imperfect information nite statespace stochastic games with nite-dimensional solutions. Applied Mathematics and Optimization, 50(2), 87–118.
  • Poznyak, A. S., Najim, K., & Gomez-Ramirez, E. (2000). Self-learning control of finite markov chains. Marcel Dekker, Inc.
  • Raghavan, T. E. S., & Filar, J. A. (1991). Algorithms for stochastic games - a survey. Methods and Models of Operations Research.
  • Sragovich, V. G. (2006). Mathematical theory of adaptive control. World Scientific Publishing Company.
  • Trejo, K. K., Clempner, J. B., & Poznyak, A. S. (2015). Computing the stackelberg/nash equilibria using the extraproximal method: Convergence analysis and implementation details for markov chains games. International Journal of Applied Mathematics and Computer Science, 25(2), 337–351.
  • Trejo, K. K., Clempner, J. B., & Poznyak, A. S. (2016). An optimal strong equilibirum solution for cooperative multi-leader-follower stackelberg markov chains games. Kybernetika, 52(2), 258–279.
  • Trejo, K. K., Clempner, J. B., & Poznyak, A. S. (2017). Computing the lp-strong nash equilibrium for markov chains games. Applied Mathematical Modelling, 41, 399–418.
  • Trejo, K. K., Juarez, R., Clempner, J. B., & Poznyak, A. S. (2020). Non-cooperative bargaining with unsophisticated agents. Computational Economics, (To be published). https://doi.org/10.1007/s10614-020-10003-7

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.