References
- Akaike, H. (1973), “Information Theory and an Extension of the Maximum Likelihood Principle,” in Proceedings of the Second International Symposium on Information Theory, eds. B. N. Petrovand and F. Csaki, Budapest: Akademiai Kiado, pp. 267–281.
- Andrieu, C., Freitas, N. D., Doucet, A., and Jordan, M. I. (2003), “An Introduction to MCMC for Machine Learning,” Machine Learning, 50, 5–43. DOI: https://doi.org/10.1023/A:1020281327116.
- Berger, J. (1985), Statistical Decision Theory and Bayesian Analysis (2nd ed.), New York: Springer-Verlag.
- Besag, J. (1989), “A Candidate’s Formula: A Curious Result in Bayesian Prediction,” Biometrika, 76, 183. DOI: https://doi.org/10.1093/biomet/76.1.183.
- Bishop, C. M. (2006), Pattern Recognition and Machine Learning, New York: Springer.
- Box, G. (1980), “Sampling and Bayes’ Inference in Scientific Modelling and Robustness” (with discussion), Journal of the Royal Statistical Society, Series A, 143, 383–430. DOI: https://doi.org/10.2307/2982063.
- Chib, S. (1995), “Marginal Likelihood From the Gibbs Output,” Journal of the American Statistical Association, 90, 1313–1321. DOI: https://doi.org/10.1080/01621459.1995.10476635.
- DeGroot, M. H. (1970), Optimal Statistical Decisions, New York: McGraw-Hill.
- Edwards, W., Lindman, H., and Savage, L. (1963), “Bayesian Statistical Inference for Psychological Research,” Psychological Review, 70, 193–242. DOI: https://doi.org/10.1037/h0044139.
- Friel, N., and Wyse, J. (2012), “Estimating the Evidence—A Review,” Statistica Neerlandica, 66, 288–308. DOI: https://doi.org/10.1111/j.1467-9574.2011.00515.x.
- Gelfand, A., and Dey, D. (1994), “Bayesian Model Choice: Asymptotic and Exact Calculations,” Journal of the Royal Statistical Society, Series B, 56, 501–514. DOI: https://doi.org/10.1111/j.2517-6161.1994.tb01996.x.
- Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., and Rubin, D. B. (2014), Bayesian Data Analysis (3rd ed.), Boca Raton, FL: Chapman and Hall/CRC.
- Gowers, T. (2002), Mathematics: A Very Short Introduction, Oxford: Oxford University Press.
- Gutmann, M. U., and Hyvärinen, A. (2012), “Noise-Contrastive Estimation of Unnormalized Statistical Models, With Applications to Natural Image Statistics,” Journal of Machine Learning Research, 13, 307–361.
- Hastie, T., Tibshirani, R., and Friedman, J. (2009), The Elements of Statistical Learning: Data Mining, Inference, and Prediction (2nd ed.), New York: Springer.
- Jefferys, W. H., and Berger, J. O. (1992), “Ockham’s Razor and Bayesian Analysis,” American Scientist, 80, 64–72.
- Le Cam, L. (1990), “Maximum Likelihood: An Introduction,” International Statistical Review, 58, 153–171. DOI: https://doi.org/10.2307/1403464.
- MacKay, D. J. C. (1992), “Bayesian Interpolation,” Neural Computation, 4, 415–447. DOI: https://doi.org/10.1162/neco.1992.4.3.415.
- MacKay, D. J. C. (2003), Information Theory, Inference, and Learning Algorithms, Cambridge: Cambridge University Press.
- Murphy, K. (2012), Machine Learning: A Probabilistic Perspective, Cambridge, MA: MIT Press.
- Robert, C., and Casella, G. (1999), Monte Carlo Statistical Methods, New York: Springer.
- Schervish, M. J. (1995), Theory of Statistics (corrected 2nd printing, 1997), New York: Springer.
- Sober, E. (2015), Ockham’s Razors: A User’s Manual, Cambridge: Cambridge University Press.