References
- Abellán, J., & Castellano, J. G. (2017). A comparative study on base classifiers in ensemble methods for credit scoring. Expert Systems with Applications, 73, 1–10. https://doi.org/https://doi.org/10.1016/j.eswa.2016.12.020
- Baesens, B., Roesch, D., & Scheule, H. (2016). Credit risk analytics: Measurement techniques, applications, and examples in SAS. John Wiley and Sons.
- Baesens, B., Van Gestel, T., Viaene, S., Stepanova, M., Suykens, J., & Vanthienen, J. (2003). Benchmarking state-of-the-art classification algorithms for credit scoring. Journal of the Operational Research Society, 54(6), 627–635. https://doi.org/https://doi.org/10.1057/palgrave.jors.2601545
- Baesens, B., Van Vlasselaer, V., & Verbeke, W. (2015). Fraud analytics using descriptive, predictive, and social network techniques: a guide to data science for fraud detection. John Wiley and Sons.
- Bahnsen, A. C., Aouada, D., Ottersten, B. (2014). Example-dependent cost-sensitive logistic regression for credit scoring. In Machine Learning and Applications (ICMLA), 2014 13th International Conference on (pp. 263–269).
- Bahnsen, A. C., Aouada, D., & Ottersten, B. (2015). Example-dependent cost-sensitive decision trees. Expert Systems with Applications, 42(19), 6609–6619. https://doi.org/https://doi.org/10.1016/j.eswa.2015.04.042
- Bellotti, T., & Crook, J. (2012). Loss given default models incorporating macroeconomic variables for credit cards. International Journal of Forecasting, 28(1), 171–182. https://doi.org/https://doi.org/10.1016/j.ijforecast.2010.08.005
- Bequé, A., Coussement, K., Gayler, R., & Lessmann, S. (2017). Approaches for credit scorecard calibration: An empirical analysis. Knowledge-Based Systems, 134, 213–227. https://doi.org/https://doi.org/10.1016/j.knosys.2017.07.034
- Breiman, L. (1996). Bagging predictors. Machine Learning, 24(2), 123–140. https://doi.org/https://doi.org/10.1023/A:1018054314350
- Breiman, L. (2001). Random forests. Machine Learning, 45(1), 5–32. https://doi.org/https://doi.org/10.1023/A:1010933404324
- Brown, I., & Mues, C. (2012). An experimental comparison of classification algorithms for imbalanced credit scoring data sets. Expert Systems with Applications, 39(3), 3446–3453. https://doi.org/https://doi.org/10.1016/j.eswa.2011.09.033
- Castermans, G., Martens, D., Gestel, T. V., Hamers, B., & Baesens, B. (2010). An overview and framework for pd backtesting and benchmarking. Journal of the Operational Research Society, 61(3), 359–373. https://doi.org/https://doi.org/10.1057/jors.2009.69
- Chawla, N. V., Bowyer, K. W., Hall, L. O., & Kegelmeyer, W. P. (2002). Smote: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research, 16, 321–357. https://doi.org/https://doi.org/10.1613/jair.953
- Chen, T., Guestrin, C. (2016). XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD ’16 (pp. 785–794.
- Crook, J., & Banasik, J. (2004). Does reject inference really improve the performance of application scoring models? Journal of Banking & Finance, 28(4), 857–874. https://doi.org/https://doi.org/10.1016/j.jbankfin.2003.10.010
- Crook, J. N., Edelman, D. B., & Thomas, L. C. (2007). Recent developments in consumer credit risk assessment. European Journal of Operational Research, 183(3), 1447–1465. https://doi.org/https://doi.org/10.1016/j.ejor.2006.09.100
- Domingos, P. (1999). Metacost: A general method for making classifiers cost-sensitive. In Proceedings of the 5th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD '99 (pp. 155-164).
- Elkan, C. (2001). The foundations of cost-sensitive learning [Paper presentation]. In Proc. Seventeenth Int. Conf. on Artificial Intelligence (pp. 973–978).
- Freund, Y., & Schapire, R. (1997). A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences, 55(1), 119–139. https://doi.org/https://doi.org/10.1006/jcss.1997.1504
- Friedman, J., Hastie, T., & Tibshirani, R. (2009). The elements of statistical learning. Springer series in statistics, New York.
- Höppner, S., Stripling, E., Baesens, B., Broucke, S. v., & Verdonck, T. (2020). Profit driven decision trees for churn prediction. European Journal of Operational Research, 284(3), 920–933. https://doi.org/https://doi.org/10.1016/j.ejor.2018.11.072
- Hurley, M., & Adebayo, J. (2017). Credit scoring in the era of big data. Yale Journal of Law and Technology, 18(1), 5.
- Lessmann, S., Baesens, B., Seow, H.-V., & Thomas, L. C. (2015). Benchmarking state-of-the-art classification algorithms for credit scoring: An update of research. European Journal of Operational Research, 247(1), 124–136. https://doi.org/https://doi.org/10.1016/j.ejor.2015.05.030
- Maldonado, S., Pérez, J., & Bravo, C. (2017). Cost-based feature selection for support vector machines: An application in credit scoring. European Journal of Operational Research, 261(2), 656–665. https://doi.org/https://doi.org/10.1016/j.ejor.2017.02.037
- Marqués, A. I., García, V., & Sánchez, J. S. (2013). On the suitability of resampling techniques for the class imbalance problem in credit scoring. Journal of the Operational Research Society, 64(7), 1060–1070. https://doi.org/https://doi.org/10.1057/jors.2012.120
- Masnadi-Shirazi, H., & Vasconcelos, N. (2011). Cost-sensitive boosting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(2), 294–309. https://doi.org/https://doi.org/10.1109/TPAMI.2010.71
- McNemar, Q. (1947). Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika, 12(2), 153–157. https://doi.org/https://doi.org/10.1007/BF02295996
- Miguéis, V. L., Benoit, D. F., & Van den Poel, D. (2013). Enhanced decision support in credit scoring using bayesian binary quantile regression. Journal of the Operational Research Society, 64(9), 1374–1383. https://doi.org/https://doi.org/10.1057/jors.2012.116
- Nikolaou, N., Edakunni, N., Kull, M., Flach, P., & Brown, G. (2016). Cost-sensitive boosting algorithms: Do we really need them? Machine Learning, 104(2-3), 359–384. https://doi.org/https://doi.org/10.1007/s10994-016-5572-x
- Oreski, S., Oreski, D., & Oreski, G. (2012). Hybrid system with genetic algorithm and artificial neural networks and its application to retail credit risk assessment. Expert Systems with Applications, 39(16), 12605–12617. https://doi.org/https://doi.org/10.1016/j.eswa.2012.05.023
- Óskarsdóttir, M., Bravo, C., Sarraute, C., Vanthienen, J., & Baesens, B. (2019). The value of big data for credit scoring: Enhancing financial inclusion using mobile phone data and social network analytics. Applied Soft Computing, 74, 26–39. https://doi.org/https://doi.org/10.1016/j.asoc.2018.10.004
- Platt, J. (1999). Probabilistic outputs for support vector machines and comparison to regularised likelihood methods. Advances in Large Margin Classifiers, 10(3), 61–74.
- Sheng, V., & Ling, C. (2006). Thresholding for making classifiers cost-sensitive. In Proc. AAAI’06, (Volume 1, pp. 476–481).
- Stripling, E. (2018). Business-Oriented Data Analytics: Advances in Profit-Driven Model Building and Fraud Detection [Doctoral Thesis]. https://lirias.kuleuven.be/2291492?limo=0
- Stripling, E., Broucke, S. v., Antonio, K., Baesens, B., & Snoeck, M. (2018). Profit maximizing logistic model for customer churn prediction using genetic algorithms. Swarm and Evolutionary Computation, 40, 116–130. https://doi.org/https://doi.org/10.1016/j.swevo.2017.10.010
- Thomas, L. C., Edelman, D. B., & Crook, J. N. (2002). Credit scoring and its applications. SIAM.
- Ting, K. (1998). Inducing cost-sensitive trees via instance weighting. In. Proc. PKDD’98, 1510, 139–147.
- Ting, K. (2000a). A comparative study of cost-sensitive boosting algorithms. In Proc. ICML’00, 983–990.
- Ting, K. (2000b). An empirical study of metacost using boosting algorithms. In Proc. ECML’00, 1810, 413–425.
- Verbeke, W., Baesens, B., & Bravo, C. (2017). Profit driven business analytics: A practitioner’s guide to transforming big data into added value. John Wiley and Sons.
- Verbeke, W., Dejaeger, K., Martens, D., Hur, J., & Baesens, B. (2012). New insights into churn prediction in the telecommunication sector: A profit driven data mining approach. European Journal of Operational Research, 218(1), 211–229. https://doi.org/https://doi.org/10.1016/j.ejor.2011.09.031
- Verbraken, T., Bravo, C., Weber, R., & Baesens, B. (2014). Development and application of consumer credit scoring models using profit-based classification measures. European Journal of Operational Research, 238(2), 505–513. https://doi.org/https://doi.org/10.1016/j.ejor.2014.04.001
- Verbraken, T., Verbeke, W., & Baesens, B. (2013). A novel profit maximizing metric for measuring classification performance of customer churn prediction models. IEEE Transactions on Knowledge and Data Engineering, 25(5), 961–973. https://doi.org/https://doi.org/10.1109/TKDE.2012.50
- Viola, P., & Jones, M. (2001). Fast and robust classification using assymetric adaboost and a detector cascade. Advances in Neural Information Processing System, 14, 1311–1318.
- Wirth, R., Hipp, J. (2000). Crisp-dm: Towards a standard process model for data mining. In Proceedings of the 4th International Conference on the Practical Applications of Knowledge Discovery and Data Mining (pp. 29–39).
- Xia, Y., Liu, C., & Liu, N. (2017). Cost-sensitive boosted tree for loan evaluation in peer-to-peer lending. Electronic Commerce Research and Applications , 24, 30–49. https://doi.org/https://doi.org/10.1016/j.elerap.2017.06.004
- Zadrozny, B., Langford, J., & Abe, N. (2003). Cost-sensitive learning by cost-proportionate example weighting. In Proc. ICDM’03, 435–442.