1,844
Views
2
CrossRef citations to date
0
Altmetric
Articles

Cost-sensitive regression learning on small dataset through intra-cluster product favoured feature selection

, , &
Pages 104-123 | Received 03 Jun 2021, Accepted 10 Aug 2021, Published online: 30 Aug 2021

References

  • Anthony, M. H. G., & Biggs, N. (1997). Computational learning theory. Cambridge University Press.
  • Bansal, G., Sinha, A. P., & Zhao, H. (2008, December). Tuning data mining methods for cost-sensitive regression: A study in loan charge-off forecasting. Journal of Management Information Systems, 25(3), 315–336. https://doi.org/10.2753/MIS0742-1222250309
  • Baumgartner, R., Somorjai, R., Bowman, C., Sorrell, T., Mountford, C., & Himmelreich, U. (2004, February). Unsupervised feature dimension reduction for classification of MR spectra. Magnetic Resonance Imaging, 22(2), 251–256. https://doi.org/10.1016/j.mri.2003.08.033
  • Boehmke, B., & Greenwell, B. (2019, November). Random forests. In Hands-on machine learning with R (pp. 203–219). Chapman and Hall/CRC.
  • Carbonneau, R. A., Kersten, G. E., & Vahidov, R. M. (2011, January). Pairwise issue modeling for negotiation counteroffer prediction using neural networks. Decision Support Systems, 50(2), 449–459. https://doi.org/10.1016/j.dss.2010.11.002
  • Chandrashekar, G., & Sahin, F. (2014, January). A survey on feature selection methods. Computers and Electrical Engineering, 40(1), 16–28. https://doi.org/10.1016/j.compeleceng.2013.11.024
  • Chang, C. J., Dai, W. L., & Chen, C. C. (2015, November). A novel procedure for multimodel development using the grey silhouette coefficient for small-data-set forecasting. Journal of the Operational Research Society, 66(11), 1887–1894. https://doi.org/10.1057/jors.2015.17
  • Chang, C. J., Li, D. C., Chen, C. C., & Chen, W. C. (2019, March). Extrapolation-based grey model for small-data-set forecasting. Economic Computation and Economic Cybernetics Studies and Research, 53(1), 171–182. https://doi.org/10.24818/18423264
  • Chang, C. J., Li, D. C., Dai, W. L., & Chen, C. C. (2014). A latent information function to extend domain attributes to improve the accuracy of small-data-set forecasting. Neurocomputing, 129, 343–349. https://doi.org/10.1016/j.neucom.2013.09.024
  • Chen, W. S., Yuen, P. C., & Huang, J. (2005, November). A new regularized linear discriminant analysis method to solve small sample size problems. International Journal of Pattern Recognition and Artificial Intelligence, 19(7), 917–935. https://doi.org/10.1142/S0218001405004344
  • Crone, S. F. (2002). Training artificial neural networks for time series prediction using asymmetric cost functions. In L. Wang, J. C. Rajapakse, K. Fukushima, S.-Y. Lee, & X. Yao (Eds.), Proceedings of the 9th international conference on Neural Information Processing, ICONIP'02 (pp. 2374–2380). Springer.
  • Crone, S. F. (2010). Neuronale netze zur prognose und disposition im handel. Gabler.
  • Crone, S. F., Lessmann, S., & Stahlbock, R. (2005). Utility based data mining for time series analysis. In Proceedings of the 1st international workshop on utility-based data mining-UBDM'05 (pp. 59–68). Knowledge Discovery and Data Mining.
  • Cui, G., Wong, M. L., & Wan, X. (2014). Cost-sensitive learning via priority sampling to improve the return on marketing and CRM investment. Journal of Management Information Systems, 29(1), 341–374. https://doi.org/10.2753/MIS0742-1222290110
  • Dai, Y., & Wang, T. (2021). Prediction of customer engagement behaviour response to marketing posts based on machine learning. Connection Science, 33(4), 891–910. https://doi.org/10.1080/09540091.2021.1912710
  • Di Nunzio, G. M. (2014, September). A new decision to take for cost-sensitive Naïve Bayes classifiers. Information Processing and Management, 50(5), 653–674. https://doi.org/10.1016/j.ipm.2014.04.008
  • Guo, Z. G., Gao, X. G., Ren, H., Yang, Y., Di, R. H., & Chen, D. Q. (2017, December). Learning bayesian network parameters from small data sets: A further constrained qualitatively maximum a posteriori method. International Journal of Approximate Reasoning, 91, 22–35. https://doi.org/10.1016/j.ijar.2017.08.009
  • Huang, C., & Moraga, C. (2004, February). A diffusion-neural-network for learning from small samples. International Journal of Approximate Reasoning, 35(2), 137–161. https://doi.org/10.1016/j.ijar.2003.06.001
  • Iranmehr, A., Masnadi Shirazi, H., & Vasconcelos, N. (2019, May). Cost-sensitive support vector machines. Neurocomputing, 343, 50–64. https://doi.org/10.1016/j.neucom.2018.11.099
  • Kim, K. J. (2003, September). Financial time series forecasting using support vector machines. Neurocomputing, 55(1–2), 307–319. https://doi.org/10.1016/S0925-2312(03)00372-2
  • Kursa, M. B., & Rudnicki, W. R. (2010). Feature selection with the Boruta Package. Journal of Statistical Software, 36(11). https://doi.org/10.18637/jss.v036.i11
  • Lessmann, S., & Voß, S. (2017, October). Car resale price forecasting: The impact of regression method, private information, and heterogeneity on forecast accuracy. International Journal of Forecasting, 33(4), 864–877. https://doi.org/10.1016/j.ijforecast.2017.04.003
  • Li, D. C., Chang, C. C., & Liu, C. W. (2012, February). Using structure-based data transformation method to improve prediction accuracies for small data sets. Decision Support Systems, 52(3), 748–756. https://doi.org/10.1016/j.dss.2011.11.021
  • Li, D. C., Chen, L. S., & Lin, Y. S. (2003, January). Using functional virtual population as assistance to learn scheduling knowledge in dynamic manufacturing environments. International Journal of Production Research, 41(17), 4011–4024. https://doi.org/10.1080/0020754031000149211
  • Li, D. C., Chen, W. C., Liu, C. W., & Lin, Y. S. (2010, August). A non-linear quality improvement model using SVR for manufacturing TFT-LCDs. Journal of Intelligent Manufacturing, 23(3), 835–844. https://doi.org/10.1007/s10845-010-0440-1
  • Li, D. C., & Liu, C. W. (2009, August). A neural network weight determination model designed uniquely for small data set learning. Expert Systems with Applications, 36(6), 9853–9858. https://doi.org/10.1016/j.eswa.2009.02.004
  • Li, D. C., & Liu, C. W. (2012, March). Extending attribute information for small data set classification. IEEE Transactions on Knowledge and Data Engineering, 24(3), 452–464. https://doi.org/10.1109/TKDE.2010.254
  • Li, D. C., Liu, C. W., & Chen, W. C. (2012, December). A multi-model approach to determine early manufacturing parameters for small-data-set prediction. International Journal of Production Research, 50(23), 6679–6690. https://doi.org/10.1080/00207543.2011.613867
  • Li, D. C., & Wen, I. H. (2014, November). A genetic algorithm-based virtual sample generation technique to improve small data set learning. Neurocomputing, 143, 222–230. https://doi.org/10.1016/j.neucom.2014.06.004
  • Li, W., Ding, S., Chen, Y., Wang, H., & Yang, S. (2018, September). Transfer learning-based default prediction model for consumer credit in China. The Journal of Supercomputing, 75(2), 862–884. https://doi.org/10.1007/s11227-018-2619-8
  • Lin, Y. F., & Chou, J. (2020). Supply chain with several competing retailers and one manufacturer. International Journal on Computer Science and Engineering, 12(1), 1–6.
  • Liu, H., & Setiono, R. (1997). Feature selection and classification: A probabilistic wrapper approach. Proceedings of 9th international conference on Industrial and Engineering Applications of AI and ES (pp. 419–424).
  • Liu, Y., Yu, X., Huang, J. X., & An, A. (2011, July). Combining integrated sampling with SVM ensembles for learning from imbalanced datasets. Information Processing and Management, 47(4), 617–631. https://doi.org/10.1016/j.ipm.2010.11.007
  • Liu, Z., Japkowicz, N., Wang, R., & Tang, D. (2019). Adaptive learning on mobile network traffic data. Connection Science, 31(2), 185–214. https://doi.org/10.1080/09540091.2018.1512557
  • Luo, H., & Paal, S. G. (2021, September). Reducing the effect of sample bias for small data sets with double–weighted support vector transfer regression. Computer-Aided Civil and Infrastructure Engineering, 36(3), 248–263. https://doi.org/10.1111/mice.12617
  • Mao, R., Zhu, H., Zhang, L., & Chen, A. (2006, October, 16–18). A new method to assist small data set neural network learning. 6th international conference on Intelligent Systems Design and Applications, Jian, China. IEEE. https://doi.org/10.1109/ISDA.2006.67
  • Micallef, L., Sundin, I., Marttinen, P., Ammad-ud din, M., Peltola, T., Soare, M., Jacucci, G., & Kaski, S. (2017, March). Interactive elicitation of knowledge on feature relevance improves predictions in small data sets. Proceedings of the 22nd international conference on Intelligent User Interfaces (pp. 547–552).
  • Mishra, N. K., & Singh, P. K. (2020, July). FS-MLC: Feature selection for multi-label classification using clustering in feature space. Information Processing and Management, 57(4), 102240–102264. https://doi.org/10.1016/j.ipm.2020.102240
  • Mohammadi, F. S., & Amiri, A. (2019). TS-WRSVM: Twin structural weighted relaxed support vector machine. Connection Science, 31(3), 215–243. https://doi.org/10.1080/09540091.2019.1573418
  • Niyogi, P., Girosi, F., & Poggio, T. (1998). Incorporating prior information in machine learning by creating virtual examples. Proceedings of the IEEE, 86(11), 2196–2209. https://doi.org/10.1109/5.726787
  • Nunzio, G. M. D. (2014). A new decision to take for cost-sensitive Naïve Bayes classifiers. Information Processing and Management: An International Journal, 50(5), 653–674. https://doi.org/10.1016/j.ipm.2014.04.008
  • Onan, A. (2017). Hybrid supervised clustering based ensemble scheme for text classification. Kybernetes the International Journal of Systems and Cybernetics, 46(2), 330–348. https://doi.org/10.1108/K-10-2016-0300
  • Onan, A. (2018a). An ensemble scheme based on language function analysis and feature engineering for text genre classification. Journal of Information Science: Principles and Practice, 44(1), 28–47. https://doi.org/10.1177/0165551516677911
  • Onan, A. (2018b). Sentiment analysis on twitter based on ensemble of psychological and linguistic feature sets. Balkan Journal of Electrical and Computer Engineering, 6(2), 69–77. https://doi.org/10.17694/bajece.419538
  • Onan, A., Bulut, H., & Korukoğlu, S. (2016). Ensemble of keyword extraction methods and classifiers in text classification. Expert Systems with Application, 57, 232–247. https://doi.org/10.1016/j.eswa.2016.03.045
  • Onan, A., & Korukoğlu, S. (2016). Exploring performance of instance selection methods in text sentiment classification. In R. Silhavy, R. Senkerik, Z. Oplatkova, P. Silhavy, & Z. Prokopova (Eds.), Artificial intelligence perspectives in intelligent systems (Vol. 464, Advances in Intelligent Systems and Computing, pp. 167–179). Springer. https://doi.org/10.1007/978-3-319-33625-1_16
  • Onan, A., & Korukoğlu, S. (2017). A feature selection model based on genetic rank aggregation for text sentiment classification. Journal of Information Science, 43(1), 25–38. https://doi.org/10.1177/0165551515613226
  • Onan, A., & Tocoglu, M. A. (2020). Satire identification in Turkish news articles based on ensemble of classifiers. Turkish Journal of Electrical Engineering and Computer Sciences, 28(2), 1086–1106. https://doi.org/10.3906/elk-1907-11
  • Phani Narasimham, M. V. S., & Sai Pragathi, Y. V. S. (2019). Development of realistic models of oil well by modeling porosity using modified ANFIS technique. International Journal on Computer Science and Engineering, 11(7), 34–39.
  • Too, J., & Abdullah, A. R. (2020). Binary atom search optimisation approaches for feature selection. Connection Science, 32(4), 406–430. https://doi.org/10.1080/09540091.2020.1741515
  • Tsou, Y. L., & Lin, H. T. (2019, April). Annotation cost-sensitive active learning by tree sampling. Machine Learning, 108(5), 785–807. https://doi.org/10.1007/s10994-019-05781-7
  • Verma, N. K., & Salour, A. (2020). Feature selection. Studies in Systems, Decision and Control, 256, 175–200. https://doi.org/10.1007/978-981-15-0512-6_5
  • Wang, S., Zhao, H., & Nai, K. (2021). Learning a maximized shared latent factor for cross-modal hashing. Knowledge-Based Systems, 228, 107252. https://doi.org/10.1016/j.knosys.2021.107252
  • Wen, G., Li, X., Zhu, Y., Chen, L., Luo, Q., & Tan, M. (2021, January). One-step spectral rotation clustering for imbalanced high-dimensional data. Information Processing and Management, 58(1), 102388–102404. https://doi.org/10.1016/j.ipm.2020.102388
  • Ye, H., & Luo, Z. C. (2020). Deep ranking based cost-sensitive multi-label learning for distant supervision relation extraction. Information Processing & Management, 57(6), 102096–102110.
  • Zhang, J., & García, J. (2015, April). Online classifier adaptation for cost-sensitive learning. Neural Computing and Applications, 27(3), 781–789. https://doi.org/10.1007/s00521-015-1896-x
  • Zhao, H., Cao, J., Xu, M., & Lu, J. (2020, January). Variational neural decoder for abstractive text summarization. Computer Science and Information Systems, 17(2), 537–552. https://doi.org/10.2298/CSIS200131012Z
  • Zhao, H., Sinha, A. P., & Bansal, G. (2011, June). An extended tuning method for cost-sensitive regression and forecasting. Decision Support Systems, 51(3), 372–383. https://doi.org/10.1016/j.dss.2011.01.003
  • Zhou, Y., Li, T., Shi, J., & Qian, Z. (2019, February). A ceemdan and XGboost-based approach to forecast crude oil prices. Complexity, 2019, 4392785. https://doi.org/10.1155/2019/4392785
  • Zhou, Z. H., & Liu, X. Y. (2006, January). Training cost-sensitive neural networks with methods addressing the class imbalance problem. IEEE Transactions on Knowledge and Data Engineering, 18(1), 63–77. https://doi.org/10.1109/TKDE.2006.17