Estimating the Time-Lapse Between Medical Insurance Reimbursement with Non-parametric Regression Models

  • Mary Akinyemi
  • Chika Yinka-BanjoEmail author
  • Ogban-Asuquo Ugot
  • Akwarandu Nwachuku
Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 1130)


Nonparametric supervised learning algorithms represent a succinct class of supervised learning algorithms where the learning parameters are highly flexible and whose values are directly dependent on the size of the training data. In this paper, we comparatively study the properties of four nonparametric algorithms, k-Nearest Neighbours (k-NNs), Support Vector Machines (SVMs), Decision trees and Random forests. The supervised learning task is a regression estimate of the time lapse in medical insurance reimbursement. Our study is concerned precisely with how well each of the nonparametric regression models fits the training data. We quantify the goodness of fit using the R-squared metric. The results are presented with a focus on the effect of the size of the training data, the feature space dimension and hyperparameter optimization. The findings suggest k-NN’s and SVM’s algorithms as better models in predicting well-defined output labels (i.e. Time lapse in days). However, overall, the decision tree model performs better because it makes a better prediction on new data points than the ballpark estimates made from likelihood models: SVMs and k-NNs.


Machine learning Supervised learning Non-parametric learning 


  1. 1.
    Tang, L., Pan, H., Yao, Y.: PANK-A financial time series prediction model integrating principal component analysis, affinity propagation clustering and nested k-nearest neighbor regression. J. Interdiscip. Math. 21, 1–12 (2018)CrossRefGoogle Scholar
  2. 2.
    Sun, J., Fujita, H., Chen, P., Li, H.: Dynamic financial distress prediction with concept drift based on time weighting combined with Adaboost support vector machine ensemble. Knowl.-Based Syst. 120, 4–14 (2017)CrossRefGoogle Scholar
  3. 3.
    Sim, D.Y.Y., Teh, C.S., Ismail, A.I.: Improved boosted decision tree algorithms by adaptive apriori and post-pruning for predicting obstructive sleep apnea. Adv. Sci. Lett. 24(3), 1680–1684 (2018)CrossRefGoogle Scholar
  4. 4.
    Wolpert, D.: The lack of a priori distinctions between learning algorithms. Neural Comput. 8, 1341–1390 (1996)CrossRefGoogle Scholar
  5. 5.
    Caruana, R., Niculescu-Mizil, A.: An empirical comparison of supervised learning algorithms using different performance metrics. In: Proceedings of the 23rd International Conference on Machine Learning, pp. 161–168. Association for Computing Machinery, New York (2006)Google Scholar
  6. 6.
    Breitenbach, M., Dieterich, W., Brennan, T., Fan, A.: Creating risk-scores in very imbalanced datasets: predicting extremely low violent crime among criminal offenders following release from prison. In: Koh, Y.S., Rountree, N. (eds.) Rare association rule mining and knowledge discovery: Technologies for infrequent and critical event detection, pp. 231–254. Information Science Reference, Hershey (2009)Google Scholar
  7. 7.
    Duwe, G., Kim, K.: Out with the old and in with the new? An empirical comparison of supervised learning algorithms to predict recidivism. Crim. Justice Policy Rev. 28(6), 570–600 (2017)CrossRefGoogle Scholar
  8. 8.
    Allyn, J., Allou, N., Augustin, P., Philip, I., Martinet, O., Belghiti, M., Provenchere, S., Montravers, P., Ferdynus, C.: A comparison of a machine learning model with EuroSCORE II in predicting mortality after elective cardiac surgery: a decision curve analysis. PLoS ONE 12(1), e0169772 (2017).
  9. 9.
    Kanchanamani, M., Varalakshmi, P.: Performance evaluation and comparative analysis of various machine learning techniques for diagnosis of breast cancer. Biomed. Res.: Int. J. Med. Sci. 17 February 2016.
  10. 10.
    Handel, B., Kolstad, J.: Health insurance for “humans”: information frictions, plan choice, and consumer welfare. Am. Econ. Rev. 105(8), 2449–2500 (2015)CrossRefGoogle Scholar
  11. 11.
    Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J.: Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12(Oct), 2825–2830 (2011)Google Scholar
  12. 12.
    Kotthoff, L., Thornton, C., Hoos, H., Hutter, F., Leyton-Brown, K.: Auto-WEKA 2.0: automatic model selection and hyperparameter optimization in WEKA. J. Mach. Learn. Res. 18(1), 826–830 (2017)MathSciNetzbMATHGoogle Scholar

Copyright information

© Springer Nature Switzerland AG 2020

Authors and Affiliations

  • Mary Akinyemi
    • 1
  • Chika Yinka-Banjo
    • 1
    Email author
  • Ogban-Asuquo Ugot
    • 1
  • Akwarandu Nwachuku
    • 1
  1. 1.University of LagosLagosNigeria

Personalised recommendations