Cross-validation is a statistical method of evaluating and comparing learning algorithms by dividing data into two segments: one used to learn or train a model and the other used to validate the model. In typical cross-validation, the training and validation sets must cross over in successive rounds such that each data point has a chance of being validated against. The basic form of cross-validation is k-fold cross-validation. Other forms of cross-validation are special cases of k-fold cross-validation or involve repeated rounds of k-fold cross-validation.
- 1.Bouckaert RR. Choosing between two learning algorithms based on calibrated tests. In: Proceedings of the 20th International Conference on Machine Learning; 2003. p. 51–8.Google Scholar
- 5.Kohavi R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In: Proceedings of the 14th International Joint Conference on AI; 1995. p. 1137–45.Google Scholar
- 8.Mosteller F, Tukey JW. Data analysis, including statistics. In: Handbook of social psychology. Reading: Addison-Wesley; 1968.Google Scholar
- 10.Refaeilzadeh P, Tang L, Liu H. On comparison of feature selection algorithms. In: Proceedings of AAAI-07 Workshop on Evaluation Methods in Machine Learning II; 2007. p. 34–9.Google Scholar
- 14.Zafarani R, Abbasi MA, Liu H. Social media mining: an introduction. Cambridge University Press, New York; 2014.Google Scholar