Abstract
When the marketing service has to contact customers to propose them a product, the probability that these customers will buy this product is calculated beforehand. This probability is calculated using a predictive model. The marketing service contacts the clients having the highest probability of buying the product. In parallel and before the commercial contact it may be interesting to realize a typology of the customers who will be contacted. The idea is to propose differentiated campaigns by group of customers. This article shows how it is possible to build such a typology so that it respects the nearness of the customers with respect to their appetency score.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
References
Blake, C.L., Merz, C.J.: UCI Repository of machine learning databases. http://www.ics.uci.edu/mlearn/MLRepository.html. http://archive.ics.uci.edu/ml/. (1998) Accessed 15 Sept 2010
Boullé, M.: Compression-based averaging of selective naive Bayes classifiers. J. Mach. Learn. Res. 8, 1659–1685 (2007)
Bradley, P.S., Mangasarian, O.L., Street, W.N.: Clustering via concave minimization. In: Advances in Neural Information Processing Systems -9, pp. 368–374. MIT Press, Cambridge (1997)AQ1
Creff, N.: Clustering à l’aide d’une représentation supervisée. Master’s thesis, Epita, rue Voltaire 94276 Kremlin Bicêtre Cedex, pp. 14–16 (2011)
Dai, W., Yang, Q., Xue, G.R., Yu, Y.: Self-taught clustering. In: Proceedings of the 25th International Conference on Machine Learning, pp. 200–207 (2008)
Féraud, R., Boullé, M., Clérot, F., Fessant, F., Lemaire, V.: The orange customer analysis platform. In: Proceedings of the 10th Industrial Conference on Data Mining, pp. 584–594. Springer Verlag, Berlin, Germany (2010)
Ferrandiz, S., Boullé, M.: Bayesian instance selection for the nearest neighbor rule. Mach. Learn. 81(3), 229–256 (2010)
Guyon, I., Lemaire, V., Boullé, M., Dror, G., Vogel, D.: Analysis of the KDD cup 2009: Fast scoring on a large orange customer database. JMLR: Workshop and Conference Proceedings 7, 1–22 (2009). Data available on http://www.kddcup-orange.com
Guyon, I., Cawley, G., Dror, G., Lemaire, V.: Results of the active learning challenge. JMLR W&CP, Workshop on Active Learning and Experimental Design, collocated with AISTATS, Sardinia, Italy, vol. 10, 1–26 (2010)
Har-peled, S., Mazumdar, S.: Coresets for k-means and k-median clustering and their applications. In: Proceedings of the 36th Annual ACM Symposium on Theory of Computing, pp. 291–300. Chicago, Illinois, USA (2003)
Huang, Z.: Clustering large data sets with mixed numeric and categorical values. In: Pacific Asia Knowledge Discovery and Data Mining Conference, pp. 21–34. World Scientific, Singapore (1997)
Huang, Z.: Extensions to the k-means algorithm for clustering large data sets with categorical values. Data Min. Knowl. Discov. 2, 283–304 (1998)
Jajuga, K.: A clustering method based on the l1-norm. Comput. Stat. Data Anal. 5(4), 357–371 (1987)
Kashima, H., Hu, J., Ray, B., Singh, M.: K-means clustering of proportional data using l1 distance. In: 19th International Conference on Pattern Recognition, ICPR 2008, pp. 1–4 (2008)
Kaufman, L., Rousseeuw, P.J.: Finding Groups in Data: An Introduction to Cluster Analysis. Wiley, New York (1990)
Langley, P., Iba, W., Thompson, K.: An analysis of Bayesian classifiers. In: Proceedings of the 10th National Conference on Artificial Intelligence, pp. 223–228. MIT Press, San Jose, California, USA (1992)
MacQueen, J.: Some methods for classification and analysis of multivariate observations. In: 5th Berkeley Symposium on Mathematical Statistics and Probability, vol. 1, pp. 81–297 (1967)
Meila, M., Heckerman, D.: An experimental comparison of several clustering and initialization methods. In: Machine Learning, pp. 386–395 (1998)
Park, H.S., Jun, C.H.: A simple and fast algorithm for k-medoids clustering. Expert Syst. Appl. 36(2), 3336–3341 (2009)
Pelleg, D., Moore, A.W.: X-means: Extending k-means with efficient estimation of the number of clusters. In: Proceedings of the 17th International Conference on Machine Learning, ICML '00, pp. 727–734. Morgan Kaufmann, San Francisco, California, USA (2000)
Quionero-Candela, J., Sugiyama, M., Schwaighofer, A., Lawrence, N.D.: Dataset shift in machine learning. MIT Press (2009)
Zliobaite, I.: Learning under concept drift: An overview. CoRR abs/1010.4784 (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Lemaire, V., Clérot, F., Creff, N. (2015). K-means Clustering on a Classifier-Induced Representation Space: Application to Customer Contact Personalization. In: Abou-Nasr, M., Lessmann, S., Stahlbock, R., Weiss, G. (eds) Real World Data Mining Applications. Annals of Information Systems, vol 17. Springer, Cham. https://doi.org/10.1007/978-3-319-07812-0_8
Download citation
DOI: https://doi.org/10.1007/978-3-319-07812-0_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-07811-3
Online ISBN: 978-3-319-07812-0
eBook Packages: Business and EconomicsBusiness and Management (R0)