Abstract
This article deals with the description of a new way to learn from multiple and heterogeneous data sets, and with the integration of this method in a multi-agent hybrid learning system. This system integrates different kinds of unsupervised classification methods and gives a set of clusterings as the result and a unifying result, representing all the other one. In this new approach, the method occurrences compare their results and automatically refine them to try to make them converge towards a unique clustering that unifies all the results. Thus, the data are not really merged but the results from their classification are compared and refined according to the results from all the other data sets. This enables to produce a set of classification hierarchies which classes are very similar, although these hierarchies were extracted from different data sets. Then it is easy to build a unifying result from all of them.
Similar content being viewed by others
References
Alpaydin E (1998) Techniques for combining multiple learners. In: Alpaydin E (ed) Proceedings of Engineering of Intelligent Systems’98 Conference, vol 2, pp 6–12
Alpaydin E, Kaynak C (1998) Cascading classifiers. Kybernetika 34(4):369–374
Bauer E, Kohavi R (1999) An empirical comparison of voting classification algorithms: bagging, boosting, and variants. Mach Learn 36:105–142
Binaghi E, Pepe M, Radice F (1997) Multistrategy fuzzy learning for multisource remote sensing classifiers. In: Proceedings of the International Society for Optical Engineering, vol 3217, pp 306–317
Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140
Breiman L (1998) Arcing classifiers. Ann Stat 26(3):801–849
Fisher DH (1987) Knowledge acquisition via incremental conceptual clustering. Mach Learn 2:139–172
Frigui H, Nasraoui O (2004) Unsupervised learning of prototypes and attribute weights. Pattern Recogn 34:567–581
Gama J (1998) Combining classifiers with constructive induction. In: Proceedings of the 10th European Conference on Machine Learning (ECML-98). LNAI, vol 1398. Springer, Berlin, pp 178–189, 21–23 April 1998
Grabmeier J, Rudolph A (2002) Techniques of cluster algorithms in data mining. Data Min Knowl Disc 6(4):303–360
Halkidi M, Batistakis Y, Vazirgiannis M (2001) On clustering validation techniques. J Intel Inf Syst 17(2,3):107–145
Halkidi M, Batistakis Y, Vazirgiannis M (2002) Cluster validity methods. SIGMOD 31(2):40–45
Hammadi-Mesmoudi F, Korczak JJ (1995) An unsupervised neural network classifier and its application in remote sensing. In: Proceedings of International Conference on Image Processing
Hartigan J (1979) Algorithm AS136: a k-means clustering algorithm. 28:100–108
He J, Tan A-H, Tan C-L, Sung S-Y (2002) On quantitative evaluation of clustering systems. In: Wu W, Xiong H (eds) Information Retrieval and Clustering. Kluwer, Boston, MA
Jain AK, Murty MN, Flynn PJ (1999) Data clustering: a review. ACM Comput Surv 31(3): 264–323
Kittler J (1998) On combining classifiers. IEEE Trans Pattern Anal Mach Intell 20(3):226–239
Kohonen T (1990) The self-organizing map. Proc IEEE 78(9):1464–1480
Korczak JJ, Blamont D, Ketterlin A (1994) Thematic image segmentation by a concept formation algorithm. In: Proceedings of the European Symposium on Satelite Remote Sensing, Rome
Korczak JJ, Rymarczyk M (1993) Application of classical clustering methods to digital image analysis. Technical report, CRI
Paredis J (1997) Coevolving cellular automata: be aware of the red queen! In: 7th Int. Conference on Genetic Algorithms (ICGA 97), pp 393–400
Quinlan JR (1996) Boosting first-order learning. In: Arikawa S, Sharma AK (eds) Proceedings of the 7th International Workshop on Algorithmic Learning Theory. LNAI, vol 1160. Springer, Berlin, pp 143–155, 23–25 October 1996
Roli F, Giacinto G, Vernazza G (2001) Methods for designing multiple classifier systems. In: Lecture notes in Computer Science, vol 2096
Schapire RE (1999) Theoretical views of boosting. In: Fischer P, Simon HU (eds) Proceedings of the 4th European Conference on Computational Learning Theory (COLT-99). LNAI, vol 1572. Springer, Berlin, pp 1–10, 29–31 March 1999
Ting KM, Low BT, Witten IH (1999) Learning from batched data: model combination versus data combination. Knowl Inf Syst 1(1):83–106, February 1999
Ting KM, Witten IH (1999) Issues in stacked generalization. J Artif Intell Res 10:271–289
Wemmert C, Gançarski P (2002) A multiview voting method to combine unsupervised classifications. In: Proceedings of the 2nd IASTED International Conference on Artificial Intelligence and Applications, AIA 2002, pp 447–453
Wemmert C, Gançarski P (2002) Urban thematical zones construction from remote sensing data by unsupervised classification. In: Proceedings of the 23rd Urban Data Management Symposium, UDMS 2002
Wemmert C, Gançarski P, Korczak JJ (2000) A collaborative approach to combine multiple learning methods. Int J Artif Intell Tools 9(1):59–78, March
Wolpert DH (1990) Stacked generalization. Technical Report LA-UR-90-3460, Complex Systems Group, Theoretical Division, and Center for Non-linear Studies, Los Alamos, NM
Xie X, Beni G (1991) A validity measure for fussy clustering. IEEE Trans Pattern Anal Mach Intell 13(4):841–847
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Gançarski, P., Wemmert, C. Collaborative multi-step mono-level multi-strategy classification. Multimed Tools Appl 35, 1–27 (2007). https://doi.org/10.1007/s11042-007-0115-x
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-007-0115-x