Abstract
The objective of this book is to establish a framework which combines two different paradigms in machine learning: global learning and local learning. The combined model demonstrates that a hybrid learning of these two different schools of approaches can outperform each isolated approach both theoretically and empirically. Global learning focuses on describing a phenomenon or modeling data in a global way. For example, a distribution over the variables is usually estimated for summarizing the data. Its output can usually reconstruct the data. This school of approaches, including Bayesian Networks [8, 13, 30], Gaussian Mixture Models [3, 21], and Hidden Markov Models [2, 25], has a long an distinguished history, which has been extensively applied in artificial intelligence [26], pattern recognition [9], and computer vision [7]. On the other hand, local learning does not intend to summarize a phenomenon, but builds learning systems by concentrating on some local parts of data. It lacks the flexibility yet surprisingly demonstrates superior performance to global learning according to recent researches [4, 16, 15]. In this book, a bridge has been established between these two different paradigms. Moreover, the resulting principled framework subsumes several important models, which respectively locate themselves into the global learning paradigm and the local learning paradigm.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Barber CB, Dobkin DP, Huhanpaa H (1996) The quickhull algorithm for convex hulls. ACM Transactions on Mathematical Software 22(4): 469–483
Baum LE, Egon JA (1967) An inequality with applications to statistical estimation for probabilistic functions of a Markov process and to a model for ecology. Bull. Amer. Meteorol. Soc. 73: 360C–363
Bozdogan H (2004) Statistical Data Mining and Knowledge Discovery. Boca Raton, Fla.: Chapman & Hall/CRC
Christopher J, Burges C (1998) A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery 2(2): 121–167
Chow CK, Liu CN (1968) Approximating discrete probability distributions with dependence trees. IEEE Trans. on Information Theory 14: 462–467
Duda R, Hart P (1973) Pattern Classification and Scene Analysis. New York, NY: John Wiley & Sons
Forsyth DA, Ponce J (2003) Computer Vision: A Modern Approach. Upper Saddle River, N. J.: Prentice Hall
Friedman N, Geiger D, Goldszmidt M (1997) Bayesian network classifiers. Machine Learning 29: 131–161
Fukunaga K (1990) Introduction to Statistical Pattern Recognition. San Diego, Academic Press, 2nd edition
Girosi F (1998) An equivalence between sparse approximation and support vector machines. Neural Computation 10(6): 1455–1480
Gonzalez MG, Thomason RC (1978) Syntactic Pattern Recognition: An Introduction. Reading, Mass.: Addison-Wesley Pub. Co., Advanced Book Program
Grzegorzewski P, Hryniewicz O, Gil M (2002) Soft Methods in Probability, Statistics and Data Analysis. Heidelberg; New York: Physica-Verlag
Hackman D, Meek C, Cooper G (1995) A tutorial on learning bayesian networks. In Tech Report MSR-TR-95-06. Microsoft Research
Huang K, King I, Lyu MR (2003) Discriminative training of Bayesian chow-liu tree multinet classifiers. In Proceedings of International Joint Conference on Neural Network (IJCNN-2003), Oregon, Portland, U.S.A. 1: 484–488
Jaakkola TS, Haussler D (1998) Exploiting generative models in discriminative classifiers. In Advances in Neural Information Processing Systems (NIPS)
Jebara T (2002) Discriminative, Generative and Imitative Learning. PhD thesis, Massachusetts Institute of Technology
Jordan MI (1998) Learning in Graphical Models. Kluwer Academic Publishers
Toussaint GT, Jaromczyk JW (1992) Relative neighborhood graphs and their relatives. Proceedings IEEE 80(9): 1502–1517
Lanckriet GRG, Ghaoui LE, Bhattacharyya C, Jordan MI (2002) A robust minimax approach to classification. Journal of Machine Learning Research 3: 555–582
Marshall AW, Olkin I (1960) Multivariate Chebyshev inequalities. Annals of Mathematical Statistics 31(4): 1001–1014
McLachlan GJ, Basford KE (1988) Mixture Models: Inference and Applications to Clustering. New York, NY: Marcel Dekker Inc
Pearl J (1988) Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. San Francisco, CA: Morgan Kaufmann
Popescu I, Bertsimas D (2001) Optimal inequalities in probability theory: A convex optimization approach. Technical Report TM62, INSEAD
Pruessner A (2003) Conic programming in GAMS. In Optimization Software-The State of the Art. INFORMS Atlanta, http://www.gamsworld.org/cone/links.htm
Rabiner LR (1989) A tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE 77(2): 257–286
Russell SJ, Norvig P (1995) Artificial Intelligence: A Modern Approach. Englewood Cliffs, N.J.: Prentice Hall
Schölkopf B, Smola A (2002) Learning with Kernels. Cambridge, MA: The MIT Press
Smola AJ, Bartlett PL, Scholkopf B, Schuurmans D (2000) Advances in Large Margin Classifiers. MA: The MIT Press
Sturm JF(1999) Using sedumi 1.02, a matlab toolbox for optimization over symmetric cones. Optimization Methods and Software 11: 625–653
Thiesson B, Meek C, Heckman D (1998). Learning mixtures of Bayesian networks. In Technique Report, MSR-TR-97-30. Microsoft Research
Vapnik VN (1998). Statistical Learning Theory. John Wiley & Sons
Weizenbaum J (1966). Eliza—a computer program for the study of natural language communication between man and machine. Communications of the Association for Computing Machinery
Yedidia J, Freeman WT, Weiss Y (2000). Generalized belief propogation. In Neural Information Processing Systems 13
Zhang W, King I (2002) A study of the relationship between support vector machine and Gabriel Graph. In Proceedings of IEEE World Congress on Computational Intelligence—International Joint Conference on Neural Networks
Rights and permissions
Copyright information
© 2008 Zhejiang University Press, Hangzhou and Springer-Verlag GmbH Berlin Heidelberg
About this chapter
Cite this chapter
(2008). Introduction. In: Machine Learning. Advanced Topics in Science and Technology in China. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-79452-3_1
Download citation
DOI: https://doi.org/10.1007/978-3-540-79452-3_1
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-79451-6
Online ISBN: 978-3-540-79452-3
eBook Packages: Computer ScienceComputer Science (R0)