Abstract
Three separate sections comprise this chapter. The first presents an overview of statistical learning theory (SLT) as applied to machine learning. The topics covered are empirical or true risk minimization, the risk minimization principle (RMP), theoretical concept of risk minimization, function f0(X) that minimizes the expected (or true) risk, asymptotic consistency or uniform convergence, an example of the generalized bound for binary classification and finally, how are learning machines formed.
The second part of the chapter covers the theory of support vector machine (SVM) learning theory and develops solutions for the following three classes of problems:
-
Linear separable systems
-
Linear non-separable systems
-
And non-linear, non-separable systems
It then introduces the topic of kernels, what they are, and how they might be chosen. A brief pointer is provided to the SVM literature available on the web.
These sections are followed by a sketch of how the SVM may be hybridized with the GA for feature subset selection and points the way to the value of further hybridization with an ensemble approach, the topic of the next chapter.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
See Chap. 3.
Abbreviations
- ERM:
-
Empirical risk minimization
- GA:
-
Genetic algorithm
- QP:
-
Quadratic programming
- RBF:
-
Radial basis function
- RMP:
-
Risk minimization principle
- SLT:
-
Statistical learning theory
- SMO:
-
Sequential minimization optimization
- SRM:
-
Structured risk minimization
- SVM:
-
Support vector machine
- VC:
-
Vapnik Chervonenkis
References
Azerman A, Bowerna EM (1964) Theoretical formulation of the potential function method in pattern recognition. Autom Remote Control, (Automat I Telemekh) 25:917–932
Mercer J (1909) Functions of positive and negative type, and their connection with the theory of integral equations. Philos Trans R Soc Lond A 209:415–446
Rozložník M (2018) Saddle point problems and their iterative solution. Springer, New York
Vapnik VN, Chervonenkis AY (1974) Teoriya raspoznavaniya obrazov: Statisticheskie problemy obucheniya. (Russian) [Theory of pattern recognition: Statistical problems of learning]. Moscow: Nauka.
Vapnik VN (1995) The nature of statistical learning theory. Springer, New York
Vapnik VN (1998) Statistical learning theory. Wiley, New York
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Land, W.H., Schaffer, J.D. (2020). The Support Vector Machine. In: The Art and Science of Machine Intelligence. Springer, Cham. https://doi.org/10.1007/978-3-030-18496-4_2
Download citation
DOI: https://doi.org/10.1007/978-3-030-18496-4_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-18495-7
Online ISBN: 978-3-030-18496-4
eBook Packages: EngineeringEngineering (R0)