Abstract
Machine learning builds mathematical models from data containing multiple attributes (i.e., variables) in order to predict some variables from others. For example, in a cancer prediction application, each data point might contain the variables obtained from running clinical tests, whereas the predicted variable might be a binary diagnosis of cancer. Such models are sometimes expressed as linear and nonlinear relationships between variables. These relationships are discovered in a data-driven manner by optimizing (maximizing) the “agreement” between the models and the observed data. This is an optimization problem.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Instead of referring to such matrices as rectangular diagonal matrices, some authors use a quotation around the word diagonal, while referring to such matrices. This is because the word “diagonal” was originally reserved for square matrices.
References
C. Aggarwal. Data mining: The textbook. Springer, 2015.
C. Aggarwal. Machine learning for text. Springer, 2018.
C. Aggarwal. Recommender systems: The textbook. Springer, 2016.
C. Aggarwal. Outlier analysis. Springer, 2017.
M. Bazaraa, H. Sherali, and C. Shetty. Nonlinear programming: theory and algorithms. John Wiley and Sons, 2013.
D. Bertsekas. Nonlinear programming. Athena scientific, 1999.
D. Bertsimas and J. Tsitsiklis. Introduction to linear optimization. Athena Scientific, 1997.
C. M. Bishop. Pattern recognition and machine learning. Springer, 2007.
C. M. Bishop. Neural networks for pattern recognition. Oxford University Press, 1995.
S. Boyd and L. Vandenberghe. Convex optimization. Cambridge University Press, 2004.
S. Boyd and L. Vandenberghe. Applied linear algebra. Cambridge University Press, 2018.
R. Duda, P. Hart, and D. Stork. Pattern classification. John Wiley and Sons, 2012.
P. Flach. Machine learning: the art and science of algorithms that make sense of data. Cambridge University Press, 2012.
G. Golub and C. F. Van Loan. Matrix computations, John Hopkins University Press, 2012.
I. Goodfellow, Y. Bengio, and A. Courville. Deep learning. MIT Press, 2016.
T. Hastie, R. Tibshirani, and J. Friedman. The elements of statistical learning. Springer, 2009.
K. Hoffman and R. Kunze. Linear algebra, Second Edition, Pearson, 1975.
D. Lay, S. Lay, and J. McDonald. Linear Algebra and its applications, Pearson, 2012.
S. Marsland. Machine learning: An algorithmic perspective, CRC Press, 2015.
T. Mitchell. Machine learning, McGraw Hill, 1997.
K. Murphy. Machine learning: A probabilistic perspective, MIT Press, 2012.
J. Nocedal and S. Wright. Numerical optimization. Springer, 2006.
J. Solomon. Numerical Algorithms: Methods for Computer Vision, Machine Learning, and Graphics. CRC Press, 2015.
G. Strang. An introduction to linear algebra, Fifth Edition. Wellseley-Cambridge Press, 2016.
G. Strang. Linear algebra and its applications, Fourth Edition. Brooks Cole, 2011.
G. Strang. Linear algebra and learning from data. Wellesley-Cambridge Press, 2019.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Aggarwal, C.C. (2020). Linear Algebra and Optimization: An Introduction. In: Linear Algebra and Optimization for Machine Learning. Springer, Cham. https://doi.org/10.1007/978-3-030-40344-7_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-40344-7_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-40343-0
Online ISBN: 978-3-030-40344-7
eBook Packages: Computer ScienceComputer Science (R0)