Abstract
In this chapter we do violence to some problems to reveal their inner structure. The focus is on problems which, at first glance, may not seem to be of the continuous linear variety yet can be marshalled into that form with a handful of creative alterations. The key is to ensure a one-to-one correspondence between the original and the altered problems so that we can retrieve a solution to the original from a solution to the alteration.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
All research mathematicians agree on the labels “convex” and its opposite, “concave,” but textbook authors for high schools in the US, ignoring thousands of papers, journals, and research monographs, insist on “concave up” and “concave down.”
- 2.
The expression “regression” comes from Francis Galton’s original paper about “Re-gression towards mediocrity” and shadows rather than highlights the technique. As for “parameter,” what, pray tell, is not a parameter?
- 3.
Carl Friedrich Gauss, Theoria Combinationis Observationum Erroribus Minimis Obnoxiae (Theory of the Combination of Observations Least Subject to Errors) (Philadelphia, PA: Society for Industrial and Applied Mathematics, 1987).
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2018 Serge Kruk
About this chapter
Cite this chapter
Kruk, S. (2018). Hidden Linear Continuous Models. In: Practical Python AI Projects. Apress, Berkeley, CA. https://doi.org/10.1007/978-1-4842-3423-5_3
Download citation
DOI: https://doi.org/10.1007/978-1-4842-3423-5_3
Published:
Publisher Name: Apress, Berkeley, CA
Print ISBN: 978-1-4842-3422-8
Online ISBN: 978-1-4842-3423-5
eBook Packages: Professional and Applied ComputingApress Access BooksProfessional and Applied Computing (R0)