Abstract
In this paper, the problem of simultaneously approximating a function and its derivatives is formulated. First, the problem is solved for a one-dimensional input space by using the least square support vector machines and introducing additional constraints in the approximation of the derivative. To optimize the regression estimation problem, we have derived an algorithm that works fast and more accuracy for moderate-size problems. The proposed method shows that using the information about derivatives significantly improves the reconstruction of the function.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Gallant, A.R., White, H.: On learning the derivatives of an unknown mapping with multilayer feedforward networks. Neural Networks 5, 129–138 (1992)
Hornik, K., Stinchcombe, M., White, H.: Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks. Neural Networks 3, 551–560 (1990)
Li, X.: Simultaneous approximations of multivariate functions and their derivatives by neural networks with one hidden layer. Neurocomputing 12, 327–343 (1996)
Nguyen-Thien, T., Tran-Cong, T.: Approximation of functions and their derivatives: a neural network implementation with applications. Neurocomputing 23, 687–704 (1999)
Lázaro, M., SantamarÃa, I., Pantaleón, C., Ibáñez, J., Vielva, L.: A regularized technique for the simultaneous reconstruction of a function and its derivatives with application to nonlinear transistor modeling. Signal Processing 83, 1859–1870 (2003)
Schölkopf, B., Smola, A.: Learning with Kernels. MIT Press, Cambridge (2002)
Vapnik, V.N.: Statistical Learning Theory. Wiley, New York (1998)
Lázaro, M., SantamarÃa, I., et al.: Support Vector Regression for the simultaneous learning of a multivariate function and its derivatives. Neurocomputing 67, 42–61 (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, R., Liu, G. (2011). Least Square Support Vector Machine for the Simultaneous Learning of a Function and Its Derivative. In: Shen, G., Huang, X. (eds) Advanced Research on Electronic Commerce, Web Application, and Communication. ECWAC 2011. Communications in Computer and Information Science, vol 143. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-20367-1_69
Download citation
DOI: https://doi.org/10.1007/978-3-642-20367-1_69
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-20366-4
Online ISBN: 978-3-642-20367-1
eBook Packages: Computer ScienceComputer Science (R0)