Abstract
The Nyström method is an efficient technique for large-scale kernel learning. It provides a low-rank matrix approximation to the full kernel matrix. The quality of Nyström approximation largely depends on the choice of landmark points. While standard method uniformly samples columns of the kernel matrix, improved sampling techniques have been proposed based on ensemble learning [1] and clustering [2]. These methods are focused on minimizing the approximation error for the original kernel. In this paper, we take a different perspective by minimizing the approximation error for the input vectors instead. We show under some restrictive condition that the new formulation is equivalent to the standard Nyström solution. This leads to a novel approach for optimizing landmark points for the Nyström approximation. Experimental results demonstrate the superior performance of the proposed landmark optimization method compared to existing Nyström methods in terms of lower approximation errors obtained.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Kumar, S., Mohri, M., Talwalkar, A.: Sampling methods for the nyström method. Journal of Machine Learning Research 13, 981–1006 (2012)
Zhang, K., Kwok, J.T.: Clustered nyström method for large scale manifold learning and dimension reduction. IEEE Transactions on Neural Networks 21(10), 1576–1587 (2010)
Scholkopf, B., Smola, A.J.: Learning with Kernels: Support Vector Machines, Regularization, Optimization and Beyond. MIT Express (2002)
Williams, C.K.I., Seeger, M.: Using the nyström method to speed up kernel machines. In: NIPS (2001)
Drineas, P., Mahoney, M.W.: On the nyström method for approximating a gram matrix for improved kernel-based learning. Journal of Machine Learning Research 6, 2153–2175 (2005)
Bonnans, J.F., Shapiro, A.: Optimization problems with pertubation: A guided tour. SIAM Review 40(2), 202–227 (1998)
Rakotomamonjy, A., Bach, F.R., Canu, S., Grandvalet, Y.: Simplemkl. Journal of Machine Learning Research 9(11), 2491–2521 (2008)
Fu, Z., Lu, G., Ting, K.M., Zhang, D.: Learning sparse kernel classifiers for multi-instance classification. IEEE Trans. Neural Networks 24(9), 1377–1389 (2013)
Bach, F.R., Jordan, M.I.: Predictive low-rank decomposition for kernel methods. In: ICML, pp. 33–40 (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Fu, Z. (2014). Optimal Landmark Selection for Nyström Approximation. In: Loo, C.K., Yap, K.S., Wong, K.W., Teoh, A., Huang, K. (eds) Neural Information Processing. ICONIP 2014. Lecture Notes in Computer Science, vol 8835. Springer, Cham. https://doi.org/10.1007/978-3-319-12640-1_38
Download citation
DOI: https://doi.org/10.1007/978-3-319-12640-1_38
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-12639-5
Online ISBN: 978-3-319-12640-1
eBook Packages: Computer ScienceComputer Science (R0)