Simultaneous Learning of Localized Multiple Kernels and Classifier with Weighted Regularization

  • Naoya Inoue
  • Yukihiko Yamashita
Part of the Lecture Notes in Computer Science book series (LNCS, volume 7626)

Abstract

Kernel classifiers have demonstrated their high performance for many classification problems. For the proper selection of kernel functions, multiple kernel learning (MKL) has been researched. Furthermore, the localized MKL (LMKL) enables to set the weights for the kernel functions at each point. However, the training of the weight functions for kernel functions is a complex nonlinear problem and a classifier can be trained separately after the weights are fixed. The iteration of the two processes are often necessary. In this paper we propose a new framework for MKL/LMKL. In the framework, not kernel functions but mappings to the feature space are combined with weights. We also propose a new learning scheme to train simultaneously weights for kernel functions and a classifier. We realize a classifier by our framework with the Gaussian kernel function and the support vector machine. Finally, we show its advantages by experimental results.

References

  1. 1.
    Bach, F.R., Lanckriet, G.R.G., Jordan, M.I.: Multiple kernel learning, conic duality, and the smo algorithm. In: ICML (July 2004)Google Scholar
  2. 2.
    Gönen, M., Alpaydin, E.: Localized multiple kernel learning. In: Proceedings of the 25th International Conference on Machine Learning, ICML 2008, pp. 352–359. ACM, New York (2008), http://doi.acm.org/10.1145/1390156.1390201 CrossRefGoogle Scholar
  3. 3.
    Gönen, M., Alpaydin, E.: Multiple kernel learning algorithm. Journal of Machine Learning Research 12, 2211–2268 (2011)Google Scholar
  4. 4.
    Lanckriet, G.R.G., Bie, T.D., Cristianini, N., Jordan, M.I., Noble, W.S.: A statistical framework for genomic data fusion. Some Fine Journal 20(16), 2626–2635 (2004)Google Scholar
  5. 5.
    Mika, S., Rätsch, G., Weston, J., Schölkopf, B., Müller, K.R.: Fisher discriminant analysis with kernels. In: Neural Networks for Signal Processing IX, pp. 41–48. IEEE (1999)Google Scholar
  6. 6.
    Pavlidis, P., Weston, J., Cai, J., Grundy, W.N.: Gene functional classification from heterogeneous data. In: Proceedings of the Fifth Annual International Conference on Computational Biology, RECOMB 2001, pp. 249–255. ACM, New York (2001), http://doi.acm.org/10.1145/369133.369228 CrossRefGoogle Scholar
  7. 7.
    Sonnenburg, S., Rätsch, G., Schäfer, C., Schölkopf, B.: Large scale multiple kernel learning. J. Mach. Learn. Res. 7, 1531–1565 (2006), http://dl.acm.org/citation.cfm?id=1248547.1248604 MathSciNetMATHGoogle Scholar
  8. 8.
    Vapnik, V.N.: Statistical Learning Theory. Wiley, New-York (1998)MATHGoogle Scholar
  9. 9.
    Wahba, G.: Spline Models for Observational Data (CBMS-NSF Regional Conference Series in Applied Mathematics). Society for Industrial and Applied Mathematics, Philadelphia (1990)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2012

Authors and Affiliations

  • Naoya Inoue
    • 1
  • Yukihiko Yamashita
    • 2
  1. 1.Planning & Development DivisionPlanex Communications Inc.Shibuya-kuJapan
  2. 2.Graduate School of Science and EngineeringTokyo Institute of TechnologyMeguro-kuJapan

Personalised recommendations