Hidden Space Neighbourhood Component Analysis for Cancer Classification
Neighbourhood component analysis (NCA) is a method for learning a distance metric which can maximize the classification performance of the K nearest neighbour (KNN) classifier. However, NCA suffers from the small size sample problem that the number of samples is much less than the number of features. To remedy this, this paper proposes a hidden space neighbourhood components analysis (HSNCA), which is a nonlinear extension of NCA. HSNCA first maps the data in the original space into a feature space by a set of nonlinear mapping functions, and then performs NCA in the feature space. Notably, the number of samples is equal to the number of features in the feature space. Thus, HSNCA can avoid the small size sample problem. Experimental results on DNA array datasets show that HSNCA is feasibility and efficiency.
KeywordsNeighbourhood components analysis Nonlinear mapping Small size sample problem Feature space Nearest neighbour
This work was supported in part by the National Natural Science Foundation of China under Grant Nos. 61373093, and 61402310, by the Natural Science Foundation of Jiangsu Province of China under Grant No. BK20140008, by the Natural Science Foundation of the Jiangsu Higher Education Institutions of China under Grant No. 13KJA520001, and by the Soochow Scholar Project.
- 5.Goldberger, J., Roweis, S., Hinton, G., Salakhutdinov, R.: Neighbourhood components analysis. In: Advances in Neural Information Processing Systems, vol. 17, pp. 513–520. MIT Press (2004)Google Scholar
- 6.Shental, N., Hertz, T., Weinshall, D., Pavel, M.: Adjustment learning and relevant component analysis. In: Proceedings of 7th European Conference on Computer Vision, London, UK, pp. 776–792 (2002)Google Scholar
- 7.Globerson, A., Roweis, S.T.: Metric learning by collapsing classes. In: Advances in Neural Information Processing Systems, vol. 18 (2005)Google Scholar
- 9.Xing, E.P., Ng, A.Y., Jordan, M.I., Russell, S.: Distance metric learning, with application to clustering with side-information. In: Advances in Neural Information Processing Systems, vol. 14, pp. 521–528. MIT Press, Cambridge (2002)Google Scholar
- 10.Chopra, S., Hadsell, R., LeCunGoldberger, Y.: Learning a similiarty metric discriminatively, with application to face verification. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, San Diego, CA, pp. 349C–356 (2005)Google Scholar
- 11.Davis, J.V., Kulis, B., Jain, P., Sra, S., Dhillon, I.S.: Information-theoretic metric learning. In: Proceedings of 24th International Conference on Machine Learning, pp. 209–216. ACM, New York (2007)Google Scholar
- 13.Yang, W., Wang, K., Zuo, W.: Neighborhood component feature selection for high-dimensional data. J. Comput. 7(1), 161–168 (2012)Google Scholar
- 15.Qin, C., Song, S., Huang, G.: Non-linear neighborhood component analysis based on constructive neural networks. In: Proceedings of 2014 IEEE International Conference on Systems, Man and Cybernetics, pp. 1997–2002. IEEE (2014)Google Scholar
- 20.Ding, C., Zhang, L., Wang, B.J.: Hidden space discriminant neighborhood embedding. In: Proceedings of 2014 International Joint Conference on Neural Networks, pp. 271–277. IEEE (2014)Google Scholar