Statistical Methods in Pattern Recognition

  • C. Robert Appledorn
Conference paper
Part of the NATO ASI Series book series (volume 39)


When measurements group together and begin to form clusters in some measurement feature space, one tends to remark that a pattern is developing. Furthermore, the size and shape of the pattern can be provided a statistical description. In a variety of applications, one is faced with the problem of using these statistical descriptions to classify a particular measurement to a specific cluster; that is, to make a decision regarding which pattern group generated the measurement.

This paper presents an overview of the mathematical considerations that statistical pattern recognition entails. Topics that receive emphasis include normalizations based upon covariance matrix eigen-factorizations, eigen-expansion feature extraction methods, linear classifier functions, and distance measurements. Particular emphasis is given to Linear algebraic techniques that lead to simple computational implementations. Lastly, estimation in a noisy environment is briefly discussed.


Covariance Matrix Measurement Vector Linear Classifier Statistical Pattern Recognition Discriminant Vector 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. Fukunaga, K. (1972). Introduction to Statistical Pattern Recognition, Academic, New York.Google Scholar
  2. Swain, P.H. and Davis, S.M. (1978). Remote Sensing: The Quantitative Approach, McGraw-Hill, New York.Google Scholar
  3. Fisher, R.A. (1936). The use of multiple measurements in taxonomic problems, Ann. Eugenics 7-II, pp. 179–188.Google Scholar
  4. Sammon, J.W. (1970). An optimal discriminant plane, IEEE Trans. Comput. C-19, pp. 826–829.Google Scholar
  5. Foley, D.H. and Sammon, J.W. (1975). An optimal set of discriminant vectors, IEEE Trans. Comput. C-24, pp. 281–289.Google Scholar
  6. Kailath, T. (1967). The divergence and Bhattacharyya distance measures in signal selection, IEEE Trans. Commun. Technol. COM-15, pp. 52–60.Google Scholar
  7. Henderson, T.L. and Lainiotis, D.G. (1969). Comments on linear feature ex-traction, IEEE Trans. Inform. Theory IT-15, pp. 728–730.Google Scholar
  8. Kanal, L. (1974). Patterns in pattern recognition: 1968–1974, IEEE Trans. Inform. Theory IT-20, pp. 697–722.Google Scholar
  9. Foley, D.H. (1972). Considerations of sample size and feature size, IEEE Trans. Inform. Theory IT-18, pp. 618–626.Google Scholar
  10. Robins, H. and Monroe, S. (1951). A stochastic approximation method, Ann. Math. Stat. 22, p. 400.CrossRefGoogle Scholar
  11. Dvoretsky, A. (1956). On stochastic approximation, Proc. 3rd Berkeley Sym-posium on Math. Stat. Prob., Univ. of California Press, Los Angeles.Google Scholar
  12. Kalayeh, H.M. and Landgrebe, D.A. (1983). Predicting the required number of training samples, IEEE Trans. Pattern Anal. Mach. Intell. PAMI-5, pp. 664–667.Google Scholar
  13. Gelb, A. (ed.) (1974). Applied Optimal Control, MIT Press, Cambridge.Google Scholar
  14. Kalman, R. (1960). A new approach to linear filtering and prediction problems, Trans. ASME J. Basic Eng. 82, pp. 34–35.Google Scholar
  15. Wiener, N. (1949). The Extrapolation, Interpolation, and Smoothing of Stationary Time Series, Wiley, New York.Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 1988

Authors and Affiliations

  • C. Robert Appledorn
    • 1
  1. 1.Indiana University Medical CenterIndianapolisUSA

Personalised recommendations