Abstract
Achieving good performance in biometrics requires matching the capacity of the classifier or a set of classifiers to the size of the available training set. A classifier with too many adjustable parameters (large capacity) is likely to learn the training set without dificulty but be unable to generalize properly to new patterns. If the capacity is too small, the training set might not be learned without appreciable error. There is thus advantage to control the capacity through a variety of methods involving not only the structure of the classifiers themselves, but also the property of the input space. Ths paper proposes an original non parametric method to combine optimaly multiple classifier responses. Highly favorable results have been obtained using the above method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
U. Beyer and F. Smieja. Learning from exemples, agent teams, and the concept of reflection. International Journal of Pattern Recognition and Artificial Intelligence, 1994.
L. Breiman. Bagging predictors. Rapport technique TR-421, Statistics Department, University of California, Berkeley, 1994.for our model.
J. Daugman. “high confidence visual recognition of persons by a test of statsitical independence”, IEEE Trans. PAMI, 1, Nov. 1993, pp. 148–160.
L. Devroye, L. Györfi and G. Lugosi. A probabilistic theory of pattern recognition, Application of Mathematics, Springer, 1991.
H. Drüker, C. Cortes, L. Jackel, Y. Lecun and V. Vapnik. Boosting and other ensemble methodes. Neural Computation, 6(6), pp. 1289–1301, 1994.
B. Efron (1979) The convex hull of a random set of points. Biometrika, 52, p331–342.
Y. Freund, H. Seung, E. Shamir and N. Tishby. Information, prediction and query commitee. Neural Information Processing Systems, S. Hanson, J. Cowan, C. Giles ed., pp. 483–490, Denver, 1993.
J. Gubner. Distributed estimation and quantization. IEEE Trans. on Information Theory, 39(4), pp. 1456–1459, 1993.
L.K. Hansen and P. Salamon. Neural network ensembles. IEEE Trans. on PAMI, 12, pp. 993–1001, 1990.
L.K. Hansen, C. Lüsberg and P. Salamon. Ensemble methods for recognition of handwritten digits. Proc. IEEE Signal Processing Workshop, S.Y. Kung, F. Fallside, J.A. Sorensen and C.A. Kamm Ed.; Piscataway, NJ, pp. 540–549, 1992.
M. I. Jordan and R.A. Jacobs. Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6, pp. 181–214, 1994.
A.K. Jain. Fundamentals of digital image processing, Prentice-Hall Inc., 1989.
R. Kallel, M. Cottrell and V. Vigneron. Bootstrap for neural model selection. Proc. of the 8th European Symposium on Artificial Neural Networks, Bruges, 22–24 Avril, 2000.
J. Kittler and M. Hatef. Improving recognition rates by classifier combination. Proc. IWFHR’96, pp. 81–101, 1996.
A.N. Kolmogorov and S.V. Fomin. Elements of the theory of functions and functional analysis, Vol. 1, Graylock Press, 1957.
J.P. Nadal, R. Legault et C. Suen. Complementary algorithms for the recognition of totally unconstrained handwritten numerals. 10th International Conference on Pattern Recognition, pp. 443–446, Atlantic cit, 1990.
I. S. Oh, J. S. Lee and C.Y. Suen. Analysis of class separation and combination of class dependent features for handwriting recognition, IEEE Trans. on PAMI, 21(10), pp. 1099–1125.
A. Hyvärinen, J. Karhunen and E. Oja. Independent component analysis, John Wiley & Sons, 2001.
J. Pearl, D. Geiger and T. Verma. Conditional independence and its representations. Kybernetika, 25(2), 1989.
B.D. Ripley. Pattern recognition and neural networks, Cambridge University Press, 1996.
C. M. Soares, C. L. Fróes da Silva, M. De Gregorio and F.M.G. França. A software implmentation of the WISARD classifier, Proceeding of Brasilian Symposium on Artificial Neural Network, Belo Horizonte, MG, december 9–11, 1998, vol. II, 225–229.
K. Tumer and J. Ghosh. A framework for estimating performance improvements in hybrid pattern classifiers. Proc. World Congress on Neural Networks, San Diego, pp. 220–225, 1994.
L. G. Valiant. A theory of the learnable. Communications of the association for computing machinery, 27, pp. 1134–1142, reprint in Shavlik & Dietterich, 1990.
D. Wolpert. Stacked regression. Neural Networks, 5(2), pp. 241–260, 1992.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Vigneron, V., Maaref, H., Lelandais, S. (2003). “Poor Man” Vote with M-ary Classifiers. Application to Iris Recognition. In: Kittler, J., Nixon, M.S. (eds) Audio- and Video-Based Biometric Person Authentication. AVBPA 2003. Lecture Notes in Computer Science, vol 2688. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44887-X_76
Download citation
DOI: https://doi.org/10.1007/3-540-44887-X_76
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40302-9
Online ISBN: 978-3-540-44887-7
eBook Packages: Springer Book Archive