Skip to main content

Invariant object recognition using fahlman and Lebiere's learning algorithm

  • Conference paper
  • First Online:
  • 255 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 686))

Abstract

A new neural network system for object recognition is proposed which is invariant to translation, scaling and rotation. The system consists of two parts. The first is a preprocessor which obtains projection from the input image such that, for any rotation and scaling of standard image, the projection results are reduced to cyclically shifted ones, and then adopts the Rapid Transform [9] which makes the projected images cyclic shift invariant. The second part is a neural net classifier which receives the outputs of preprocessing part as the input signals. The most attractive feature of this system is that, by using only a simple shift invariant transformation (Rapid Transform) in conjunction with the projection of the input image plane, invariancy is achieved and the system is reasonably small. Experiments with six geometrical objects with different degree of scaling and rotation show that the proposed system performs excellent when the neural net classifier is trained by Fahlman and Lebiere's learning algorithm [8].

This is a preview of subscription content, log in via an institution.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. C. T. Zahn and R. Z. Roskies, “Fourier descriptions for plane closed curves,” IEEE Trans. Computers, vol. C-21, pp. 269–281, 1972.

    Google Scholar 

  2. S. A. Dudani, K. J. Breeding and R. B. McGhee, “Aircraft identification by moment invariants,” IEEE Trans. Computers, vol. C-26, pp. 39–45, 1977.

    Google Scholar 

  3. R. L. Kashyap and R. Chellappa, “Stochastic models for close boundary analysis: representation and reconstruction,” IEEE Trans. Information Theory, vol. IT-27, no. 5, pp. 627–637, 1981.

    Google Scholar 

  4. B. Widrow, R. G. Winter and R. A. Baxter, “Layered neural nets for pattern recognition,” IEEE Trans. Acoustics, Speech, and Signal Processing, vol. ASSP-36, no.7, pp. 1109–1118, 1988.

    Google Scholar 

  5. K. Fukushima, S. Miyake and T. Ito, “Neocognitron: a neural network model for a mechanism of visual pattern recognition,” IEEE Trans. Syst., Man, and Cyber., vol. SMC-13, pp. 826–834, 1983.

    Google Scholar 

  6. L. Spirkovska and M. B. Reid, “Connectivity strategies for higher-order neural networks,” in Proc. International Joint Conference on Neural Networks, IJCNN'90, San Diego, vol. 1, pp. 21–26, 1990.

    Google Scholar 

  7. S. D. You and G. E. Ford, “Object recognition based on projection,” in Proc. International Joint Conference on Neural Networks, Baltimore, vol. 4, pp. 31–36, 1992.

    Google Scholar 

  8. S. E. Fahlman and C. Lebiere, “The Cascade-correlation learning architecture,” in D. S. Touretzky (ed.), Advances in Neural Information Processing Systems, vol. 2, pp. 524–532, Morgan Kaufmann, 1990.

    Google Scholar 

  9. H. Reitboeck and T. P. Brody, “A transformation with Invariance under cyclic permutation for applications in pattern recognition”, Information and Control, vol. 15, pp. 130–154, 1969.

    Google Scholar 

  10. D. E. Rumelhart, J. L. McClelland and the PDP Research Group, “Parallel Distributed Processing,” vol. 1, M.I.T. Press, 1986.

    Google Scholar 

  11. M. Hamamoto, J. Kamruzzaman and Y. Kumagai, “Generalization ability of artificial neural network using Fahlman and Lebiere's learning algorithm,” in Proc. of IEEE/INNS Int. Joint Conf. on Neural Networks, IJCNN'92, Baltimore, vol. I, pp. 613–618, 1992.

    Google Scholar 

  12. M. Hamamoto, J. Kamruzzaman and Y. Kumagai, “Network synthesis and generalization properties of artificial neural network using Fahlman and Lebiere's learning algorithm”, to be published in Proc. of IEEE 35tli midwest symp. on circuits and systems, MWSCAS'92, Washington DC, 1992.

    Google Scholar 

  13. M. Hamamoto, J. Kamruzzaman, Y. Kumagai and H. Hikita, “Generalization ability of feedforward neural network trained by Fahlman and Lebiere's learning algorithm,” IEICE Tran. Fundamentals of Electronics, Communications & Computer Science, Japan, vol. E75-A, no. 11, pp. 1597–1601, 1992.

    Google Scholar 

  14. M. Hamamoto, J. Kamruzzaman, Y. Kumagai and H. Hikita, “Incremental learning and generalization ability of artificial neural network using Fahlman and Lebiere's learning algorithm,” IEICE Tran. Fundamentals of Electronics, Communications & Computer Science, Japan, vol. E76-A, no. 2, pp. 242–247, 1993.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

José Mira Joan Cabestany Alberto Prieto

Rights and permissions

Reprints and permissions

Copyright information

© 1993 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Ito, K., Hamamoto, M., Kamruzzaman, J., Kumagai, Y. (1993). Invariant object recognition using fahlman and Lebiere's learning algorithm. In: Mira, J., Cabestany, J., Prieto, A. (eds) New Trends in Neural Computation. IWANN 1993. Lecture Notes in Computer Science, vol 686. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-56798-4_153

Download citation

  • DOI: https://doi.org/10.1007/3-540-56798-4_153

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-56798-1

  • Online ISBN: 978-3-540-47741-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics