Skip to main content

Optimizing Classifiers for Handwritten Digits by Genetic Algorithms

  • Conference paper
Artificial Neural Nets and Genetic Algorithms

Abstract

We present the first large real-world application for the neural network optimizing genetic algorithm Enzo. Nets had several thousands links and the training data up to over 200,000 patterns. We evolved nets for a classification task that have an order of magnitude free parameters less than commonly used polynomial classifiers while maintaining the same performance.

To achieve this we implemented some significant enhancements and minor improvements of the original algorithm.

It is also shown how to use Enzo as an efficient tool to create nets satisfying task-specific constraints.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Heinrich Braun and Joachim Weisbrod. Evolving feedforward neural networks. Proceedings of the International Conference on Artificial Neural Nets and Genetic Algorithms, 1993.

    Google Scholar 

  2. Heinrich Braun and Peter Zagorski. Enzo-M: A hybrid approach for optimizing neural networks by evolution and learning. In Proceedings of the third parallel problem solving from Nature, Jerusalem, Israel, 1994.

    Google Scholar 

  3. R. Allan Wilkinson et al The first census optical character recognition systems conference. NIST ir4912. Available at the NIST-Archive: sequoya.ucsl.nist.gov.

    Google Scholar 

  4. John McDonell and Don Waagen. Neural structure design by evolutionary programming, 1993. NCCOSC, RDT & E Division, San Diego, CA 92152.

    Google Scholar 

  5. Martin Riedmiller and Heinrich Braun. A direct adaptive method for faster backpropagation learning: The Rprop algorithm. In Proceedings of the ICNN 93, San Francisco, 1993.

    Google Scholar 

  6. D.E. Rumelhart, G. Hinton, and R. Williams. Learning internal representations by error propagation. In D.E. Rumelhart and J.L. McClelland, editors, Parallel Distributed Processing, Vol. I Foundations, pages 318–362. MIT Press, Cambridge, MA, 1986.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 1995 Springer-Verlag/Wien

About this paper

Cite this paper

Schäfer, J., Braun, H. (1995). Optimizing Classifiers for Handwritten Digits by Genetic Algorithms. In: Artificial Neural Nets and Genetic Algorithms. Springer, Vienna. https://doi.org/10.1007/978-3-7091-7535-4_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-7091-7535-4_5

  • Publisher Name: Springer, Vienna

  • Print ISBN: 978-3-211-82692-8

  • Online ISBN: 978-3-7091-7535-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics