Skip to main content

Abstract

This chapter describes an evolutionary approach to deep learning networks. We first explain neuroevolution approach, which can adaptively learn a network structure and size appropriate to the task. A typical example of neuroevolution is NEAT. NEAT has demonstrated performance superior to that of conventional methods in a large number of problems. Then, several studies on deep neural networks with evolutionary optimization are explained, such as Genetic CNNs, hierarchical feature construction using GP, and Differentiable pattern-producing network (DPPSN).

A sonnet written by a machine would be better appreciated by another machine.

(Alan Turing)

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    A columnar structure in the cerebrum where neurons with similar properties concentrate. In mouse perceptual fields, there are believed to be columns corresponding to each individual whisker.

  2. 2.

    This image was created by LGPC for Art, a simulator created with reference to Sbart. This tool can be used to “nurture” to influence their creation. See our Web site for information on installing and using LGPC for Art.

  3. 3.

    Here hypot gives the Euclidean distance on two dimensions, defined as \(hypot(x,y)=\sqrt{x^2+y^2}\).

  4. 4.

    http://gigazine.net/news/20150616-mari-o/.

  5. 5.

    During training, each layer’s inputs are normalized across the current minibatch to the Gaussian distributions (usually zero mean and unit variance). It has been shown to have several benefits, e.g., faster convergence, easier to escape from local optima, more robust network.

  6. 6.

    A function outputting 0 when the input is 0 or less and outputting the input as it is when the input is greater than 1.

  7. 7.

    BCE is a loss function \(\mathcal {L}\) commonly used for a binary classification, which is a special case of multiclass cross-entropy. The definition is given as follows: \(\mathcal {L}(\theta )= -\frac{1}{n}\sum _{i=1}^n \left[ y_i \log (p_i) + (1-y_i) \log (1-p_i)\right] \), where n is the number of samples, \(y_i\) is the sample label of ith sample, and \(p_i\) is the prediction for the ith sample. Smaller values indicate a better prediction.

  8. 8.

    http://yann.lecun.com/exdb/mnist/index.html.

  9. 9.

    1623 different handwritten characters from 50 different alphabets. See https://github.com/brendenlake/omniglot for details.

References

  1. Fernando, C., Banarse, D., Reynolds, M., Besse, F., Pfau, D., Jaderberg, M., Lanctot, M., Wierstra, D.: Convolution by evolution–differentiable pattern producing networks. In: Proceedings of the Genetic and Evolutionary Computation Conference 2016. (GECCO16), pp. 109–116 (2016)

    Google Scholar 

  2. Hausknecht, M., Khandelwal, P., Miikkulainen, R., Stone, P.: HyperNEAT-GGP: A HyperNEAT-based Atari general game player. In: Proceedings of the Genetic and Evolutionary Computation Conference (GECCO 2012), pp. 217–224 (2012)

    Google Scholar 

  3. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)

    Google Scholar 

  4. Huang, G., Liu, Z., Weinbergerz, K.: Densely connected convolutional networks. In: Proceedings of Computer Vision and Pattern Recognition (CVPR2017) (2017)

    Google Scholar 

  5. Iba, H.: Agent-Based Modeling and Simulation with Swarm. Chapman and Hall/CRC, London (2013)

    Google Scholar 

  6. Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd International Conference on Machine Learning, vol. 37, pp. 448–456 (2015)

    Google Scholar 

  7. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the 3rd International Conference on Learning Representations (ICLR2015)

    Google Scholar 

  8. Krizhevsky, A., Hinton, G.: Learning multiple layers of features from tiny images. Technical report 1. Computer Science Department, University of Toronto (2009)

    Google Scholar 

  9. Krizhevsky, A., Sutskerver, I. Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems 25 (NIPS), pp. 1097–1105 (2012)

    Google Scholar 

  10. Miikkulainen, R., Liang, J., Meyerson, E., Rawal, A., Fink, D., Francon, O., Raju, B., Shahrzad, H., Navruzyan, A., Duffy, N., Hodjat, B.: Evolving deep neural networks (2017). arXiv:1703.00548

  11. Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., Ng, A.: Reading digits in natural images with unsupervised feature learning. In: Proceedings of NIPS Workshop on Deep Learning and Unsupervised Feature Learning (2011)

    Google Scholar 

  12. Rozenberg, G. (ed.): The Book of L. Springer, Berlin (1986)

    Google Scholar 

  13. Schrum, J., Miikkulainen, R.: Evolving multimodal behavior with modular neural networks in Ms. Pac-Man. In: Proceedings of the Genetic and Evolutionary Computation Conference (GECCO 2014), pp. 325–332 (2014)

    Google Scholar 

  14. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: Proceedings of International Conference on Learning Representations (2014)

    Google Scholar 

  15. Stanley, K.O.: Compositional pattern producing networks: a novel abstraction of development. Genet. Program. Evolvable Mach. (Special Issue on Dev. Syst.) 8(2), 131–162 (2007)

    Google Scholar 

  16. Stanley, K.O., Miikkulainen, R.: Evolving neural networks through augmenting topologies. Evol. Comput. 10(2), 99–127 (2002)

    Google Scholar 

  17. Stanley, K.O., D’Ambrosio, D.B., Gauci, J.: A hypercube-based encoding for evolving large-scale neural networks. Artif. Life 15(2), 185–212 (2009)

    Google Scholar 

  18. Suganuma, M., Shirakawa, S., Nagao, T.: A genetic programming approach to designing convolutional neural network architectures. In: Proceedings of the Genetic and Evolutionary Computation. Conference 2017 (GECCO2017), pp. 497–504 (2017)

    Google Scholar 

  19. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: Proceedings of Computer Vision and Pattern Recognition (CVPR2016) (2016)

    Google Scholar 

  20. Unemi, T.: SBART2.4: Breeding 2D CG images and movies, and creating a type of collage. In: Proceedings of The Third International Conference on Knowledge-based Intelligent Information Engineering Systems, pp. 288–291 (1999)

    Google Scholar 

  21. Xie, L., Yuille, A.: Genetic CNN. In: Proceedings of IEEE International Conference on Computer Vision (ICCV)

    Google Scholar 

  22. Yao, X.: A review of evolutionary artificial neural networks. Int. J. Intell. Syst. 8, 539–567 (1993)

    Google Scholar 

  23. Zagoruyko, S., Komodakis, N.: Wide residual networks (2016). arXiv: 1605.07146

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hitoshi Iba .

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Singapore Pte Ltd.

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Iba, H. (2018). Evolutionary Approach to Deep Learning. In: Evolutionary Approach to Machine Learning and Deep Neural Networks. Springer, Singapore. https://doi.org/10.1007/978-981-13-0200-8_3

Download citation

  • DOI: https://doi.org/10.1007/978-981-13-0200-8_3

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-13-0199-5

  • Online ISBN: 978-981-13-0200-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics