Skip to main content

Conclusions, Contributions, and Future Work

  • Chapter
  • First Online:
Embedded Deep Learning

Abstract

This dissertation has focused on techniques to minimize the energy consumption of deep learning algorithms for embedded applications on battery-constrained wearable edge devices. Although SotA in many typical machine-learning tasks, deep learning algorithms are also very costly in terms of energy consumption, due to their large amount of required computations and huge model sizes. Because of this, deep learning applications on battery-constrained wearables have only been possible through wireless connections with a resourceful cloud. This setup has several drawbacks. First, there are privacy concerns. This setup requires users to share their raw data—images, video, locations, and speech—with a remote system. As most users are not willing to share all of this, large-scale applications cannot yet be developed. Second, the cloud-setup requires users to be connected all the time, which is unfeasible given current cellular coverage. Furthermore, real-time applications require low latency connections, which cannot be guaranteed using the current communication infrastructure. Finally, this wireless connection is very inefficient—requiring too much energy per transferred bit for real-time data transfer on energy constrained platforms. All these issues—privacy, latency/connectivity, and costly wireless connections—can be resolved by moving towards computing on the edge.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 139.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Ando K, Ueyoshi K, Orimo K, Yonekawa H, Sato S, Nakahara H, Ikebe M, Asai T, Takamaeda-Yamazaki S, Kuroda T, et al (2017) Brein memory: a 13-layer 4.2 k neuron/0.8 m synapse binary/ternary reconfigurable in-memory deep neural network accelerator in 65 nm CMOS. In: Symposium on VLSI circuits, 2017. IEEE, pp C24–C25

    Google Scholar 

  • Andri R, Cavigelli L, Rossi D, Benini L (2016) Yodann: an ultra-low power convolutional neural network accelerator based on binary weights. In: IEEE computer society annual symposium on VLSI (ISVLSI), 2016. IEEE, pp 236–241

    Google Scholar 

  • Bong K, Choi S, Kim C, Kang S, Kim Y, Yoo HJ (2017) 14.6 a 0.62 mw ultra-low-power convolutional-neural-network face-recognition processor and a CIS integrated with always-on haar-like face detector. In: IEEE International Solid-State Circuits Conference (ISSCC), 2017. IEEE, pp 248–249

    Google Scholar 

  • Esser S, Merolla P, Arthur J, Cassidy A, et al (2016) Convolutional networks for fast, energy-efficient neuromorphic computing. In: Proceedings of the national academy of sciences

    Google Scholar 

  • Han S, Mao H, Dally WJ (2016) Deep compression: compressing deep neural network with pruning, trained quantization and Huffman coding. In: International Conference on Learning Representations (ICLR)

    Google Scholar 

  • Huang G, Che D, Li T, Wu F, van der Maaten L, Weinberger K (2017) Multi-scale dense networks for resource efficient image classification. arXiv preprint arXiv:170309844, submitted to ICLR 2018

    Google Scholar 

  • Lee J, Kim C, Kang S, Shin D, Kim S, Yoo HY (2018) Unpu: a 50.6 tops/w unified deep neural network accelerator with 1b-to-16b fully-variable weight bit-precision. In: International Solid-State Circuits Conference (ISSCC)

    Google Scholar 

  • Moons B, Verhelst M (2016) A 0.3-2.6 tops/w precision-scalable processor for real-time large-scale convnets. In: Proceedings of the IEEE symposium on VLSI circuits, pp 178–179

    Google Scholar 

  • Moons B, Uytterhoeven R, Dehaene W, Verhelst M (2017) Envision: a 0.26-to-10 tops/w subword-parallel dynamic-voltage-accuracy-frequency-scalable convolutional neural network processor in 28nm FDSOI. In: International Solid-State Circuits Conference (ISSCC)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Moons, B., Bankman, D., Verhelst, M. (2019). Conclusions, Contributions, and Future Work. In: Embedded Deep Learning. Springer, Cham. https://doi.org/10.1007/978-3-319-99223-5_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-99223-5_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-99222-8

  • Online ISBN: 978-3-319-99223-5

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics