Skip to main content

Deep Style Transfer

  • Living reference work entry
  • First Online:
Computer Vision
  • 151 Accesses

Synonyms

Deep image stylization; Deep visual stylization; Neural style transfer

Related Concepts

Definition

Deep style transfer is an optimization technique, which is characterized by its use of deep neural networks (deep learning), used to manipulate digital images, or videos, to adopt the appearance or visual style of another image. As shown in Fig. 1, given a content image I, a style reference image S (such as an artwork by a famous painter), the output stylized image O will blend texture details from S and the local structure of I, to look like the content image, but “painted” in the style of the style image S”.

Fig. 1
figure 1

Style transfer is to migrate the style/texture from the right style image to the left content image

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

References

  1. Gooch B, Gooch A (2001) Non-photorealistic rendering. AK Peters/CRC Pres, Natick

    Book  Google Scholar 

  2. Strothotte T, Schlechtweg S (2002) Non-photorealistic computer graphics: modeling, rendering, and animation. Morgan Kaufmann, San Francisco

    Google Scholar 

  3. Hertzmann A, Jacobs CE, Oliver N, Curless B, Salesin DH (2001) Image analogies. In: Proceedings of the 28th annual conference on computer graphics and interactive techniques. ACM, pp 327–340

    Google Scholar 

  4. Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2414–2423

    Google Scholar 

  5. Li C, Wand M (2016) Combining markov random fields and convolutional neural networks for image synthesis. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2479–2486

    Google Scholar 

  6. Liao J, Yao Y, Yuan L, Hua G, Kang SB (2017) Visual attribute transfer through deep image analogy. arXiv preprint arXiv:1705.01088

    Google Scholar 

  7. Gu S, Chen C, Liao J, Yuan L (2018) Arbitrary style transfer with deep feature reshuffle. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8222–8231

    Google Scholar 

  8. Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for real-time style transfer and super-resolution. In: European conference on computer vision, pp 694–711. Springer

    Google Scholar 

  9. Ulyanov D, Lebedev V, Vedaldi A, Lempitsky VS (2016) Texture networks: feed-forward synthesis of textures and stylized images. In: ICML, vol 1, p 4

    Google Scholar 

  10. Li C, Wand M (2016) Precomputed real-time texture synthesis with Markovian generative adversarial networks. In: European conference on computer vision, pp 702–716. Springer (2016)

    Google Scholar 

  11. Dumoulin V, Shlens J, Kudlur M (2017) A learned representation for artistic style. In: Proceedings of the ICLR, vol 2

    Google Scholar 

  12. Chen D, Yuan L, Liao J, Yu N, Hua G (2017) Stylebank: an explicit representation for neural image style transfer. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1897–1906

    Google Scholar 

  13. Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M-H (2017) Diversified texture synthesis with feed-forward networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3920–3928

    Google Scholar 

  14. Ulyanov D, Vedaldi A, Lempitsky V (2017) Improved texture networks: Maximizing quality and diversity in feed-forward stylization and texture synthesis. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6924–6932

    Google Scholar 

  15. Chen TQ, Schmidt M (2016) Fast patch-based style transfer of arbitrary style. arXiv preprint arXiv:1612.04337

    Google Scholar 

  16. Huang X, Belongie S (2017) Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE international conference on computer vision, pp 1501–1510

    Google Scholar 

  17. Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M-H (2017) Universal style transfer via feature transforms. In: Advances in neural information processing systems, pp 386–396

    Google Scholar 

  18. Ruder M, Dosovitskiy A, Brox T (2016) Artistic style transfer for videos. In: German conference on pattern recognition, pp 26–36. Springer

    Google Scholar 

  19. Chen D, Liao J, Yuan L, Yu N, Hua G (2017) Coherent online video style transfer. In: Proceedings of the IEEE international conference on computer vision, pp 1105–1114

    Google Scholar 

  20. Chen D, Yuan L, Liao J, Yu N, Hua G (2018) Stereoscopic neural style transfer. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6654–6663

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Section Editor information

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this entry

Check for updates. Verify currency and authenticity via CrossMark

Cite this entry

Chen, D., Yuan, L., Hua, G. (2020). Deep Style Transfer. In: Computer Vision. Springer, Cham. https://doi.org/10.1007/978-3-030-03243-2_863-1

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-03243-2_863-1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-03243-2

  • Online ISBN: 978-3-030-03243-2

  • eBook Packages: Springer Reference Computer SciencesReference Module Computer Science and Engineering

Publish with us

Policies and ethics