Skip to main content

Cycle Generative Adversarial Network for Unpaired Sketch-to-Character Translation

  • Conference paper
  • First Online:
Emerging Trends in Intelligent Computing and Informatics (IRICT 2019)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1073))

Abstract

This research investigated the capabilities of Cycle-Consistent adversarial network (Cycle GAN) in the application of stick figure sketches to cartoon character translation. Few studies focused on the task of generating a variety of poses and facial expression of cartoon characters from simple sketches of stick figures, based on unpaired dataset samples. Furthermore, existing studies showed low performance in detecting rare pose features. In this research, two datasets have been created which consists of paired and unpaired images of manually drawn sketches and cartoon characters. The performance of Cycle GAN has been compared against a paired based model, Pix2Pix, by using qualitative and quantitative evaluation measurements. Results show that Pix2Pix outperforms Cycle GAN in accurately mapping cartoon characters to stick figures. Despite that, the Cycle GAN still managed to produce competing results.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Goodfellow, I.J., Pouget-abadie, J., Mirza, M., Xu, B., Warde-farley, D.: Generative adversarial nets, pp. 1–9 (2014)

    Google Scholar 

  2. Zhu, J., Park, T., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks (2018)

    Google Scholar 

  3. Jin, Y., Zhang, J., Li, M., Tian, Y., Zhu, H.: Towards the high-quality anime characters generation with generative adversarial networks. In: Machine Learning for Creativity and Design Workshop at NIPS Work, Long Beach, California, USA, pp. 1–13 (2017)

    Google Scholar 

  4. Kataoka, Y., Matsubara, T., Uehara, K.: Automatic manga colorization with color style by generative adversarial nets. In: IEEE SNPD, Kanazawa, Japan, vol. 495, pp. 495–499 (2017)

    Google Scholar 

  5. Saito, M., Matsui, Y.: Illustration2Vec: a semantic vector representation of illustrations. In: SIGGRAPH Asia, no. 2 (2015)

    Google Scholar 

  6. Zhang, S., Ji, R., Hu, J., Lu, X., Li, X.: Face sketch synthesis by multidomain adversarial learning. IEEE Trans. Neural Netw. Learn. Syst. 1–10 (2018)

    Google Scholar 

  7. Huang, Y., Khan, S.M.: DyadGAN : generating facial expressions in dyadic interactions. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops Work, DyadGAN, pp. 2259–2266 (2017)

    Google Scholar 

  8. Philip, C., Jong, L.H.: Face sketch synthesis using conditional adversarial networks. In: ICTC, pp. 373–378 (2017)

    Google Scholar 

  9. Wang, L., Sindagi, V.A., Patel, V.M.: High-quality facial photo-sketch synthesis using multi-adversarial networks. In: 13th IEEE International Conference on Automatic Face and Gesture Recognition High-Quality, pp. 83–90 (2018)

    Google Scholar 

  10. Royer, A., et al.: XGAN: unsupervised Image-to-image translation for many-to-many mappings, pp. 1–19 (2017)

    Google Scholar 

  11. Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, pp. 5967–5976 (2017)

    Google Scholar 

  12. Fukumoto, Y., Shimizu, D., Shibata, C.: Generation of character illustrations from stick figures using a modification of generative adversarial network. In: 2018 IEEE 42nd Annual Computer Software and Applications Conference, vol. 01, pp. 183–186 (2018)

    Google Scholar 

  13. Selseng, S.: Guiding the training of generative adversarial networks (2017)

    Google Scholar 

  14. Fang, F., Yamagishi, J., Echizen, I., Lorenzo-trueba, J.: High-quality nonparallel voice conversion based on cycle-consistent adversarial network national. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 5279–5283 (2018)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Leena Alsaati or Siti Zaiton Mohd Hashim .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Alsaati, L., Hashim, S.Z.M. (2020). Cycle Generative Adversarial Network for Unpaired Sketch-to-Character Translation. In: Saeed, F., Mohammed, F., Gazem, N. (eds) Emerging Trends in Intelligent Computing and Informatics. IRICT 2019. Advances in Intelligent Systems and Computing, vol 1073. Springer, Cham. https://doi.org/10.1007/978-3-030-33582-3_31

Download citation

Publish with us

Policies and ethics