Advertisement

Cantonese porcelain classification and image synthesis by ensemble learning and generative adversarial network

  • Steven Szu-Chi Chen
  • Hui Cui
  • Ming-han Du
  • Tie-ming Fu
  • Xiao-hong Sun
  • Yi JiEmail author
  • Henry Duh
Article
  • 4 Downloads

Abstract

Accurate recognition of modern and traditional porcelain styles is a challenging issue in Cantonese porcelain management due to the large variety and complex elements and patterns. We propose a hybrid system with porcelain style identification and image recreation modules. In the identification module, prediction of an unknown porcelain sample is obtained by logistic regression of ensembled neural networks of top-ranked design signatures, which are obtained by discriminative analysis and transformed features in principal components. The synthesis module is developed based on a conditional generative adversarial network, which enables users to provide a designed mask with porcelain elements to generate synthesized images of Cantonese porcelain. Experimental results of 603 Cantonese porcelain images demonstrate that the proposed model outperforms other methods relative to precision, recall, area under curve of receiver operating characteristic, and confusion matrix. Case studies on image creation indicate that the proposed system has the potential to engage the community in understanding Cantonese porcelain and promote this intangible cultural heritage.

Key words

Cantonese porcelain Classification Generative adversarial network Creative arts 

CLC number

TP751 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Notes

References

  1. Bao H, Liang Y, Liu HZ, et al., 2010. A novel algorithm for extraction of the scripts part in traditional Chinese painting images. Proc 2nd Int Conf on Software Technology and Engineering, p.V2-26–V2-30.  https://doi.org/10.1109/ICSTE.2010.5608756
  2. Buades A, Coll B, Morel JM, 2005. A non-local algorithm for image denoising. IEEE Computer Society Conf on Computer Vision and Pattern Recognition, p.60–65.  https://doi.org/10.1109/CVPR.2005.38
  3. Chen KH, 2019. Image Operations with cGAN. http://www.k4ai.com/imageops/index.html
  4. China Intangible Cultural Heritage Network, 2008. Cantonese Porcelain Inheritance Project. http://www.ihchina.cn/project_details/14453/ [Accessed on July 16, 2019] (in Chinese).
  5. Cochran WG, 1954. Some methods for strengthening the common χ2 tests. Int Biom Soc, 10(4):417–451.  https://doi.org/10.2307/3001616 MathSciNetzbMATHGoogle Scholar
  6. Dirvanauskas D, Maskeliūnas R, Raudonis V, et al., 2019. HEMIGEN: human embryo image generator based on generative adversarial networks. Sensors, 19(16):3578.  https://doi.org/10.3390/s19163578 CrossRefGoogle Scholar
  7. Efros AA, Freeman WT, 2001. Image quilting for texture synthesis and transfer. Proc 28th Annual Conf on Computer Graphics and Interactive Techniques, p.341–346.  https://doi.org/10.1145/383259.383296
  8. El Hattami A, Pierre-Doray É, Barsalou Y, 2019. Background removal using U-net, GAN and image matting. https://github.com/eti-p-doray/unet-gan-matting [Accessed on July 14, 2019].
  9. Emami H, Dong M, Nejad-Davarani SP, et al., 2018. Generating synthetic CTs from magnetic resonance images using generative adversarial networks. Med Phys, 45(8): 3627–3636.  https://doi.org/10.1002/mp.13047 CrossRefGoogle Scholar
  10. Goodfellow IJ, Pouget-Abadie J, Mirza M, et al., 2014. Generative adversarial nets. Proc 27th Int Conf on Neural Information Processing Systems, p.2672–2680.Google Scholar
  11. Hertzmann A, Jacobs CE, Oliver N, et al., 2001. Image analogies. Proc 28th Annual Conf on Computer Graphics and Interactive Techniques, p.327–340.  https://doi.org/10.1145/383259.383295
  12. Iizuka S, Simo-Serra E, Ishikawa H, 2016. Let there be color!: joint end-to-end learning of global and local image priors for automatic image colorization with simultaneous classification. ACM Trans Graph, 35(4), Article 110.  https://doi.org/10.1145/2897824.2925974 CrossRefGoogle Scholar
  13. Isola P, Zhu JY, Zhou TH, et al., 2017. Image-to-image translation with conditional adversarial networks. IEEE Conf on Computer Vision and Pattern Recognition, p.1125–1134.  https://doi.org/10.1109/CVPR.2017.632
  14. Ji Y, Tan P, Chen SC, et al., 2019. Kansei engineering for E-commerce Cantonese porcelain selection in China. 21st Int Conf on Human-Computer Interaction, p.463–474.  https://doi.org/10.1007/978-3-030-22646-6_34 CrossRefGoogle Scholar
  15. Jiang SQ, Huang QM, Ye QX, et al., 2006. An effective method to detect and categorize digitized traditional Chinese paintings. Patt Recogn Lett, 27(7):734–746.  https://doi.org/10.1016/j.patrec.2005.10.017 CrossRefGoogle Scholar
  16. Kira K, Rendell LA, 1992. A practical approach to feature selection. Machine Learning Proc, p.249–256.  https://doi.org/10.1016/B978-1-55860-247-2.50037-1 CrossRefGoogle Scholar
  17. Kurin R, 2004. Safeguarding intangible cultural heritage in the 2003 UNESCO convention: a critical appraisal. Museum Int, 56(1–2):66–77.  https://doi.org/10.1111/j.1350-0775.2004.00459.x CrossRefGoogle Scholar
  18. Larsson G, Maire M, Shakhnarovich G, 2016. Learning representations for automatic colorization. Proc 14th European Conf on Computer Vision, p.577–593.  https://doi.org/10.1007/978-3-319-46493-0_35 CrossRefGoogle Scholar
  19. Lecoutre A, Négrevergne B, Yger F, 2017. Recognizing art style automatically in painting with deep learning. Proc 9th Asian Conf on Machine Learning, p.327–342.Google Scholar
  20. Li WB, Zhang PC, Zhang L, et al., 2019. Object-driven text-to-image synthesis via adversarial training. https://arxiv.org/abs/1902.10740
  21. Lin TY, Maire M, Belongie S, et al., 2014. Microsoft COCO: common objects in context. Proc 13th European Conf on Computer Vision, p.740–755.  https://doi.org/10.1007/978-3-319-10602-1_48 CrossRefGoogle Scholar
  22. Liu YF, Qin ZC, Wan T, et al., 2018. Auto-painter: cartoon image generation from sketch by using conditional Wasserstein generative adversarial networks. Neurocomputing, 311:78–87.  https://doi.org/10.1016/j.neucom.2018.05.045 CrossRefGoogle Scholar
  23. Lowenthal D, 2005. Natural and cultural heritage. Int J Herit Stud, 11(1):81–92.  https://doi.org/10.1080/13527250500037088 CrossRefGoogle Scholar
  24. Mao XF, Wang SH, Zheng LY, et al., 2018. Semantic invariant cross-domain image generation with generative adversarial networks. Neurocomputing, 293:55–63.  https://doi.org/10.1016/j.neucom.2018.02.092 CrossRefGoogle Scholar
  25. Meng QY, Zhang HH, Zhou MQ, et al., 2018. The classification of traditional Chinese painting based on CNN. Proc 4th Int Conf on Cloud Computing and Security, p.232–241.  https://doi.org/10.1007/978-3-030-00009-7_22 CrossRefGoogle Scholar
  26. Połap D, Woźniak M, Wei W, et al., 2018. Multi-threaded learning control mechanism for neural networks. Fut Gener Comput Syst, 87:16–34.  https://doi.org/10.1016/j.future.2018.04.050 CrossRefGoogle Scholar
  27. Quinlan JR, 1986. Induction of decision trees. Mach Learn, 1(1):81–106.  https://doi.org/10.1007/BF00116251 Google Scholar
  28. Smith L, Akagawa N, 2008. Intangible Heritage. Routledge, London, UK.CrossRefGoogle Scholar
  29. Yu L, Liu H, 2003. Feature selection for high-dimensional data: a fast correlation-based filter solution. Proc 20th Int Conf on Machine Learning, p.856–863.Google Scholar
  30. Zhang R, Isola P, Efros AA, 2016. Colorful image colorization. Proc 14th European Conf on Computer Vision, p.649–666.  https://doi.org/10.1007/978-3-319-46487-9_40 CrossRefGoogle Scholar
  31. Zhu JY, Park T, Isola P, et al., 2017. Unpaired image-to-image translation using cycle-consistent adversarial networks. IEEE Int Conf on Computer Vision, p.2223–2232.  https://doi.org/10.1109/ICCV.2017.244
  32. Zujovic J, Gandy L, Friedman S, et al., 2009. Classifying paintings by artistic genre: an analysis of features & classifiers. IEEE Int Workshop on Multimedia Signal Processing, p.1–5.  https://doi.org/10.1109/MMSP.2009.5293271

Copyright information

© Zhejiang University and Springer-Verlag GmbH Germany, part of Springer Nature 2019

Authors and Affiliations

  1. 1.Department of Computer Science and Information TechnologyLa Trobe UniversityMelbourneAustralia
  2. 2.School of Art and DesignGuangdong University of TechnologyGuangzhouChina

Personalised recommendations