skip to main content
research-article
Free Access
Just Accepted

Color Hint-guided Ink Wash Painting Colorization with Ink Style Prediction Mechanism

Authors Info & Claims
Online AM:11 April 2024Publication History
Skip Abstract Section

Abstract

We propose an end-to-end generative adversarial network that allows for controllable ink wash painting generation from sketches by specifying the colors via color hints. To the best of our knowledge, this is the first study for interactive Chinese ink wash painting colorization from sketches. To help our network understand the ink style and artistic conception, we introduced an ink style prediction mechanism for our discriminator, which enables the discriminator to accurately predict the style with the help of a pre-trained style encoder. We also designed our generator to receive multi-scale feature information from the feature pyramid network for detail reconstruction of ink wash painting. Experimental results and user study show that ink wash paintings generated by our network have higher realism and richer artistic conception than existing image generation methods.

References

  1. Elad Aharoni-Mack, Yakov Shambik, and Dani Lischinski. 2017. Pigment-based recoloring of watercolor paintings. In Proceedings of the Symposium on Non-Photorealistic Animation and Rendering. 1–11.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Mikołaj Bińkowski, Sutherland Danica J, Arbel Michael, and Gretton Arthur. 2018. Demystifying mmd gans. arXiv preprint arXiv:1801.01401(2018), 13–28.Google ScholarGoogle Scholar
  3. Caroline Chan, Durand Fredo, and Isola Phillip. 2022. Learning to generate line drawings that convey geometry and semantics. arXiv preprint arXiv:2203.12691(2022), 2054–2068.Google ScholarGoogle Scholar
  4. Nelson S-H Chu and Chiew-Lan Tai. 2005. Moxi: real-time ink dispersion in absorbent paper. ACM Transactions on Graphics (TOG) 24, 3 (2005), 504–511.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Jia Deng, Wei Dong, Socher Richard, Lijia Li, Kai Li, and Feifei Li. 2009. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition. Ieee, 248–255.Google ScholarGoogle ScholarCross RefCross Ref
  6. Lixing Dong, Shufang Lu, and Xiaogang Jin. 2014. Real-time image-based Chinese ink painting rendering. Multimedia tools and applications 69, 3 (2014), 605–620.Google ScholarGoogle Scholar
  7. Zhi Dou, Ning Wang, Baopu Li, Zhihui Wang, Haojie Li, and Bin Liu. 2021. Dual Color Space Guided Sketch Colorization. IEEE Transactions on Image Processing 30 (2021), 7292–7304.Google ScholarGoogle ScholarCross RefCross Ref
  8. Feifei Fu, Jiancheng Lv, Chenwei Tang, and Mao Li. 2021. Multi-style Chinese art painting generation of flowers. IET Image Processing 15, 3 (2021), 746–762.Google ScholarGoogle ScholarCross RefCross Ref
  9. Chengying Gao, Qi Liu, Qi Xu, Limin Wang, Jianzhuang Liu, and Changqing Zou. 2020. Sketchycoco: Image generation from freehand scene sketches. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 5174–5183.Google ScholarGoogle ScholarCross RefCross Ref
  10. Bin He, Feng Gao, Daiqian Ma, Boxin Shi, and Lingyu Duan. 2018. Chipgan: A generative adversarial network for chinese ink wash painting style transfer. In Proceedings of the 26th ACM international conference on Multimedia. 1172–1180.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Xun Huang, Mingyu Liu, Belongie Serge, and Kautz Jan. 2018. Multimodal unsupervised image-to-image translation. In Proceedings of the European conference on computer vision (ECCV). 172–189.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Xun Huang and Belongie Serge. 2017. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE international conference on computer vision. 1501–1510.Google ScholarGoogle ScholarCross RefCross Ref
  13. Johnson Justin, Alahi Alexandre, and Feifei Li. 2016. Perceptual losses for real-time style transfer and super-resolution. In European conference on computer vision. Springer, 694–711.Google ScholarGoogle Scholar
  14. Simonyanb Karen and Zisserman Andrew. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556(2014), 567–587.Google ScholarGoogle Scholar
  15. Jintae Lee. 2001. Diffusion rendering of black ink paintings using new paper and ink models. Computers & Graphics 25, 2 (2001), 295–308.Google ScholarGoogle ScholarCross RefCross Ref
  16. Junsoo Lee, Eungyeup Kim, Yunsung Lee, Dongjun Kim, Jaehyuk Chang, and Jaegul Choo. 2020. Reference-based sketch image colorization using augmented-self reference and dense semantic correspondence. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5801–5810.Google ScholarGoogle ScholarCross RefCross Ref
  17. Bo Li, Caiming Xiong, Tianfu Wu, Yu Zhou, Lun Zhang, and Rufeng Chu. 2018. Neural abstract style transfer for chinese traditional painting. In Asian Conference on Computer Vision. Springer, 212–227.Google ScholarGoogle Scholar
  18. Chengze Li, Xueting Liu, and Tientsin Wong. 2017. Deep extraction of manga structural lines. ACM Transactions on Graphics (TOG) 36, 4 (2017), 1–12.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Daoyu Lin, Yang Wang, Guangluan Xu, Jun Li, and Kun Fu. 2018. Transform a simple sketch to a chinese painting by a multiscale deep neural network. Algorithms 11, 1 (2018), 4.Google ScholarGoogle ScholarCross RefCross Ref
  20. Tsungyi Lin, Dollár Piotr, Girshick Ross, Kaiming He, Hariharan Bharath, and Belongie Serge. 2017. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2117–2125.Google ScholarGoogle ScholarCross RefCross Ref
  21. Thomas Lindemeier, Jörg Marvin Gülzow, and Oliver Deussen. 2018. Painterly rendering using limited paint color palettes. In Proceedings of the Conference on Vision, Modeling, and Visualization. 135–145.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Bingchen Liu, Kunpeng Song, Yizhe Zhu, and Elgammal Ahmed. 2020. Sketch-to-art: Synthesizing stylized art images from sketches. In Proceedings of the Asian Conference on Computer Vision. 125–149.Google ScholarGoogle Scholar
  23. Heusel Martin, Ramsauer Hubert, Unterthiner Thomas, Nessler Bernhard, and Hochreiter Sepp. 2017. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems 30 (2017), 678–689.Google ScholarGoogle Scholar
  24. Taesung Park, Mingyu Liu, Tingchun Wang, and Junyan Zhu. 2019. Semantic image synthesis with spatially-adaptive normalization. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2337–2346.Google ScholarGoogle ScholarCross RefCross Ref
  25. Isola Phillip, Junyan Zhu, Tinghui Zhou, and Efros Alexei A. 2017. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition. 1125–1134.Google ScholarGoogle Scholar
  26. Shaham Tamar Rott, Gharbi Michaël, Richard Zhang, Shechtman Eli, and Michaeli Tomer. 2021. Spatially-adaptive pixelwise networks for fast image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 14882–14891.Google ScholarGoogle Scholar
  27. Hicsonmez Samet, Samet Nermin, Akbas Emre, and Duygulu Pinar. 2021. Adversarial Segmentation Loss for Sketch Colorization. In 2021 IEEE International Conference on Image Processing (ICIP). IEEE, 2403–2407.Google ScholarGoogle Scholar
  28. Meijun Sun, Jizhou Sun, and Bin Yun. 2005. Physical modeling of” Xuan” paper in the simulation of Chinese ink-wash drawing. In International Conference on Computer Graphics, Imaging and Visualization (CGIV’05). IEEE, 317–322.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Chungming Wang and Renjie Wang. 2007. Image-based color ink diffusion rendering. IEEE Transactions on Visualization and Computer Graphics 13, 2(2007), 235–246.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Tingchun Wang, Mingyu Liu, Junyan Zhu, Andrew Tao, Kautz Jan, and Catanzaro Bryan. 2018. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE conference on computer vision and pattern recognition. 8798–8807.Google ScholarGoogle ScholarCross RefCross Ref
  31. Yuan Wang, Weibo Zhang, and Peng Chen. 2019. Chinastyle: A mask-aware generative adversarial network for chinese traditional image translation. In SIGGRAPH Asia 2019 Technical Briefs. 5–8.Google ScholarGoogle Scholar
  32. Der-Lor Way, Shengwen Huang, and Zenchung Shih. 2003. Physical-based Model of Ink Diffusion in Chinese Paintings. In WSCG. 215–226.Google ScholarGoogle Scholar
  33. Ning Xie, Laga Hamid, Saito Sugru, and Nakajima Masayuki. 2010. IR2s: interactive real photo to Sumi-e. In Proceedings of the 8th international symposium on non-photorealistic animation and rendering. 63–71.Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Tianchen Xu, Lijie Yang, and Enhua Wu. 2012. Stroke-based real-time ink wash painting style rendering for geometric models. In SIGGRAPH Asia 2012 Technical Briefs. 1–4.Google ScholarGoogle Scholar
  35. Alice Xue. 2021. End-to-end chinese landscape painting creation using generative adversarial networks. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 3863–3871.Google ScholarGoogle ScholarCross RefCross Ref
  36. Jinhui Yu, Guoming Luo, and Qunsheng Peng. 2003. Image-based synthesis of Chinese landscape painting. Journal of Computer Science and Technology 18, 1 (2003), 22–28.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Jiajing Zhang, Jinhui Yu, Yongwei Miu, and Ren Peng. 2021. Self-Adaptive Computational Aesthetic Evaluation of Chinese Ink Paintings Based on Deep Learning. Journal of Computer-Aided Design & Computer Graphics/Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao 33, 9 (2021), 16–25. (in Chinese).Google ScholarGoogle Scholar
  38. Lvmin Zhang, Yi Ji, Xin Lin, and Chunping Liu. 2017. Style transfer for anime sketches with enhanced residual u-net and auxiliary classifier gan. In 2017 4th IAPR Asian conference on pattern recognition (ACPR). IEEE, 506–511.Google ScholarGoogle ScholarCross RefCross Ref
  39. Lvmin Zhang, Chengze Li, Tientsin Wong, Yi Ji, and Chunping Liu. 2018. Two-stage sketch colorization. ACM Transactions on Graphics (TOG)(2018), 1–14.Google ScholarGoogle Scholar
  40. Richard Zhang, Isola Phillip, Efros Alexei A, Shechtman Eli, and Oliver Wang. 2018. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition. 586–595.Google ScholarGoogle ScholarCross RefCross Ref
  41. Mingtian Zhao and Songchun Zhu. 2013. Abstract painting with interactive control of perceptual entropy. ACM Transactions on Applied Perception (TAP) 10, 1 (2013), 1–21.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Chengyu Zheng and Yuan Zhang. 2018. Two-stage color ink painting style transfer via convolution neural network. In 2018 15th International Symposium on Pervasive Systems, Algorithms and Networks (I-SPAN). IEEE, 193–200.Google ScholarGoogle ScholarCross RefCross Ref
  43. Le Zhou, Qiufeng Wang, Kaizhu Huang, and Chenghung Lo. 2019. An interactive and generative approach for chinese shanshui painting document. In 2019 International Conference on Document Analysis and Recognition (ICDAR). IEEE, 819–824.Google ScholarGoogle ScholarCross RefCross Ref
  44. Junyan Zhu, Taesung Park, Phillip Isola, and Efros Alexei A. 2017. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision. 2223–2232.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Color Hint-guided Ink Wash Painting Colorization with Ink Style Prediction Mechanism

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Applied Perception
      ACM Transactions on Applied Perception Just Accepted
      ISSN:1544-3558
      EISSN:1544-3965
      Table of Contents

      Copyright © 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Online AM: 11 April 2024
      • Accepted: 31 March 2024
      • Revised: 15 February 2024
      • Received: 26 May 2023
      Published in tap Just Accepted

      Check for updates

      Qualifiers

      • research-article
    • Article Metrics

      • Downloads (Last 12 months)41
      • Downloads (Last 6 weeks)41

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader