Abstract
Stroke based rendering methods have recently become a popular solution for the generation of stylized paintings. However, the current research in this direction is focused mainly on the improvement of final canvas quality, and thus often fails to consider the intelligibility of the generated painting sequences to actual human users. In this work, we motivate the need to learn more human-intelligible painting sequences in order to facilitate the use of autonomous painting systems in a more interactive context (e.g. as a painting assistant tool for human users or for robotic painting applications). To this end, we propose a novel painting approach which learns to generate output canvases while exhibiting a painting style which is more relatable to human users. The proposed painting pipeline Intelli-Paint consists of 1) a progressive layering strategy which allows the agent to first paint a natural background scene before adding in each of the foreground objects in a progressive fashion. 2) We also introduce a novel sequential brushstroke guidance strategy which helps the painting agent to shift its attention between different image regions in a semantic-aware manner. 3) Finally, we propose a brushstroke regularization strategy which allows for \(\sim \)60–80% reduction in the total number of required brushstrokes without any perceivable differences in the quality of generated canvases. Through both quantitative and qualitative results, we show that the resulting agents not only show enhanced efficiency in output canvas generation but also exhibit a more natural-looking painting style which would better assist human users express their ideas through digital artwork.
This is a preview of subscription content, access via your institution.
Buying options




Notes
- 1.
For simplicity, we primarily use \(L=2\) in the main paper. Further details on extending progressive layering to \(L>2\) are provided in Appendix A.2.
References
Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: International Conference on Machine Learning, pp. 214–223. PMLR (2017)
Bangham, J.A., Gibson, S.E., Harvey, R.W.: The art of scale-space. In: BMVC. pp. 1–10. Citeseer (2003)
Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. arXiv preprint arXiv:1706.05587 (2017)
Chen, W., Hays, J.: SketchyGAN: towards diverse and realistic sketch to image synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9416–9425 (2018)
Collomosse, J.P., Hall, P.M.: Genetic paint: a search for salient paintings. In: Rothlauf, F., et al. (eds.) EvoWorkshops 2005. LNCS, vol. 3449, pp. 437–447. Springer, Heidelberg (2005). https://doi.org/10.1007/978-3-540-32003-6_44
Collomosse, J.P., Hall, P.M.: Salience-adaptive painterly rendering using genetic search. Int. J. Artif. Intell. Tools 15(04), 551–575 (2006)
Collomosse, J., Hall, P.: Painterly rendering using image salience. In: Proceedings 20th Eurographics UK Conference, pp. 122–128. IEEE (2002)
Frans, K., Cheng, C.Y.: Unsupervised image to sequence translation with canvas-drawer networks. arXiv preprint arXiv:1809.08340 (2018)
Frans, K., Soros, L., Witkowski, O.: Clipdraw: exploring text-to-drawing synthesis through language-image encoders. arXiv preprint arXiv:2106.14843 (2021)
Gagniuc, P.A.: Markov Chains: From Theory to Implementation and Experimentation. John Wiley & Sons, Hoboken (2017)
Ganin, Y., Kulkarni, T., Babuschkin, I., Eslami, S., Vinyals, O.: Synthesizing programs for images using reinforced adversarial learning. arXiv preprint arXiv:1804.01118 (2018)
Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414–2423 (2016)
Graves, A.: Generating sequences with recurrent neural networks. arXiv preprint arXiv:1308.0850 (2013)
Ha, D., Eck, D.: A neural representation of sketch drawings. arXiv preprint arXiv:1704.03477 (2017)
Haeberli, P.: Paint by numbers: abstract image representations. In: Proceedings of the 17th Annual Conference on Computer Graphics and Interactive Techniques, pp. 207–214 (1990)
Hertzmann, A.: Painterly rendering with curved brush strokes of multiple sizes. In: Proceedings of the 25th Annual Conference on Computer Graphics and Interactive Techniques, pp. 453–460 (1998)
Hertzmann, A., et al.: A survey of stroke-based rendering. IEEE Comput. Graph. Appl. 23, 70–81. Institute of Electrical and Electronics Engineers (2003)
Huang, Z., Heng, W., Zhou, S.: Learning to paint with model-based deep reinforcement learning. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 8709–8718 (2019)
Jia, B., Brandt, J., Mech, R., Kim, B., Manocha, D.: LPaintB: learning to paint from self-supervision arXiv preprint arXiv:1906.06841 (2019)
Jocher, G., et al.: ultralytics/yolov5: v6.0 - YOLOv5n ’Nano’ models, Roboflow integration, TensorFlow export, OpenCV DNN support, October 2021. https://doi.org/10.5281/zenodo.5563715
Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time Style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694–711. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_43
Kite-Powell, J.: This AI robot will paint a canvas at sxsw 2021 (March 2021), https://www.forbes.com/sites/jenniferhicks/2021/03/10/this-ai-robot-will-paint-a-canvas-at-sxsw-2021/?sh=5b1f0d1ab449
Kotovenko, D., Wright, M., Heimbrecht, A., Ommer, B.: Rethinking style transfer: from pixels to parameterized brushstrokes. arXiv preprint arXiv:2103.17185 (2021)
Krause, J., Stark, M., Deng, J., Fei-Fei, L.: 3D object representations for fine-grained categorization. In: 4th International IEEE Workshop on 3D Representation and Recognition (3dRR-13). Sydney, Australia (2013)
Lee, C.H., Liu, Z., Wu, L., Luo, P.: MaskGAN: towards diverse and interactive facial image manipulation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2020)
Litwinowicz, P.: Processing images and video for an impressionist effect. In: Proceedings of the 24th Annual Conference on Computer Graphics and Interactive Techniques, pp. 407–414 (1997)
Liu, B., Gould, S., Koller, D.: Single image depth estimation from predicted semantic labels. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1253–1260. IEEE (2010)
Liu, S., et al.: Paint transformer: feed forward neural painting with stroke prediction. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6598–6607 (2021)
Mellor, J.F., et al.: Unsupervised doodling and painting with improved spiral. arXiv preprint arXiv:1910.01007 (2019)
Nakano, R.: Neural painters: a learned differentiable constraint for generating brushstroke paintings. arXiv preprint arXiv:1904.08410 (2019)
Nemire, B.: Ai painting robot, May 2017. https://developer.nvidia.com/blog/ai-painting-robot/
Qin, X., Zhang, Z., Huang, C., Dehghan, M., Zaiane, O., Jagersand, M.: U2-Net: going deeper with nested u-structure for salient object detection. Pattern Recogn. 106, p. 107404 (2020)
Reyner, N.: How to paint with layers - in acrylic & oil, December 2017. https://nancyreyner.com/2017/12/25/what-is-layering-for-painting/
Shiraishi, M., Yamaguchi, Y.: An algorithm for automatic painterly rendering based on local source image approximation. In: Proceedings of the 1st International Symposium on Non-photorealistic Animation and Rendering, pp. 53–58 (2000)
Singh, J., Zheng, L.: Combining semantic guidance and deep reinforcement learning for generating human level paintings. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (2021)
Singh, J., Zheng, L., Smith, C., Echevarria, J.: Paint2pix: interactive painting based progressive image synthesis and editing. In: European Conference on Computer Vision. Springer, Cham (2022)
Sochorová, Š, Jamriška, O.: Practical pigment mixing for digital painting. ACM Trans. Graph. 40(6), 1–11 (2021)
Teece, D.: 3D painting for non-photorealistic rendering. In: ACM SIGGRAPH 98 Conference Abstracts and Applications, p. 248 (1998)
Treavett, S., Chen, M.: Statistical techniques for the automated synthesis of non-photorealistic images. In: Proceedings of 15th Eurographics UK Conference, pp. 201–210 (1997)
Turk, G., Banks, D.: Image-guided streamline placement. In: Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, pp. 453–460 (1996)
Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The Caltech-UCSD Birds-200-2011 Dataset. Tech. Rep. CNS-TR-2011-001, California Institute of Technology (2011)
Wikipedia contributors: Ai-da (robot) – Wikipedia, the free encyclopedia (2022). https://en.wikipedia.org/w/index.php?title=Ai-Da_(robot) &oldid=1070639724. Accessed 7 Mar 2022
Xiang, X., Liu, D., Yang, X., Zhu, Y., Shen, X., Allebach, J.P.: Adversarial open domain adaptation for sketch-to-photo synthesis. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 1434–1444 (2022)
Xie, N., Hachiya, H., Sugiyama, M.: Artist agent: a reinforcement learning approach to automatic stroke generation in oriental ink painting. IEICE Trans. Inf. Syst. 96(5), 1134–1144 (2013)
Yang, S., Wang, Z., Liu, J., Guo, Z.: Controllable sketch-to-image translation for robust face synthesis. IEEE Trans. Image Process. 30, 8797–8810 (2021)
Zeng, K., Zhao, M., Xiong, C., Zhu, S.C.: From image parsing to painterly rendering. ACM Trans. Graph. 29(1), 1–2 (2009)
Zhao, A., Balakrishnan, G., Lewis, K.M., Durand, F., Guttag, J.V., Dalca, A.V.: Painting many pasts: Synthesizing time lapse videos of paintings. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8435–8445 (2020)
Zheng, N., Jiang, Y., Huang, D.: Strokenet: A neural painting environment. In: International Conference on Learning Representations (2018)
Zou, Z., Shi, T., Qiu, S., Yuan, Y., Shi, Z.: Stylized neural painting. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15689–15698 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Singh, J., Smith, C., Echevarria, J., Zheng, L. (2022). Intelli-Paint: Towards Developing More Human-Intelligible Painting Agents. In: Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T. (eds) Computer Vision – ECCV 2022. ECCV 2022. Lecture Notes in Computer Science, vol 13676. Springer, Cham. https://doi.org/10.1007/978-3-031-19787-1_39
Download citation
DOI: https://doi.org/10.1007/978-3-031-19787-1_39
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-19786-4
Online ISBN: 978-3-031-19787-1
eBook Packages: Computer ScienceComputer Science (R0)