Abstract
The objective of this paper is to develop an effective photographic style transfer method while preserving the semantic correspondence between the style and content images for both scenery and portrait images. A semantic correspondence guided photographic style transfer algorithm is developed, which is to ensure that the semantic structure of the content image has not been changed while the color of the style images is being migrated. The semantic correspondence is constructed in large scale regions based on image segmentation and also in local scale patches using Nearest-neighbor Field Search in the deep feature domain. Based on the semantic correspondence, a matting optimization is utilized to optimize the style transfer result to ensure the semantic accuracy and transfer faithfulness. The proposed style transfer method is further extended to automatically retrieve the style images from a database to make style transfer more-friendly. The experimental results show that our method could successfully conduct the style transfer while preserving semantic correspondence between diversity of scenes. A user study also shows that our method outperforms state-of-the-art photographic style transfer methods.
Similar content being viewed by others
References
Arbelot B, Hurtut T, Hurtut T, Thollot J (2016) Automatic texture guided color transfer and colorization. Joint symposium on computational aesthetics and sketch based interfaces and modeling and non-photorealistic animation and rendering, pp 21–32
Barnes C, Shechtman E, Finkelstein A, Goldman DB (2009) Patchmatch: A randomized correspondence algorithm for structural image editing. ACM Trans Graph 28(3):1–11. Proc. of SIGGRAPH
Champandard A J (2016) Semantic style transfer and turning two-bit doodles into fine artworks. arXiv:1603.01768
Dong X, Shen J (2018) Triplet loss in siamese network for object tracking. Proceedings of the European conference on computer vision, pp 459–474
Dong X, Shen J, Liu Y, et al. (2019) Quadruplet network with one-shot learning for fast visual object tracking. IEEE Trans Image Process 28(7):3516–3527
Dong X, Shen J, Wang W, et al. (2018) Hyperparameter optimization for tracking with continuous deep q-learning. Proceedings of the IEEE conference on computer vision and pattern recognition, pp 518–527
Efros A A, Freeman W T (2001) Image quilting for texture synthesis and transfer. Conference on computer graphics and interactive techniques, pp 341–346
Efros A A, Leung T K (2002) Texture synthesis by non-parametric sampling. IEEE international conference on computer vision, pp 1033
Gatys L A, Ecker A S, Bethge M (2015) A neural algorithm of artistic style. arXiv:1508.06576
Hertzmann A, Jacobs C E, Oliver N, Curless B, Salesin D H (2001) Image analogies. Conference on computer graphics and interactive techniques, pp 327–340
Hristova H, Meur O L, Bouatouch K (2015) Style-aware robust color transfer. The workshop on computational aesthetics, pp 67–77
Hwang S J, Kapoor A, Kang S B (2012) Context-based automatic local image enhancement. European conference on computer vision, pp 569–582
Johnson J, Alahi A, Li F (2016) Perceptual losses for real-time style transfer and super-resolution. European conference on computer vision, pp 694–711
Kang S B (2017) Visual attribute transfer through deep image analogy. ACM Trans Graph 36(4):120
Kaur P, Zhang H, Dana K J (2017) Photo-realistic facial texture transfer. Computer Science, IEEE international conference on computer vision and pattern recognition
Levin A, Lischinski D, Weiss Y (2008) A closed-form solution to natural image matting. IEEE Trans Pattern Anal Mach Intell 30(2):228–242
Li C, Wand M (2016) Combining markov random fields and convolutional neural networks for image synthesis. IEEE international conference on computer vision and pattern recognition, pp 2479–2486
Lin T Y, Dollar P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. arXiv:1612.03144
Liu J, Yang W, Sun X, Zeng W (2016) Photo stylistic brush: Robust style transfer via superpixel-based bipartite graph. IEEE Transactions on Multimedia pp 1C–1
Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. IEEE international conference on computer vision and pattern recognition, pp 3431–3440
Luan F, Paris S, Shechtman E, Bala K (2017) Deep photo style transfer. IEEE international conference on computer vision and pattern recognition, pp 6997–7005
Pitie F, Kokaram A C, Dahyot R (2005) N-dimensional probability density function transfer and its application to color transfer. IEEE international conference on computer vision, pp 1434–1439
Pitie F, Kokaram A C, Dahyot R (2007) Automated colour grading using colour distribution transfer. Comput Vis Image Underst 107(1):123–237
Qin X, Shen J, Mao X, et al. (2015) Structured-patch optimization for dense correspondence. IEEE Trans. Multimedia 17(3):295–306
Rabin J, Ferradans S, Papadakis N (2015) Adaptive color transfer with relaxed optimal transport. IEEE international conference on image processing, pp 4852–4856
Reinhard E, Adhikhmin M, Gooch B, Shirley P (2002) Color transfer between images. IEEE Comput Graph Appl 5:34–41
Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. International conference on computing and artificial intelligence, vol 9351, pp 234–241
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556
Song X, Feng F, Han X, et al. (2018) Neural compatibility modeling with attentive knowledge distillation. The 41st international ACM SIGIR conference on research and development in information retrieval, pp 5–14
Song X, Feng F, Liu J, et al. (2017) Neurostylist: Neural compatibility modeling for clothing matching. Proceedings of the 25th ACM international conference on multimedia, pp 753–761
Wang W, Shen J (2018) Deep visual attention prediction. IEEE Trans Image Process 27(5):2368–2378
Wang W, Shen J, Ling H (2018) A deep network solution for attention and aesthetics aware photo cropping. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1–1
Wang W, Shen J, Shao L (2017) Video salient object detection via fully convolutional networks. IEEE Transactions on Image Processing, 1–1
Wang W, Shen J, Shao L, et al. (2016) Correspondence driven saliency transfer. IEEE Trans Image Process 25(11):5025–5034
Wu F, Dong W, Kong Y, Mei X, Paul J C, Zhang X (2013) Content-based colour transfer. Computer graphics forum. Wiley Online Library, pp 190–203
Yang Y, Zhao H, You L, Tu R, Wu X (2017) Semantic portrait color transfer with internet images. Multimed Tools Appl 76(1):523–541
Zhang H, Dana K (2017) Multi-style generative network for real-time transfer. arXiv:1703.06953
Acknowledgments
The authors wish to acknowledge the financial support from: (i) Chinese Natural Science Foundation under the Grant No. 61602313, 61620106008; (ii) Shenzhen Commission of Scientific Research and Innovations under the Grant No. JCYJ20170302153632883, JCYJ20160422151736824; (iii) Startup Foundation for Advanced Talents, Shenzhen; (iv) The Natural Science Foundation of Guangdong Province No. 2016A030310053, 2017A030310521.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhang, X., Zhang, X. & Xiao, Z. Deep photographic style transfer guided by semantic correspondence. Multimed Tools Appl 78, 34649–34672 (2019). https://doi.org/10.1007/s11042-019-08099-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-019-08099-7