Skip to main content
Log in

Artistic font generation network combining font style and glyph structure discriminators

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Artistic font plays an important practical value in advertising media and graphic design. The font is rendered to present a unique text effect, which is more ornamental and attractive. In order to explore more efficient font design method, this paper proposes artistic font generation network combining font style and glyph structure discriminators (ArtFontNet). We adopt the idea of generative adversarial network to obtain artistic fonts. The generator uses residual dense module to generate artistic fonts, and font style discriminator and glyph structure discriminator guide the generator together. The font style discriminator supervises the color and texture information of the entire font image. The glyph structure discriminator extracts the glyph structure and texture distribution of the generated image through the Canny edge detection operator. For the task of artistic font generation, our approach achieves significant performance compared to other existing methods. Compared with the font generation methods in the experiment, the PSNR value of the generated image in this paper is increased by 2.95 dB on average. The SSIM value is increased by 0.03 on average. The VIF value improved by 0.025 on average. The UV quantization results are maintained at 85%-90%. From both visual and objective evaluations, ArtFontNet enhances the detail fidelity and style accuracy of the generated artistic fonts.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig.10
Fig.11
Fig.12
Fig.13
Fig.14
Fig. 15
Fig. 16

Similar content being viewed by others

Data availability

The dataset used in this article is available at http://www.icst.pku.edu.cn/struct/Projects/TETGAN.html.

References

  1. Arjovsky M, Chintala S, Bottou L (2017) Wasserstein GAN https://doi.org/10.48550/arXiv.1701.07875

  2. Azadi S, Fisher M, Kim V G, Wang Z, Shechtman E, Darrell T (2018) Multi-content gan for few-shot font style transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition 7564–7573 https://doi.org/10.48550/arXiv.1712.00516

  3. Baluja S (2017) Learning typographic style: from discrimination to synthesis. Mach Vis Appl 551–568 https://doi.org/10.1007/s00138-017-0842-6

  4. Chang B, Zhang Q, Pan S, Meng L (2018) Generating handwritten chinese characters using cyclegan. In 2018 IEEE winter conference on applications of computer vision 199–207 https://doi.org/10.1109/wacv.2018.00028

  5. Gao Y, Guo Y, Lian Z, Tang Y, Xiao J (2019) Artistic glyph image synthesis via one-stage few-shot learning. ACM Trans Graphics 38(6):1–12. https://doi.org/10.1145/3355089.3356574

    Article  Google Scholar 

  6. Gatys LeonA, Ecker AlexanderS, Bethge M (2015) A Neural Algorithm of Artistic Style. https://doi.org/10.48550/arXiv.1508.06576

  7. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139–144. https://doi.org/10.1145/3422622

    Article  MathSciNet  Google Scholar 

  8. Gu X, Liu J, Zou X, Kuang P (2017). Using checkerboard rendering and deconvolution to eliminate checkerboard artifacts in images generated by neural networks. In 2017 14th Int Comp Conf Wavelet Active Media Technol Inform Process 197–200 https://doi.org/10.1109/ICCWAMTIP.2017.8301478

  9. Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville A C (2017) Improved training of wasserstein gans. Adv Neural Inform Process Syst 30 https://doi.org/10.48550/arXiv.1704.00028

  10. He K, Zhang X, Ren S, Sun J (2016). Deep residual learning for image recognition. In Proc IEEE Conf Comp Vis Pattern Recog 770–778 https://doi.org/10.48550/arXiv.1512.03385

  11. Huang G, Liu Z, Van Der Maaten L, Weinberger K Q (2017) Densely connected convolutional networks. In Proc IEEE Conf Comp Vis Pattern Recog 4700–4708 https://doi.org/10.48550/arXiv.1608.06993

  12. Isola P, Zhu J Y, Zhou T, Efros A A (2017) Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, 1125–1134 https://doi.org/10.48550/arXiv.1611.07004

  13. Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for real-time style transfer and super-resolution. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016. Proc Part II 14:694–711. https://doi.org/10.1007/978-3-319-46475-6_43

    Article  Google Scholar 

  14. Lyu P, Bai X, Yao C, Zhu Z, Huang T, Liu W (2017) Auto-encoder guided GAN for Chinese calligraphy synthesis. In 2017 14th IAPR International Conference on Document Analysis and Recognition 1 1095–1100 https://doi.org/10.1109/icdar.2017.181

  15. Mirza M, Osindero S (2014) Conditional generative adversarial nets https://doi.org/10.48550/arXiv.1411.1784

  16. Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In Proc IEEE Conf Comp Vis Pattern Recog 7794–7803 https://doi.org/10.48550/arXiv.1711.07971

  17. Wang W, Liu J, Yang S, Guo Z (2019) Typography with decor: Intelligent text style transfer. Proc IEEE/CVF Conf Comp Vis Pattern Recogn 5889–5897 https://doi.org/10.1109/cvpr.2019.00604

  18. Yang S, Liu J, Lian Z, Guo Z (2017) Awesome typography: Statistics-based text effects transfer. Proc IEEE Conf Comp Vis Pattern Recog 7464–7473 https://doi.org/10.48550/arXiv.1611.09026

  19. Yang S, Liu J, Lian Z, Guo Z (2018) Text effects transfer via distribution-aware texture synthesis. Comput Vis Image Underst 174:43–56. https://doi.org/10.1016/j.cviu.2018.07.004

    Article  Google Scholar 

  20. Yang S, Liu J, Wang W, Guo Z (2019) TET-GAN: Text effects transfer via stylization and destylization. In Proc AAAI Conf Artif Intell 33(01):1238–1245. https://doi.org/10.1609/aaai.v33i01.33011238

    Article  Google Scholar 

  21. Zhang L, Ji Y, Lin X, Liu C (2017) Style transfer for anime sketches with enhanced residual u-net and auxiliary classifier gan. In 2017 4th IAPR Asian conference on pattern recognition 506–511 https://doi.org/10.1109/acpr.2017.61

  22. Zhang H, Goodfellow I, Metaxas D, Odena A (2019) Self-attention generative adversarial networks. In Int Conf Mach Learn 7354–736. https://doi.org/10.48550/arXiv.1805.08318

  23. Zhang Y, Tian Y, Kong Y, Zhong B, Fu Y (2018) Residual dense network for image super-resolution. In Proc IEEE Conf Comp Vis Pattern Recog 2472–2481 https://doi.org/10.1109/cvpr.2018.00262

  24. Zhang XY, Yin F, Zhang YM, Liu CL, Bengio Y (2017) Drawing and recognizing chinese characters with recurrent neural network. IEEE Trans Pattern Anal Mach Intell 40(4):849–862. https://doi.org/10.1109/TPAMI.2017.2695539

    Article  PubMed  Google Scholar 

  25. Zhu J Y, Park T, Isola P, Efros A A (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision 2223–2232 https://doi.org/10.1109/iccv.2017.244

Download references

Acknowledgements

This work was supported by the two funds. They are Research on the Inheritance Technology of Ancient Inscription Calligraphy Culture Based on Artificial Intelligence, 62076200, Chinese National Natural Science Foundation. And Research and Application of Key Technologies for Vectorization of Traditional Chinese Calligraphy Based on Artificial Intelligence, 2023-YBGY-149, key research and development Project in Shaanxi Province.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Huanhuan Jia.

Ethics declarations

Confict of interest

The authors declare that they have no confict of interest.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Miao, Y., Jia, H. & Tang, K. Artistic font generation network combining font style and glyph structure discriminators. Multimed Tools Appl 83, 21883–21903 (2024). https://doi.org/10.1007/s11042-023-16396-5

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-023-16396-5

Keywords

Navigation