Abstract
There are various font styles in the world. Different styles give different impressions and readability. This paper analyzes the relationship between font styles and contextual factors that might affect font style selection with large-scale datasets. For example, we will analyze the relationship between font style and its surrounding object (such as “bus”) by using about 800,000 words in the Open Images dataset. We also use a book cover dataset to analyze the relationship between font styles with book genres. Moreover, the meaning of the word is assumed as another contextual factor. For these numeric analyses, we utilize our own font-style feature extraction model and word2vec. As a result of co-occurrence-based relationship analysis, we found several instances of specific font styles being used for specific contextual factors.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
- 3.
We tried padding to unify the width of the images, but this concatenation-based method was better for style feature extraction.
References
Baek, Y., Lee, B., Han, D., Yun, S., Lee, H.: Character region awareness for text detection. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9365–9374 (2019)
Chen, G., Yang, J., Jin, H., Brandt, J., Shechtman, E., Agarwala, A., Han, T.X.: Large-scale visual font recognition. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3598–3605 (2014)
Choi, S., Matsumura, S., Aizawa, K.: Assist users’ interactions in font search with unexpected but useful concepts generated by multimodal learning. In: Proceedings of the International Conference on Multimedia Retrieval, pp. 235–243 (2019)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171–4186 (2019)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778 (2016)
Ikoma, M., Iwana, B.K., Uchida, S.: Effect of text color on word embeddings. In: Bai, X., Karatzas, D., Lopresti, D. (eds.) DAS 2020. LNCS, vol. 12116, pp. 341–355. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-57058-3_24
Jiang, S., Wang, Z., Hertzmann, A., Jin, H., Fu, Y.: Visual font pairing. IEEE Trans. Multimedia 22(8), 2086–2097 (2019)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Krasin, I., et al.: OpenImages: a public dataset for large-scale multi-label and multi-class image classification. Dataset available from https://storage.googleapis.com/openimages/web/index.html (2017)
Kulahcioglu, T., De Melo, G.: Predicting semantic signatures of fonts. In: Proceedings of the International Conference on Semantic Computing (ICSC), pp. 115–122. IEEE (2018)
Kulahcioglu, T., De Melo, G.: Paralinguistic recommendations for affective word clouds. In: Proceedings of the International Conference on Intelligent User Interfaces, pp. 132–143 (2019)
Kulahcioglu, T., De Melo, G.: Fonts like this but happier: a new way to discover fonts. In: Proceedings of the International Conference on Multimedia, pp. 2973–2981 (2020)
Matsumura, S., Choi, S., Aizawa, K.: Font search across various languages based on multimodal learning. In: Proceedings of the Conference on Multimedia Information Processing and Retrieval (MIPR), pp. 173–176. IEEE (2020)
Mikolov, T., Chen, K., Corrado, G.S., Dean, J.: Efficient estimation of word representations in vector space. In: International Conference on Learning Representations (2013)
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Proceedings of the Advances in Neural Information Processing Systems (NIPS) 26 (2013)
O’Donovan, P., Lībeks, J., Agarwala, A., Hertzmann, A.: Exploratory font selection using crowdsourced attributes. ACM Trans. Graph. 33(4), 1–9 (2014)
Shinahara, Y., Karamatsu, T., Harada, D., Yamaguchi, K., Uchida, S.: Serif or sans: visual font analytics on book covers and online advertisements. In: Proceedings of the International Conference on Document Analysis and Recognition (ICDAR), pp. 1041–1046. IEEE (2019)
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)
Takeshita, K., Shioyama, J., Uchida, S.: Label or message: a large-scale experimental survey of texts and objects co-occurrence. In: Proceedings of the International Conference on Pattern Recognition (ICPR), pp. 6227–6234. IEEE (2021)
Tibshirani, R., Walther, G., Hastie, T.: Estimating the number of clusters in a data set via the gap statistic. J. Roy. Statist. Soc. Ser. B (Statist. Methodol.) 63(2), 411–423 (2001)
Tsuji, K., Uchida, S., Iwana, B.K.: Using robust regression to find font usage trends. In: Barney Smith, E.H., Pal, U. (eds.) ICDAR 2021. LNCS, vol. 12917, pp. 126–141. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-86159-9_9
Wang, Z., et al.: DeepFont: identify your font from an image. In: Proceedings of the International Conference on Multimedia, pp. 451–459 (2015)
Yim, M., Kim, Y., Cho, H.-C., Park, S.: SynthTIGER: synthetic text image GEneratoR towards better text recognition models. In: Lladós, J., Lopresti, D., Uchida, S. (eds.) ICDAR 2021. LNCS, vol. 12824, pp. 109–124. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-86337-1_8
Zhao, N., Cao, Y., Lau, R.W.: Modeling fonts in context: Font prediction on web designs. In: Computer Graphics Forum, pp. 385–395. Wiley Online Library (2018)
Acknowledgment
This work was supported in part by JSPS KAKENHI Grant Numbers JP22H00540 and JP21H03511.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Yasukochi, N., Hayashi, H., Haraguchi, D., Uchida, S. (2023). Analyzing Font Style Usage and Contextual Factors in Real Images. In: Fink, G.A., Jain, R., Kise, K., Zanibbi, R. (eds) Document Analysis and Recognition - ICDAR 2023. ICDAR 2023. Lecture Notes in Computer Science, vol 14189. Springer, Cham. https://doi.org/10.1007/978-3-031-41682-8_21
Download citation
DOI: https://doi.org/10.1007/978-3-031-41682-8_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-41681-1
Online ISBN: 978-3-031-41682-8
eBook Packages: Computer ScienceComputer Science (R0)