Abstract
Various fonts give us various impressions, which are often represented by words. This paper proposes Impressions2Font (Imp2Font) that generates font images with specific impressions. Imp2Font is an extended version of conditional generative adversarial networks (GANs). More precisely, Imp2Font accepts an arbitrary number of impression words as the condition to generate the font images. These impression words are converted into a soft-constraint vector by an impression embedding module built on a word embedding technique. Qualitative and quantitative evaluations prove that Imp2Font generates font images with higher quality than comparative methods by providing multiple impression words or even unlearned words.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In this paper, we use the term “impression” in a broader meaning; some impression is described by words that relate more to font shapes, such as sans-serif, rather than subjective impression.
- 2.
These two differences make it very difficult to fairly compare Wang et al. [20] and our proposed method.
- 3.
As noted later, each impression word is converted to a semantic vector by word2vec [13]. Therefore, we remove too rare impression words that are not included even in the 3-million English vocabulary for training word2vec. This results in \(K=1,574\) impression words that we used in the following. Note that an impression word with hyphenation is split into sub-words, and then its semantic vector is derived by taking the sum of the semantic vectors of the sub-words.
- 4.
“HERONS” is a common word to check the font style since it contains sufficient variations of stroke shapes.
References
Azadi, S., Fisher, M., Kim, V.G., Wang, Z., Shechtman, E., Darrell, T.: Multi-content GAN for few-shot font style transfer. In: CVPR (2018)
Cha, J., Chun, S., Lee, G., Lee, B., Kim, S., Lee, H.: Few-shot compositional font generation with dual memory. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 735–751. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58529-7_43
Chen, T., Wang, Z., Xu, N., Jin, H., Luo, J.: Large-scale tag-based font retrieval with generative feature learning. In: ICCV (2019)
Davis, R.C., Smith, H.J.: Determinants of feeling tone in type faces. J. Appl. Psychol. 17(6), 742–764 (1933)
Goodfellow, I.J., et al.: Generative adversarial networks. arXiv preprint arXiv:1406.2661 (2014)
Hayashi, H., Abe, K., Uchida, S.: GlyphGAN: style-consistent font generation based on generative adversarial networks. Knowledge-Based Syst. 186, 104927 (2019)
Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: NIPS (2017)
Ikoma, M., Iwana Brian, K., Uchida, S.: Effect of text color on word embeddings. In: DAS (2020)
Jiang, Y., Lian, Z., Tang, Y., Xiao, J.: DCFont: an end-to-end deep Chinese font generation system. In: SIGGRAPH Asia (2017)
Kaneko, T., Ushiku, Y., Harada, T.: Class-distinct and class-mutual image generation with GANs. In: BMVC (2019)
Lyu, P., Bai, X., Yao, C., Zhu, Z., Huang, T., Liu, W.: Auto-encoder guided GAN for Chinese calligraphy synthesis. In: ICDAR, vol. 1, pp. 1095–1100 (2017)
Mao, Q., Lee, H.Y., Tseng, H.Y., Ma, S., Yang, M.H.: Mode seeking generative adversarial networks for diverse image synthesis. In: CVPR (2019)
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: NIPS (2013)
Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)
Odena, A., Olah, C., Shlens, J.: Conditional image synthesis with auxiliary classifier GANs. In: ICML (2017)
O’Donovan, P., Lībeks, J., Agarwala, A., Hertzmann, A.: Exploratory font selection using crowdsourced attributes. ACM Trans. Graph. 33(4), 92 (2014)
Poffenberger, A.T., Franken, R.: A study of the appropriateness of type faces. J. Appl. Psychol. 7(4), 312–329 (1923)
Shirani, A., Dernoncourt, F., Echevarria, J., Asente, P., Lipka, N., Solorio, T.: Let me choose: from verbal context to font selection. In: ACL (2020)
Vijayakumar, A., Vedantam, R., Parikh, D.: Sound-Word2Vec: learning word representations grounded in sounds. In: EMNLP (2017)
Wang, Y., Gao, Y., Lian, Z.: Attribute2font: creating fonts you want from attributes. ACM Trans. Graph. 39(4), 69 (2020)
Wang, Z., et al.: DeepFont: identify your font from an image. In: ACM Multimedia (2015)
Zhu, A., Lu, X., Bai, X., Uchida, S., Iwana, B.K., Xiong, S.: Few-shot text style transfer via deep feature similarity. IEEE Trans. Image Proc. 29, 6932–6946 (2020)
Zramdini, A., Ingold, R.: Optical font recognition using typographical features. IEEE Trans. Patt. Anal. Mach. Intell. 20(8), 877–882 (1998)
Acknowledgment
This work was supported by JSPS KAKENHI Grant Number JP17H06100.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Matsuda, S., Kimura, A., Uchida, S. (2021). Impressions2Font: Generating Fonts by Specifying Impressions. In: Lladós, J., Lopresti, D., Uchida, S. (eds) Document Analysis and Recognition – ICDAR 2021. ICDAR 2021. Lecture Notes in Computer Science(), vol 12823. Springer, Cham. https://doi.org/10.1007/978-3-030-86334-0_48
Download citation
DOI: https://doi.org/10.1007/978-3-030-86334-0_48
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-86333-3
Online ISBN: 978-3-030-86334-0
eBook Packages: Computer ScienceComputer Science (R0)