SSNet: Structure-Semantic Net for Chinese typography generation based on image translation

Abstract The abundant complex Chinese characters often lead to the high cost of time and labor in its typography production, which cannot meet various demands of typohraphies in daily life. Image translation methods are becoming the mainstream of typography generation to facilitate typography production. Nevertheless, current translation methods do not take the Chinese semantics and structure into account. In this paper, we propose a method called Structure-Semantic Net (SSNet) for Chinese typography generation, which utilizes disentangled stroke features from the structure module, pre-trained semantic features from the semantic module to generate target typographies. Furthermore, a novel loss called dual-masked Hausdorff distance is proposed to punish the incorrectly generated pixels to regularize the character contours, stabilizing the training process. Qualitative and quantitative results show that the proposed SSNet surpasses existing image translation methods in image quality, and the ablation study of SSNet verifies the effectiveness of each module and loss function.

[1]  Jonathon Shlens,et al.  Conditional Image Synthesis with Auxiliary Classifier GANs , 2016, ICML.

[2]  Tien-Ruey Hsiang,et al.  Generating Chinese Typographic and Handwriting Fonts from a Small Font Sample Set , 2018, 2018 International Joint Conference on Neural Networks (IJCNN).

[3]  Zunlei Feng,et al.  Neural Style Transfer: A Review , 2017, IEEE Transactions on Visualization and Computer Graphics.

[4]  Hao Jiang,et al.  Automatic Generation of Personal Chinese Handwriting by Capturing the Characteristics of Personal Handwriting , 2009, IAAI.

[5]  Paul Smolensky,et al.  Information processing in dynamical systems: foundations of harmony theory , 1986 .

[6]  Ronald N. Perry,et al.  An improved representation for stroke-based fonts , 2006, SIGGRAPH '06.

[7]  Michael G. Strintzis,et al.  Optimized transmission of JPEG2000 streams over wireless channels , 2006, IEEE Transactions on Image Processing.

[8]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[9]  Antoni B. Chan,et al.  FlexyFont: Learning Transferring Rules for Flexible Typeface Synthesis , 2015, Comput. Graph. Forum.

[10]  Zunlei Feng,et al.  Stroke Controllable Fast Style Transfer with Adaptive Receptive Fields , 2018, ECCV.

[11]  Jianguo Xiao,et al.  DCFont: an end-to-end deep chinese font generation system , 2017, SIGGRAPH Asia Technical Briefs.

[12]  Alexei A. Efros,et al.  Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[13]  Daniel P. Huttenlocher,et al.  Comparing Images Using the Hausdorff Distance , 1993, IEEE Trans. Pattern Anal. Mach. Intell..

[14]  Wenyu Liu,et al.  Auto-Encoder Guided GAN for Chinese Calligraphy Synthesis , 2017, 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR).

[15]  Roger D. Hersch,et al.  Perceptually tuned generation of grayscale fonts , 1995, IEEE Computer Graphics and Applications.

[16]  Li Fei-Fei,et al.  Perceptual Losses for Real-Time Style Transfer and Super-Resolution , 2016, ECCV.

[17]  Alan C. Bovik,et al.  Image information and visual quality , 2006, IEEE Trans. Image Process..

[18]  Wei Guo,et al.  An Automatic Chinese Font Library Generation Method by Modifying Vector Contour Curves , 2014, 2014 Tenth International Conference on Intelligent Information Hiding and Multimedia Signal Processing.

[19]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[20]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[21]  Zunlei Feng,et al.  Interpretable Partitioned Embedding for Customized Multi-item Fashion Outfit Composition , 2018, ICMR.

[22]  Chun Chen,et al.  Painterly Rendering with Vector Field Based Feature Extraction , 2006, ICAT.

[23]  Geoffrey E. Hinton,et al.  Deep Boltzmann Machines , 2009, AISTATS.

[24]  Yi-Chao Wu,et al.  Chinese Handwriting Generation by Neural Network Based Style Transformation , 2017, ICIG.

[25]  Masaki Nakagawa,et al.  Generating realistic Kanji character images from on-line patterns , 2001, Proceedings of Sixth International Conference on Document Analysis and Recognition.

[26]  Alexei A. Efros,et al.  Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Hongwei Lin,et al.  Outline Font Generating from Images of Ancient Chinese Calligraphy , 2011, Trans. Edutainment.

[28]  Takeo Igarashi,et al.  Example-Based Automatic Font Generation , 2010, Smart Graphics.

[29]  Qiong Zhang,et al.  Generating Handwritten Chinese Characters Using CycleGAN , 2018, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).

[30]  Bo Zhao,et al.  Automatic generation of large-scale handwriting fonts via style learning , 2016, SIGGRAPH Asia Technical Briefs.

[31]  Jie Li,et al.  Font generation based on least squares conditional generative adversarial nets , 2017, Multimedia Tools and Applications.

[32]  Trevor Darrell,et al.  Multi-content GAN for Few-Shot Font Style Transfer , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[33]  Shinichiro Omachi,et al.  Automatic Generation of Typographic Font From Small Font Subset , 2017, IEEE Computer Graphics and Applications.

[34]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[35]  Jan Kautz,et al.  Learning a manifold of fonts , 2014, ACM Trans. Graph..

[36]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.