This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

End-to-End Rubbing Restoration Using Generative Adversarial Networks

Gongbo Sun
Beijing National Day School
Beijing, China
   Zijie Zheng
Beijing National Day School
Beijing, China
   Ming Zhang
School of Computer Science,
Peking University
mzhang_cs@pku.edu.cn
Corresponding author.
Abstract

Rubbing restorations are significant for preserving world cultural history. In this paper, we propose the RubbingGAN model for restoring incomplete rubbing characters. Specifically, we collect characters from the Zhang Menglong Bei and build up the first rubbing restoration dataset. We design the first generative adversarial network for rubbing restoration.111The dataset and codes are available at https://github.com/qingfengtommy/RubbingGAN Based on the dataset we collect, we apply the RubbingGAN to learn the Zhang Menglong Bei font style and restore the characters. The results of experiments show that RubbingGAN can repair both slightly and severely incomplete rubbing characters fast and effectively.

1 Introduction

Rubbings are ink-on-paper copies of engraved or cast inscriptions and designs on what are mostly cultural objects of metal, stone, and other firm substances [29]. The rubbing images are shown in Figure 1.

Refer to caption
(a) Slightly incomplete Rubbing
Refer to caption
(b) Severely incomplete Rubbing
Figure 1: Rubbing characters from ZhangMenglong Bei

The history of rubbings is about 1400 years and rubbings play an important role in providing supportive evidence and clues for cultural relics restoration. However, due to the weathering process in a long time, the characters left on the rubbings usually are incomplete, as shown in Figure 1(b). Therefor, repairing rubbings is significant for both the Asian history preservation and the world culture inheritance. In traditional rubbings restoration, artists need to confirm the acidity first, and then based on the breakage level to determine corresponding methods. For instance, Li and his colleagues repair Xuan Mi Ta Bei by using a method called bracket mounting (tuo biao fa) [9]. However, the traditional rubbing restoration method requires a lot of efforts from artists, which is time consuming and impossible for large-scale artworks repairing.

In contrast to the traditional manual way of repairing cultural relics, the computer scientists have proposed a digital method called inpainting [17]. Digital inpainting is a way of reconstructing missing or damaged part in the digital images or videos, which can be divided into three categories: structural inpainting, textural inpainting, and combined structural and textural inpainting [1]. Even though the current inpainting methods can achieve delightful performance in repairing a number of relics such as images [37, 6, 12], they cannot be directly applied in rubbing restoration. There are two possible reasons: 1) the structure and texture of the Asian characters on the rubbings are complex and diverse. 2) some Asian characters are similar and there is a slight difference between them.

Another field relevant to rubbing restoration is the font style transfer problem. The font style transfer or the character font generation is a problem which developed from the image style transfer question. Recently, there are some dedicated works on Chinese fonts generation. The most representative work is zi2zi [32].

However, methods in traditional Chinese font style transfer cannot be applied in the rubbing restoration task directly. First, the main objective for rubbing restoration is to restore the incomplete characters, which indicates that we do not know which character needed to generate, and have to recognize it firstly. Second, characters in the datasets used in Chinese font style transfer are usually digital designed characters or just the standard printed fonts, such as KaiTi_GB2312. The digital or printed font have already be processed elaborately during the designing period of font libraries. And it is relatively easy to transfer the style between two different domain of printed fonts due to the consistency and continuity of the font style. However, in the rubbing restoration question, the characters from the dataset are more aesthetic and unique. The difficulties of the rubbing restoration task include repairing the rubbing characters and recognizing the aesthetic and diverse font style in the rubbings. In other words, we must use the characters which come from the rubbings to realize the job of repairing the incomplete characters and preserving the correspond rubbing style.

In this paper, we propose a method named as Rubbing Generative Adversarial Network (RubbingGAN) that can restore the incomplete rubbing characters. Specifically, we collect the characters from a rubbing called Zhang Menglong Bei to the dataset. Zhang Menglong Bei has a very high artistic value in the field of calligraphy and has a number of incomplete characters to repair [33, 10]. Based on the dataset we collect, we apply the RubbingGAN to restore the characters.

In short, our main contributions are summarized as follows.

  • We collect the first rubbing restoration dataset of a Chinese rubbing called Zhang Menglong Bei, including 337 complete characters as the training set, 75 complete characters as the testing set, and 34 incomplete characters as the restoration set.

  • We design a RubbingGAN to synthesize the rubbing fonts and repair the incomplete characters based on the training set. As far as we know, it is the first machine learning model in restoring the Chinese rubbing.

  • The experiments corroborate that the RubbingGAN can learn the rubbing fonts in Zhang Menglong Bei, and can repair both the slightly and severely incomplete characters.

2 Related works

For the Rubbing restoration problem, we find out that the recognition for the incomplete characters is not the most difficult task. The main difficulty in rubbing restoration is to preserve the aesthetic value of the artworks. Therefor, we mainly focus on artistic style transferring and we regard the rubbing restoration as the image-to-image translation problem. The related works of image-to-image translation are discussed in the below.

2.1 Generative Adversarial Networks

The generative adversarial networks is proposed by Ian J. Goodfellow and his colleagues in 2014 [15]. The main objective of the GAN is a min-max game. The generator try to deceive the discriminator by generating the high similarity pictures with the source datasets. On the other hand, the discriminator try to recognize the fake and real images. With the training, generator can generate high similarity images with the original input source. Based on the Ian J. Goodfellow’s work, the M. Mirza and S. Osindero propose a method that directs the data generation by adding the label information into the model [25]. And the DCGAN first introduced a deep convolutional network to improve the image quality [24]. However, the GAN also has some problems. First, the GAN may encounters the mode collapses, which due to the unbalance ability between the generator and the discriminator. Second, the GAN can not converge easily in the training period. Therefor, there are some works try to solve these problems. For example, the researchers propose WGAN which using the Wasserstein distance to train the GAN model. And based on the WGAN [2], researchers propose the BEGAN model that aims to match auto-encoder loss distribution using a loss derived from the Wasserstein distance.

2.2 Image-to-Image translation

Image-to-image translation is a task of translating one possible representation of a scene into another, given sufficient training data [18]. The representative work of translation is a conditional GAN-based model called pix2pix. It performs image-to-image translation on paired data. It uses a "U-Net" [27] based architecture as its generator and uses a convolutional "PatchGAN" auto encoder decoder as its discriminator [21]. In our work, the generation of rubbing style characters can be regarded as a kind of image-to-image translation. The limitation of the pix2pix model is only supporting the paired datasets [18]. Therefor, the researchers propose the model such as CylcleGAN that achieving the cross-domain style transfer [7]. However, the model like CylcleGAN can not apply into the rubbing restoration task. Because the precondition of the CycleGAN is that we have the enough data from the different domain. However, this is impractical in the rubbing restoration task. Because there are huge amounts of artworks only have the characters around hundred level, which is not enough for training the CycleGAN model.

2.3 Character Fonts Generation

There are a plenty of relative works on neural style transfer or on the character generation. Although the works on the Chinese characters generation are fewer than the English characters, there are some pioneer researches in this field. It has been studied since the beginning of the digital age. The problem can be formulated either as a personal handwritten characters generation problem [20], or an automatic printed font generation problem [23]. In the early study of glyph synthesis, researchers focused on the geometric modeling, which relied on the hierarchical representation of simple strokes [34, 35, 22, 36, 7]. They decompose the characters into a hierarchy of reusable components and then to compose the target character in the personalized style. In fact, these methods only focuses on the local characters’ features instead of the whole style, in other words, they are limited to particular glyph topology. With the development and popularity of deep neural networks, the topic of style transfer is frequently discussed. Researchers have designed new models which could add the artistic style to image content [13, 14, 3] successfully. And there are some works used the idea of neural style transfer to do the character generation problem [3, 39, 4]. This topic gradually evolves to image-to-image translation problem [32, 26, 19, 31]. The image-to-image translation problem focuses on converting objects from one domain to another domain. However, most existing methods require an impractical large-size dataset to train the network. For example, [32, 26, 30, 38] require 3000 paired characters which costs huge labor resources. In [8], researchers reduce the paired data number to 775 samples. However, this quantities still exceed the total number of the characters in most rubbings. And the dataset they adopt is either font-rendered character images or brush-written character images. Compared with these character images, the rubbing images are more irregular and incomplete. To the best of our knowledge, there is no available collection of rubbing dataset and there is no existing paper using rubbing dataset to conduct experiment.

3 Dataset

In this section, We first briefly introduce ZhangMenglong Bei from an artistic perspective. Then we describe the method for data collection and provide the basic information of our collected dataset.

3.1 A brief introduction of ZhangMenglong Bei

As we have mentioned above, the existed dataset of Chinese font style are usually generated by the computer, such as KaiTi_GB2312. Comparing with the traditional rubbing font style characters, these standard printed font style have a very short history and lack of the value for preserving traditional culture. In our task, we care about characters that written on rubbings in ancient ages. Along the history of Chinese calligraphy, there are a huge number of excellent art works that were created during different dynasties. The categories of traditional calligraphy can be divided into five main styles following the chronological order of appearance: Seal Script, Clerical Script, Cursive Script, Semi-Cursive Script and Standard Script (Kai Ti). Between the adjacent two styles, there is a transition style that assimilates the previous core idea and guides the after font style development. The Wei Bei is a famous transitional calligraphy system linking the past and the future. Wei Bei calligraphy refers to the Chinese characters engraved on the tablet during the Wei, Jin, and the Northern and Southern Dynasties. Wei Bei calligraphy has a great influence on the formation of Standard Script in the Later Sui and Tang dynasties. That is to say, Wei Bei calligraphy can be regarded as a predecessor of Standard Script, which has with a high artistic, cultural and research value [11]. However, due to the long history, most Wei Bei tablets are usually seriously damaged. A number of characters in each Wei Bei are severely incomplete. Thus, it is necessary to design a method to learn the font style of Wei Bei and restore the incomplete characters. Zhang Menglong Bei is the most prestigious work in the Northern Wei Dynasty and receives high praises from famous calligraphers such as Kang Youwei, Yang Shoujin, Qi gong.

3.2 Dataset collection

ZhangMenglong Bei has various versions online, and the version we used to extract characters is from the website. We use python to write a web crawler that automatically downloading Zhang Menglong Bei, which is shown in Figure 2. Then, we can find that Zhang Menglong Bei consist of both the complete characters and the incomplete characters.

Refer to caption
Refer to caption
Refer to caption
Figure 2: The images of ZhangMenglong Bei downloaded from the web.

To train the model, we first crop the complete characters. The cropping progress is achieved manually. The datasets is illustrated in Figure 3.

Refer to caption
(a) Training Dataset
Refer to caption
(b) Restoration Dataset
Figure 3: Dataset Overview

Since the characters we have cropped are clear, we label the ZhangMenglong style characters through using the same character in another font and joint it horizontally. The training dataset consists of 337 images and the size of the each pair data is 256×128256\times 128 with 3 color channels. The validation dataset consists of 75 images and the image information is as same as the training dataset. In addition there is a reconstruction or restoration dataset, the total number of images in this dataset is 34. The size of images in this dataset is 128×128128\times 128.

In the training process, we crop the pair data first and get two images that the size are 128×128×3128\times 128\times 3 [standard font image and Zhang Menglong font image]. During the training period, we resize both the standard font characters and the Zhang Menglong font characters in pair data from the size of 128×128×3128\times 128\times 3 to 256×256×3256\times 256\times 3 by using bilinear interpolation. And we convert the image into correspond tensor. The input tensor size for Generator is 256×256×3256\times 256\times 3 and the output tensor size is 256×256×3256\times 256\times 3. For Discriminator1, we concatenate the standard image tensor with the Zhang Menglong font image tensor or with the generated image tensor to get the input tensor. The input tensor size for Discriminator1 is 256×256×6256\times 256\times 6. For Discriminator2, we input the real Zhang Menglong font tensor or the generated font tensor, which size are both 256×256×3256\times 256\times 3.

For the rest characters in ZhangMenglong Bei, they are slightly or severely damaged, as shown in Fig. 3(b). We collect 34 characters and form a dataset to save these incomplete characters. Each character is exhibited as an image with 128×128128\times 128 pixels. This dataset is named as the incomplete character dataset. We use the incomplete character dataset to corroborate the model can repair the characters and keep the ZhangMenglong Bei font style.

4 Proposed Method

In this section, we will introduce the main architecture of the RubbingGAN in details. The objective of RubbingGAN is to learn a mapping from an original font style source to a specific rubbing font style. We denote the original font KaiTi_GB2312 as {X}\{X\}, and the target ZhangMenglong Bei font the {Y}\{Y\}. The RubbingGAN is to learn a mapping G:XYG:X\to Y.

4.1 The Architecture of Rubbing GAN

We propose a framework for Rubbing characters restoration that transferred standard font image xx into Zhang Menglong font yy. Our framework is designed based on the pix2pix [18] and BEGAN [5]. The architecture of our model is shown in Figure 4.

Figure 4: RubbingGAN architecture
Refer to caption

The RubbingGAN includes one generator denoted by G and two discriminators, denoted by D1 and D2.

Generator.

From previous work [32] and [24], we adopt "U-Net" structure as our generator which adds skip connections between layers. The Generator structure is shown in Figure 5. In the rubbing restoration problem, the core problem is to learn the structure changes and the correspond textures. The "U-net" architecture provide high structure semantics with the low-level information.

Refer to caption
Figure 5: The architecture of U-net. The input size of the image is 256×256×3256\times 256\times 3, the output of the Generator is a Zhangfont style image with size of 256×256×3256\times 256\times 3

Discriminator1.

We designed Discriminator1 based on the PatchGAN [21], which encourages GAN to focus both on the high-frequencies and low-frequencies structure features loss. Discriminator1 only penalizes structure at the scale of patches, which can be seen as a type of style/font loss as mentioned in pix2pix.

Discriminator2.

The discriminator2 is an auto encoder-decoder architecture. We designed the structure of Discriminator2 in terms of the BEGAN literature [5]. The input size of the image for Discriminator2 is 256×256×3256\times 256\times 3, the output size of the Discriminator is also 256×256×3256\times 256\times 3. The structure of Discriminator2 is really easy, and plays a good effect in image generation. Another advantage of using Discriminator 2 is that in BEGAN researchers propose a method that measured the convergence, which solve the problem of difficulty of training a traditional GAN. Therefore, we would like to use the Discriminator2 to improve generated images quality and also avoid the possibility of mode collapse. The architecture of Discriminator2 is shown in Figure 6.

Refer to caption
Figure 6: The architecture of Discriminator2.

The Figure 7 shows a whole procedure of our rubbing restoration, which recognized the characters first, and then generate the rubbing font characters.

Figure 7: The whole process of our Rubbing restoration
Refer to caption

Losses Our model losses mainly consist of three parts: the adversarial losses from the traditional conditional GAN, the loss from the L1 distance distance, and the loss from the reconstruction. We use the conditional adversarial loss to generate the realistic image. The loss can be expressed as:

cGAN(G,D1)=logD1(x,y)+log(1D1(x,y^))\begin{split}\mathcal{L}_{cGAN}(G,D_{1})=logD_{1}(x,y)+log(1-D_{1}(x,\hat{y}))\end{split} (1)

In order to improve the quality of the generated images, we add the L1 distance into the model. The L1 loss is expressed as:

L1(G)=𝔼x,y[yG(x)]\mathcal{L}_{L1}(G)=\mathbb{E}_{x,y}[||{y-G(x)}||] (2)

In order to solve the problem of mode collapses, we use idea from the BEGAN to balance the ability of the generator and discriminator 2. We use an equilibrium term in RubbingGAN:

𝔼[(G(z))]=γ𝔼[(x)]\mathbb{E}\left[\mathcal{L}(G(z))\right]=\gamma\mathbb{E}\left[\mathcal{L}(x)\right] (3)

where the hype-parameter λ[0,1]\lambda\in[0,1] is used to control the output diversity of the Generator. The ()\mathcal{L}(\cdot) represents the reconstruction loss function for the auto-encoder. The loss function for the generator and discriminator2 can be expressed as:

{D2=(x)kt.(G(x))for θD2G=(G(x))for θGkt+1=kt+λk(γ(x)(G(xG)))for each training step\begin{cases}\mathcal{L}_{D2}=\mathcal{L}(x)-k_{t}.\mathcal{L}(G(x))&\textrm{for }\theta_{D2}\\ \mathcal{L}_{G}=\mathcal{L}(G(x))&\textrm{for }\theta_{G}\\ k_{t+1}=k_{t}+\lambda_{k}(\gamma\mathcal{L}(x)-\mathcal{L}(G(x_{G})))&\textrm{for each training step}\end{cases}

To sum up, the objective of our model can be expressed as:

G=argmin 𝐺max D1D2cGAN(G,D1)+λL1(G)+(G,D2){G^{*}}={arg}\underset{G}{\text{min }}\underset{D_{1}\ D_{2}}{\text{max }}\mathcal{L}_{cGAN}(G,D_{1})+\lambda\mathcal{L}_{L1}(G)+\mathcal{L}(G,D_{2})\\ (4)

where D1D_{1} represents the PatchGAN discriminator and the D2D_{2} refers to the auto-encoder decoder discriminator.

5 Experiments

In this section, we discuss the training procedure and evaluate our model performance on the test dataset. Then, we estimate the font style learning accuracy for the characters based on different criteria. In addition, we investigate the restoration effect by using the criteria.

5.1 Training Process

Parameters and Optimization To optimize RubbingGAN networks, we refer to the steps in pix2pix and BEGAN. We train the Discriminator1 with the Generator alternately. We train the Discriminator1 with Generator first, and then train the Discriminator2 with Generator. The method we trained Discriminator1 and Generator is similar as mentioned in pix2pix. We train the network with the gradient descent step on D1 and then train the Generator with same method. We use the Adam Optimizer which designed from the SGD to train our Generator and D1. The learning rate we adopt is 0.00020.0002, with the momentum parameters β1=0.5,β2=0.999\beta_{1}=0.5,\beta_{2}=0.999. After training the D1 and Generator, we train the D2 and Generator. For the D2 and Generator, we use the same parameters between the D1 and Generator. We analysis the model loss and convergence quantitatively.

5.2 Evaluation

To evaluate the restoration effects, we first measure the font style transfer similarity by using two criteria. Then, we research the restoration effects on our incomplete characters dataset.

5.3 Font Style Similarity in Testing Set

In this part, we use the testing dataset to show that the RubbingGAN can learn the font style of ZhangMenglong Bei. We use the paired data to train the RubbingGAN. To evaluate the accuracy of the font style transfer, we adopt the a method called Style Similarity Error. The formula of style similarity error is shown in (3). In detail, the evaluation is achieved based on the complete characters in Zhang Menglong font. We measure the difference between the generated Zhang font with the real Zhang font images. In mathematics, we represent generated font as XX, the real font as YY. And XX and YY are two metrics, which dimension are 256×256256\times 256. Our method for evaluation is to calculate square value of every pixel difference in one image.

1m1m(Xi,jYi,j)2\displaystyle{\frac{1}{m}\sum_{1}^{m}({X_{i,j}-Y_{i,j}})}^{2} (5)

We use the testing dataset as input for the Generator and validate its performance on rubbing style transfer. The results are shown in Figure 8.

Refer to caption
(a) The original font image.
Refer to caption
(b) The Zhang font image.
Refer to caption
(c) The generated Zhang font image. The characters in first two lines are trained on 9800 iterations. The rest of characters is trained on 4000 iterations.
Figure 8: Rubbing style transfer result

We use the formula mentioned in the previous section to calculate the similarity between the generated image and original Zhang font image. The result is shown in the Table 1.

Table 1: RubbingGAN results for Zhang fonts style transfer
Model pix2pixpix2pix RubbingGANRubbingGAN zi2zizi2zi
MSE 0.0006128 0.0005846 0.0005028

We find out that RubbingGAN model has a better performance than the pix2pix and a little behind of the zi2zi model.

We also use the FID index to evaluate our model performance. The FID is an index used for evaluating the generating images quality [16]. We use the pytorch version to test our model [28]. The result is shown in the Table 2.

Table 2: RubbingGAN results for image generation quality
Model pix2pixpix2pix RubbingGANRubbingGAN zi2zizi2zi
FID 218.6381 204.9481 270.6007

We find out our model RubbingGAN has the best performance (the lower the better), which suggests our model has a good ability on image generation.

5.4 Slightly incomplete characters reconstruction

In this experiment, we test the performance on slightly incomplete characters. We first recognize the characters in the slightly incomplete characters dataset. Then, we input the correspond integrated character in standard font through the RubbingGAN and we get the Zhangfont style character. The results are shown in Figure 9.

Refer to caption
Figure 9: The characters with the black background are the little incomplete characters in Zhang Menglong bei. The characters with the white background are the generated Zhang font characters training on 4000 epochs.

We still use the method in experiment A to evaluate our results. However, due to the characters are incomplete, we can’t get result quantitatively. However, we can still use the previous formula to consider this problem. If the character are restored successfully, the correspond incomplete parts should be different with the original part. Based on this, we assume that if the difference of pixel value between the generated image and the incomplete image is bigger than 253, then we think this pixel should be in the incomplete part (the value 253 is chose empirically) in the Zhang Menglong Bei. The results are shown in Figure 10.

Refer to caption
Refer to caption
Refer to caption
Figure 10: To calculate the difference between the characters, we first reverse all the pixel value of the incomplete characters in ZhangMenglong Bei. Then if the two image pixel difference is bigger than 253, we mark the pixel with blue color.

5.5 Severe Incomplete Characters Reconstruction

In the previous experiments, we test the model performance on the style transfer, however, the effect of restorations still unknown. Therefore, we use the dataset to generate the Zhang’s font image. And we evaluate the restoration effect. The results are shown in Figure 11.

Refer to caption
Figure 11: The sever incomplete characters reconstruction

6 Conclusion

Repairing rubbings is significant for preserving Chinese history and for protecting traditional culture. In this paper, we propose a method named as Rubbing Generative Adversarial Network (RubbingGAN) that can restore the incomplete rubbing characters. Specifically, we collect the characters from a rubbing called Zhang Menglong Bei to the dataset. Based on the dataset we collect, we apply the RubbingGAN to learn the Zhang Menglong Bei font style and restore the characters. In the training process, we prove that our model can converge. The results of our experiment show that generator is good at generating rubbing font characters in the test set and can repair both the slightly and severely incomplete characters.

References

  • [1] https://en.wikipedia.org/wiki/Inpainting.
  • [2] M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein generative adversarial networks. In International conference on machine learning, pages 214–223. PMLR, 2017.
  • [3] G. Atarsaikhan, B. K. Iwana, A. Narusawa, K. Yanai, and S. Uchida. Neural font style transfer. In 2017 14th IAPR international conference on document analysis and recognition (ICDAR), volume 5, pages 51–56. IEEE, 2017.
  • [4] S. Azadi, M. Fisher, V. G. Kim, Z. Wang, E. Shechtman, and T. Darrell. Multi-content gan for few-shot font style transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7564–7573, 2018.
  • [5] D. Berthelot, T. Schumm, and L. Metz. Began: Boundary equilibrium generative adversarial networks. arXiv preprint arXiv:1703.10717, 2017.
  • [6] A. Bugeau and M. Bertalmio. Combining texture synthesis and diffusion for image inpainting. In VISAPP 2009-Proceedings of the Fourth International Conference on Computer Vision Theory and Applications, pages 26–33, 2009.
  • [7] B. Chang, Q. Zhang, S. Pan, and L. Meng. Generating handwritten chinese characters using cyclegan. In 2018 IEEE winter conference on applications of computer vision (WACV), pages 199–207. IEEE, 2018.
  • [8] J. Chang, Y. Gu, Y. Zhang, Y.-F. Wang, and C. Innovation. Chinese handwriting imitation with hierarchical generative adversarial network. In BMVC, page 290, 2018.
  • [9] L. ChangHong. Traditional rubbing restoration techniques viewed from the restoration case of mystic pagoda monument. pages 104–105, 2020.
  • [10] Z. Chao and V. Xiong. Stone inscriptions of the wei-jin nanbeichao period. Early Medieval China, 1994(1):84–96, 1994.
  • [11] D.Chen. Saturared of powerful and incomparably graceful the study of zhang menglong tablet. 2019.
  • [12] O. Elharrouss, N. Almaadeed, S. Al-Maadeed, and Y. Akbari. Image inpainting: A review. Neural Processing Letters, 51(2):2007–2028, 2020.
  • [13] L. A. Gatys, A. S. Ecker, and M. Bethge. A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576, 2015.
  • [14] L. A. Gatys, A. S. Ecker, and M. Bethge. Image style transfer using convolutional neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2414–2423, 2016.
  • [15] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. Advances in neural information processing systems, 27, 2014.
  • [16] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
  • [17] G. Hu and L. Xiong. Criminisi-based sparse representation for image inpainting. In 2017 IEEE Third International Conference on Multimedia Big Data (BigMM), pages 389–393. IEEE, 2017.
  • [18] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1125–1134, 2017.
  • [19] Y. Jiang, Z. Lian, Y. Tang, and J. Xiao. Scfont: Structure-guided chinese font generation via deep stacked networks. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 4015–4022, 2019.
  • [20] Y. Lei, L. Zhou, T. Pan, H. Qian, and Z. Sun. Learning and generation of personal handwriting style chinese font. In 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), pages 1909–1914. IEEE, 2018.
  • [21] C. Li and M. Wand. Precomputed real-time texture synthesis with markovian generative adversarial networks. In European conference on computer vision, pages 702–716. Springer, 2016.
  • [22] P. Liu, S. Xu, and S. Lin. Automatic generation of personalized chinese handwriting characters. In 2012 Fourth International Conference on Digital Home, pages 109–116. IEEE, 2012.
  • [23] S.-Y. Lu and T.-R. Hsiang. Generating chinese typographic and handwriting fonts from a small font sample set. In 2018 International Joint Conference on Neural Networks (IJCNN), pages 1–8. IEEE, 2018.
  • [24] P. Lyu, X. Bai, C. Yao, Z. Zhu, T. Huang, and W. Liu. Auto-encoder guided gan for chinese calligraphy synthesis. In 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), volume 1, pages 1095–1100. IEEE, 2017.
  • [25] M. Mirza and S. Osindero. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784, 2014.
  • [26] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
  • [27] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015.
  • [28] M. Seitzer. pytorch-fid: FID Score for PyTorch. https://github.com/mseitzer/pytorch-fid, August 2020. Version 0.2.1.
  • [29] K. Starr. Black Tigers: A Grammar of Chinese Rubbings. University of Washington Press, 2017.
  • [30] D. Sun, T. Ren, C. Li, H. Su, and J. Zhu. Learning to write stylized chinese characters by reading a handful of examples. arXiv preprint arXiv:1712.06424, 2017.
  • [31] D. Sun, Q. Zhang, and J. Yang. Pyramid embedded generative adversarial network for automated font generation. In 2018 24th International Conference on Pattern Recognition (ICPR), pages 976–981. IEEE, 2018.
  • [32] Y. Tian. zi2zi: Master chinese calligraphy with conditional adversarial networks. https://kaonashi-tyc.github.io/2017/04/06/zi2zi.html, 2018.
  • [33] F. Wang. Of master hongyi’s last calligraphic work. 2014.
  • [34] S. Xu, T. Jin, H. Jiang, and F. C. Lau. Automatic generation of personal chinese handwriting by capturing the characteristics of personal handwriting. In Twenty-First IAAI Conference, 2009.
  • [35] S. Xu, F. C. Lau, W. K. Cheung, and Y. Pan. Automatic generation of artistic chinese calligraphy. IEEE Intelligent Systems, 20(3):32–39, 2005.
  • [36] S. Xu, F. C. Lau, W. K. Cheung, and Y. Pan. Automatic generation of artistic chinese calligraphy. IEEE Intelligent Systems, 20(3):32–39, 2005.
  • [37] C. Yang, X. Lu, Z. Lin, E. Shechtman, O. Wang, and H. Li. High-resolution image inpainting using multi-scale neural patch synthesis. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6721–6729, 2017.
  • [38] Y. Zhang, Y. Zhang, and W. Cai. Separating style and content for generalized style transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8447–8455, 2018.
  • [39] A. Zhu, Q. Zhang, X. Lu, and S. Xiong. Character image synthesis based on selected content and referenced style embedding. In 2019 IEEE International Conference on Multimedia and Expo (ICME), pages 374–379. IEEE, 2019.