End-to-End Rubbing Restoration Using Generative Adversarial Networks
Abstract
Rubbing restorations are significant for preserving world cultural history. In this paper, we propose the RubbingGAN model for restoring incomplete rubbing characters. Specifically, we collect characters from the Zhang Menglong Bei and build up the first rubbing restoration dataset. We design the first generative adversarial network for rubbing restoration.111The dataset and codes are available at https://github.com/qingfengtommy/RubbingGAN Based on the dataset we collect, we apply the RubbingGAN to learn the Zhang Menglong Bei font style and restore the characters. The results of experiments show that RubbingGAN can repair both slightly and severely incomplete rubbing characters fast and effectively.
1 Introduction
Rubbings are ink-on-paper copies of engraved or cast inscriptions and designs on what are mostly cultural objects of metal, stone, and other firm substances [29]. The rubbing images are shown in Figure 1.


The history of rubbings is about 1400 years and rubbings play an important role in providing supportive evidence and clues for cultural relics restoration. However, due to the weathering process in a long time, the characters left on the rubbings usually are incomplete, as shown in Figure 1(b). Therefor, repairing rubbings is significant for both the Asian history preservation and the world culture inheritance. In traditional rubbings restoration, artists need to confirm the acidity first, and then based on the breakage level to determine corresponding methods. For instance, Li and his colleagues repair Xuan Mi Ta Bei by using a method called bracket mounting (tuo biao fa) [9]. However, the traditional rubbing restoration method requires a lot of efforts from artists, which is time consuming and impossible for large-scale artworks repairing.
In contrast to the traditional manual way of repairing cultural relics, the computer scientists have proposed a digital method called inpainting [17]. Digital inpainting is a way of reconstructing missing or damaged part in the digital images or videos, which can be divided into three categories: structural inpainting, textural inpainting, and combined structural and textural inpainting [1]. Even though the current inpainting methods can achieve delightful performance in repairing a number of relics such as images [37, 6, 12], they cannot be directly applied in rubbing restoration. There are two possible reasons: 1) the structure and texture of the Asian characters on the rubbings are complex and diverse. 2) some Asian characters are similar and there is a slight difference between them.
Another field relevant to rubbing restoration is the font style transfer problem. The font style transfer or the character font generation is a problem which developed from the image style transfer question. Recently, there are some dedicated works on Chinese fonts generation. The most representative work is zi2zi [32].
However, methods in traditional Chinese font style transfer cannot be applied in the rubbing restoration task directly. First, the main objective for rubbing restoration is to restore the incomplete characters, which indicates that we do not know which character needed to generate, and have to recognize it firstly. Second, characters in the datasets used in Chinese font style transfer are usually digital designed characters or just the standard printed fonts, such as KaiTi_GB2312. The digital or printed font have already be processed elaborately during the designing period of font libraries. And it is relatively easy to transfer the style between two different domain of printed fonts due to the consistency and continuity of the font style. However, in the rubbing restoration question, the characters from the dataset are more aesthetic and unique. The difficulties of the rubbing restoration task include repairing the rubbing characters and recognizing the aesthetic and diverse font style in the rubbings. In other words, we must use the characters which come from the rubbings to realize the job of repairing the incomplete characters and preserving the correspond rubbing style.
In this paper, we propose a method named as Rubbing Generative Adversarial Network (RubbingGAN) that can restore the incomplete rubbing characters. Specifically, we collect the characters from a rubbing called Zhang Menglong Bei to the dataset. Zhang Menglong Bei has a very high artistic value in the field of calligraphy and has a number of incomplete characters to repair [33, 10]. Based on the dataset we collect, we apply the RubbingGAN to restore the characters.
In short, our main contributions are summarized as follows.
-
•
We collect the first rubbing restoration dataset of a Chinese rubbing called Zhang Menglong Bei, including 337 complete characters as the training set, 75 complete characters as the testing set, and 34 incomplete characters as the restoration set.
-
•
We design a RubbingGAN to synthesize the rubbing fonts and repair the incomplete characters based on the training set. As far as we know, it is the first machine learning model in restoring the Chinese rubbing.
-
•
The experiments corroborate that the RubbingGAN can learn the rubbing fonts in Zhang Menglong Bei, and can repair both the slightly and severely incomplete characters.
2 Related works
For the Rubbing restoration problem, we find out that the recognition for the incomplete characters is not the most difficult task. The main difficulty in rubbing restoration is to preserve the aesthetic value of the artworks. Therefor, we mainly focus on artistic style transferring and we regard the rubbing restoration as the image-to-image translation problem. The related works of image-to-image translation are discussed in the below.
2.1 Generative Adversarial Networks
The generative adversarial networks is proposed by Ian J. Goodfellow and his colleagues in 2014 [15]. The main objective of the GAN is a min-max game. The generator try to deceive the discriminator by generating the high similarity pictures with the source datasets. On the other hand, the discriminator try to recognize the fake and real images. With the training, generator can generate high similarity images with the original input source. Based on the Ian J. Goodfellow’s work, the M. Mirza and S. Osindero propose a method that directs the data generation by adding the label information into the model [25]. And the DCGAN first introduced a deep convolutional network to improve the image quality [24]. However, the GAN also has some problems. First, the GAN may encounters the mode collapses, which due to the unbalance ability between the generator and the discriminator. Second, the GAN can not converge easily in the training period. Therefor, there are some works try to solve these problems. For example, the researchers propose WGAN which using the Wasserstein distance to train the GAN model. And based on the WGAN [2], researchers propose the BEGAN model that aims to match auto-encoder loss distribution using a loss derived from the Wasserstein distance.
2.2 Image-to-Image translation
Image-to-image translation is a task of translating one possible representation of a scene into another, given sufficient training data [18]. The representative work of translation is a conditional GAN-based model called pix2pix. It performs image-to-image translation on paired data. It uses a "U-Net" [27] based architecture as its generator and uses a convolutional "PatchGAN" auto encoder decoder as its discriminator [21]. In our work, the generation of rubbing style characters can be regarded as a kind of image-to-image translation. The limitation of the pix2pix model is only supporting the paired datasets [18]. Therefor, the researchers propose the model such as CylcleGAN that achieving the cross-domain style transfer [7]. However, the model like CylcleGAN can not apply into the rubbing restoration task. Because the precondition of the CycleGAN is that we have the enough data from the different domain. However, this is impractical in the rubbing restoration task. Because there are huge amounts of artworks only have the characters around hundred level, which is not enough for training the CycleGAN model.
2.3 Character Fonts Generation
There are a plenty of relative works on neural style transfer or on the character generation. Although the works on the Chinese characters generation are fewer than the English characters, there are some pioneer researches in this field. It has been studied since the beginning of the digital age. The problem can be formulated either as a personal handwritten characters generation problem [20], or an automatic printed font generation problem [23]. In the early study of glyph synthesis, researchers focused on the geometric modeling, which relied on the hierarchical representation of simple strokes [34, 35, 22, 36, 7]. They decompose the characters into a hierarchy of reusable components and then to compose the target character in the personalized style. In fact, these methods only focuses on the local characters’ features instead of the whole style, in other words, they are limited to particular glyph topology. With the development and popularity of deep neural networks, the topic of style transfer is frequently discussed. Researchers have designed new models which could add the artistic style to image content [13, 14, 3] successfully. And there are some works used the idea of neural style transfer to do the character generation problem [3, 39, 4]. This topic gradually evolves to image-to-image translation problem [32, 26, 19, 31]. The image-to-image translation problem focuses on converting objects from one domain to another domain. However, most existing methods require an impractical large-size dataset to train the network. For example, [32, 26, 30, 38] require 3000 paired characters which costs huge labor resources. In [8], researchers reduce the paired data number to 775 samples. However, this quantities still exceed the total number of the characters in most rubbings. And the dataset they adopt is either font-rendered character images or brush-written character images. Compared with these character images, the rubbing images are more irregular and incomplete. To the best of our knowledge, there is no available collection of rubbing dataset and there is no existing paper using rubbing dataset to conduct experiment.
3 Dataset
In this section, We first briefly introduce ZhangMenglong Bei from an artistic perspective. Then we describe the method for data collection and provide the basic information of our collected dataset.
3.1 A brief introduction of ZhangMenglong Bei
As we have mentioned above, the existed dataset of Chinese font style are usually generated by the computer, such as KaiTi_GB2312. Comparing with the traditional rubbing font style characters, these standard printed font style have a very short history and lack of the value for preserving traditional culture. In our task, we care about characters that written on rubbings in ancient ages. Along the history of Chinese calligraphy, there are a huge number of excellent art works that were created during different dynasties. The categories of traditional calligraphy can be divided into five main styles following the chronological order of appearance: Seal Script, Clerical Script, Cursive Script, Semi-Cursive Script and Standard Script (Kai Ti). Between the adjacent two styles, there is a transition style that assimilates the previous core idea and guides the after font style development. The Wei Bei is a famous transitional calligraphy system linking the past and the future. Wei Bei calligraphy refers to the Chinese characters engraved on the tablet during the Wei, Jin, and the Northern and Southern Dynasties. Wei Bei calligraphy has a great influence on the formation of Standard Script in the Later Sui and Tang dynasties. That is to say, Wei Bei calligraphy can be regarded as a predecessor of Standard Script, which has with a high artistic, cultural and research value [11]. However, due to the long history, most Wei Bei tablets are usually seriously damaged. A number of characters in each Wei Bei are severely incomplete. Thus, it is necessary to design a method to learn the font style of Wei Bei and restore the incomplete characters. Zhang Menglong Bei is the most prestigious work in the Northern Wei Dynasty and receives high praises from famous calligraphers such as Kang Youwei, Yang Shoujin, Qi gong.
3.2 Dataset collection
ZhangMenglong Bei has various versions online, and the version we used to extract characters is from the website. We use python to write a web crawler that automatically downloading Zhang Menglong Bei, which is shown in Figure 2. Then, we can find that Zhang Menglong Bei consist of both the complete characters and the incomplete characters.



To train the model, we first crop the complete characters. The cropping progress is achieved manually. The datasets is illustrated in Figure 3.


Since the characters we have cropped are clear, we label the ZhangMenglong style characters through using the same character in another font and joint it horizontally. The training dataset consists of 337 images and the size of the each pair data is with 3 color channels. The validation dataset consists of 75 images and the image information is as same as the training dataset. In addition there is a reconstruction or restoration dataset, the total number of images in this dataset is 34. The size of images in this dataset is .
In the training process, we crop the pair data first and get two images that the size are [standard font image and Zhang Menglong font image]. During the training period, we resize both the standard font characters and the Zhang Menglong font characters in pair data from the size of to by using bilinear interpolation. And we convert the image into correspond tensor. The input tensor size for Generator is and the output tensor size is . For Discriminator1, we concatenate the standard image tensor with the Zhang Menglong font image tensor or with the generated image tensor to get the input tensor. The input tensor size for Discriminator1 is . For Discriminator2, we input the real Zhang Menglong font tensor or the generated font tensor, which size are both .
For the rest characters in ZhangMenglong Bei, they are slightly or severely damaged, as shown in Fig. 3(b). We collect 34 characters and form a dataset to save these incomplete characters. Each character is exhibited as an image with pixels. This dataset is named as the incomplete character dataset. We use the incomplete character dataset to corroborate the model can repair the characters and keep the ZhangMenglong Bei font style.
4 Proposed Method
In this section, we will introduce the main architecture of the RubbingGAN in details. The objective of RubbingGAN is to learn a mapping from an original font style source to a specific rubbing font style. We denote the original font KaiTi_GB2312 as , and the target ZhangMenglong Bei font the . The RubbingGAN is to learn a mapping .
4.1 The Architecture of Rubbing GAN
We propose a framework for Rubbing characters restoration that transferred standard font image into Zhang Menglong font . Our framework is designed based on the pix2pix [18] and BEGAN [5]. The architecture of our model is shown in Figure 4.

The RubbingGAN includes one generator denoted by G and two discriminators, denoted by D1 and D2.
Generator.
From previous work [32] and [24], we adopt "U-Net" structure as our generator which adds skip connections between layers. The Generator structure is shown in Figure 5. In the rubbing restoration problem, the core problem is to learn the structure changes and the correspond textures. The "U-net" architecture provide high structure semantics with the low-level information.

Discriminator1.
We designed Discriminator1 based on the PatchGAN [21], which encourages GAN to focus both on the high-frequencies and low-frequencies structure features loss. Discriminator1 only penalizes structure at the scale of patches, which can be seen as a type of style/font loss as mentioned in pix2pix.
Discriminator2.
The discriminator2 is an auto encoder-decoder architecture. We designed the structure of Discriminator2 in terms of the BEGAN literature [5]. The input size of the image for Discriminator2 is , the output size of the Discriminator is also . The structure of Discriminator2 is really easy, and plays a good effect in image generation. Another advantage of using Discriminator 2 is that in BEGAN researchers propose a method that measured the convergence, which solve the problem of difficulty of training a traditional GAN. Therefore, we would like to use the Discriminator2 to improve generated images quality and also avoid the possibility of mode collapse. The architecture of Discriminator2 is shown in Figure 6.

The Figure 7 shows a whole procedure of our rubbing restoration, which recognized the characters first, and then generate the rubbing font characters.

Losses Our model losses mainly consist of three parts: the adversarial losses from the traditional conditional GAN, the loss from the L1 distance distance, and the loss from the reconstruction. We use the conditional adversarial loss to generate the realistic image. The loss can be expressed as:
(1) |
In order to improve the quality of the generated images, we add the L1 distance into the model. The L1 loss is expressed as:
(2) |
In order to solve the problem of mode collapses, we use idea from the BEGAN to balance the ability of the generator and discriminator 2. We use an equilibrium term in RubbingGAN:
(3) |
where the hype-parameter is used to control the output diversity of the Generator. The represents the reconstruction loss function for the auto-encoder. The loss function for the generator and discriminator2 can be expressed as:
To sum up, the objective of our model can be expressed as:
(4) |
where represents the PatchGAN discriminator and the refers to the auto-encoder decoder discriminator.
5 Experiments
In this section, we discuss the training procedure and evaluate our model performance on the test dataset. Then, we estimate the font style learning accuracy for the characters based on different criteria. In addition, we investigate the restoration effect by using the criteria.
5.1 Training Process
Parameters and Optimization To optimize RubbingGAN networks, we refer to the steps in pix2pix and BEGAN. We train the Discriminator1 with the Generator alternately. We train the Discriminator1 with Generator first, and then train the Discriminator2 with Generator. The method we trained Discriminator1 and Generator is similar as mentioned in pix2pix. We train the network with the gradient descent step on D1 and then train the Generator with same method. We use the Adam Optimizer which designed from the SGD to train our Generator and D1. The learning rate we adopt is , with the momentum parameters . After training the D1 and Generator, we train the D2 and Generator. For the D2 and Generator, we use the same parameters between the D1 and Generator. We analysis the model loss and convergence quantitatively.
5.2 Evaluation
To evaluate the restoration effects, we first measure the font style transfer similarity by using two criteria. Then, we research the restoration effects on our incomplete characters dataset.
5.3 Font Style Similarity in Testing Set
In this part, we use the testing dataset to show that the RubbingGAN can learn the font style of ZhangMenglong Bei. We use the paired data to train the RubbingGAN. To evaluate the accuracy of the font style transfer, we adopt the a method called Style Similarity Error. The formula of style similarity error is shown in (3). In detail, the evaluation is achieved based on the complete characters in Zhang Menglong font. We measure the difference between the generated Zhang font with the real Zhang font images. In mathematics, we represent generated font as , the real font as . And and are two metrics, which dimension are . Our method for evaluation is to calculate square value of every pixel difference in one image.
(5) |
We use the testing dataset as input for the Generator and validate its performance on rubbing style transfer. The results are shown in Figure 8.



We use the formula mentioned in the previous section to calculate the similarity between the generated image and original Zhang font image. The result is shown in the Table 1.
Model | |||
---|---|---|---|
MSE | 0.0006128 | 0.0005846 | 0.0005028 |
We find out that RubbingGAN model has a better performance than the pix2pix and a little behind of the zi2zi model.
We also use the FID index to evaluate our model performance. The FID is an index used for evaluating the generating images quality [16]. We use the pytorch version to test our model [28]. The result is shown in the Table 2.
Model | |||
---|---|---|---|
FID | 218.6381 | 204.9481 | 270.6007 |
We find out our model RubbingGAN has the best performance (the lower the better), which suggests our model has a good ability on image generation.
5.4 Slightly incomplete characters reconstruction
In this experiment, we test the performance on slightly incomplete characters. We first recognize the characters in the slightly incomplete characters dataset. Then, we input the correspond integrated character in standard font through the RubbingGAN and we get the Zhangfont style character. The results are shown in Figure 9.

We still use the method in experiment A to evaluate our results. However, due to the characters are incomplete, we can’t get result quantitatively. However, we can still use the previous formula to consider this problem. If the character are restored successfully, the correspond incomplete parts should be different with the original part. Based on this, we assume that if the difference of pixel value between the generated image and the incomplete image is bigger than 253, then we think this pixel should be in the incomplete part (the value 253 is chose empirically) in the Zhang Menglong Bei. The results are shown in Figure 10.



5.5 Severe Incomplete Characters Reconstruction
In the previous experiments, we test the model performance on the style transfer, however, the effect of restorations still unknown. Therefore, we use the dataset to generate the Zhang’s font image. And we evaluate the restoration effect. The results are shown in Figure 11.

6 Conclusion
Repairing rubbings is significant for preserving Chinese history and for protecting traditional culture. In this paper, we propose a method named as Rubbing Generative Adversarial Network (RubbingGAN) that can restore the incomplete rubbing characters. Specifically, we collect the characters from a rubbing called Zhang Menglong Bei to the dataset. Based on the dataset we collect, we apply the RubbingGAN to learn the Zhang Menglong Bei font style and restore the characters. In the training process, we prove that our model can converge. The results of our experiment show that generator is good at generating rubbing font characters in the test set and can repair both the slightly and severely incomplete characters.
References
- [1] https://en.wikipedia.org/wiki/Inpainting.
- [2] M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein generative adversarial networks. In International conference on machine learning, pages 214–223. PMLR, 2017.
- [3] G. Atarsaikhan, B. K. Iwana, A. Narusawa, K. Yanai, and S. Uchida. Neural font style transfer. In 2017 14th IAPR international conference on document analysis and recognition (ICDAR), volume 5, pages 51–56. IEEE, 2017.
- [4] S. Azadi, M. Fisher, V. G. Kim, Z. Wang, E. Shechtman, and T. Darrell. Multi-content gan for few-shot font style transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7564–7573, 2018.
- [5] D. Berthelot, T. Schumm, and L. Metz. Began: Boundary equilibrium generative adversarial networks. arXiv preprint arXiv:1703.10717, 2017.
- [6] A. Bugeau and M. Bertalmio. Combining texture synthesis and diffusion for image inpainting. In VISAPP 2009-Proceedings of the Fourth International Conference on Computer Vision Theory and Applications, pages 26–33, 2009.
- [7] B. Chang, Q. Zhang, S. Pan, and L. Meng. Generating handwritten chinese characters using cyclegan. In 2018 IEEE winter conference on applications of computer vision (WACV), pages 199–207. IEEE, 2018.
- [8] J. Chang, Y. Gu, Y. Zhang, Y.-F. Wang, and C. Innovation. Chinese handwriting imitation with hierarchical generative adversarial network. In BMVC, page 290, 2018.
- [9] L. ChangHong. Traditional rubbing restoration techniques viewed from the restoration case of mystic pagoda monument. pages 104–105, 2020.
- [10] Z. Chao and V. Xiong. Stone inscriptions of the wei-jin nanbeichao period. Early Medieval China, 1994(1):84–96, 1994.
- [11] D.Chen. Saturared of powerful and incomparably graceful the study of zhang menglong tablet. 2019.
- [12] O. Elharrouss, N. Almaadeed, S. Al-Maadeed, and Y. Akbari. Image inpainting: A review. Neural Processing Letters, 51(2):2007–2028, 2020.
- [13] L. A. Gatys, A. S. Ecker, and M. Bethge. A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576, 2015.
- [14] L. A. Gatys, A. S. Ecker, and M. Bethge. Image style transfer using convolutional neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2414–2423, 2016.
- [15] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. Advances in neural information processing systems, 27, 2014.
- [16] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
- [17] G. Hu and L. Xiong. Criminisi-based sparse representation for image inpainting. In 2017 IEEE Third International Conference on Multimedia Big Data (BigMM), pages 389–393. IEEE, 2017.
- [18] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1125–1134, 2017.
- [19] Y. Jiang, Z. Lian, Y. Tang, and J. Xiao. Scfont: Structure-guided chinese font generation via deep stacked networks. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 4015–4022, 2019.
- [20] Y. Lei, L. Zhou, T. Pan, H. Qian, and Z. Sun. Learning and generation of personal handwriting style chinese font. In 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), pages 1909–1914. IEEE, 2018.
- [21] C. Li and M. Wand. Precomputed real-time texture synthesis with markovian generative adversarial networks. In European conference on computer vision, pages 702–716. Springer, 2016.
- [22] P. Liu, S. Xu, and S. Lin. Automatic generation of personalized chinese handwriting characters. In 2012 Fourth International Conference on Digital Home, pages 109–116. IEEE, 2012.
- [23] S.-Y. Lu and T.-R. Hsiang. Generating chinese typographic and handwriting fonts from a small font sample set. In 2018 International Joint Conference on Neural Networks (IJCNN), pages 1–8. IEEE, 2018.
- [24] P. Lyu, X. Bai, C. Yao, Z. Zhu, T. Huang, and W. Liu. Auto-encoder guided gan for chinese calligraphy synthesis. In 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), volume 1, pages 1095–1100. IEEE, 2017.
- [25] M. Mirza and S. Osindero. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784, 2014.
- [26] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
- [27] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015.
- [28] M. Seitzer. pytorch-fid: FID Score for PyTorch. https://github.com/mseitzer/pytorch-fid, August 2020. Version 0.2.1.
- [29] K. Starr. Black Tigers: A Grammar of Chinese Rubbings. University of Washington Press, 2017.
- [30] D. Sun, T. Ren, C. Li, H. Su, and J. Zhu. Learning to write stylized chinese characters by reading a handful of examples. arXiv preprint arXiv:1712.06424, 2017.
- [31] D. Sun, Q. Zhang, and J. Yang. Pyramid embedded generative adversarial network for automated font generation. In 2018 24th International Conference on Pattern Recognition (ICPR), pages 976–981. IEEE, 2018.
- [32] Y. Tian. zi2zi: Master chinese calligraphy with conditional adversarial networks. https://kaonashi-tyc.github.io/2017/04/06/zi2zi.html, 2018.
- [33] F. Wang. Of master hongyi’s last calligraphic work. 2014.
- [34] S. Xu, T. Jin, H. Jiang, and F. C. Lau. Automatic generation of personal chinese handwriting by capturing the characteristics of personal handwriting. In Twenty-First IAAI Conference, 2009.
- [35] S. Xu, F. C. Lau, W. K. Cheung, and Y. Pan. Automatic generation of artistic chinese calligraphy. IEEE Intelligent Systems, 20(3):32–39, 2005.
- [36] S. Xu, F. C. Lau, W. K. Cheung, and Y. Pan. Automatic generation of artistic chinese calligraphy. IEEE Intelligent Systems, 20(3):32–39, 2005.
- [37] C. Yang, X. Lu, Z. Lin, E. Shechtman, O. Wang, and H. Li. High-resolution image inpainting using multi-scale neural patch synthesis. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6721–6729, 2017.
- [38] Y. Zhang, Y. Zhang, and W. Cai. Separating style and content for generalized style transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8447–8455, 2018.
- [39] A. Zhu, Q. Zhang, X. Lu, and S. Xiong. Character image synthesis based on selected content and referenced style embedding. In 2019 IEEE International Conference on Multimedia and Expo (ICME), pages 374–379. IEEE, 2019.