Date of publication August 19, 2020, date of current version August 31, 2020. 10.1109/ACCESS.2020.3017881
*Doyeon Kim and Donggyu Joo contributed equally to this work.
Contact: doyeon_kim@kaist.ac.kr, jdg105@kaist.ac.kr
Corresponding author: Junmo Kim (e-mail: junmo.kim@kaist.ac.kr).
TiVGAN: Text to Image to Video Generation with Step-by-Step Evolutionary Generator
Abstract
Advances in technology have led to the development of methods that can create desired visual multimedia. In particular, image generation using deep learning has been extensively studied across diverse fields. In comparison, video generation, especially on conditional inputs, remains a challenging and less explored area. To narrow this gap, we aim to train our model to produce a video corresponding to a given text description. We propose a novel training framework, Text-to-Image-to-Video Generative Adversarial Network (TiVGAN), which evolves frame-by-frame and finally produces a full-length video. In the first phase, we focus on creating a high-quality single video frame while learning the relationship between the text and an image. As the steps proceed, our model is trained gradually on more number of consecutive frames. This step-by-step learning process helps stabilize the training and enables the creation of high-resolution video based on conditional text descriptions. Qualitative and quantitative experimental results on various datasets demonstrate the effectiveness of the proposed method.
Index Terms:
Computer Vision, Deep Learning, Generative Adversarial Networks, Video Generation, Text-to-Video Generation,=-15pt
I Introduction
In the last few years, there has been intensive research on generative models. In particular, the recent developments of variational auto encoders(VAEs) [1] and generative adversarial networks(GANs) [2] represent the forefront of rapid, abundant, and high-quality progress. Further, since the deep convolutional GAN [3], which employs a convolutional neural network(CNN), succeeded in generating a realistic output using a GAN framework, many studies have reported impressive results using deep networks. It is now possible to produce photo-realistic images that are difficult to discriminate from real images, even for humans [4].
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=1]access_fig1.pdf Simple overview of our TiVGAN framework and generated videos. (a) The generator starts with producing a single frame and gradually evolves to create longer frames with the given text. (b) Generated sample videos using our framework TiVGAN.
However, the number of studies on video generation is significantly lower than those on image generation, because the video generation is a considerably more challenging task than image generation. Image creation is only concerned with the completeness of a single frame, whereas videos also need to consider the connectivity between frames. Even if each image has good quality, well-crafted videos cannot be generated if the continuity between adjacent frames is not guaranteed. In addition, nearly all public video datasets are extremely diverse and unaligned, thereby further complicating the video generation process.
Conditional video generation has received little research attention, whereas the generation of conditional output for a wide variety of inputs has been widely studied in the image generation field [5, 6, 7]. For example, a simple one-hot vector can be used as a control code to manipulate the attributes of a resulting image [5], and there is also a network that creates photo-realistic images corresponding to a given text [8]. However, studies regarding text-to-video generation are lacking and are generally performed on a low resolution compared to text-to-image generation. Therefore, to broaden the field of video generation, we focused on generating a conditional video that has not yet been investigated thoroughly in this domain. This study introduces a new scheme for text-to-video generation tasks with GANs.
We propose a novel network that generates a video corresponding to a given description. The learning framework of our network is established on the basic concept that connected frames of a video have substantial continuity. If we can create one high-quality video frame, it will be easier to create a linked frame because they are related. Rather than first finding a mapping function between the text and all video frames, we train our network with respect to one image and gradually extend it to longer frames (Figure I). Our model call this scheme as TiVGAN, which stands for Text-to-Image-to-Video Generative Adversarial Network framework. In the process of progressively learning to generate a large number of adjacent frames, TiVGAN can learn to create long consecutive scenes. Our extensive experimental results show that our model not only produces an accurate video for a given text but also produces qualitatively and quantitatively sharper and better results than those presented in other comparable works.
II Related Works
Generative image models have been studied actively in the past few years. Kingma et al. [1] suggested a re-parametrization trick to derive a variational lower limit of data likelihood. GAN shows promising results with the use of adversarial training between discriminators and generators. The discriminator is trained to distinguish between fake and real distribution, and the generator attempts to create realistic data to deceive the discriminator. Since then, there have been creation tasks for various datasets, such as human faces, furniture, animals, and others [9, 10, 3].
Following the research on the generation of images based on GANs, studies regarding conditional GANs have also been published with various kinds of conditional inputs. InfoGAN [5] uses a one-dimensional vector as a condition to control output image by concatenating the code into input noise. Maximizing the mutual information between the code and the generated image enables the network to learn interpretable representations. Furthermore, Reed et al. [11] demonstrated networks that can create text-based images by learning text feature representations and using them to synthesize images. StackGAN [8] extends this structure to stage- and stage-, which enables the generation of photo-realistic images after the generation of low-resolution images. Moreover, there have been many studies on the whole-image translations, such as image domain transfer [12, 13, 14] and image manipulation [15, 16, 17].
In contrast, very few experiments have been conducted on video synthesis. Vondrick et al. [18] untangled the background and foreground of the scenes with two streams using 2D spatial convolution and 3D spatio-temporal convolutions for each scene. TGAN [19] exploits two different generators for temporal vector sampling and multiple frame creation based on the acquired vectors. The developers of MoCoGAN [20] suggested the decomposition of motion and content space for effective video generation. They used a recurrent neural network for sampling from a motion subspace and concatenated the sampled features with the content vector to generate continuous frames. Our goal of text-to-video generation, on the other hand, has rarely been attempted. Li et al. [21] used a conditional VAE to generate a ‘gist’, which refers to the video background color and object layer; the video content and motion are created based on the gist and text. Pan et al. [22] proposed a new architecture for the text-to-video task by using 3D convolutions on their network and different types of losses. Balaji et al. [23] suggested a multi-scale text conditioning scheme with GAN to generate desired video frames with the given text. However, despite these examples, the number of studies on text-to-video remains small. Therefore, we present a new architecture suitable for video generation conditioned on the text description.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=1]access_fig2.pdf Full architecture of our proposed network, TiVGAN, and the training stages. We first start with training for generating a single image at the text-to-image generation stage, and we make consecutive frames in an evolutionary way through further stages. Although the text-to-image generation stage only uses an image discriminator , the evolutionary generation stage uses both an image discriminator and a step-discriminator .
III Method
As we have emphasized in the above sections, GANs have been proven for their ability to create sharp images. From this point of view, starting our text-to-video network training with a text-to-image stage may result in more effective video generation. Based on these ideas, we decomposed the training process into two stages: Text-to-Image Generation and Evolutionary Generation. The overall flow of our proposed architecture is described in Figure II. We begin with the learning of text-to-single image generation, gradually increase the number of produced images, and repeat this training process until the desired video length is achieved. This is our key paradigm. These two stages will be described in detail in subsections III-A and III-B, followed by explanations of the techniques we used to stabilize the learning.
III-A Text-to-Image Generation
We aim to generate a realistic fake video that matches with the given text description using a recurrent unit and a generator where and is each frame of video . At this stage, we only focus on the text-to-image generation task without considering the image sequence. Then, our goal is simplified to the generation of the single realistic image from .
To train the model with text, we must first transform the text string into an encoded feature vector. We adopted the pre-trained skip-thoughts vector network [24] to encode text into a -dimensional vector. Since the encoded vector is high-dimensional, we used principal component analysis (PCA) to derive meaningful features and reduce its dimensionality. We defined this embedded vector as .
We start with a single GRU cell and a generator to create one frame. Given the embedded text vector , the recurrent unit outputs the vector , where and are random noises from . The noise is a initial hidden state, and is an input of . This from is the source input vector for , and it creates the resulting image with the same size as the real frame. To ensure that matches with the provided text description and follows the distribution of real data, we trained the generator and the image discriminator adversarially using a GAN framework with slightly modified losses consisting of real, wrong, and fake pair as similar to those used by Reed et al. [11]. Overall losses will be explained in Section III-E.
When training the image discriminator , the real image is randomly selected from frames of the real video . Therefore, and aim not only to create the corresponding image for the given text but also to model the image distribution of the frames in the given video dataset. After the training of the text-to-image generation, can generate various images if some appropriate input vector is received. Therefore, if provides a meaningful sequence of vectors, can easily generate consecutive frames, which leads to realistic video generation. could act as an instructor that teaches to synthesize consecutive frames. This initial stage is the basis of the whole training process and plays a significant role in generating a series of frames.
III-B Evolutionary Generation
The evolutionary generation stage begins after the completion of the initial stage training described in the above section. We now create a series of consecutive frames based on the model trained in the previous stage. Since has the ability to generate various frames and is a recurrent unit that can output a series of meaningful vectors, the extension of the text-to-image generation can lead to successful consecutive frames generation.
This stage consists of the process from step to , and the goal of each step is to generate consecutive frames stably. As the step proceeds, the number of created frames increases, and we can finally reach the video-level generation we desired. In each step , images can be obtained by iterative operations of and learned in the previous stage. Let’s look at an example of the generation in step . After the text-to-image generation, we forward once more with as a hidden state and the (, ) as an input where is randomly sampled noise from . Then the next latent vector is obtained from . This vector is again delivered through the same to create another image, .
Unlike the text-to-image generation stage, the temporal consistency of the generated frames should be managed in this stage. At each step , step-discriminator is newly added to discriminate the sequence of frames. receives the fake input and the real input where are randomly selected connected frames from the real video . Since the real input has temporal information (images are concatenated with original order), the fake input should be generated to have temporal information correctly. After this training step converges, we move on to the further step . Then, is removed, and training proceeds with a new step-discriminator .
To summarize, we use different step-discriminator on individual steps. For , repeats times, and generates a sequence of images from input vector . is initialized at the beginning of step to discriminate the images, and it is only used for the step . Meanwhile, is used through all steps to maintain high-quality image output. This step-by-step training converges quickly and stably.
When all the training finishes, we can easily generate the images by forwarding single and for times, and it forms the video with lengths. The detailed training procedure and algorithm is described in Section III-E and Algorithm 1.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=0.99]CVPR_step_D.pdf Illustration of the first convolution layer of step-discriminators when step changes. At the beginning of step of evolutionary generation stage, step-discriminator is initialized using .
III-C Step-discriminator Initialization
Each step-discriminator is newly added at the beginning of every step of the evolutionary generation stage. Since is already trained through the steps, there could be an imbalance in the training progress of discriminators and . Therefore, simply applying the adversarial learning of and can even disrupt the learning of . To prevent this, we present an advanced scheme which initializes the to a better state rather than noise.
For all , the weight of is initialized with the previous step-discriminator . Image discriminator is used for initialization of the first step-discriminator . We designed all the step-discriminators to have the same architecture except for the number of input channels in the first layer. The only difference is that receives images as input, while receives images. Let be the weights connected to the input channel of the layer of step-discriminator . Then, our step-discriminator initialization can be defined as:
(1) |
All layers except the first layer are initialized with same weight of the step-discriminator . However, there is a slight variation only in the first layer, which is that and are initialized to for all . This is illustrated in Figure III-B. After the evolution in step to step , the appearance of the generated images within each step should be similar, but the number of the resulting images are twice as long. To retain the output of the discriminator when the step changes, we divide the weight value by . We believe that this initialization technique can assist the training of our framework to maintain stability even with sudden step changes. The effect of this method will be discussed further in the results section.
III-D Independent Samples Pairing
One of the main failures in training GAN is mode collapse. To mitigate this phenomenon, independent samples pairing (ISP) is applied in our training process. Our model produces two output images and with one fixed text description and two different randomly sampled input noises and when generating the frame (). These two independently generated images are paired by concatenation in the channel dimension and form the fake pair. To make the generator create various examples corresponding to the same text , we train the discriminator to distinguish between and , where and are two real dissimilar images associated with the same text description. Since and are dissimilar, if a mode collapsed generator generates very similar and , it will be easily detected as fake by the discriminator. Thus, the generator attempts to create different images with the same to deceive the discriminator. This independent samples pairing technique is only used for image discriminator .
III-E Training Procedures
We retain the adversarial training framework of the generator and discriminator. Unlike in a conventional GAN, however, we include a slight perturbation by having two branches on the discriminator similar to [8]. At first, the discriminator passes several convolution layers to acquire a high-level feature map. Then, one branch calculates the text-image match loss by concatenation with , and the other branch performs patch discrimination without text concatenation. The operations of these two independent branches ensure that the image matches well with the text while improving the image quality. We use three kinds of text-image losses to train the model for text matching, the same as those used by Reed et al. [11]. The losses are obtained by a real pair (, ), fake pair (, ), and wrong pair (, ), where is one of the embedded text vector that is not identical to .
The overall loss for , , , and are as follows:
(2) |
where
(3) |
and
(4) |
and are randomly selected frames from real and fake videos, respectively. is an image loss without text conditioning, and is the pair loss with text conditioning. At evolutionary step , represents . Here, is a consecutive images set randomly selected from the real video, and is generated fake images set with images. They are concatenated in channel dimensions. is only used at the evolutionary generation stage, and is used through all processes. Our overall training algorithm is described in Algorithm 1.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=1]access_kth.pdf Qualitative results of the models trained on KTH dataset. (a) Our generation results. (b) Comparative results with previous works.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=1]Paper_MUG.pdf Qualitative results for MUG dataset. The images above and below are parts of generated video frames created from the same text description. Generated video frames are well matched with given text description.
IV Experiments
To prove the effectiveness of our proposed method, we conduct several experiments on three diverse datasets: KTH Action, MUG, and Kinetics. To compare with the previous text-to-video method, we reproduced [22] and make experiments on three datasets. Since other methods [21, 23] have experimented on the Kinetics dataset, we directly compare results for the Kinetics. Due to the lack of published papers in the text-to-video area, we also employ several other existing video generation methods (TGAN [19], MoCoGAN [20]), and trained on each dataset using the same settings. For a fair comparison, we attempted to balance our model and video generation model by adding text conditioning to each method to yield TGAN++ and MoCoGAN++.
For all experiments, we used steps, which implies that we generated a 16-frame video. The detailed structure of the network is given in the supplementary material. In the training of the text-to-image generation stage, 30k iterations are performed, and in the evolutionary generation stage, we perform 15k iterations for each step. For PCA, we reduce the dimension of the vector to , i.e., .
IV-A KTH Action
The KTH Action dataset [25] contains six types of human actions: walking, jogging, running, boxing, hand waving, and hand clapping. Of these, we use the jogging class. In each clip, a man is jogging from left to right or right to left on two backgrounds. We extracted 16 frames from each video sequence and reshaped it into . For training, a total of 200 videos (3,200 frames) are used. Our qualitative results are shown in Figure 1. We can see that the person in the generated video is moving exactly as described in the text description while maintaining the image quality with high resolution.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=1]Paper_Kinetics_edit_ECCV.pdf Example results of text-to-video generation trained on Kinetics dataset.
Quantitative results: For a quantitative evaluation, we use frechet inception distance(FID) [26] to measure the quality of the generated images. FID measures the similarity between two image sets. We collect the video frames generated from each method, and then calculate the FID between each image set with the same number of video frames in the training dataset. Our TiVGAN shows the best performance as shown in Table I.
IV-B MUG Facial Expression
MUG is a human facial expression database [27]. There are tens of people in the dataset, and each person shows seven types of facial expressions: ‘happy’, ‘disgust’, ‘sad’, ‘neutral’, ‘surprise’, ‘fear’, and ‘anger’. We reshaped and used resolution frames, and the models are trained on a total of videos. Result images on the MUG dataset and their given captions are shown in Figure 1. We observe that our model generates sharp images corresponding to the given text.
without text | with text | |
---|---|---|
TGAN [19] | 3.50 | 4.63 |
MoCoGAN [20] | 3.31 | 4.92 |
TGANs-C [22] | - | 4.65 |
TiVGAN | - | 5.34 |
Quantitative results: The inception score is used for the quantitative evaluation of the MUG dataset results. It is a measurement proposed by Salimans et al. [28] that evaluates the quality of a GAN by observing the diversity and classification confidence of its generated images. In this experiment, a simple 5-layers 3D convolutional neural network is used instead of an inception network owing to the limitations of the number of data and classes; each video is classified into seven classes representing human facial expressions. Table II shows a comparison of the results using the MUG dataset. Our results showed the highest inception score among the studied methods.
IV-C Kinetics
Kinetics is a large-scale, human-focused video dataset from YouTube [29]. The dataset comprises thousands of video URLs covering human action classes. We used six classes from this dataset: ‘snow bike’, ‘swimming’, ‘sailing’, ‘golf’, ‘kite surfing’, and ‘water ski’, which are similar to those used in previous works [21]. We reshaped and used a frame size, and every model is trained on a total of videos. For Kinetics, the recent text-to-video generation method proposed by Li et al.(T2V) [21] is also compared qualitatively.
The generated qualitative results are shown in Figure IV-A. From the results, we can easily see that our method produces a much higher-quality video. Our generated images are much clearer because of their higher resolution, and they can also capture more distinctive features of given text, apart from the resolution differences.
In-set | T2V [21] | TFGAN [23] | TiVGAN(Ours) | |
Acc. (%) | 78.1 | 42.6 | 76.2 | 77.8 |
Quantitative results: First, the inception score is again used for the evaluation of the Kinetics dataset. Each video is trained on the 6-class classification by using a 5-layers 3D convolutional neural network. Table III shows a comparison of the results on the Kinetics dataset. Our result again shows the highest inception score among the compared methods. Next, the video classification accuracy of the generated results is recorded in Table IV following the settings of previous text-to-video works [23, 21]. Our TiVGAN achieves the highest performance which is very close to the in-set accuracy.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=0.99]Figure_NN.pdf Ablation study on nearest neighbors. Left images are generated samples, and right images are corresponding nearest neighbors in training dataset.
V Ablation studies
We conduct ablation studies to analyze the effectiveness of the proposed architecture. All experiments are tested on the Kinetics dataset.
V-A Nearest Neighbors
To address that our model does not simply memorize the dataset, we present the nearest neighbor image in Figure IV. We can observe that our generated results are different from the nearest neighbors in the training set.
V-B Step-by-Step Generation
When we aim to create frames of video, our network starts with generating a single frame () and gradually double the number of images to create (). To show the advantage of our step-by-step evolutionary generation framework, we perform an ablation study with various cases. Several steps are omitted in the comparison experiment, but total training time in all cases is same for fair comparison.
Number of generated frames | Inception Score | ||||
1 | 2 | 4 | 8 | 16 | |
✓ | ✓ | ✓ | ✓ | ✓ | 5.55 |
- | ✓ | ✓ | ✓ | ✓ | 5.13 |
✓ | - | ✓ | - | ✓ | 5.47 |
- | - | - | - | ✓ | 5.25 |
As shown in Table V, TiVGAN yields the highest inception score than other experiments. The results of the second row and fourth row demonstrate that initial text-to-image generation plays a significant role in the final step. Also, the result of third row indicates that skipping a few steps decrease the performance of the model. From the experimental results, it can be seen that our step-by-step generation is critical to producing high-quality video.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=0.99]Abl_init_new_2.pdf Ablation study on step discriminator initialization. Training loss is used as a measure to demonstrate the effectiveness of step-discriminator initialization compared to random initialization. The points where the blue line rises abruptly indicates the time when the step changes.
[!t](topskip=0pt, botskip=0pt, midskip=0pt)[width=0.99]access_abl_isp.pdf Ablation study on Independent Samples Pairing. Two different video clips are independently generated from one text input ‘swimming in the swimming pool’ with different noises. The left (with ISP) generated a completely different but appropriate video following the text description. However, without ISP, two samples are identically generated even with the different input noise.
V-C Step-discriminator Initialization
In Sec III-C, we described our step discriminator initialization as our training strategy to enhance the network. In this experiment, we record the training loss of for the cases with and without our initialization. The results are shown in Figure V. With random noisy initialization, shows an unstable loss graph at the beginning of every step. Since random initialization does not utilize the previous learning, the loss rises rapidly when the discriminator is newly added. The step-discriminator initialization indicates that is not affected by step change. This means that the model can reliably handle the generation of a larger number of images owing to the suggested initialization for the step-discriminator.
V-D Independent Samples Pairing
We employ Independent Samples Pairing to prevent the mode-collapse of the generator. The effects of ISP can be visualized in Figure V. Without ISP, the generator often produces identical outputs when the same input text with different noise are given. However, we verify that our network generates various videos when the same text description and different random noise are given.
VI Conclusion
In this paper, we proposed a new effective learning paradigm for text-to-video generation. Beginning with the creation of a single image, our network evolves progressively to synthesize a video clip of a desired length. Additionally, several techniques were used for stabilizing the training. Experimental results on the KTH, MUG, and Kinetics datasets support that our model can accomplish the given task under various situations. Conditional video generation is still a less explored field, but we believe it will be actively researched in the near future. We hope that our work will invite more interest in this field.
References
- [1] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” arXiv preprint arXiv:1312.6114, 2013.
- [2] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in neural information processing systems, 2014, pp. 2672–2680.
- [3] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” arXiv preprint arXiv:1511.06434, 2015.
- [4] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive growing of gans for improved quality, stability, and variation,” arXiv preprint arXiv:1710.10196, 2017.
- [5] X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel, “Infogan: Interpretable representation learning by information maximizing generative adversarial nets,” in Advances in neural information processing systems, 2016, pp. 2172–2180.
- [6] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
- [7] A. Odena, C. Olah, and J. Shlens, “Conditional image synthesis with auxiliary classifier gans,” in Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 2017, pp. 2642–2651.
- [8] H. Zhang, T. Xu, H. Li, S. Zhang, X. Wang, X. Huang, and D. N. Metaxas, “Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 5907–5915.
- [9] D. Berthelot, T. Schumm, and L. Metz, “Began: Boundary equilibrium generative adversarial networks,” arXiv preprint arXiv:1703.10717, 2017.
- [10] T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida, “Spectral normalization for generative adversarial networks,” arXiv preprint arXiv:1802.05957, 2018.
- [11] S. Reed, Z. Akata, X. Yan, L. Logeswaran, B. Schiele, and H. Lee, “Generative adversarial text to image synthesis,” arXiv preprint arXiv:1605.05396, 2016.
- [12] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1125–1134.
- [13] Z. Yi, H. Zhang, P. Tan, and M. Gong, “Dualgan: Unsupervised dual learning for image-to-image translation,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2849–2857.
- [14] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2223–2232.
- [15] D. Joo, D. Kim, and J. Kim, “Generating a fusion image: One’s identity and another’s shape,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 1635–1643.
- [16] W. Shen and R. Liu, “Learning residual images for face attribute manipulation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 4030–4038.
- [17] L. Tran, X. Yin, and X. Liu, “Disentangled representation learning gan for pose-invariant face recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 1415–1424.
- [18] C. Vondrick, H. Pirsiavash, and A. Torralba, “Generating videos with scene dynamics,” in Advances In Neural Information Processing Systems, 2016, pp. 613–621.
- [19] M. Saito, E. Matsumoto, and S. Saito, “Temporal generative adversarial nets with singular value clipping,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2830–2839.
- [20] S. Tulyakov, M.-Y. Liu, X. Yang, and J. Kautz, “Mocogan: Decomposing motion and content for video generation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 1526–1535.
- [21] Y. Li, M. R. Min, D. Shen, D. Carlson, and L. Carin, “Video generation from text,” in Thirty-Second AAAI Conference on Artificial Intelligence, 2018.
- [22] Y. Pan, Z. Qiu, T. Yao, H. Li, and T. Mei, “To create what you tell: Generating videos from captions,” in Proceedings of the 25th ACM international conference on Multimedia. ACM, 2017, pp. 1789–1798.
- [23] Y. Balaji, M. R. Min, B. Bai, R. Chellappa, and H. P. Graf, “Conditional gan with discriminative filter generation for text-to-video synthesis,” in IJCAI, 2019, pp. 1995–2001.
- [24] R. Kiros, Y. Zhu, R. R. Salakhutdinov, R. Zemel, R. Urtasun, A. Torralba, and S. Fidler, “Skip-thought vectors,” in Advances in neural information processing systems, 2015, pp. 3294–3302.
- [25] I. Laptev, B. Caputo et al., “Recognizing human actions: a local svm approach,” in null. IEEE, 2004, pp. 32–36.
- [26] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “Gans trained by a two time-scale update rule converge to a local nash equilibrium,” in Advances in Neural Information Processing Systems, 2017, pp. 6626–6637.
- [27] N. Aifanti, C. Papachristou, and A. Delopoulos, “The mug facial expression database,” in 11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS 10. IEEE, 2010, pp. 1–4.
- [28] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen, “Improved techniques for training gans,” in Advances in neural information processing systems, 2016, pp. 2234–2242.
- [29] W. Kay, J. Carreira, K. Simonyan, B. Zhang, C. Hillier, S. Vijayanarasimhan, F. Viola, T. Green, T. Back, P. Natsev et al., “The kinetics human action video dataset,” arXiv preprint arXiv:1705.06950, 2017.
![]() |
Doyeon Kim received the B.S. degree in computer science from Korea University, Seoul, South Korea, in 2016, and the M.S. degree in robotics program from Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea, in 2018. She is currently pursuing the Ph.D. degree of electrical engineering with KAIST. Her research interests include computer vision, deep learning, and machine learning. |
![]() |
donggyu joo received the B.S. and M.S. degrees in electrical engineering from Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea, in 2016 and 2018, respectively. He is currently pursuing the Ph.D. degree of electrical engineering with KAIST. His research interests include computer vision, deep learning, and machine learning. |
![]() |
Junmo Kim (S’01-M’05) received the B.S. degree from Seoul National University, Seoul, Korea, in 1998, and the M.S. and Ph.D. degrees from the Massachusetts Institute of Technology (MIT), Cambridge, in 2000 and 2005, respectively. From 2005 to 2009, he was with the Samsung Advanced Institute of Technology (SAIT), Korea, as a Research Staff Member. He joined the faculty of KAIST in 2009, where he is currently an Associate Professor of electrical engineering. His research interests are in image processing, computer vision, statistical signal processing, machine learning, and information theory. |