This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Enhance Gender and Identity Preservation in Face Aging Simulation for Infants and Toddlers

Yao Xiao1, Yijun Zhao1 ©2020 This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible. 1 Computer and Information Science Department, Fordham University, New York, NY, USA
Abstract

Realistic age-progressed photos provide invaluable biometric information in a wide range of applications. In recent years, deep learning-based approaches have made remarkable progress in modeling the aging process of the human face. Nevertheless, it remains a challenging task to generate accurate age-progressed faces from infant or toddler photos. In particular, the lack of visually detectable gender characteristics and the drastic appearance changes in early life contribute to the difficulty of the task. We propose a new deep learning method inspired by the successful Conditional Adversarial Autoencoder (CAAE, 2017) model. In our approach, we extend the CAAE architecture to 1) incorporate gender information, and 2) augment the model’s overall architecture with an identity-preserving component based on facial features. We trained our model using the publicly available UTKFace dataset and evaluated our model by simulating up to 100 years of aging on 1,156 male and 1,207 female infant and toddler face photos. Compared to the CAAE approach, our new model demonstrates noticeable visual improvements. Quantitatively, our model exhibits an overall gain of 77.0% (male) and 13.8% (female) in gender fidelity measured by a gender classifier for the simulated photos across the age spectrum. Our model also demonstrates a 22.4% gain in identity preservation measured by a facial recognition neural network.

Index Terms:
Deep learning, age progression/regression, gender consistency, identity preservation, generative adversarial networks, conditional adversarial autoencoder, image generation, face recognition.

I Introduction

Face progression, also known as face aging, is the process of rendering face images to different age groups to simulate an identity-preserving aging effect. Generating accurate prediction images from a reference photo facilitates many practical applications to help create a safe and secure society. For example, the quality of age-progressed photos plays a critical role in finding long lost children, identifying fugitives, as well as developing age-invariant security systems. Furthermore, being described as ”half art and half science”, face progression also benefits the entertainment, cosmetology, and biometrics industries.

In recent years, deep learning-based approaches have made remarkable progress in modeling the aging process of the human face [1, 2, 3, 4, 5, 6]. Nevertheless, generating accurate age-progressed faces is still a challenging task. This is particularly true when the original face photos are of young children. For example, infants and toddlers often lack salient gender characteristics. Thus, a machine learning model may guess a child’s sex incorrectly and produce an inaccurate simulation. Furthermore, even with a correct guess, such models can fail to capture the drastic physical changes which occur during early childhood and adolescence. As a result, predicting future appearance based on a two-year-old face photo is significantly harder than extrapolating from a 20-year-old one.

In this paper, we introduce a new deep learning model to address the two challenges mentioned above. Our approach is inspired by the highly successful Conditional Adversarial Autoencoder (CAAE) [5] model. First, we incorporate additional gender information into the latent vector of the CAAE architecture. We evaluate the gender consistency of the generated face images using a gender classifier based on a deep neural network model [7]. Our study suggests that the CAAE model is prone to generate female faces when sex characteristics are not evident on the original photo. Providing additional gender information is essential in helping the model make proper adjustments.

The second contribution of our study is to enhance identity preservation during the aging simulation for young faces. Before entering early adulthood, the human body undergoes rapid physical growth and puberty development. These physiological changes (e.g., height, weight, and voice changes) are accompanied by substantial changes in facial features. For example, compared to the archetypal adult face, an infant’s face tends to have bigger eyes, shorter and flatter eyebrows, and a smaller and turned-up nose. As a result, it is challenging to preserve identity from baby photos in simulating the aging process. To address this issue, we leverage the popular VGG[8] architecture to maximize the similarity of high-level facial features between the original and simulated faces. Compared to the baseline CAAE model, our approach demonstrates significant improvement in identity preservation as measured by a face recognition deep neural network developed by Schroff et al. [9].

The rest of the paper is organized as follows. We present a brief literature survey of related work in Section II. We then illustrate our proposed model in Section III. We present and discuss our experimental methods and results in Section IV. Lastly, we conclude in Section V.

II Related Work

II-A Face Age Progression/Regression

Human face progression/regression has always been an active area of research due to its broad applications across various disciplines [10, 11, 12, 13]. In its early stages, face progression techniques simulated biomechanical aspects of human skin (e.g., anisotropy, visco-elasticity, etc.) to artificially produce wrinkles, creases, and folds [14]. These simulations accounted for many specific aging factors, including the facial muscles [15, 16], cranium [17], facial skin [18, 14, 19, 20], depth of wrinkles [21, 22], facial structure [23, 24], etc. While successful, these methods involved parameter tweaking and domain-specific knowledge. As the state-of-the-arts transitioned to image data driven approaches, methods which modeled face aging as a mapping from young to old faces became more common. For example, Park et al. [25] used sequential face data to approximate the aging process as a Markov process. Shu et al. presented an efficient and effective Kinship-Guided Age Progression (KinGAP) approach [26], which can automatically generate personalized aging images with guidance of the senior kinship face. One limitation of these methods is the difficulty of collecting a large dataset with chronologically sequential faces for each individual.

Another data-based approach is the prototype-based method, which first divides all the faces into different age groups. The technique then estimates the average face within a predefined age group denoted as the prototype. Differences between prototypes form the axes, along which faces can smoothly transition across the aging spectrum. Prototype-based methods are simple, straightforward, and fast. However, since the generated faces are averaged, high frequency details such as wrinkles are not accurately captured. To address this issue, researchers developed models [27, 28, 29], which captured person-specific facial features via sparse representation.

II-B Deep Learning-based Approaches

In recent years, deep learning techniques have made remarkable progress in the quality of face progression/regression using large datasets of raw face images. Wang et al.[30] employed a recurrent neural network [31] model to capture the common transition patterns among all age groups. Leveraging the slowly evolving intermediate images, their model offers more realistic and smoothly progressing faces than other non-temporal models. The limitation of their approach is that it can only perform one directional (i.e., from young to old) simulation.

Another class of popular deep learning-based methods is derived from the novel generative adversarial networks (GANs)[32] framework, in which a model trains a generator and a discriminator simultaneously. While the generator strives to generate realistic faces, the discriminator acts as the adversary to distinguish fake (i.e., generated) vs. real faces. At equilibrium, the generator is forced to produce photo-realistic faces to fool the discriminator. Generative approaches facilitate both face progression and rejuvenation on a face photo of any age.

One variant of the GAN architecture is the conditional generative adversarial network (C-GAN) [33]. C-GANs models provide the generator with additional information to control the scope (e.g., a specific age) of the generated images. These models [1, 34, 30] are effective, but they require paired images in training their models, i.e., pictures of the same person across the entire age spectrum, which is difficult to collect. Recent work by Zhang et al. [5] overcomes this difficulty by combining the concepts of C-GAN and autoencoder[35]. Their conditional adversarial autoencoder (CAAE) model embeds age-specific information in the latent vector of the autoencoder. Additionally, the authors employed an extra discriminator to impose a uniform prior on the latent distribution of the input data. We give a more detailed introduction to the CAAE model in Section III-A because our model is an extension to its architecture.

II-C Our Contribution

Most of the existing studies focus on the photo-realistic quality of the generated faces, including richer face texture (e.g., wrinkles), smooth transitions, and lesser ghosting effects. Evaluations are typically conducted using visual examinations centered on adult and senior faces. In our study, we focus on face aging simulation on baby photos in which the gender information is not detectable and the facial features undergo drastic changes over the years. To the best of our knowledge, there is no prior work addressing this task that is both challenging and of significant practical value. We also establish rigorous quantitative metrics to evaluate the efficacy of our approach.

Refer to caption
Figure 1: Architecture of our Proposed Model

III Methods

In this section, we first give a brief introduction to the conditional adversarial autoencoder (CAAE) model [5]. We then present our proposed approach which extends the CAAE architecture with gender and identity preserving components.

III-A Conditional Adversarial Autoencoder

As illustrated within the dashed bounding box in Fig. 1, a CAAE [5] model consists of three separate neural networks: a convolutional autoencoder (AE, i.e., Encoder + Generator) and two adversarial discriminators (i.e., DzD_{z} and DimgD_{img}). For the AE component, the encoder (EE) maps the input face to a vector zz which can be interpreted as the personality of a particular face input xx. The generator (GG) regenerates the input photo based on the latent vector zz and an additional age label (ll). Thus, a concatenated latent vector [z,l][z,l] serves as the input to the generator. Incorporating specific age information into the latent space facilitates clustering face images by age groups. Consequently, the model produces smooth age transitions and forces the output face to be photo-realistic and plausible for a given age.

The second component is the discriminator DzD_{z} which regularizes the latent space zz using a uniform distribution. The uniform prior forces zz to evenly populate the latent space without large gaps. Aassuming that images of different personalities and ages form a 2D manifold (MM) along the two corresponding directions, generating age-progressed photos can be viewed as traversing the MM along the age direction. Thus, continuity in the latent space is desirable because it ensures sufficient intermediate data points to guide the model along the traversal.

The last component DimgD_{img} is an inherited concept from the effective GAN[32] methodology. In particular, while the generator strives to produce as realistic as possible photos for a given age, DimgD_{img} acts as an adversary to discriminate between real and generated images. During the training process, DimgD_{img} contributes to the loss function based on how well it detects generated images, while the generator is evaluated based on how well its constructed images can fool the discriminator. Thus, this adversarial process forces the generator to produce high quality images that are indistinguishable from the real ones.

The overall loss of the CAAE model is the sum of corresponding losses from its three components. Following the same notation as in [5], the objective function can be formulated as follows:

minE,G\displaystyle\min_{E,G} maxDz,Dimgλ(x,G(E(x),l))+γTV(G(E(x),l))\displaystyle\max_{D_{z},D_{img}}\lambda\mathcal{L}(x,G(E(x),l))+\gamma TV(G(E(x),l)) (1)
+𝔼zp(z)[log Dz(z)]\displaystyle+\mathbb{E}_{z^{*}\sim p(z)}[\text{log }D_{z}(z^{*})]
+𝔼xpdata(x)[log (1Dz(E(x)))]\displaystyle+\mathbb{E}_{x\sim p_{data}(x)}[\text{log }(1-D_{z}(E(x)))]
+𝔼x,lpdata(x,l)[log Dimg(x,l)]\displaystyle+\mathbb{E}_{x,l\sim p_{data}(x,l)}[\text{log }D_{img}(x,l)]
+𝔼x,lpdata(x,l)[log (1Dimg(G(E(x),l)))]\displaystyle+\mathbb{E}_{x,l\sim p_{data}(x,l)}[\text{log }(1-D_{img}(G(E(x),l)))]

where (,)\mathcal{L}(\cdot,\cdot) denotes L2L_{2} norm, pdata(x)p_{data}(x) denotes the distribution of the training data, and zp(z)z^{*}\sim p(z) denotes the random sampling process from the prior distribution of zz. The first line of equation (1) is the loss of the CAE component. Specifically, (,)\mathcal{L}(\cdot,\cdot) and TV()TV(\cdot) represent the image reconstruction loss and total variation respectively. The coefficients λ\lambda and γ\gamma balance the smoothness and high resolution. The remaining four terms are the standard adversarial loss for discriminators DzD_{z} and DimgD_{img} respectively.

III-B Our Approach

Fig. 1 presents the structure of our proposed model. The components outside the dashed box are extensions to the CAAE model.

III-B1 Integrating Gender Information

One limitation of the CAAE model is that it does not account for gender differences in the input data. While gender can generally be inferred from adult photos, this is not the case with infant or toddler pictures. We demonstrate in our experimental results that the CAAE model tends to age-progress a baby photo toward the female direction. To address this issue, we augment the latent space zz with additional gender information. As illustrated in Fig. 1, the input to our generator GG is [z,l,s][z,l,s] where s{male,female}s\in\{male,female\}. We can further interpret our model as a 3D extension to the 2D manifold of the CAAE model with gender being the third dimension.

We evaluate the efficacy of integrating the new gender indicator using simulated aging images constructed from 0-5 years old photos. Besides visual examinations, we quantify the generated images with a gender score measured by a gender classifier. We present our gender classifier and the evaluation results in Section IV.

III-B2 Enhancing Identify Preservation

Our second improvement is to leverage a VGG [8] model to preserve each individual’s identity in the face aging process. The VGG network was introduced by Simonyan and Zisserman in 2014. It employs deep convolutional layers and is one of the most popular architectures for image feature extraction. In our approach, we utilize a pre-trained VGG19 (i.e., 19 convolutional layers) model to minimize the differences in high-level facial features between the input and generated images. Specifically, we apply the VGG19 model to both the input and the generated images and compute the L2L_{2} distance between the two feature maps after the last convolutional layer. The difference is added to the total training loss to force the model to maintain the same personalities as the original faces in its aging simulation process.

We evaluate the efficacy of this new identity preservation component using a Face Recognition (FR) score from the FaceNet system[9] developed by Google in 2015. Specifically, the FaceNet system maps a pair of images to an embedded Euclidean space, and the distance (e.g., L2L_{2}) in the embedded space measures the similarity of two identities. We present the FR score calculation and the evaluation of the VGG component in Section IV.

III-B3 Objective Function

We make two modifications to the objective function in Equation (1) to serve the two new components in our model. First, we augment the terms of adversarial loss of DimgD_{img} with additional gender information (ss). Formally, the last two lines in Equation (1) are adjusted as follows:

𝔼x,l,spdata(x,l,s)[log Dimg(x,l,s)]+𝔼x,l,spdata(x,l,s)[log (1Dimg(G(E(x),l,s)))]\mathbb{E}_{x,l,s\sim p_{data}(x,l,s)}[\text{log }D_{img}(x,l,s)]+\\ \hskip 28.45274pt\mathbb{E}_{x,l,s\sim p_{data}(x,l,s)}[\text{log }(1-D_{img}(G(E(x),l,s)))]\\

Next we add a new term to Equation (1) representing the feature map differences between the input and simulated images. Formally, the loss can be expressed as:

(FM(x),FM(G(E(x)),l,s))\mathcal{L}(FM(x),FM(G(E(x)),l,s))

where FM()FM(\cdot) denotes the feature map vector of the last convolutional layer after applying VGG19 to an image. ss is the gender indicator. All other symbols have been kept the same as in Equation (1).

IV Experiments

In this section, we first describe the data we used to conduct our study. We then introduce two quantitative measures to evaluate the quality of gender and identify preservation for the age-progressed photos. Lastly, we present the experimental results of our model in comparison to the baseline CAAE model.

IV-A Data Acquisition

We conduct our experiments using the UTKFace dataset[36], a large-scale face dataset with a long age span (from 0 to 116 years old). The dataset consists of 12,391 male and 11,317 female face images with annotations of age, gender, and ethnicity. We divided the photos into ten groups (i.e., 0-5, 6-10, 11-15, 16-20, 21-30, 31-40, 41-50, 51-60, 61-70, and ¿70) according to the stages of life in which the human face goes through significant changes. Additionally, the face images are cropped to 228×\times228 and aligned to make training more tractable.

IV-B Quantitative Metrics

Besides visually examining the quality of generated photos, we introduce ”gender score” and ”FR (Face Recognition) score” to assess the sex and identity consistencies between the original and simulated images.

IV-B1 Gender Score

To quantitatively evaluate gender fidelity during the face aging process, we employ a binary gender classifier (CC) to measure masculinity and femininity in the generated photos. We define the male/female gender score for a given age group as the accuracy when applying CC to all of the expected male/female photos in the group. For example, the male gender score for the 16-20-year old group is the accuracy of applying CC to those photos in the group that are simulated from male input. Thus, a higher gender score is more desirable because it indicates higher consistency with the expected sex.

Our gender classifier is based on the architecture of the deep neural network model introduced by Levi and Hassner [7] in 2015. We trained our model using a total of 8,659 male and 7,936 female adult instances from the UTKFace dataset with a 70%, 15%, 15% split for model training, validation, and testing respectively. Table I presents the gender scores of each age group in the test data. We observe that our gender classifier has an over 90% average accuracy for both classes. Furthermore, the model has a more balanced performance in the middle section (21 to 70 years old) of the age spectrum. Since we are interested in the relative performance of various models within a specific class and age group, the imbalanced performance at the two tails is insignificant because the same biases will apply to all models.

TABLE I: Performance of Gender Classifier CC
Age Group Male Female
050-5 0.77 (138/179)* 0.70 (128/183)
6106-10 0.75 (45/60) 0.96 (75/78)
111511-15 0.72 (29/40) 0.93 (51/55)
162016-20 0.88 (52/59) 0.96 (102/106)
213021-30 0.93 (471/504) 0.96 (684/714)
314031-40 0.97 (373/383) 0.92 (226/245)
415041-50 0.98 (171/174) 0.94 (84/89)
516051-60 0.98 (236/240) 0.90 (84/93)
617061-70 0.98 (122/125) 0.91 (49/54)
>> 70 0.95 (86/91) 0.73 (61/84)
Overall accuracy 0.93 (1723/1855) 0.91 (1544/1701)

*: (M/N) denotes the correctly classified instances M over the total instances N for a particular gender category and age group.

IV-B2 FR (Face Recognition) Score

We institute a second quantitative measure to evaluate the identity-preserving effect in the age-progressed photos. To this end, we utilize a pre-trained FaceNet classifier[9], a state-of-the-art face recognition system developed by Google in 2015. The system achieved record-high accuracy on a range of face recognition benchmark datasets, including 99.63%99.63\% accuracy in the Wild (LFW) dataset[37] for face verification. Conceptually, FaceNet learns a direct mapping from face images to a compact Euclidean space where distances correspond to a measure of face similarity.

Refer to caption
Figure 2: Visual Assessment of Age-progressed Photos Across Three Models. CAAE-G denotes a modified CAAE model with additional gender information. CAAE-GV denotes our model, a modified CAAE model with both gender and VGG components.

We define the FR score for a particular age group to be the accuracy of applying the FaceNet classifier to all generated and corresponding input image pairs within the group. A higher FR score indicates an overall higher likelihood of being the same person; thus, it can assess the identity-preserving effects of age-progressed photos. Clearly, the FaceNet model’s performance is dependent on the distance threshold that the classifier uses to decide if two faces are considered from the same person.

TABLE II: Gender Score Comparison for Four Models Across Ten Age Groups
Age Group Male Samples Female Samples
CAAE CAAE-G CAAE-V CAAE-GV CAAE CAAE-G CAAE-V CAAE-GV
050-5 0.650.65 0.770.77 (19.5%)* 0.650.65 (-0.1%) 0.680.68 (5.2%) 0.610.61 0.660.66 (8.0%) 0.640.64 (5.3%) 0.700.70 (14.7%)
6106-10 0.530.53 0.740.74 (40.3%) 0.540.54 (3.5%) 0.660.66 (25.6%) 0.680.68 0.730.73 (7.1%) 0.710.71 (3.9%) 0.790.79 (15.0%)
111511-15 0.440.44 0.710.71 (61.2%) 0.430.43 (-2.5%) 0.610.61 (38.2%) 0.750.75 0.790.79 (5.6%) 0.770.77 (3.6%) 0.830.83 (11.0%)
162016-20 0.340.34 0.780.78 (129.4%) 0.360.36 (6.1%) 0.680.68 (100.6%) 0.810.81 0.810.81 (0.1%) 0.810.81 (-0.1%) 0.870.87 (6.4%)
213021-30 0.310.31 0.780.78 (152.2%) 0.280.28 (-7.6%) 0.700.70 (126.4%) 0.830.83 0.870.87 (6.0%) 0.840.84 (2.0%) 0.910.91 (10.2%)
314031-40 0.360.36 0.840.84 (133.9%) 0.350.35 (-2.3%) 0.760.76 (113.1%) 0.800.80 0.860.86 (7.1%) 0.800.80 (0.2%) 0.910.91 (13.5%)
415041-50 0.420.42 0.890.89 (110.2%) 0.400.40 (-6.8%) 0.830.83 (95.4%) 0.760.76 0.830.83 (9.5%) 0.770.77 (1.5%) 0.880.88 (16.0%)
516051-60 0.440.44 0.910.91 (106.1%) 0.420.42 (-5.6%) 0.860.86 (93.2%) 0.730.73 0.810.81 (11.6%) 0.770.77 (5.1%) 0.890.89 (21.3%)
617061-70 0.450.45 0.900.90 (101.2%) 0.430.43 (-4.3%) 0.850.85 (89.7%) 0.710.71 0.800.80 (11.8%) 0.760.76 (6.0%) 0.850.85 (18.8%)
>>70 0.470.47 0.920.92 (94.0%) 0.450.45 (-5.9%) 0.870.87 (83.0%) 0.700.70 0.680.68 (-3.1%) 0.750.75 (7.5%) 0.770.77 (10.9%)
Average 0.440.44 0.82 (94.8%) 0.430.43 (-2.5%) 0.75 (77.0%) 0.740.74 0.780.78 (6.4%) 0.760.76 (3.5%) 0.84 (13.8%)

*: Numbers in parenthesis are the percentage gains over the CAAE model
           CAAE-G: CAAE model + gender component
           CAAE-V: CAAE model + VGG component
           CAAE-GV: CAAE model + both gender and VGG components

IV-C Experimental Results

Our model consists of two extensions to the baseline CAAE model, each of which serves a different purpose. To examine the individual and combined effects of these two components, we present and analyze the results of four models, CAAE, CAAE-G, CAAE-V, and CAAE-GV, which denote the baseline, baseline+gender, baseline+VGG, and baseline+gender+VGG respectively. Each model is trained using the UTKFace dataset and applied to 1156 male and 1207 female infant and toddler photos to generate simulated faces in ten age groups.

Fig. 2 presents sample simulated faces across different models and the age spectrum. We selected three test samples from each of the male and female categories. For each test photo, we display a simulation block of three rows, each of which consists of predicted faces by the indicated model for the 0-5, 16-20, 31-40, 51-60, and >>70 age buckets. Note that the CAAE-V model was excluded in Fig. 2 because we found that the effects of adding the VGG component alone are not visually pronounced. Nevertheless, we analyze its impact quantitatively using the gender and FR scores.

IV-C1 Visual Evaluation of the Gender Component

We examine the visual effects of the augmented gender indicator by comparing the faces in the top two rows in each simulation block in Fig. 2. We observe that, for male samples, there is a pronounced difference in masculinity between the two rows of simulated faces. Furthermore, the discrepancies are more salient as age progresses. The baseline CAAE model (i.e., top row) tends to project the faces toward the female direction, and the CAAE-G model (i.e., second row) corrects the mistake and generates the faces along the correct gender trajectory.

For the female samples, the difference in gender characteristics between the CAAE and CAAE-G models is not as notable as that of the male ones. This is expected if the baseline model is biased toward females. Nevertheless, we do observe the differences and it is arguable that the second rows exhibit more femininity compared to the first rows. To establish a more rigorous evaluation, we employ the ”gender score” (described in Section IV-B1) to quantify the fidelity of the sex category in the simulated photos compared to the ground-truth.

The third row in each of the simulation blocks presents the results of the CAAE-GV model. We observe that the gender advantage of the CAAE-G model is preserved and the faces are more realistic due to the additional VGG components. We quantitatively confirm this observation in Section IV-C4.

IV-C2 Quantitative Evaluation of the Gender Component

Tables II present the gender scores of different models over the age spectrum for both male and female categories. For the male samples, we observe that the CAAE-G model demonstrates an average gain of 94.8% across all age groups. This confirms our visual observation discussed in the above section. We further observe that the improvements are consistent across all age groups and there is a monotonic upward trend until the 31-40 group. The trend then reverts itself afterwards to decrease monotonically. There is a sharp gain (i.e., from 61.2% to 129.4%) when entering the 16-20 group. This is consistent with physical development during adolescence when gender features become apparent in facial images.

For the female category, we observe that the CAAE-G model demonstrates an overall 6.4% gain for the female test samples over the baseline model. Compared to the male case, the gains are consistent but at a much smaller scale. They are also more evenly distributed over the age spectrum. We conclude that the additional gender information also helps to produce more sex consistent photos in female samples in all age groups. Furthermore, compared to the male category, the considerably higher gender scores in the CAAE column for the female samples suggests that the baseline model has a bias towards the female gender in its simulations.

The CAAE-V columns in Table II demonstrate that the VGG component does not have a high influence on the gender characteristics in the simulated faces. This is plausible because the purpose of the VGG discriminator is for preserving the facial features that are essential for identity verification. In our study, all input photos are from infants and toddlers whose gender features are not fully developed on their faces. Indeed, the CAAE-V model exhibits a marginal improvement (3.5%) for the female class and a marginal negative impact (-2.5%) for the male class over the baseline model. However, we present next that the VGG component leads to a significant improvement in preserving individual identities in the face aging simulations.

TABLE III: Distance (L2L_{2}) Statistics Between Original and Simulated Faces
CAAE CAAE-G CAAE-V CAAE-GV
min 0.330.33 0.290.29 (10.8%) 0.230.23 (31.6%) 0.17 (49.8%)
max 5.395.39 4.614.61 (14.5%) 4.564.56 (15.4%) 4.81 (10.7%)
mean 1.881.88 1.831.83 (2.3%) 1.861.86 (0.8%) 1.77 (5.8%)
SD 0.700.70 0.730.73 (-3.1%) 0.730.73 (-3.1%) 0.760.76 (-8.2%)
10-PCTL 1.021.02 0.960.96 (6.3%) 0.98 (4.0%) 0.85 (16.8%)
20-PCTL 1.241.24 1.191.19 (4.4%) 1.191.19 (4.4%) 1.07 (14.1%)
30-PCTL 1.451.45 1.371.37 (5.3%) 1.391.39 (3.9%) 1.27 (12.3%)
40-PCTL 1.631.63 1.551.55 (4.8%) 1.601.60 (1.9%) 1.47 (10.1%)
50-PCTL 1.811.81 1.761.76 (2.9%) 1.791.79 (1.2%) 1.68 (7.2%)
60-PCTL 2.012.01 1.961.96 (2.3%) 2.002.00 (0.4%) 1.91 (4.9%)
70-PCTL 2.202.20 2.192.19 (0.8%) 2.222.22 (-0.7%) 2.14 (3.0%)
80-PCTL 2.462.46 2.452.45 (0.4%) 2.482.48 (-0.7%) 2.44 (0.7%)
90-PCTL 2.832.83 2.832.83 (0.2%) 2.862.86 (-0.9%) 2.79 (1.5%)

*: Numbers in parenthesis are the percentage gains over the CAAE model

IV-C3 Visual Evaluation of the VGG Component

The effect of the VGG component can be visualized by comparing the second and third rows in each of the simulation blocks in Fig. 2. While both models have integrated the gender indicator, the CAAE-GV model (i.e., third row) has the extra VGG loss (described in Section III-B2) incorporated into its model training process.

We observe that the third row has more vivid facial expressions and better skin tones in general compared to the second row. For all examples, faces in the third row also develop more natural wrinkles and nasolabial folds in the older groups. Consequently, the CAAE-GV model produces more realistic simulations compared to the CAAE-G model. We confirm our observation more rigorously using the ”FR score” in the next section.

IV-C4 Quantitative Evaluation of the VGG Component

As described in Section IV-B2, we institute an ”FR (Face Recognition) score” to quantify the degree of identity preservation throughout the face aging process. Because the FR score relies on the distance measure in the embedded Euclidean space, we first examine the statistics of the L2L_{2} distance between the original image and the age-progressed faces for different models.

Table III presents, for each model, the distribution of distances between the test samples and their corresponding generated images. The smaller the distance, the more likely the two faces belong to the same person. We observe that the CAAE-GV model produces the smallest mean (1.77) among the four distributions with a 5.8% gain over the baseline model, while adding gender information or the VGG component alone results in a 2.3% and 0.8% gain respectively. Similarly, for the min/max statistics, the CAAE-GV model demonstrates 49.8%/10.7% improvements over the baseline model, while the CAAE-G and CAAE-V models exhibit 10.8%/14.5% and 31.6%/15.4% gains respectively. Furthermore, the advantage of the CAAE-GV model over the other three models is consistent across all percentile groups.

Next, we compare the FR score of the four models. Recall that an FR score represents the FaceNet model’s classification accuracy in determining if the simulated face is the same as the original person’s. This accuracy depends on the distance threshold we use to make the verification decision. A lower threshold (i.e., closer distance) requires a higher similarity between the paired images to be identified as the same person. On the other hand, a higher threshold relaxes the verification criteria. In the extreme case, when the threshold is set above the max distance of the distributions in Table III, all FR scores will degenerate to 1. Thus, the FR score improvement with a lower threshold is more substantial than that with a higher threshold, even though lower thresholds are associated with lower FR scores.

Table IV presents the FR score for different models using three distance thresholds. We observe that in each model’s corresponding column, the FR score increases as expected when we relax the identity verification criteria. When we set a high verification standard (threshold=1.6), the CAAE-GV model exhibits a 22.4% gain over the baseline CAAE model. The CAAE-G and CAAE-V models offer a 12.9% and 5.3% gain, respectively. As we lower the criteria (i.e., increase the threshold), the advantage of our model reduces due to the expected diminishing FR score differences among all four models.

TABLE IV: Face Recognition (FR) Score Comparison of Four Models
Threshold CAAE CAAE-G CAAE-V CAAE-GV
1.6 0.38 0.43 (12.9%)* 0.40 (5.3%) 0.47 (22.4%)
2.02.0 0.600.60 0.620.62 (3.9%) 0.600.60 (0.8%) 0.650.65 (8.2%)
2.52.5 0.820.82 0.820.82 (-0.2%) 0.810.81 (-1.0%) 0.820.82 (0.3%)

*: Numbers in parenthesis are the percentage gains over the CAAE model

V Conclusion

In this paper, we focused on the challenging task of age progression/regression for photos of infants and toddlers. We proposed two enhancements to the existing CAAE architecture to help ensure gender and identity consistencies in the face aging process. In our approach, we augmented the input vector of the generator with gender information since young faces lack salient gender characteristics. We further strengthened the model with a new identity preservation component based on facial features extracted by the VGG19 convolutional neural network. Our experimental results demonstrate significant visual and quantitative improvements over the CAAE model for our particular task. Our methods and findings can be adopted by other deep learning approaches in the face progression and regression studies.

References

  • [1] S. Liu, Y. Sun, D. Zhu, R. Bao, W. Wang, X. Shu, and S. Yan, “Face aging with contextual generative adversarial nets,” in Proceedings of the 25th ACM international conference on Multimedia, 2017, pp. 82–90.
  • [2] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” arXiv preprint arXiv:1511.06434, 2015.
  • [3] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2223–2232.
  • [4] G. Antipov, M. Baccouche, and J.-L. Dugelay, “Face aging with conditional generative adversarial networks,” in 2017 IEEE international conference on image processing (ICIP).   IEEE, 2017, pp. 2089–2093.
  • [5] Z. Zhang, Y. Song, and H. Qi, “Age progression/regression by conditional adversarial autoencoder,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 5810–5818.
  • [6] J. Song, T. He, L. Gao, X. Xu, A. Hanjalic, and H. T. Shen, “Binary generative adversarial networks for image retrieval,” in Thirty-Second AAAI Conference on Artificial Intelligence, 2018.
  • [7] G. Levi and T. Hassner, “Age and gender classification using convolutional neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2015, pp. 34–42.
  • [8] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
  • [9] F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 815–823.
  • [10] X. Shu, G. S. Xie, Z. Li, and J. Tang, “Age progression: Current technologies and applications,” Neurocomputing, vol. 208, pp. 249–261, 2016.
  • [11] N. Ramanathan, R. Chellappa, and S. Biswas, “Computational methods for modeling facial aging: A survey,” Journal of Visual Languages & Computing, vol. 20, no. 3, pp. 131–144, 2009.
  • [12] N. Ramanathan, R. Chellappa, S. Biswas, et al., “Age progression in human faces: A survey,” Journal of Visual Languages and Computing, vol. 15, pp. 3349–3361, 2009.
  • [13] Y. Fu, G. Guo, and T. S. Huang, “Age synthesis and estimation via faces: A survey,” IEEE transactions on pattern analysis and machine intelligence, vol. 32, no. 11, pp. 1955–1976, 2010.
  • [14] L. Boissieux, G. Kiss, N. M. Thalmann, and P. Kalra, “Simulation of skin aging and wrinkles with cosmetics insight,” in Computer Animation and Simulation 2000.   Springer, 2000, pp. 15–27.
  • [15] J. Suo, X. Chen, S. Shan, W. Gao, and Q. Dai, “A concatenational graph evolution aging model,” IEEE transactions on pattern analysis and machine intelligence, vol. 34, no. 11, pp. 2083–2096, 2012.
  • [16] A. C. Berg and S. C. Justo, “Aging of orbicularis muscle in virtual human faces,” in Proceedings on Seventh International Conference on Information Visualization, 2003. IV 2003.   IEEE, 2003, pp. 164–168.
  • [17] J. T. Todd, L. S. Mark, R. E. Shaw, and J. B. Pittenger, “The perception of human growth,” Scientific american, vol. 242, no. 2, pp. 132–145, 1980.
  • [18] Y. Bando, T. Kuratate, and T. Nishita, “A simple method for modeling wrinkles on human skin,” in 10th Pacific Conference on Computer Graphics and Applications, 2002. Proceedings.   IEEE, 2002, pp. 166–175.
  • [19] W.-S. Lee, Y. Wu, and N. Magnenat-Thalmann, “Cloning and aging in a vr family,” in Proceedings IEEE Virtual Reality (Cat. No. 99CB36316).   IEEE, 1999, pp. 61–68.
  • [20] Y. Wu, P. Beylot, and N. M. Thalmann, Skin aging estimation by facial simulation.   IEEE, 1999.
  • [21] N. Ramanathan and R. Chellappa, “Modeling shape and textural variations in aging faces,” in 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition.   IEEE, 2008, pp. 1–8.
  • [22] J. Suo, S.-C. Zhu, S. Shan, and X. Chen, “A compositional and dynamic model for face aging,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 3, pp. 385–401, 2009.
  • [23] N. Ramanathan and R. Chellappa, “Modeling age progression in young faces,” in 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), vol. 1.   IEEE, 2006, pp. 387–394.
  • [24] A. Lanitis, C. J. Taylor, and T. F. Cootes, “Toward automatic simulation of aging effects on face images,” IEEE Transactions on pattern Analysis and machine Intelligence, vol. 24, no. 4, pp. 442–455, 2002.
  • [25] U. Park, Y. Tong, and A. K. Jain, “Age-invariant face recognition,” IEEE transactions on pattern analysis and machine intelligence, vol. 32, no. 5, pp. 947–954, 2010.
  • [26] X. Shu, J. Tang, H. Lai, Z. Niu, and S. Yan, “Kinship-guided age progression,” Pattern Recognition, vol. 59, pp. 156–167, 2016.
  • [27] X. Shu, J. Tang, Z. Li, H. Lai, L. Zhang, S. Yan, et al., “Personalized age progression with bi-level aging dictionary learning,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 4, pp. 905–917, 2017.
  • [28] H. Yang, D. Huang, Y. Wang, H. Wang, and Y. Tang, “Face aging effect simulation using hidden factor analysis joint sparse representation,” IEEE Transactions on Image Processing, vol. 25, no. 6, pp. 2493–2507, 2016.
  • [29] W. Wang, Y. Yan, S. Winkler, and N. Sebe, “Category specific dictionary learning for attribute specific feature selection,” IEEE Transactions on Image Processing, vol. 25, no. 3, pp. 1465–1478, 2016.
  • [30] W. Wang, Z. Cui, Y. Yan, J. Feng, S. Yan, X. Shu, and N. Sebe, “Recurrent face aging,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2378–2386.
  • [31] M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural networks,” IEEE transactions on Signal Processing, vol. 45, no. 11, pp. 2673–2681, 1997.
  • [32] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in neural information processing systems, 2014, pp. 2672–2680.
  • [33] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
  • [34] X. Shu, J. Tang, H. Lai, L. Liu, and S. Yan, “Personalized age progression with aging dictionary,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 3970–3978.
  • [35] G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” science, vol. 313, no. 5786, pp. 504–507, 2006.
  • [36] J. Geralds, “Utkface large scale face dataset,” github.com.
  • [37] G. B. Huang, M. Mattar, T. Berg, and E. Learned-Miller, “Labeled faces in the wild: A database for studying face recognition in unconstrained environments,” 2008.