This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

ConsistentAvatar: Learning to Diffuse Fully Consistent Talking Head Avatar with Temporal Guidance

Haijie Yang Nanjing University of Science and TechnologyNanjingChina yanghaijie@njust.edu.cn Zhenyu Zhang Nanjing UniversityNanjingChina zhangjesse@foxmail.com Hao Tang School of Computer Science, Peking UniversityBeijingChina bjdxtanghao@gmail.com Jianjun Qian Nanjing University of Science and TechnologyNanjingChina csjqian@njust.edu.cn  and  Jian Yang Nanjing University of Science and TechnologyNanjingChina csjyang@mail.njust.edu.cn
(2018)
Abstract.

Diffusion models have shown impressive potential on talking head generation. While plausible appearance and talking effect are achieved, these methods still suffer from temporal, 3D or expression inconsistency due to the error accumulation and inherent limitation of single-image generation ability. In this paper, we propose ConsistentAvatar, a novel framework for fully consistent and high-fidelity talking avatar generation. Instead of directly employing multi-modal conditions to the diffusion process, our method learns to first model the temporal representation for stability between adjacent frames. Specifically, we propose a Temporally-Sensitive Detail (TSD) map containing high-frequency feature and contours that vary significantly along the time axis. Using a temporal consistent diffusion module, we learn to align TSD of the initial result to that of the video frame ground truth. The final avatar is generated by a fully consistent diffusion module, conditioned on the aligned TSD, rough head normal, and emotion prompt embedding. We find that the aligned TSD, which represents the temporal patterns, constrains the diffusion process to generate temporally stable talking head. Further, its reliable guidance complements the inaccuracy of other conditions, suppressing the accumulated error while improving the consistency on various aspects. Extensive experiments demonstrate that ConsistentAvatar outperforms the state-of-the-art methods on the generated appearance, 3D, expression and temporal consistency.

3D head avatars, Diffusion model, Consistent avatars
*Corresponding author
copyright: acmlicensedjournalyear: 2018doi: 10.1145/3664647.3680619.conference: Make sure to enter the correct conference title from your rights confirmation email; October 28 - November 1, 2024; Melbourne, Australia.isbn: 979-8-4007-0686-8/24/10ccs: Computing methodologies Reconstruction
Refer to caption
Figure 1. For short monocular RGB videos, our method synthesizes avatars demonstrating full consistency in 3D (Figure (b)), temporal (Figure (d)), and emotional aspects (Figure (c)), while maintaining high quality (Figure (a)). Compared to state-of-the-art methods, ours effectively addresses all three consistency issues simultaneously. Notably, for temporal consistency, we quantify the degree of change between adjacent frames using optical flow. Comparative analysis with real data shows significant mitigation of temporal inconsistencies by our method. For the final video demonstration and details about optical flow, please refer to the supplementary materials.
\Description

Enjoying the baseball game from the third-base seats. Ichiro Suzuki preparing to bat.

1. Introduction

In the realm of virtual reality and its associated applications, the creation of lifelike and animatable character avatars poses a significant challenge. Early approaches (Deng et al., 2020b; Feng et al., 2021; Sanyal et al., 2019; Zielonka et al., 2023; Zheng et al., 2022) often involve generating head avatars by fitting 3D Morphable Models (3DMMs) (Blanz and Vetter, 1999), utilizing a parameterized representation to describe the shape, texture, and details of the human face. By adjusting these parameters, precise control over facial features can be achieved, thereby demonstrating relatively strong 3D controllability. However, 3DMMs typically rely on certain assumptions, such as linear shape spaces, which may not fully capture the complex variations in facial shape, texture, and details. Consequently, the resulting facial renderings may lack realism (see Fig. 1 (a)).

To address these issues, some methods (Siarohin et al., 2020; Zhao et al., 2021; Tewari et al., 2020; Karras et al., 2020; Gu et al., 2019; Zhang et al., 2023a) combine StyleGANs (Karras et al., 2019) with the priors of 3DMMs to improve the fidelity of generated avatars. For example, SadTalker (Zhang et al., 2023a) generates 3D motion coefficients of 3DMMs from audio and implicitly adjusts a novel 3D-aware face rendering. Thanks to the powerful generative capabilities of GANs, these methods often yield realistic images. However, these methods fundamentally rely on a single-image 2D generation process, where the image rendering process is intricately entangled and facial fitting capability is limited. Consequently, these methods encounter difficulties in achieving consistent 3D control (see Fig. 1 (b)). To overcome the limitations of 2D GAN-based methods, recent research have tended to utilize Neural Radiance Field (NeRF) or similar implicit representations to generate face avatars (Gafni et al., 2021; Zheng et al., 2022; Nguyen-Phuoc et al., 2020; Noguchi et al., 2022; Khakhulin et al., 2022), or integrate NeRF into GANs for 3D-aware dynamic face synthesis (Khakhulin et al., 2022; Noguchi et al., 2022; Nguyen-Phuoc et al., 2020). Typically, these methods learn dynamic NeRF based on input expressions as the conditions to represent the deformed 3D space. For example, IMAvatar (Zheng et al., 2022) optimizes the implicit function on the mesh template to achieve more accurate 3D control. However, as the volume rendering involves manual approximation on the real-world imaging process, it has potential to lose fidelity on the rendered images. Additionally, NeRF struggles to decouple temporal information from 3D representation, resulting in temporal inconsistencies in such methods. As illustrated in Fig. 1 (a), INSTA (Zielonka et al., 2023) and IMAvatar (Zheng et al., 2022) suffer from inaccurate expressions or appearance degradation. With the development of diffusion models, more recent studies (Han et al., 2023; Shen et al., 2023; Kim et al., 2023a, b) have adopted this strategy to enhance the quality of face generation and editing. For example, DiffTalk (Shen et al., 2023) models the generation of talking heads as a denoising process driven by audio. Due to the powerful generation capabilities of diffusion models, these methods often achieve highly realistic results. However, these methods are essentially 2D-aware, so that they also lack sufficient 3D consistency.
Based on the above discussion, it seems that using more widely-covered conditions lead to fully consistent generated avatars in diffusion models. Actually, several efforts (Kirschstein et al., 2023; Ding et al., 2023) have integrated 3D-aware conditions like face normal or depth to guide their diffusion models. To further analyse relative strategies, we build a baseline diffusion model to generate avatars, conditioned on the low-resolution face image / normal generated from pre-trained INSTA (Zielonka et al., 2023) and emotion label embedding from CLIP (Radford et al., 2021). As illustrated in Fig. 1 (d), we observe that DiffusionRig (Ding et al., 2023) and our baseline model both suffer from high temporal error and instability, revealing that employing more conditions for diffusion models may not always lead to superior results. We argue the reason behind are two-fold: 1) the inaccuracy within these conditions disturbs the diffusion process and accumulates to the final result; 2) the based diffusion model (Rombach et al., 2022) is image but not video generation model, and thus temporal consistency cannot be modeled or guaranteed. Note that, given a video as training data, the ground truth of temporal patterns has been implicitly contained. This motivates us to learn these patterns and constrains other noisy conditions for talking avatar generation.
To this end, we introduce ConsistentAvatar, a novel framework that combines diffusion models with the temporal, 3D-aware, and emotional conditions, for generating fully consistent and high-fidelity head avatars. Instead of directly generating avatar, we learn to first generate temporally consistent representation as a constraint to guide other conditions. Concretely, our method starts from the efficient INSTA method (Zielonka et al., 2023) and utilize its outputs as the initial results. In order to model the temporal consistency, we propose a Temporally-Sensitive Detail (TSD) generated by Fourier transformation, containing high-frequency information and contours that change significantly between frames. We extract TSD from coarse RGB output of INSTA and the target video frame, and propose a temporal consistency diffusion model to align the input TSD to the precise one. Besides, we utilize the coarse normal output of INSTA as the 3D-aware condition, and propose an emotion selection module to generate emotion embedding for each frame in an unsupervised manner. With the aligned TSD, normal, and emotion embedding as conditions, we then propose a fully consistent diffusion model to generate the final avatars. In this way, we guarantee the temporal consistency during the avatar generation, and complement other conditions to contribute to a fully consistent and high-quality avatar generation result. In summary, our contributions are as follows:

  • We propose ConsistentAvatar, a diffusion-based neural renderer that generates temporal, 3D, and expression consistent talking head avatars.

  • We learn to align a novel Temporally-Sensitive Details (TSD) to maintain the stability between generated frames, and complement rough normal and emotion conditions for high-fidelity generation.

  • Extensive experiments demonstrate that ConsistentAvatar outperforms the state-of-the-art methods on the generated appearance quality, details, expression and temporal consistency.

2. Related Work

2.1. 3D Face Animation

Early methods (Zhang et al., 2023a; Ren et al., 2021; Deng et al., 2020a) use 3DMM priors to instruct the generator. SadTalker (Zhang et al., 2023a), for instance, generates head poses and expressions of 3DMMs from audio and implicitly modulate a 3D-aware face synthesis model. Recent methods (Zheng et al., 2022; Grassal et al., 2022; Zielonka et al., 2023; Gafni et al., 2021; Athar et al., 2022; Peng et al., 2023), combined 3D representation techniques with 3DMMs to control the expressions and poses of avatars. For example, NHA (Grassal et al., 2022) and IMAvatar (Zheng et al., 2022) refine the mesh topology of FLAME (Li et al., 2017) to achieve more realistic mesh-based avatars. INSTA (Zielonka et al., 2023) and RigNeRF (Athar et al., 2022) utilize 3DMMs to construct radiance fields. Additionally, other 3D representation methods , such as GaussianAvatars (Qian et al., 2023), construct dynamic 3D representations based on 3D Gaussian splats rigged to a parametric morphable face model. More recent works (Kirschstein et al., 2023; Shen et al., 2023; Kim et al., 2023a; Ding et al., 2023) have leveraged the powerful generation capabilities of 2D diffusion models. For instance, DiffusionAvatars (Kirschstein et al., 2023) utilizes the 3D priors of the recent Neural Parametric Head Model (NPHM) (Giebenhain et al., 2023) combines with a 2D rendering network for high-quality image synthesis. In our work, we integrate 3D representation with diffusion models, simultaneously leveraging 2D and 3D priors for consistent head avatar synthesis.

2.2. Controllable face generation with 2D Diffusion

Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020) integrate image generation with a sequential denoising process of isotropic Gaussian noise. In this process, the model is trained to anticipate noise levels from the input image. Due to their remarkable ability to generate 2D content, they have gradually found application in face-related tasks. Many studies (Ding et al., 2023; Shen et al., 2023; Kirschstein et al., 2023; Kim et al., 2023a), refine pre-trained diffusion models by introducing additional control factors like landmarks or depth. For example, DiffusionRig (Ding et al., 2023) suggests conditioning the diffusion model on rasterized grids, considering factors such as normals and albedo. Conversely, DiffTalk (Shen et al., 2023) utilizes audio-driven and landmark-based conditioning to improve identity generalization. Despite the good controllability and visual quality of these methods, they often lack accurate 3D representations or precise 2D detail priors. As a result, they struggle to maintain consistent 3D rendering across various viewpoints and temporal consistency in video rendering simultaneously.

2.3. Emotional Talking Video Portraits

Recently, there has been a growing effort to incorporate emotions into the synthesis process for better control over the final output. For instance, EMOCA (Danecek et al., 2022) introduces a novel deep perceptual emotion consistency loss, ensuring alignment between the reconstructed 3D expression and the depicted emotion in the input image. Meanwhile, GMTalker (Xia et al., 2023) proposes a Gaussian mixture-based Expression Generator (GMEG), enabling the creation of a continuous and multimodal latent space for more versatile emotion manipulation. However, approaches like EVP (Ji et al., 2021) and EMMN (Tan et al., 2023), which directly infer emotions from labeled audio, may encounter accuracy issues due to the inherent complexity of emotional expression. In contrast, methods such as MEAD (Wang et al., 2020) and the work by Sinha et al. (Sinha et al., 2022) implicitly learn the intrinsic relationship between emotions and facial expressions through the use of emotion labels. In our methodology, we leverage the MEAD (Wang et al., 2020) dataset to assign emotion labels to the experimental dataset based on emotional similarities, facilitating precise control over the emotions of the resulting portraits.

3. Preliminary

INSTA. INSTA (Zielonka et al., 2023) is based on a dynamic neural radiance field composed of neural graphics primitives embedded around a parametric face model. It is capable of reconstructing photorealistic digital avatars instantaneously in less than 10 minutes, while allowing for interactive rendering of novel poses and expressions. We choose INSTA to get a proxy of talking head under a target expression and head pose. The proxy, i.e., initial result used as input for our method is 512×512512\times 512 rendered by volume rendering process, containing limited appearance quality and expression accuracy. Our method is capable of lifting the proxy to a high-fidelity and consistent avatar.

Denoising Diffusion Probabilistic Models. Denoising Diffusion Probabilistic Models (DDPMs) (Ranzato et al., 2021; Nichol and Dhariwal, 2021; Ho et al., 2020) belong to a class of generative models that take random noise images as input and progressively denoise them to produce photorealistic images. This process can be viewed as the reverse of the diffusion process, which gradually adds noise to images. The core component of DDPMs is a denoising network denoted as fθf_{\theta}, During training, it receives a noisy image xtx_{t} and a timestep t(1tT)t\ (1\leq t\leq T), and predicts the noise at time t:ϵtt:\epsilon_{t}. More formally, the predicted noise at time tt is ϵt^=fθ(xt,t)\hat{\epsilon_{t}}=f_{\theta}\left(x_{t},t\right), where xt=αtx0+1αt2ϵtx_{t}=\alpha_{t}x_{0}+\sqrt{1-\alpha_{t}^{2}}\epsilon_{t}, ϵt\epsilon_{t} is a random, normally distributed noise image, and αt\alpha_{t} is a hyperparameter that gradually increases the noise level of xtx_{t} with each step of the forward process. The loss is computed based on the distance between ϵt\epsilon_{t} and ϵt^\hat{\epsilon_{t}}. Therefore, the trained model can generate images by taking a random noise image as input and progressively denoising it to achieve photorealism.

4. Method

In this section, we introduce ConsistentAvatar, a framework specifically designed for generating high-quality avatars while maintaining full consistency. An overview of this framework is illustrated in Fig. 2. Ensuring temporal consistency in generating portraits remains a challenge as the diffusion model struggles to directly learn time-related information from images. Therefore, we learn to generate temporally consistent representations as constraints to guide other conditions. We define Temporally-Sensitive Details (TSD), a detail map ITSDI_{TSD} generated by Fourier transformation from the coarse RGB output of INSTA (Zielonka et al., 2023) and the target video frame. Furthermore, our experiments reveal that directly using the extracted TSD as a condition does not yield satisfactory temporal consistency, we propose a temporal consistency diffusion model to align the input TSD to the precise one (Sec. 4.1). After ensuring temporal consistency during the avatar generation process, we utilize the coarse normal output of INSTA and emotional text embeddings as conditions to construct a fully consistent diffusion model (Sec. 4.2). Finally, to expedite and further optimize our model, we draw inspiration from LCM (Luo et al., 2023) and SDXL (Podell et al., 2023), significantly reducing the necessary inference steps and further enhancing the quality of image generation (Sec. 4.3).

Refer to caption
Figure 2. Overview. ConsistentAvatar begins with the implementation of the highly efficient INSTA (Zielonka et al., 2023) method, leveraging its outputs as initial results (Stage 1). To address temporal consistency, we introduce a concept termed as Temporally-Sensitive Detail (TSD), derived through Fourier transformation. Extracting TSD from the coarse RGB output of INSTA and the target video frame, we develop a temporal consistency diffusion model to accurately align the input TSD with the precise one (Stage 2). Subsequently, we employ the coarse normal output of INSTA as a parameter for 3D perception and introduce an emotion selection module to generate emotion embeddings for each frame. By integrating aligned TSD, normal, and emotion embeddings as conditioning factors, we propose a fully consistent diffusion model to generate the final avatars (Stage 3).

4.1. Temporally Consistent Module

As mentioned above, to ensure temporal consistency in generating portraits, we extract Temporally-Sensitive details (TSD) generated by Fourier transformation from the coarse RGB output of INSTA (Zielonka et al., 2023) and the target video frame. We then align these details through a diffusion model. This approach complements other conditions and contributes to achieving fully consistent and high-quality avatar generation results.

Given a monocular RGB video containing KK frames {\{I1,I2,,IK{I_{1},I_{2},\ldots,I_{K}}}\} , camera pose PcP_{c}, tracked FLAME (Li et al., 2017) meshes 𝑴={Mi}\boldsymbol{M}=\{M_{i}\} with corresponding expressions 𝑬={Ei}1×100\boldsymbol{E}=\{E_{i}\}\in\mathbb{R}^{1\times 100} and poses 𝑷={Pi}1×6\boldsymbol{P}=\{P_{i}\}\in\mathbb{R}^{1\times 6}. We obtain RGB output and normal using INSTA (Zielonka et al., 2023). The process is described as follows:

(1) insta(Ii,Pc,Mi,Ei,Pi)Irgbi,Inormali,\displaystyle\mathcal{F}_{insta}(I_{i},P_{c},M_{i},E_{i},P_{i})\to I_{rgb_{i}},I_{normal_{i}},

where insta\mathcal{F}_{insta} is the INSTA model. After obtaining the RGB output, we utilize Fourier transform and its inverse transform to extract TSD from the RGB output. We apply the Fourier transform to shift an image from its spatial domain to the frequency domain, revealing various frequency components. Retaining these high-frequency elements enables the Fourier transform to efficiently extract image details, which are represented as:

(2) Fi=+Irgbiejwtdt={=0,wW0,wW,\displaystyle F_{i}=\int_{-\infty}^{+\infty}I_{rgb_{i}}e^{-jwt}dt=\left\{\begin{matrix}=0,&w\notin W\\ \neq 0,&w\in W\end{matrix}\right.,

where the equation represents content extraction at frequency ww from the image IrgbiI_{rgb_{i}}, with ejwte^{-jwt} as the orthogonal basis and WW as the frequency set of the image. Experimentally, this paper sets ww==1010. Then performing inverse Fourier transform to convert the frequency domain back to images, which are represented as:

(3) ITSDi=+Fiejwt𝑑t.\displaystyle I_{TSD_{i}}=\int_{-\infty}^{+\infty}F_{i}e^{jwt}dt.

As illustrated in Fig. 2, the TSD of a video frame contains high-frequency information and contours that represent crucial feature of expression, head pose and details for generating the avatar. However, TSD also varies significantly between adjacent frames. This inspires us to predict stable TSD of a frame from inaccurate ITSDiI_{TSD_{i}}. Performing the same operation on the ground truth video frames allows us to obtain the ground truth of TSD. Therefore, we propose employing a Temporally-Sensitive Diffusion Model (TSDM) to align ITSDiI_{TSD_{i}}. Specifically, inspired by (Ding et al., 2023), we first encode the obtained ITSDiI_{TSD_{i}} 512×512×3\in\mathbb{R}^{512\times 512\times 3} to obtain the global latent code (ITSDi)1×256\mathcal{E}(I_{TSD_{i}})\in\mathbb{R}^{1\times 256}. Then we pass PiP_{i} and EiE_{i} through a linear layer to obtain the same scale as (ITSDi)\mathcal{E}(I_{TSD_{i}}) and concatenate them to get fc3×256f_{c}\in\mathbb{R}^{3\times 256}. We add new cross-attention layers to the U-Net, following IPAdapter (Ye et al., 2023). Let ZZ be the intermediate feature map computed by an existing cross-attention operation in the pre-trained LDM (Rombach et al., 2022): Z=ATTENTION(Q,K,V)Z=ATTENTION(Q,K,V). Then, we perform direct conditioning by adding another cross-attention layer:

(4) ZZ=ATTENTION(Q,Wkfct,Wvfct).\displaystyle Z\leftarrow Z=ATTENTION(Q,{W^{k}}{f_{c}}^{t},{W^{v}}{f_{c}}^{t}).

The more specific denoising process can be represented as follows:

(5) ϵ^t=([xt,fct],t,x0),\displaystyle{\hat{\epsilon}}_{t}=\mathcal{F}([x_{t},{f_{c}}^{t}],t,x_{0}),

where xtx_{t} is the noisy image at timestep t{1,2,,T}t\in\{1,2,\ldots,T\}, x0x_{0} is the ground truth latent image, ϵ^t{\hat{\epsilon}}_{t} is the predicted noise, and \mathcal{F} represents the denoising model. The optimization objective of the entire denoising process is as follows:

(6) TSD=ϵt^ϵt22.\displaystyle\mathcal{L}_{TSD}=\left\|\hat{\epsilon_{t}}-\epsilon_{t}\right\|_{2}^{2}.

x~t1=xtϵ^t\tilde{x}_{t-1}=x_{t}-{\hat{\epsilon}}_{t} is the denoising result of xtx_{t} at time step t. The final denoised result x~0\tilde{x}_{0} is then upsampled to the pixel space with the pretrained image decoder ITSDiI_{TSD_{i}}^{\prime} = 𝒟TSD\mathcal{D}_{TSD}(x~0\tilde{x}_{0}), where ITSDi512×512×3I_{TSD_{i}}^{\prime}\in\mathbb{R}^{512\times 512\times 3} is the reconstructed TSD image. In this way, we align ITSDiI_{TSD_{i}} to an accurate one, providing the key information of real portrait changes along time axis. Further, ITSDiI_{TSD_{i}}^{\prime} can be used as a reliable guidance for temporally-consistent avatar generation.

4.2. Fully Consistent Module

In Sec. 4.1, after passing TSD through the diffusion model, we obtain precise and temporally stable TSD. This compensates for other conditions that may not be particularly accurate. Therefore, by adding normal condition and emotion condition, we alleviate 3D consistency and expression consistency.

Emotion Condition: For emotion condition, we propose utilizing the MEAD (Wang et al., 2020) dataset to acquire emotion labels for our experimental dataset. As described in Fig. 3, the MEAD comprises numerous monocular video clips {\{V1,V2,,VM{V_{1},V_{2},\ldots,V_{M}}}\}, each clips corresponds to an emotional label Lm{L_{m}\in\{L1,L2,,LM{L_{1},L_{2},\ldots,L_{M}}}\} (such as anger, happy, etc.). Leveraging this rich emotion dataset, we construct a facial expression database. Subsequently, employing DECA (Feng et al., 2021), we compute facial expression vectors for each frame in every clip EmnE_{mn}, nn represents the frame index of the clip. And then determine the similarity between the current frame’s facial expression vector EiE_{i} and those in the database. Finally, we assign the emotion label corresponding to the facial expression vector with the highest similarity in the database as the emotion label for the current frame. The process is described as follows:

(7) 𝒪(Ii)=L(argmax(cs(deca(Ii),deca(Vmn)))),\displaystyle\mathcal{E}\mathcal{M}\mathcal{O}(I_{i})={L}({argmax}(cs(deca(I_{i}),{deca}(V_{mn})))),

where IiI_{i} is an image from our dataset, and VmnV_{mn} is a frame from the MEAD dataset, cs()cs() represents the cosine similarity function, which is used to measure the similarity between two expression vectors. argmax()argmax() retrieves the image corresponding to the maximum similarity, which belongs to the MEAD dataset. After acquiring emotional labels corresponding to image through 𝒪(Ii)\mathcal{E}\mathcal{M}\mathcal{O}(I_{i}), we use the text encoder in CLIP (Radford et al., 2021) to extract textual features of emotion labels, obtaining the final emotion code Ti1×768T_{i}\in\mathbb{R}^{1\times 768}. After acquiring STD, normal, and emotion conditions, we perform the same linear operation on TiT_{i} as on EiE_{i} and PiP_{i}, and then concatenate them to obtain fd4×256f_{d}\in\mathbb{R}^{4\times 256}. At that time, it can be utilized as a condition through cross-attention operation. Finally, we construct a Fully Consistent Stable Diffusion model (FCSD). Specifically, we utilize the pretrained Stable Diffusion model (SD) (Rombach et al., 2022) as the backbone network to expedite training and enhance generation quality. Additionally, we incorporate a fully consistency module, effectively integrating consistency-related information into the backbone network, to construct a ControlNet (Zhang et al., 2023b). The final rendering is achieved through an iteratively denoising the full noise xTx_{T} with our fully-consistent neural renderer 𝒮\mathcal{S} conditioned on zcz_{c}:

(8) ϵ^t=𝒮([xt,fdt],t,x0,C(xt,zc)),\displaystyle{\hat{\epsilon}}_{t}=\mathcal{S}([x_{t},f_{d}^{t}],t,x_{0},C(x_{t},z_{c})),

where C is the ControlNet architecture and zcz_{c} is the concatenation of the normal condition and the TSD condition. During training, we minimize the following loss:

(9) portrait=ϵt^ϵt22.\displaystyle\mathcal{L}_{portrait}=\left\|\hat{\epsilon_{t}}-\epsilon_{t}\right\|_{2}^{2}.

Similar to Sec. 4.1, the final denoised result x~0\tilde{x}_{0} is then upsampled to the pixel space with the pretrained image decoder IgeniI_{gen_{i}} = 𝒟portrait\mathcal{D}_{portrait}(x~0\tilde{x}_{0}), where Igeni512×512×3I_{gen_{i}}\in\mathbb{R}^{512\times 512\times 3} is the reconstructed face image. This process provides the network with additional multimodal conditions, helping it control the 3D and expression consistency of the characters under the well-guided aligned TSD, thereby assisting in generating portrait animations with more precise motion.

Refer to caption
Figure 3. Emotion text selection module diagram.

4.3. Optimization

It is indisputable that both generation speed and the quality of portrait generation are pivotal criteria for task assessment. We closely follow the development of diffusion models and, based on this foundation, further accelerate efficiency and enhance image quality.

Latent Consistency Model (LCM). To ensure our model maintains high-quality generation while minimizing the required steps, we employ LCM (Luo et al., 2023) to expedite the generation process. The core concept of this model lies in redefining the fundamental logic of traditional diffusion models (such as DDPM (Ho et al., 2020)) in the generation process. Traditional diffusion models generate final results by gradually reducing noise in images, typically through iterative and time-consuming processes. In contrast, LCM (Luo et al., 2023) transforms traditional numerical ordinary differential equation (ODE) solvers into neural network-based solvers, enabling direct prediction of the final clear image and thereby reducing intermediate steps, significantly enhancing efficiency. By employing LCM (Luo et al., 2023) to enhance our model, we have reduced the required steps for predicting the final image from around 1000 steps to approximately 10 steps.

Refiner. Inspired by Stable Diffusion XL (SDXL) (Podell et al., 2023), we have further elevated the generation quality of our model. SDXL represents the latest optimized version of Stable Diffusion, comprising a two-stage cascaded diffusion model consisting of a Base model and a Refiner model. Integrating the Refiner model with our own, after the Base model generates latent features of the image, we utilize the Refiner model to conduct minor noise reduction and enhance detail quality on these latent features.

5. Experiment

5.1. Setup

Dataset: In our experiments, we utilize four datasets. Initially, we employ the dataset released by INSTA (Zielonka et al., 2023) to train the primary framework. This dataset comprises 10 monocular videos, each capturing the performance of an individual actor. These videos undergo cropping and resizing, achieving a resolution of 5122{512^{2}}, effectively removing extraneous elements from the facial region through background subtraction. Additionally, for equitable comparison with other methods, we utilize two datasets from IMAvatar (Zheng et al., 2022) and NerFace (Gafni et al., 2021) which share the same data format as INSTA (Zielonka et al., 2023). Finally, during the emotional labeling phase, we utilize the Multi-view Emotional Audio-visual Dataset (MEAD) (Wang et al., 2020), which includes a dialogue video corpus featuring 60 actors conversing with 8 different emotions. High-quality audio-visual clips from seven different perspectives are captured in a strictly controlled environment.

Evaluation Protocol: We utilize the final 350 frames of each video in our dataset for testing purposes, while the remaining frames are allocated for training. To assess image quality, we employ metrics consistent with state-of-the-art methods (Zielonka et al., 2023; Zheng et al., 2022), including L2 loss, structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), and the perceptual metric LPIPS. Additionally, to evaluate the temporal consistency of the generated videos, we utilize optical flow to compute the degree of change between adjacent frames.

Implementation Details: Our framework was implemented using PyTorch on a machine equipped with an RTX 3090 GPU. The training process is divided into three stages. In the first stage, all experimental configurations match those of INSTA (Zielonka et al., 2023), yielding RGB and normal outputs for the portraits. Moving to the second stage, we utilize the temporally unstable TSDs obtained from the RGB outputs of the first stage to train an TSDs generator with ground truth supervision. We utilize the Adam optimizer (Kingma and Ba, 2015) with a learning rate of 10510^{-5} and conduct 3000 iterations with a batch size of 4. Finally, in the third stage, we train a fully consistent portrait generator based on the outputs of the first two stages. Once again, we use the Adam optimizer with a learning rate of 10410^{-4} and conduct 15000 iterations with a batch size of 4.

Baseline setting: To illustrate the role of TSD in temporal consistency more clearly, we use Stage 1 combined with Stage 3 as our baseline. Here, TSD is not learned through Stage 2, but is directly used as a condition for ControlNet (Zhang et al., 2023b).

5.2. Comparison with the State-of-the-art Methods

Image quality evaluation: We conduct comprehensive experiments on our dataset, assessing the quality and consistency of the synthetic digital human avatars generated by our method, and comparing them with state-of-the-art methods such as IMAvatar (Zheng et al., 2022), DiffusionRig (Ding et al., 2023) and INSTA (Zielonka et al., 2023). As shown in Fig. 4, our approach yields more realistic results. We capture facial details such as wrinkles and eyeglass frames well. In contrast, other methods like INSTA (Zielonka et al., 2023) fail to capture wrinkle information and other details. Tab. 1 presents quantitative results, demonstrating that our method achieves state-of-the-art performance across three different datasets. Particularly noteworthy is the substantial improvement in both PSNR and L2 loss compared to INSTA (Zielonka et al., 2023). Taken together, these metrics indicate that our generated portraits are closer to ground truth and more realistic. Additionally, our ability to better capture facial expressions contributes to the overall enhancement of the results. Additional. We are well aware of the importance of real-time performance, so as mentioned in Sec. 1, we have focused on improving speed by introducing LCM (Luo et al., 2023) on top of our model. As shown in the data in the Tab. 1, we have significantly reduced the time. Note that the above results were obtained with a time step setting of 10 during the denoising process, while the default diffusion model step size is set to 1000. The incorporation of LCM (Luo et al., 2023) significantly reduces the time step.

Refer to caption
Figure 4. Qualitative Results. Clearly, the facial avatars reconstructed by our method exhibit accurate and lifelike details, including intricate features such as wrinkles and eyes. Other methods produce excessively smooth results.
Method Dataset L2 \downarrow PSNR \uparrow SSIM \uparrow LPIPS \downarrow Time(s) \downarrow
NHA (Grassal et al., 2022) INSTA 0.0022 27.71 0.95 0.040 0.63
NeRFace (Gafni et al., 2021) 0.0018 29.28 0.95 0.070 9.68
IMAvatar (Zheng et al., 2022) 0.0023 27.62 0.94 0.060 12.34
MAVavatar (Xu et al., 2023) 0.0027 25.76 0.93 0.070 1.10
INSTA (Zielonka et al., 2023) 0.0018 28.97 0.95 0.050 0.05
DiffusionRig (Ding et al., 2023) 0.0016 31.34 0.96 0.047 4.10
w/o stage2(Baseline) 0.0010 33.78 0.97 0.040 2.30
w/o LCM 0.0008 34.05 0.97 0.038 8.20
Ours 0.0008 34.05 0.97 0.038 2.40
PointAvatar (Zheng et al., 2023) PointAvatar 0.0027 26.04 0.88 0.147 0.80
w/o stage2(Baseline) 0.0012 32.42 0.93 0.044 2.30
w/o LCM 0.0011 32.72 0.95 0.040 8.20
Ours 0.0011 32.72 0.95 0.040 2.40
NeRFace (Gafni et al., 2021) NeRFace 0.0016 26.85 0.95 0.060 9.68
w/o stage2(Baseline) 0.0014 32.80 0.96 0.045 2.30
w/o LCM 0.0010 33.35 0.96 0.040 8.20
Ours 0.0010 33.35 0.96 0.040 2.40
Table 1. Quantitative analyses between our method and state-of-the-art models.

3D consistency results: 3D consistency measures whether the character maintains the correct appearance from different viewpoints. As shown in Fig. 1 (b), methods like SadTalker (Zhang et al., 2023a) do not naturally exhibit good 3D consistency since they do not involve 3D operations during learning and rely on a single-image 2D generation process. From Fig. 5, we can also observe the advantage of our method, with the angles of head rotation being closer to the ground truth and the results appearing more realistic. For quantitative results, we utilize DECA (Feng et al., 2021) to estimate the pose coefficients of the generated portraits, calculating the error between the predicted pose coefficients and the ground truth, referred to as Pose Error (PE), measured by the Euclidean distance. As shown in Tab. 2, our method significantly outperforms SadTalker, with minimal error compared to the ground truth.

Method PE \downarrow Method EE \downarrow
SadTalker (Zhang et al., 2023a) 0.0755 INSTA (Zielonka et al., 2023) 1.9236
w/o aligned TSD 0.02328 Diffusionrig (Ding et al., 2023) 1.7665
w/o normal 0.03653 w/o aligned TSD 1.4355
Ours 0.0096 w/o emotion text 1.1922
Ours 0.8452
Table 2. Quantitative results and ablation results compared to state-of-the-art methods.
Refer to caption
Figure 5. Comparison with SadTalker in terms of 3D consistency.

Expression consistency results: The expression consistency measures how well the generated results fit the target expressions. As shown in Fig. 6, our method is capable of achieving more accurate expressions. For quantitative analysis, similar to PE, we employ DECA (Feng et al., 2021) to evaluate the expression coefficients of the generated portraits, calculating the disparity between the predicted expression coefficients and the ground truth, referred to as Expression Error (EE). As depicted in Tab. 2, our approach surpasses both Diffusionrig (Ding et al., 2023) and INSTA (Zielonka et al., 2023), exhibiting minimal error relative to the ground truth.

Refer to caption
Figure 6. Comparison with different state-of-the-art methods in terms of expression consistency.

Temporal consistency results: The evaluation of temporal consistency evaluates the stability of the generated video, focusing on the absence of jitter or significant fluctuations between frames, ensuring smooth playback. We compare our method with state-of-the-art approaches like INSTA(Zielonka et al., 2023), SadTalker(Zhang et al., 2023a) and DiffusionRig (Ding et al., 2023) in various categories. We utilize optical flow to compute the degree of change between two frames. Naturally, we consider the normal variation between video frames. Therefore, we compare it with the ground truth as well. From the Fig. 7, it’s evident that our approach closely approximates the ground truth.

Refer to caption
Figure 7. Comparison with different state-of-the-art methods in terms of temporal consistency.

5.3. Ablation Studies

We conduct a series of ablation studies to analyze the different components of our method. Specifically, we focus on 1) the impact of unaligned coarse TSD versus aligned TSD through the diffusion model on other conditions and on temporal consistency; 2) the impact of the emotion condition on the final results; 3) the impact of the normal condition on the final 3D consistency;

Aligned TSD and Unaligned TSD. From Fig. 1 (d), it is evident that the temporal consistency of our baseline is inferior to that of our final method. Since the baseline directly adopts unaligned TSD conditions, this proves that aligned TSD can provide better temporal consistency. From Fig. 8 (a) and (c), it can be observed that aligned TSD conditions offer more accurate expression and better 3D consistency.

Emotion Condition. With the aligned TSD as a condition, we compared the results with and without the addition of the emotion condition. As shown in Fig. 8 (b), when the emotion condition is included, it helps the model fine-tune facial expressions, resulting in outcomes closer to the ground truth.

Normal Condition. With the aligned TSD as a condition, we compare the results with and without the addition of the normal condition. As shown in Fig. 8 (d), when the normal condition is not included, it fails to maintain good 3D consistency, resulting in artifacts and other noise.

Refer to caption
Figure 8. The results of the ablation experiments on TSD condition, normal condition, and emotion condition.

6. Limitation:

Although our work achieves realistic and fully consistent portraits through the diffusion model and the utilization of TSD, normal, and emotion conditions, there are still limitations. Our method often lacks accurate modeling of teeth and eyeball due to the lack of geometric constraints, as depicted in Fig. 9.

Refer to caption
Figure 9. The results of our method’s limitations.

7. Conclusion

ConsistentAvatar presents a novel framework for generating talking avatars with full consistency and high fidelity. We introduce a Temporally-Sensitive Detail (TSD) map containing high-frequency features and contours that exhibit significant variation over time. Utilizing a temporal consistent diffusion module, we align the TSD of the initial result with the ground truth of the video frame. The final avatar is generated using a fully consistent diffusion module, conditioned on the aligned TSD, rough head normal, and emotion prompt embedding. Aligned TSD, representing temporal patterns, guides the diffusion process to produce temporally stable talking heads. Its reliable guidance supplements the inaccuracies of other conditions, thereby reducing accumulated errors and enhancing consistency across various aspects.

8. Acknowledgments

This work was supported by the National Science Fund of China under Grant No. 62361166670 and No. 62376121.

References

  • (1)
  • Athar et al. (2022) ShahRukh Athar, Zexiang Xu, Kalyan Sunkavalli, Eli Shechtman, and Zhixin Shu. 2022. RigNeRF: Fully Controllable Neural 3D Portraits. arXiv:2206.06481 [cs.CV]
  • Blanz and Vetter (1999) Volker Blanz and Thomas Vetter. 1999. A morphable model for the synthesis of 3D faces. In Proceedings of the 26th annual conference on Computer graphics and interactive techniques. 187–194.
  • Danecek et al. (2022) Radek Danecek, Michael J. Black, and Timo Bolkart. 2022. EMOCA: Emotion Driven Monocular Face Capture and Animation. arXiv:2204.11312 [cs.CV]
  • Deng et al. (2020a) Yu Deng, Jiaolong Yang, Dong Chen, Fang Wen, and Xin Tong. 2020a. Disentangled and Controllable Face Image Generation via 3D Imitative-Contrastive Learning. arXiv:2004.11660 [cs.CV]
  • Deng et al. (2020b) Yu Deng, Jiaolong Yang, Sicheng Xu, Dong Chen, Yunde Jia, and Xin Tong. 2020b. Accurate 3D Face Reconstruction with Weakly-Supervised Learning: From Single Image to Image Set. arXiv:1903.08527 [cs.CV]
  • Ding et al. (2023) Zheng Ding, Xuaner Zhang, Zhihao Xia, Lars Jebe, Zhuowen Tu, and Xiuming Zhang. 2023. DiffusionRig: Learning Personalized Priors for Facial Appearance Editing. arXiv:2304.06711 [cs.CV]
  • Feng et al. (2021) Yao Feng, Haiwen Feng, Michael J. Black, and Timo Bolkart. 2021. Learning an Animatable Detailed 3D Face Model from In-The-Wild Images. arXiv:2012.04012 [cs.CV]
  • Gafni et al. (2021) Guy Gafni, Justus Thies, Michael Zollhöfer, and Matthias Nießner. 2021. Dynamic Neural Radiance Fields for Monocular 4D Facial Avatar Reconstruction. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, June 19-25, 2021. 8649–8658.
  • Giebenhain et al. (2023) Simon Giebenhain, Tobias Kirschstein, Markos Georgopoulos, Martin Rünz, Lourdes Agapito, and Matthias Nießner. 2023. Learning Neural Parametric Head Models. arXiv:2212.02761 [cs.CV]
  • Grassal et al. (2022) Philip-William Grassal, Malte Prinzler, Titus Leistner, Carsten Rother, Matthias Nießner, and Justus Thies. 2022. Neural Head Avatars from Monocular RGB Videos. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022. IEEE, 18632–18643.
  • Gu et al. (2019) Kuangxiao Gu, Yuqian Zhou, and Thomas Huang. 2019. FLNet: Landmark Driven Fetching and Learning Network for Faithful Talking Facial Animation Synthesis. arXiv:1911.09224 [cs.CV]
  • Han et al. (2023) Yue Han, Jiangning Zhang, Junwei Zhu, Xiangtai Li, Yanhao Ge, Wei Li, Chengjie Wang, Yong Liu, Xiaoming Liu, and Ying Tai. 2023. A Generalist FaceX via Learning Unified Facial Representation. arXiv:2401.00551 [cs.CV]
  • Ho et al. (2020) Jonathan Ho, Ajay Jain, and Pieter Abbeel. 2020. Denoising Diffusion Probabilistic Models. arXiv:2006.11239 [cs.LG]
  • Ji et al. (2021) Xinya Ji, Hang Zhou, Kaisiyuan Wang, Wayne Wu, Chen Change Loy, Xun Cao, and Feng Xu. 2021. Audio-Driven Emotional Video Portraits. arXiv:2104.07452 [cs.CV]
  • Karras et al. (2019) Tero Karras, Samuli Laine, and Timo Aila. 2019. A Style-Based Generator Architecture for Generative Adversarial Networks. arXiv:1812.04948 [cs.NE]
  • Karras et al. (2020) Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. 2020. Analyzing and Improving the Image Quality of StyleGAN. arXiv:1912.04958 [cs.CV]
  • Khakhulin et al. (2022) Taras Khakhulin, Vanessa Sklyarova, Victor Lempitsky, and Egor Zakharov. 2022. Realistic One-shot Mesh-based Head Avatars. arXiv:2206.08343 [cs.CV]
  • Kim et al. (2023b) Gyeongman Kim, Hajin Shim, Hyunsu Kim, Yunjey Choi, Junho Kim, and Eunho Yang. 2023b. Diffusion Video Autoencoders: Toward Temporally Consistent Face Video Editing via Disentangled Video Encoding. arXiv:2212.02802 [cs.CV]
  • Kim et al. (2023a) Minchul Kim, Feng Liu, Anil Jain, and Xiaoming Liu. 2023a. DCFace: Synthetic Face Generation with Dual Condition Diffusion Model. arXiv:2304.07060 [cs.CV]
  • Kingma and Ba (2015) Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, Yoshua Bengio and Yann LeCun (Eds.).
  • Kirschstein et al. (2023) Tobias Kirschstein, Simon Giebenhain, and Matthias Nießner. 2023. DiffusionAvatars: Deferred Diffusion for High-fidelity 3D Head Avatars. arXiv:2311.18635 [cs.CV]
  • Li et al. (2017) Tianye Li, Timo Bolkart, Michael J. Black, Hao Li, and Javier Romero. 2017. Learning a model of facial shape and expression from 4D scans. ACM Trans. Graph. 36, 6 (2017), 194:1–194:17.
  • Luo et al. (2023) Simian Luo, Yiqin Tan, Longbo Huang, Jian Li, and Hang Zhao. 2023. Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference. arXiv:2310.04378 [cs.CV]
  • Nguyen-Phuoc et al. (2020) Thu Nguyen-Phuoc, Christian Richardt, Long Mai, Yong-Liang Yang, and Niloy J. Mitra. 2020. BlockGAN: Learning 3D Object-aware Scene Representations from Unlabelled Images. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (Eds.).
  • Nichol and Dhariwal (2021) Alex Nichol and Prafulla Dhariwal. 2021. Improved Denoising Diffusion Probabilistic Models. arXiv:2102.09672 [cs.LG]
  • Noguchi et al. (2022) Atsuhiro Noguchi, Xiao Sun, Stephen Lin, and Tatsuya Harada. 2022. Unsupervised Learning of Efficient Geometry-Aware Neural Articulated Representations. arXiv:2204.08839 [cs.CV]
  • Peng et al. (2023) Ziqiao Peng, Haoyu Wu, Zhenbo Song, Hao Xu, Xiangyu Zhu, Jun He, Hongyan Liu, and Zhaoxin Fan. 2023. EmoTalk: Speech-Driven Emotional Disentanglement for 3D Face Animation. arXiv:2303.11089 [cs.CV]
  • Podell et al. (2023) Dustin Podell, Zion English, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas Müller, Joe Penna, and Robin Rombach. 2023. SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis. arXiv:2307.01952 [cs.CV]
  • Qian et al. (2023) Shenhan Qian, Tobias Kirschstein, Liam Schoneveld, Davide Davoli, Simon Giebenhain, and Matthias Nießner. 2023. GaussianAvatars: Photorealistic Head Avatars with Rigged 3D Gaussians. arXiv:2312.02069 [cs.CV]
  • Radford et al. (2021) Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. arXiv:2103.00020 [cs.CV]
  • Ranzato et al. (2021) Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (Eds.). 2021. Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual.
  • Ren et al. (2021) Yurui Ren, Ge Li, Yuanqi Chen, Thomas H. Li, and Shan Liu. 2021. PIRenderer: Controllable Portrait Image Generation via Semantic Neural Rendering. arXiv:2109.08379 [cs.CV]
  • Rombach et al. (2022) Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. 2022. High-resolution image synthesis with latent diffusion models. In CVPR. 10684–10695.
  • Sanyal et al. (2019) Soubhik Sanyal, Timo Bolkart, Haiwen Feng, and Michael J. Black. 2019. Learning to Regress 3D Face Shape and Expression from an Image without 3D Supervision. arXiv:1905.06817 [cs.CV]
  • Shen et al. (2023) Shuai Shen, Wenliang Zhao, Zibin Meng, Wanhua Li, Zheng Zhu, Jie Zhou, and Jiwen Lu. 2023. DiffTalk: Crafting Diffusion Models for Generalized Audio-Driven Portraits Animation. arXiv:2301.03786 [cs.CV]
  • Siarohin et al. (2020) Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. 2020. First Order Motion Model for Image Animation. arXiv:2003.00196 [cs.CV]
  • Sinha et al. (2022) Sanjana Sinha, Sandika Biswas, Ravindra Yadav, and Brojeshwar Bhowmick. 2022. Emotion-Controllable Generalized Talking Face Generation. arXiv:2205.01155 [cs.CV]
  • Tan et al. (2023) Shuai Tan, Bin Ji, and Ye Pan. 2023. EMMN: Emotional Motion Memory Network for Audio-driven Emotional Talking Face Generation. In 2023 IEEE/CVF International Conference on Computer Vision (ICCV). 22089–22099.
  • Tewari et al. (2020) Ayush Tewari, Mohamed Elgharib, Gaurav Bharaj, Florian Bernard, Hans-Peter Seidel, Patrick Pérez, Michael Zollhöfer, and Christian Theobalt. 2020. StyleRig: Rigging StyleGAN for 3D Control Over Portrait Images. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020. Computer Vision Foundation / IEEE, 6141–6150.
  • Wang et al. (2020) Kaisiyuan Wang, Qianyi Wu, Linsen Song, Zhuoqian Yang, Wayne Wu, Chen Qian, Ran He, Yu Qiao, and Chen Change Loy. 2020. MEAD: A Large-Scale Audio-Visual Dataset for Emotional Talking-Face Generation. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XXI (Lecture Notes in Computer Science, Vol. 12366), Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm (Eds.). Springer, 700–717.
  • Xia et al. (2023) Yibo Xia, Lizhen Wang, Xiang Deng, Xiaoyan Luo, and Yebin Liu. 2023. GMTalker: Gaussian Mixture based Emotional talking video Portraits. arXiv:2312.07669 [cs.CV]
  • Xu et al. (2023) Yuelang Xu, Lizhen Wang, Xiaochen Zhao, Hongwen Zhang, and Yebin Liu. 2023. AvatarMAV: Fast 3D Head Avatar Reconstruction Using Motion-Aware Neural Voxels. In ACM SIGGRAPH 2023 Conference Proceedings, SIGGRAPH 2023, Los Angeles, CA, USA, August 6-10, 2023, Erik Brunvand, Alla Sheffer, and Michael Wimmer (Eds.). ACM, 47:1–47:10.
  • Ye et al. (2023) Hu Ye, Jun Zhang, Sibo Liu, Xiao Han, and Wei Yang. 2023. IP-Adapter: Text Compatible Image Prompt Adapter for Text-to-Image Diffusion Models. arXiv:2308.06721 [cs.CV]
  • Zhang et al. (2023b) Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. 2023b. Adding Conditional Control to Text-to-Image Diffusion Models. arXiv:2302.05543 [cs.CV]
  • Zhang et al. (2023a) Wenxuan Zhang, Xiaodong Cun, Xuan Wang, Yong Zhang, Xi Shen, Yu Guo, Ying Shan, and Fei Wang. 2023a. SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation. arXiv:2211.12194 [cs.CV]
  • Zhao et al. (2021) Ruiqi Zhao, Tianyi Wu, and Guodong Guo. 2021. Sparse to Dense Motion Transfer for Face Image Animation. arXiv:2109.00471 [cs.CV]
  • Zheng et al. (2022) Yufeng Zheng, Victoria Fernández Abrevaya, Marcel C. Bühler, Xu Chen, Michael J. Black, and Otmar Hilliges. 2022. I M Avatar: Implicit Morphable Head Avatars from Videos. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022. IEEE, 13535–13545.
  • Zheng et al. (2023) Yufeng Zheng, Wang Yifan, Gordon Wetzstein, Michael J. Black, and Otmar Hilliges. 2023. PointAvatar: Deformable Point-based Head Avatars from Videos. arXiv:2212.08377 [cs.CV]
  • Zielonka et al. (2023) Wojciech Zielonka, Timo Bolkart, and Justus Thies. 2023. Instant Volumetric Head Avatars. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 4574–4584.