Next Article in Journal
Research on a Hybrid Intelligent Method for Natural Gas Energy Metering
Previous Article in Journal
Research on the Flow Parameters of Waste Motion in a Rotary Kiln with the Use of the Tracer Method
Previous Article in Special Issue
An Optimized Convolutional Neural Network for the 3D Point-Cloud Compression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TED-Face: Texture-Enhanced Deep Face Reconstruction in the Wild

1
Institute of Virtual Reality and Intelligent Systems, Hangzhou Normal University, Hangzhou 311121, China
2
Department of Computer and Information Sciences, Northumbria University, Newcastle-upon-Tyne NE1 8ST, UK
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(14), 6525; https://doi.org/10.3390/s23146525
Submission received: 13 June 2023 / Revised: 12 July 2023 / Accepted: 13 July 2023 / Published: 19 July 2023

Abstract

:
We present TED-Face, a new method for recovering high-fidelity 3D facial geometry and appearance with enhanced textures from single-view images. While vision-based face reconstruction has received intensive research in the past decades due to its broad applications, it remains a challenging problem because human eyes are particularly sensitive to numerically minute yet perceptually significant details. Previous methods that seek to minimize reconstruction errors within a low-dimensional face space can suffer from this issue and generate close yet low-fidelity approximations. The loss of high-frequency texture details is a key factor in their process, which we propose to address by learning to recover both dense radiance residuals and sparse facial texture features from a single image, in addition to the variables solved by previous work—shape, appearance, illumination, and camera. We integrate the estimation of all these factors in a single unified deep neural network and train it on several popular face reconstruction datasets. We also introduce two new metrics, visual fidelity (VIF) and structural similarity (SSIM), to compensate for the fact that reconstruction error is not a consistent perceptual metric of quality. On the popular FaceWarehouse facial reconstruction benchmark, our proposed system achieves a VIF score of 0.4802 and an SSIM score of 0.9622, improving over the state-of-the-art Deep3D method by 6.69% and 0.86%, respectively. On the widely used LS3D-300W dataset, we obtain a VIF score of 0.3922 and an SSIM score of 0.9079 for indoor images, and the scores for outdoor images are 0.4100 and 0.9160, respectively, which also represent an improvement over those of Deep3D. These results show that our method is able to recover visually more realistic facial appearance details compared with previous methods.

1. Introduction

Reconstructing 3D faces from single-view images has received years of research attention in computer vision and computer graphics due to its broad applications in facial image editing, virtual reality [1], and entertainment. However, it remains a challenging problem as estimating facial geometry and appearance from a single image is inherently an inverse problem that may not admit a unique solution. Nevertheless, human eyes are particularly good at comparing a pair of faces (original and reconstructed) and judging their visual similarity, thereby serving as an oracle quantifying the quality of facial recovery solutions. One subtle yet powerful cue that influences human judgment is facial skin texture [2], which determines the interaction of lights and geometry to produce visual perception of faces.
To address the ill-posed nature of image-based facial reconstruction, most of previous work represents both geometry and texture in low-dimensional spaces [3,4,5,6,7,8,9,10], in which the unique solution is found through minimizing the discrepancy between an input facial image and the rendering of an estimated face. This approach, albeit popular and effective in practice, can suffer from two main flaws. First, encoding spatially varying facial textures in a reduced space can incur information loss on high-frequency residuals around facial part boundaries and as a result lead to over-smoothed, plastic-like results. Those residuals, as shown in Section 3.4, are visually sparse and localized, which in theory cannot be captured using low-frequency basis functions of limited bandwidth (the uncertainty principle in Fourier analysis) [11]. Second, computer rendering of reconstructed faces as an essential step in deriving the reconstruction error to be minimized can bias the estimation because it acts as low-pass filters when a less sophisticated illumination model (e.g., with the spherical harmonic representation of lighting in existing work) is used for light transport computation [12]. This can produce faces lacking fine colors, too, as shown in Figure 1.
There are some algorithms that have been proposed to circumvent the limitation of low-dimensional facial texture models by using the UV space instead as a representation of facial textures [13,14,15]. However, estimating the dense UV map when a face is partially occluded is considerably unreliable, and learning the UV mapping of the whole head is not an easy task either [13,16,17]. Hence, the robustness issue of these methods can be a limiting factor when applied to unconstrained facial images in the wild.
Figure 1. Reconstruction error does not equal visual similarity. While the root mean squared error (RMSE) between the input image (left) and our reconstructed output (middle) is 1.1677, which is slightly higher than that between the input and the Deep3D [17] output (right) of 1.1377, our result resembles the original face more closely in terms of skin tone and texture.
Figure 1. Reconstruction error does not equal visual similarity. While the root mean squared error (RMSE) between the input image (left) and our reconstructed output (middle) is 1.1677, which is slightly higher than that between the input and the Deep3D [17] output (right) of 1.1377, our result resembles the original face more closely in terms of skin tone and texture.
Sensors 23 06525 g001
We propose two simple yet effective techniques—sparse texture enhancement and radiance residual learning—to address the two aforementioned difficulties with existing methods separately. For the first issue, it arises because the standard pixel-to-pixel reconstruction error between input image and rendered face averages the contribution of each pixel’s discrepancy with equal weight and therefore fails to prioritize facial areas that exhibit sharp variations in color intensity. Indeed, it is thoroughly demonstrated in [18] that using image gradients to capture such first-order texture information is critical for high-quality image editing. Motivated by this, we propose to extract facial image gradients using the discrete Prewitt operator, which enables us to re-weight each pixel with its gradient magnitude for computing a texture-enhanced reconstruction loss to be jointly minimized with the standard one. Benefiting from the natural sparsity of image gradient intensities, the recovery algorithm can be effectively guided to produce sufficient details in high-frequency texture areas overlooked by previous work, thereby improving the visual appearance of generated faces.
The source of the second difficulty differs from the first one in that it concerns properly shading an estimated face to be rendered and compared with the input. Due to the low-pass nature of simplified illumination models, increasing the expressiveness of lighting functions may not be able to fully resolve this issue as there will always be residuals of radiance across the geometry of a facial shape. Drawing inspirations from [19], we propose to learn these residuals by directly regressing dense radiance compensations from the input image, which are then added to the base shading results for final rendering. While estimating full dense radiance without going through an analytic shading process can lead to serious over-fitting, our technique combines the strength of base shading and residual learning to recover more detailed facial colors.
We integrate the two above techniques into a unified deep learning architecture (Figure 2), along with the inference of facial geometry (using the 3D morphable model), facial texture, lighting functions, and camera parameters, for joint end-to-end training. We train and evaluate our system on a variety of public face reconstruction datasets, including CelebA [20], 300W-LP [21], I-JBA [22], FFHQ [23], LFW [24], and LS3D [25]. In addition to the commonly reported reconstruction error metric, we also adopt the structural similarity index (SSIM) [26] and visual fidelity index (VIF) [27] to quantify the similarity of original and recovered faces in a way that is more consistent with human judgment. The results show that our method consistently outperforms the state of the art on all evaluated datasets.
The contributions of this paper include the following:
  • A new way to enhance the recovery of high-frequency facial textures by re-weighting facial reconstruction error using sparse image gradient intensities.
  • A technique to directly learn dense radiance residuals for compensating the low-pass filtering of facial rendering.
The structure of this paper is as follows. We survey the existing methods for human face reconstruction in Section 2, through which we point out the unique advantage of our method for texture enhancement. We then describe our system in Section 3 and present the experimental results in Section 4. We conclude the paper in Section 5.

2. Related Work

2.1. 3D Morphable Model of Faces

To solve the inverse problem of 3D face reconstruction from single images, some low-dimensional parametric face models such as the 3D morphable model (3DMM) [3] have been proposed to constrain the solution process, playing a vital role in the field of vision-based face recovery and tracking. Early 3DMM typically construct linear low-dimensional facial geometry and texture spaces using a reduced number of eigen-basis functions from principal component analysis (PCA) decomposition on facial scan data [28,29,30,31,32]. Subsequent methods improve the model by using more training data and adding expression basis functions to the model for encoding faces of more variations [33]. While such a reduced model can greatly improve the efficiency and stability of face estimation and potentially reduce over-fitting, it is not easy to directly synthesize highly detailed, fast-changing facial texture variations, which can be a limiting factor in producing faces that agree with human judgment of visual similarity.

2.2. Deep Learning for 3DMM

Deep neural networks, as a powerful learning tool, have achieved tremendous success for vision-based 3D face reconstruction. In [34], they propose a model with encoder and decoder based on deep neural networks to directly learn the mapping from 2D images to 3D faces. Some works also use neural networks to learn the parameters of a 3DMM model for face recovery, which combines the advantage of neural networks in parameter fitting and the 3DMM model in recovering stable geometric and textural representations [6,7]. The advancement of deep learning solutions, however, comes with the requirement of more training data [35,36,37,38,39]. Due to the high cost of acquiring 3D face scanning data, some methods that rely on weekly-supervised learning have also been proposed to reconstruct faces through synthesis by analysis [6,9,17,34]. They generally work by projecting an estimated face back to an image plane via a differentiable rendering process and then comparing the rendered face with the original one to provide a supervision signal for training. Such an approach can effectively improve data efficiency and enhance the robustness of deep neural networks [40,41,42]. During training, these methods seek to minimize the reconstruction error that typically treats each facial location with equal importance, which can lead to over-smoothed estimation when the underlying low-dimensional face model (e.g., the 3DMM) has a limited representational capacity. We address this challenge by using image gradient intensities to re-weight each reconstruction location proportionally so that the learning process can be induced to focus more on numerically small yet visually prominent facial areas.

2.3. Facial Texture Representation

As the low-dimensional linear basis functions of 3DMM models can limit the visual fidelity of generated faces, techniques such as the UV mapping has been adopted to improve the quality of 3D face reconstruction [33,34]. In [33], they define an additional UV map on top of the 3DMM model to describe facial textures. The work of [34] implements an inverse projection from 2D image to 3D texture by constructing an UV space mapping relationship. In [43], a face texture extraction method via matrix factorization is proposed. While the method of [44] can directly regress depth values for each face pixel, it cannot recover the radiance intensity of color surface and requires an offline alignment step. In recent years, the neural radiance field (NeRF) has received a lot of attention [45,46,47,48]. Although the reconstruction fidelity is amazing, NeRF relies on voxel representation, which may not generalize well for face reconstruction. Computer rendering with a simplified shading model can limit the visual perception of estimated faces also through implicit low-pass filtering [12]. Inspired by the work of [19], we propose to directly regress radiance residuals from input image to be added on top of computer-generated face rendering. Albeit simple, we show that this technique works well for enhancing facial colors in reconstruction.

3. TED-Face: Our Proposed Approach

3.1. Overview of System

Our proposed system (Figure 2) is an end-to-end deep neural network architecture that accepts an input facial image (of arbitrary identity, pose, and expression) and estimates its corresponding 3D face model that contains geometry and texture as two essential outputs. During the testing phase, the input image is first processed by a backbone convolutional neural network (CNN) to extract deep features for estimating the 3DMM face parameters (geometry, texture), head pose orientation, lighting functions, and radiance residuals. Motivated by similar lines of work, we use the popular standard ResNet-50 [19] as our backbone network, which has sufficient capacity to extract texture-rich image features that are highly informative for facial recovery. These features are then used to infer the 3DMM parameters that allow us to reconstruct a full 3D textured face using the pre-learned geometry and texture basis functions. The face is subsequently shaded using the lighting functions and enhanced by the radiance residuals to be eventually rendered on the image plane with the estimated head pose. As the shading and rendering processes are both differentiable, the whole architecture can be trained provided with suitable loss functions. For training, we design the following losses:
  • Standard Reconstruction Loss: measuring the pixel-to-pixel discrepancy between input and rendered faces with equal importance weight for each pixel.
  • Texture-Enhanced Reconstruction Loss: calculating the same error as above but with an adaptive importance weight applied to each individual pixel using the corresponding image gradient intensity.
  • Landmark Prediction Loss: measuring the distance from a set of detected 2D facial landmark points on the image and the corresponding set of estimated 3D key points after projection (using the head pose information).
  • Face Perception Loss: computing the cosine dissimilarity between deep CNN feature embeddings (pre-trained for face recognition) of input and reconstructed facial images.
The two contributions of our architecture are the learning of radiance residuals for compensating for the shading process (Section 3.3), as well as re-weighting the standard reconstruction error with image gradient intensities to recover texture details (Section 3.4). The face model we use is detailed in Section 3.2, and the neural network architecture is presented in Section 4.1.

3.2. 3D Morphable Face Model

Following existing work, we adopt the 3D morphable model of faces [3] to represent the geometry and texture of a 3D face using linear combinations of their corresponding basis functions (i.e., subspace) as follows:
S = S ¯ + B i d α + B e x p β
T = T ¯ + B t e x γ
where S ¯ and T ¯ are the mean face geometry and texture, respectively. The variations in geometry are encoded using a set of linear basis functions, B i d and B e x p , which are constructed from PCA decomposition on a 3D face scan dataset to explain the identity and expression factors separately. Similarly, the variations in texture are compactly represented using the principal components B t e x extracted from vertex colors of facial meshes on the same dataset. The unknown shape coefficients ( α , β ), which are to be inferred from an input facial image using our system (Figure 2), allow us to reconstruct a 3D face model using Equation (1). The texture parameters γ are estimated jointly with the geometry ones for facial color reconstruction using Equation (2). To instantiate the model, we use the popular Basel face space [33], which includes the pre-trained S ¯ , T ¯ , B i d , and B t , with the latter two having 80 components (i.e., 80 dimensions). We estimate the facial expression subspace B e x p from the FaceWarehouse dataset [49]. As a result, each 3D face without the ears and the neck part has 36,000 vertices.

3.3. Radiance Residual Learning

Once a 3D textured face is constructed using the technique in Section 3.2, the next step of our system is shading the face and then projecting it down to the image plane for comparison with the input. One of the contributions we make here is learning radiance residuals directly from the input image to compensate for the lack of fine details of shading.
Face Illumination. Following previous work, we assume that each face geometry is a Brownian surface (i.e., the reflected light is the same in all directions and does not absorb any incident light) and estimate face illumination using a set of spherical harmonic functions [50,51]. The radiance intensity at each vertex s i is calculated directly using its surface normal vector n i and the corresponding skin texture t i according to the following:
C ( i | σ ) = t i · b = 1 B 2 σ b Φ b ( n i )
Φ b : spherical harmonic basis function
σ b : spherical harmonic basis coefficient
We choose B for three bands of frequency [9,17] to approximate lighting, resulting in nine spherical harmonic coefficients σ to be estimated from the input image for face shading.
Radiance Compensation. As indicated in Equation (3), the shading process is essentially a low-pass filter that only relies on several bands of low-frequency spherical harmonic functions while discarding the high-frequency components. This may not suffice for high-fidelity face reconstruction since fine facial texture colors cannot be adequately captured.
We tackle this challenge using a deep residual learning [19] technique as follows:
L i = C ( i | σ ) + φ i radiance compensation
in which φ i is the radiance residual we directly regress from the input image for each facial vertex i. We sum the base shading result C ( i | σ ) and the residual component φ i that cannot be reconstructed through shading alone to form the full estimation of face illumination. As shown in Figure 3, this technique can enhance the visual appearance of reconstructed faces in certain facial areas.
Rendering. Once shading is computed in (3) and compensated in (6), we use a camera model of perspective projection with an empirically chosen focal length for rendering an inferred 3D face on the image plane for comparison with the input facial image. The head pose parameters including a rotation matrix R S O ( 3 ) and a translation vector t R 3 are estimated from the input image [17], which needs to be applied on the 3D face before rendering. The projection process is differentiable and therefore permits continuous optimization of our system using gradient descent.

3.4. Sparse Texture Enhancement

As previewed in Section 3.1, our face reconstruction system requires several loss functions for training, among which the texture-enhanced reconstruction error is another contribution of this paper to enrich texture details.
Standard Reconstruction Loss. As our approach works in a weakly supervised synthesis-by-analysis manner, we need the standard reconstruction loss to measure how close an input facial image to the estimated face after shading and rendering for providing supervision signals [3,17]. To eliminate the effect of facial whiskers and hairs on face reconstruction, we use a pre-trained Bayesian classifier [17] to predict a face mask M, in which each element M i = 1 , i f P i > 0.5 P i , o t h e r w i s e indicates the likelihood of the corresponding image pixel i belonging to facial skin, with P i being the raw probability output of the classifier. Now, we can calculate the pixel-to-pixel distance between an input face I and its rendered version I as follows [3,6,9,52]:
L standard_reconstruction = i M M i · I i I i 2 i M M i
While the standard loss (7) widely adopted in existing work works well for bringing the estimated facial colors closer to the original ones on average, it does not differentiate between visually salient and non-salient areas, which can lead to quality loss of important facial texture details.
Texture-Enhanced Reconstruction Loss. We propose to address the above-mentioned difficulty with the standard reconstruction loss by introducing a texture-sensitive importance weight for each pixel to more effectively guide the recovery process towards detail synthesis. Motivated by the success of gradient information for high-quality image editing [18], we adopt the simple yet effective Prewitt operator to compute image gradients and use their intensities to re-weight the reconstruction loss as follows:
L enhanced_reconstruction = i M M i · I i I i 2 · i i M M i texture enhancement using gradient
where i is the 2D gradient vector (in both horizontal and vertical directions) at pixel i as derived from convolving the 3 × 3 Prewitt kernel with the input image I. It can be seen in Figure 4 that these importance weights clearly highlight visually prominent facial regions that undergo sharp texture variations, such as the eyes, wrinkles, nasolabial folds, and mouth. These areas contain relatively high-frequency contents, and we use those weights to prioritize their recovery, which can produce faces of higher fidelity.
Landmark Prediction Loss. To improve the geometric alignment of input and reconstructed faces, we also introduce the landmark prediction loss function that penalizes the deviation of projected 3D key facial points from their ground-truth counterparts on the input image. We use a state-of-the-art facial landmark detection method to localize 68 key points on a face image [6,9,17]. We then track the indices of those points on the 3D face model to obtain their positions on the 2D image after projection for loss calculation:
L landmark_prediction = 1 N n = 1 N ω n v n v n 2
where N = 68 is the number of landmark points, v i corresponds to the i-th point on the input image, and v i is the 2D projection of the corresponding 3D key point on the face model. We set the weight ω i to 20 for points at visually sensitive facial areas, such as the mouth and nose, while weighting the remaining points with 1.
Face Perception Loss. In addition to the above losses that work in the raw image domain, we also incorporate a perceptual loss that can enforce the specificity of generated faces by pulling their feature representations closer to that of the input in a semantically meaningful space. To this end, we employ a pre-trained face recognition network FaceNet [53] to extract deep identity-relevant features for input and rendered faces, giving rise to the following loss:
L face_perception = 1 f ( I ) , f ( I ) f ( I ) · f ( I )
where f ( ) denotes the feature extraction function of FaceNet and , computes the inner product of two feature vectors. Since the features were originally learned for solving identity recognition, the perception loss (10) can enhance the visual similarity of original and recovered faces.
Total Loss. Summing up all loss functions together, we arrive at the following total loss that includes an additional parameter regularization term to prevent degradation of facial geometry and texture recovery:
L total = ω s L standard_reconstruction + ω e L enhanced_reconstruction + ω l L landmark_reconstruction + ω f L face_perception + ω α α 2 + ω β β 2 + ω γ γ 2 parameter regularization term
where { ω s , ω e , ω l , ω f , ω α , ω β , ω γ } are the weights for each loss term that need to be chosen.

4. Experimental Results

4.1. Implementation Details

The backbone convolutional neural network of our system is the standard ResNet-50 network [19], which we modify to allow its final fully connected layer to produce the following predictions from an input facial image I (cropped and resized to 224 × 224 in our experiments):
  • α : 80 identity coefficients of facial geometry in (1);
  • β : 80 expression coefficients of facial geometry in (1);
  • γ : 80 coefficients of facial texture in (3);
  • σ : 9 spherical harmonic coefficients in (3);
  • R: 9 parameters in a 3 × 3 rotation matrix in Section 3.3;
  • t: 3 translation amounts in Section 3.3;
  • φ : 3 × 36 k radiance with one for each vertex in (6).
We train our system using the Adam optimizer [54] with the default learning rates. Our network is initialized by the model pre-trained on the ImageNet dataset [55]. The weight for each loss term in (11) is set as ω s = 1.92 , ω e = 1.92 , ω l = 1.6 e 3 , ω f = 0.2 , ω α = 1.0 , ω β = 0.8 , and ω γ = 10 . The epochs and batch size are set to 20 and 16, respectively. We find that these hyper-parameters work well in practice. Our training and testing are performed on a PC with an i9 9900k CPU, 32GB RAM, and an 2080ti graphics card. It is noted that in [56] the authors propose the use of intermediate deep features for object representation, which can be cheaper to compute and therefore enable faster inferences. This inspires us to explore in future work a similar idea to speed up facial reconstruction, especially for real-time applications.

4.2. Datasets and Evaluation Metrics

We evaluate our proposed face reconstruction method on six widely used facial image datasets, namely CelebA [20], 300W-LP [21], I-JBA [22], FFHQ [23], LFW [24], and LS3D [25]. These datasets contain a large quantity of facial images of a wide-range of identifiers, expressions, and poses, which are sufficiently diverse and challenging to benchmark different methods. Following previous work, we balance the pose and race distributions to randomly select 320K face images for training and 6K images for testing.
Evaluation Metrics. For the evaluation of facial geometry reconstruction, we use the root mean squared error (RMSE) between two rigidly aligned faces (one is input and the other is estimation) to measure the geometric reconstruction error. For the comparison of visual fidelity, we utilize the structural similarity index (SSIM) [26] and visual fidelity index (VIF) [27], which are both popular metrics for image quality assessment that take human perception into account, in contrast to pixel-level reconstruction errors that are weakly correlated with human judgment of visual resemblance.
In the following, we first compare our method with several state-of-the-art methods on recovering facial color appearance (Section 4.3) and then present evaluation results on the quality of estimated 3D facial geometry (Section 4.4). Afterwards, we show ablation study results to validate the effectiveness of our two key design choices (Section 4.5): sparse texture enhancement and radiance residual learning.

4.3. Evaluation of Facial Appearance Reconstruction

Controlled Scenarios. Here, we compare our method with several state-of-the-art methods for face reconstruction including Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59], on the FaceWarehouse dataset [49] that was collected under well-controlled settings. In Table 1, we use the 300 images from the FaceWarehouse dataset [49] as the validation set and calculate the SSIM and VIF scores between the original and reconstructed facial images. The results show that our method improves upon the competitors on both metrics. Particularly, the VIF index is improved by over 6% compared with the score obtained by Deep3D, which is the second highest in the table.
Low-Resolution Images. Figure 5 shows the comparison with state-of-the-art methods on the low-resolution AFLW-2000 facial image dataset [21]. Due to the lack of high-frequency image information, these methods struggle with estimating accurate facial texture and illumination details, leading to unsatisfactory reconstruction results in some highlighted areas. In contrast, our method alleviates the challenge of low-resolution inputs by guiding texture reconstruction using image gradients and enhancing lighting with estimated radiance residuals. This gives rise to visually clearer and higher-quality recovered faces.
High-Resolution Images. Similarly, we perform evaluation and comparison on high-resolution facial images. Figure 6 shows our validation results on the high-resolution FFHQ (1024 × 1024) dataset. In contrast to the AFLW-2000 facial image dataset [21], the FFHQ dataset consists of facial images that contain more accurate appearance and scene illumination cues. Benefiting from the texture enhancing features of our method, it can be seen that our reconstruction around facial skin folds matches that of the input images much more closely in comparison with the results generated by the other methods. This, together with the low-resolution evaluation, shows that our method is able to process both poor and high-quality input images, which can be desirable for applications that involve both low and high-end face capturing devices.
Indoor-Outdoor Scenes. We also conducted experiments on the LS3D-300W face dataset [25], which contains facial images captured in both indoor and outdoor environments. Noted that all the compared methods in our study have been replicated, and there is no case of copying the results of other papers. The results in Table 2 were obtained using the LS3D-300W dataset as the test set, which contains many low-resolution facial images captured indoors and outdoors. If the test set is input into the Detailed 3D method, due to the low quality of the images, there are many cases where it cannot generate a facial mesh or obtain corresponding facial textures, so Detailed 3D is not included in Table 2 as a comparison method. As shown in Table 2, our method makes considerable improvements upon the state-of-the-art methods on both the SSIM and VIF evaluation metrics for indoor and outdoor scenes. We show the visual comparison of reconstructed faces in Figure 7, along with the comparison of image color histograms focused on the facial areas. Despite the variety of the scene lighting conditions, it can be seen that our method produces more visually accurate facial texture representations compared with the other methods. By inspecting the color histograms, it is also validated that our estimated facial color distributions are numerically closer to that of the input images.

4.4. Evaluation of Facial Geometry Reconstruction

Our method focuses on enhancing the final visual presentation of reconstructed faces (as measured by the SSIM and VIF metrics) while ensuring that the quality of generated 3D face geometric models is on par with that of the current state-of-the-art methods in terms of geometric reconstruction error. To this end, we further conducted experiments on two face scan datasets: the MICC Florence 3D face dataset [60] and the FaceWarehouse dataset [61], both of which contain ground-truth facial scan data. The MICC dataset contains 53 subjects and 3 video sequences captured in well-controlled, indoor and outdoor scenes. For FaceWarehouse, we use nine subjects with 20 expressions for evaluation.
In Table 3, we present the reconstruction error results for the 3D face models generated using our and the state-of-the-art methods, among which our baseline is the Deep3D method [17]. In recent years, various methods based on neural radiance fields have been proposed for face reconstruction. However, these methods are sensitive to data and have limitations in the scope of application scenarios, making them unable to serve as a general solution for parameterized face modeling. Parametric models are still the main solution for general face modeling. This paper mainly aims to address the issue of insufficient expressive power of parametric face models. Therefore, the methods we used for comparison are also based on the 3DMM parametric face model. Currently, the state-of-the-art method for 3D face reconstruction based on 3DMM is still Deep3D. Most researchers in the field of 3DMM-based research compare their methods with Deep3D. Other parametric face model methods are different from the 3DMM approach in terms of the construction of the face parameter structure. We only compare methods within the same class to demonstrate the performance improvement of the proposed method. It can be seen that the accuracy of facial geometries estimated by our method is similar to that generated by Deep3D while outperforming MGCNet [58], Face Rec [59], and Detailed 3D [57] to a large extent. Figure 8 visualizes the geometric errors of face reconstruction results generated by each method using heat map, with the error metric displayed below each rendered image. We omit the results of Detailed 3D to avoid color map distortion because they are not comparable with the others due to their much larger errors. It can be seen that our method maintains a similar quality of facial geometry estimation as that of Deep3D. In the meantime, the enhancement of visual appearance of reconstructed faces brought by our method is visually clear and numerically large, as experimentally validated in Section 4.3.

4.5. Evaluation of Design Choices

To validate the effectiveness of our design choices, we also perform ablation studies on the FaceWarehouse dataset [49] using 150 different individuals, each with 20 varying expressions. Table 4 shows the effects of adding each module design. The results indicate that our image-gradient-guided texture enhancement method significantly improves the final visual fidelity of reconstructed faces, while the radiance compensation technique further enhances the visual representation of the results. Figure 9 presents several visual and quantitative comparisons of faces generated using our system with different design choices. It can be see that our full system with both texture enhancement and radiance compensation produces the highest-quality face recovery results, compared with the baseline versions.

5. Conclusions and Future Work

We have proposed a new method, TED-Face, for reconstructing 3D facial geometry and texture from singe view images using deep learning. Our method works in an analysis-by-synthesis manner similar to previous work on vision-based 3D face reconstruction while improving over them on two important technical aspects to produce faces of higher visual fidelity. Accordingly, we have made two main contributions in this paper:
  • We have proposed a new, simple yet effective technique, sparse texture enhancement, that can guide the inverse estimation process to learn to recover high-frequency facial texture details by re-weighting the standard per-pixel reconstruction loss with image gradient intensities extracted using the Prewitt operator. This was inspired by the Poisson image editing approach of [18], and it was shown in Table 4 to enhance the visual quality of estimated faces as measured by the SSIM and VIF metrics.
  • We have also proposed a new technique named radiance residual learning, which can regress dense radiance residuals from input images to fill in facial appearance details that cannot be reproduced by a reduced texture space and shading model. The technique is straightforward to implement on top of facial reconstruction methods (including ours) that use the 3DMM model for surface estimation and shading. The results in Table 4 show that it further improves the visual quality of recovered faces.
We have evaluated our proposed system on two popular facial recovery datasets: FaceWarehouse [49] and LS3D-300W [25]. The results show that on the former, our system achieves a VIF score of 0.4802 and an SSIM score of 0.9622, improving over the state-of-the-art Deep3D method [17] by 6.69% and 0.86%, respectively. On the latter, our system obtains a VIF score of 0.3922 and an SSIM score of 0.9079 for indoor images, and the scores for outdoor images are 0.4100 and 0.9160 respectively, which also represent an improvement over that of Deep3D.
Due to the use of the low-dimensional 3DMM model for facial representation, our method has certain limitations that may prevent it from recovering accurate faces in challenging scenarios. First, the 3DMM model assumes a linear subspace of facial geometry, which greatly alleviates the difficulty of the inverse estimation problem but can fail to capture fine facial geometric details (e.g., wrinkles due to aging). As shown in the second column of Figure 8, the geometric reconstruction errors around the nose and mouth regions are generally higher than that in other flat areas. Second, our method uses spherical harmonic functions to approximate the facial shading process (Equation (3)). This process is differentiable and therefore enables gradient-based optimization. However, it can introduce approximation errors when the ground-truth lighting condition is overly complex and cannot be captured using a limited set of basis functions. Also, the shading is only calculated based on the surface normals and textures of a face as in the 3DMM model, which means that it cannot reproduce any sub-surface scattering phenomenon that can often be found when thin skins are lit by strong directional lights.
To address the aforementioned shortcoming of the 3DMM surface shading component, as an important line of future work, we plan to replace the currently used facial texture model with a physically more realistic one that can account for more intricate illumination effects such as subsurface scattering. The new model may contain a layered structure that represents the heterogeneous properties (e.g., albedo, scattering coefficient, phase coefficient) of skin, muscle, and fat, respectively, each of which would account for different illumination effects. This would allow us to reproduce facial appearance with richer details as human skin is not entirely opaque as implied by the 3DMM representation, especially when the skin is thin and the lighting is strongly directional. Differentiable volume rendering coupled with Monte Carlo sampling techniques for sub-surface lighting calculations can be a promising solution to this goal.
Due to the prevalent applications of facial reconstruction in virtual reality and mobile platforms, it is important that the reconstruction process can be run in real-time to account for dynamic contents. Our system is deep-learning-based and can be run quickly on powerful graphics cards. However, it is still heavy for commodity platforms, and, as a result, we intend to conduct further research to reduce the computational overhead via network pruning and quantization. While we have conducted experiments on several popular datasets in this work, the generalization ability of our method compared to other datasets remains to be seen, especially for human faces in the wild under challenging lighting conditions. Application to video face datasets is also worth consideration given their dynamic nature. Another line of future research is exploring data augmentation techniques to improve the generalization performance of trained models. One example could be using neural radiance field methods to relight facial images to create more diverse training data, which we plan to investigate.

Author Contributions

Conceptual design and methodology: Y.H.; experimental verification and analysis: L.F. and S.H.; writing—draft, writing—review and editing: S.H. and Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant Nos. 92146001 and 61673151), the Major Project of The National Social Science Fund of China (Grant No. 19ZDA324), the Scientific Research Foundation for Scholars of HZNU (2020QDL005), and the Fundamental Research Funds for the Central Universities.

Institutional Review Board Statement

This article does not involve ethical research and does not require ethical approval.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Guo, Y.; Yang, Z.; Zhang, J. Prior-Guided Multi-View 3D Head Reconstruction. IEEE Trans. Multimed. 2022, 24, 4028–4040. [Google Scholar] [CrossRef]
  2. Meinhardt-Injac, B.; Persike, M.; Meinhardt, G. Holistic face processing is induced by shape and texture. Perception 2013, 42, 716–732. [Google Scholar] [CrossRef] [PubMed]
  3. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3D faces. In Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 187–194. [Google Scholar]
  4. Gecer, B.; Lattas, A.; Ploumpis, S.; Deng, J.; Papaioannou, A.; Moschoglou, S.; Zafeiriou, S. Synthesizing coupled 3d face modalities by trunk-branch generative adversarial networks. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 415–433. [Google Scholar]
  5. Gecer, B.; Ploumpis, S.; Kotsia, I.; Zafeiriou, S. Ganfit: Generative adversarial network fitting for high fidelity 3d face reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1155–1164. [Google Scholar]
  6. Tewari, A.; Zollhöfer, M.; Garrido, P.; Bernard, F.; Kim, H.; Pérez, P.; Theobalt, C. Self-supervised multi-level face model learning for monocular reconstruction at over 250 hz. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2549–2559. [Google Scholar]
  7. Tran, L.; Liu, X. On learning 3d face morphable model from in-the-wild images. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 157–171. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Zhu, W.; Wu, H.; Chen, Z.; Vesdapunt, N.; Wang, B. Reda: Reinforced differentiable attribute for 3d face reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4958–4967. [Google Scholar]
  9. Tewari, A.; Zollhofer, M.; Kim, H.; Garrido, P.; Bernard, F.; Perez, P.; Theobalt, C. Mofa: Model-based deep convolutional face autoencoder for unsupervised monocular reconstruction. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 1274–1283. [Google Scholar]
  10. Wang, S.; Wang, S.; Yang, W.; Zhang, X.; Wang, S.; Ma, S.; Gao, W. Towards Analysis-Friendly Face Representation With Scalable Feature and Texture Compression. IEEE Trans. Multimed. 2022, 24, 3169–3181. [Google Scholar] [CrossRef]
  11. Havin, V.; Jöricke, B. The Uncertainty Principle in Harmonic Analysis; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 28. [Google Scholar]
  12. Soler, C.; Molazem, R.; Subr, K. A Theoretical Analysis of Compactness of the Light Transport Operator. In Proceedings of the ACM SIGGRAPH 2022 Conference Proceedings, Vancouver, BC, Canada, 7–11 August; Association for Computing Machinery: New York, NY, USA, 2022. [Google Scholar]
  13. Deng, J.; Cheng, S.; Xue, N.; Zhou, Y.; Zafeiriou, S. Uv-gan: Adversarial facial uv map completion for pose-invariant face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7093–7102. [Google Scholar]
  14. Lin, J.; Yuan, Y.; Shao, T.; Zhou, K. Towards high-fidelity 3D face reconstruction from in-the-wild images using graph convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5891–5900. [Google Scholar]
  15. Kim, J.; Yang, J.; Tong, X. Learning High-Fidelity Face Texture Completion without Complete Face Texture. In Proceedings of the IEEE/CVF International Conference on Computer, Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 13990–13999. [Google Scholar]
  16. Lattas, A.; Moschoglou, S.; Gecer, B.; Ploumpis, S.; Triantafyllou, V.; Ghosh, A.; Zafeiriou, S. AvatarMe: Realistically Renderable 3D Facial Reconstruction “in-the-wild”. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 760–769. [Google Scholar]
  17. Deng, Y.; Yang, J.; Xu, S.; Chen, D.; Jia, Y.; Tong, X. Accurate 3d face reconstruction with weakly-supervised learning: From single image to image set. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  18. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. In Proceedings of the SIGGRAPH 2003, Special Interest Group on Computer Graphics and Interactive Techniques, San Diego, CA, USA, 27–31 July 2003; pp. 313–318. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. Liu, Z.; Luo, P.; Wang, X.; Tang, X. Deep learning face attributes in the wild. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3730–3738. [Google Scholar]
  21. Zhu, X.; Lei, Z.; Liu, X.; Shi, H.; Li, S.Z. Face alignment across large poses: A 3d solution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 146–155. [Google Scholar]
  22. Klare, B.F.; Klein, B.; Taborsky, E.; Blanton, A.; Cheney, J.; Allen, K.; Grother, P.; Mah, A.; Jain, A.K. Pushing the frontiers of unconstrained face detection and recognition: Iarpa janus benchmark a. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1931–1939. [Google Scholar]
  23. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4401–4410. [Google Scholar]
  24. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Workshop on Faces in’Real-Life’Images: Detection, Alignment, and Recognition; Springer: Marseille, France, 2008. [Google Scholar]
  25. Bulat, A.; Tzimiropoulos, G. How far are we from solving the 2d & 3d face alignment problem? (and a dataset of 230,000 3d facial landmarks). In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1021–1030. [Google Scholar]
  26. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef] [PubMed]
  28. Moccozet, L.; Thalmann, N.M. Dirichlet Free-Form Deformations and Their Application to Hand Simulation; IEEE: Piscataway, NJ, USA, 1997. [Google Scholar]
  29. DeCarlo, D.; Metaxas, D.; Stone, M. An anthropometric face model using variational techniques. In Proceedings of the 25th Annual Conference on Computer Graphics and Interactive Techniques, Orlando, FL, USA, 19–24 July 1998; pp. 67–74. [Google Scholar]
  30. Beymer, D.; Poggio, T. Image representations for visual learning. Science 1996, 272, 1905–1909. [Google Scholar] [CrossRef] [PubMed]
  31. Choi, C.S.; Okazaki, T.; Harashima, H.; Takebe, T. A system of analyzing and synthesizing facial images. In Proceedings of the 1991 IEEE International Symposium on Circuits and Systems (ISCAS), Singapore, 11–14 June 1991; pp. 2665–2668. [Google Scholar]
  32. Lanitis, A.; Taylor, C.J.; Cootes, T.F. Automatic interpretation and coding of face images using flexible models. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 743–756. [Google Scholar] [CrossRef]
  33. Paysan, P.; Knothe, R.; Amberg, B.; Romdhani, S.; Vetter, T. A 3D face model for pose and illumination invariant face recognition. In Proceedings of the 2009 Sixth IEEE International Conference on Advanced Video and Signal Based Surveillance, Genova, Italy, 2–4 September 2009; pp. 296–301. [Google Scholar]
  34. Tran, L.; Liu, X. Nonlinear 3d face morphable model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7346–7355. [Google Scholar]
  35. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  36. Huang, Y.; Hu, S.; Zhang, Z. Structured Spatial Reasoning for Human Pose Estimation. In Proceedings of the 33rd British Machine Vision Conference, London, UK, 21–24 November 2022; p. 797. [Google Scholar]
  37. Huang, Y.; Shum, H.; Ho, E.; Aslam, N. High-speed multi-person pose estimation with deep feature transfer. Computer Vision and Image Understanding. 2020, 197–198, 103010. [Google Scholar] [CrossRef]
  38. Huang, Y.; Zhuang, J.; Qin, Z. Multi-Level Network for High-Speed Multi-Person Pose Estimation. In Proceedings of the 2019 IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; pp. 589–593. [Google Scholar]
  39. Huang, Y.; Sun, B.; Kan, H.; Zhuang, J.; Qin, Z. FollowMeUp Sports: New Benchmark for 2D Human Keypoint Recognition. In Proceedings of the Pattern Recognition and Computer Vision—Second Chinese Conference, Xi’an, China, 8–11 November 2019; pp. 110–121. [Google Scholar]
  40. Zheng, M.; Wang, F.; You, S.; Qian, C.; Zhang, C.; Wang, X.; Xu, C. Weakly supervised contrastive learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10042–10051. [Google Scholar]
  41. Joulin, A.; Maaten, L.v.d.; Jabri, A.; Vasilache, N. Learning visual features from large weakly supervised data. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 67–84. [Google Scholar]
  42. Saito, S.; Yang, J.; Ma, Q.; Black, M.J. SCANimate: Weakly supervised learning of skinned clothed avatar networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2886–2897. [Google Scholar]
  43. Moschoglou, S.; Ververas, E.; Panagakis, Y.; Nicolaou, M.A.; Zafeiriou, S. Multi-attribute robust component analysis for facial uv maps. IEEE J. Sel. Top. Signal Process. 2018, 12, 1324–1337. [Google Scholar] [CrossRef] [Green Version]
  44. Blanz, V.; Basso, C.; Poggio, T.; Vetter, T. Reanimating faces in images and video. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2003; Volume 22, pp. 641–650. [Google Scholar]
  45. Hong, Y.; Peng, B.; Xiao, H.; Liu, L.; Zhang, J. Headnerf: A real-time nerf-based parametric head model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 20374–20384. [Google Scholar]
  46. Bühler, M.C.; Meka, A.; Li, G.; Beeler, T.; Hilliges, O. VariTex: Variational Neural Face Textures. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 13890–13899. [Google Scholar]
  47. Deng, Y.; Yang, J.; Chen, D.; Wen, F.; Tong, X. Disentangled and controllable face image generation via 3d imitative-contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5154–5163. [Google Scholar]
  48. Ghosh, P.; Gupta, P.S.; Uziel, R.; Ranjan, A.; Black, M.J.; Bolkart, T. GIF: Generative interpretable faces. In Proceedings of the 2020 International Conference on 3D Vision (3DV), Fukuoka, Japan, 25–28 November 2020; pp. 868–878. [Google Scholar]
  49. Cao, C.; Weng, Y.; Zhou, S.; Tong, Y.; Zhou, K. Facewarehouse: A 3d facial expression database for visual computing. IEEE Trans. Vis. Comput. Graph. 2013, 20, 413–425. [Google Scholar]
  50. Ramamoorthi, R.; Hanrahan, P. An efficient representation for irradiance environment maps. In Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 12–17 August 2001; pp. 497–500. [Google Scholar]
  51. Ramamoorthi, R.; Hanrahan, P. A signal-processing framework for inverse rendering. In Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 12–17 August 2001; pp. 117–128. [Google Scholar]
  52. Thies, J.; Zollhofer, M.; Stamminger, M.; Theobalt, C.; Nießner, M. Face2face: Real-time face capture and reenactment of rgb videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2387–2395. [Google Scholar]
  53. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  54. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
  55. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  56. Bayraktar, E.; Wang, Y.; DelBue, A. Fast re-OBJ: Real-time object re-identification in rigid scenes. Mach. Vis. Appl. 2022, 33, 97. [Google Scholar] [CrossRef]
  57. Chen, Y.; Wu, F.; Wang, Z.; Song, Y.; Ling, Y.; Bao, L. Self-supervised learning of detailed 3d face reconstruction. IEEE Trans. Image Process. 2020, 29, 8696–8705. [Google Scholar] [CrossRef] [PubMed]
  58. Shang, J.; Shen, T.; Li, S.; Zhou, L.; Zhen, M.; Fang, T.; Quan, L. Self-supervised monocular 3d face reconstruction by occlusion-aware multi-view geometry consistency. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 53–70. [Google Scholar]
  59. Ju, Y.J.; Lee, G.H.; Hong, J.H.; Lee, S.W. Complete face recovery gan: Unsupervised joint face rotation and de-occlusion from a single-view image. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 3711–3721. [Google Scholar]
  60. Booth, J.; Antonakos, E.; Ploumpis, S.; Trigeorgis, G.; Panagakis, Y.; Zafeiriou, S. 3d face morphable models" in-the-wild". In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 48–57. [Google Scholar]
  61. Feng, Y.; Wu, F.; Shao, X.; Wang, Y.; Zhou, X. Joint 3d face reconstruction and dense alignment with position map regression network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 534–551. [Google Scholar]
Figure 2. Overview of our TED-Face reconstruction system (Section 3.1). Given an input facial image, a deep convolutional neural network is used to predict the various coefficients of 3D face (Section 3.2), including the parameters of the coarse-grained 3DMM model (geometry, texture), illumination functions, face pose, and dense radiance residuals (Section 3.3). Multiple loss functions including our proposed texture-enhanced reconstruction loss (Section 3.4) are jointly used to train the whole system end-to-end.
Figure 2. Overview of our TED-Face reconstruction system (Section 3.1). Given an input facial image, a deep convolutional neural network is used to predict the various coefficients of 3D face (Section 3.2), including the parameters of the coarse-grained 3DMM model (geometry, texture), illumination functions, face pose, and dense radiance residuals (Section 3.3). Multiple loss functions including our proposed texture-enhanced reconstruction loss (Section 3.4) are jointly used to train the whole system end-to-end.
Sensors 23 06525 g002
Figure 3. Radiance residuals enhance the visual fidelity of reconstructed faces. Estimating facial color appearance using a reduced-dimensional texture space and simplified shading can result in over-smoothed faces (middle) due to lack of fine details. Directly learning dense radiance residuals from an input image to compensate for the loss of details improves the appearance of the reconstructed faces (right).
Figure 3. Radiance residuals enhance the visual fidelity of reconstructed faces. Estimating facial color appearance using a reduced-dimensional texture space and simplified shading can result in over-smoothed faces (middle) due to lack of fine details. Directly learning dense radiance residuals from an input image to compensate for the loss of details improves the appearance of the reconstructed faces (right).
Sensors 23 06525 g003
Figure 4. Image gradient magnitudes indicate high-frequency salient textures. Applying the 3 × 3 Prewitt convolutional kernel on the input image (left) produces the gradient map (middle, visualized as intensities) that more clearly highlight visually prominent areas such as the eyes, wrinkles, and nasolabial folds, producing consistent log spectrum shapes (right, zooming in to see details) that contain a large proportion of high-frequency components. We use these gradient intensities as importance weights in our system to guide the recovery of facial texture details.
Figure 4. Image gradient magnitudes indicate high-frequency salient textures. Applying the 3 × 3 Prewitt convolutional kernel on the input image (left) produces the gradient map (middle, visualized as intensities) that more clearly highlight visually prominent areas such as the eyes, wrinkles, and nasolabial folds, producing consistent log spectrum shapes (right, zooming in to see details) that contain a large proportion of high-frequency components. We use these gradient intensities as importance weights in our system to guide the recovery of facial texture details.
Sensors 23 06525 g004
Figure 5. Comparison with state-of-the-arts on the low-resolution AFLW-2000 test set [21]. From left to right are the input faces, our reconstruction results, and the faces recovered by Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59]. Certain facial areas are zoomed in to highlight the improvement of our method for reconstructing fine facial texture details.
Figure 5. Comparison with state-of-the-arts on the low-resolution AFLW-2000 test set [21]. From left to right are the input faces, our reconstruction results, and the faces recovered by Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59]. Certain facial areas are zoomed in to highlight the improvement of our method for reconstructing fine facial texture details.
Sensors 23 06525 g005
Figure 6. Comparison with state-of-the-art methods on the high-resolution FFHQ dataset [23]. From left to right are the input faces, our reconstruction results, and the faces recovered by Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59]. Certain facial areas are zoomed in to highlight the improvement of our method for reconstructing fine facial texture details.
Figure 6. Comparison with state-of-the-art methods on the high-resolution FFHQ dataset [23]. From left to right are the input faces, our reconstruction results, and the faces recovered by Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59]. Certain facial areas are zoomed in to highlight the improvement of our method for reconstructing fine facial texture details.
Sensors 23 06525 g006
Figure 7. Comparison with state-of-the-art methods on the indoor-outdoor LS3D-300W validation dataset [25]. From left to right are the input faces, our reconstruction results, and the faces recovered by Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59]. The last column shows the comparison of color histograms of facial images produced by these methods (zoom in to see details).
Figure 7. Comparison with state-of-the-art methods on the indoor-outdoor LS3D-300W validation dataset [25]. From left to right are the input faces, our reconstruction results, and the faces recovered by Deep3D [17], Detailed 3D [57], MGCNet [58], and Face Rec [59]. The last column shows the comparison of color histograms of facial images produced by these methods (zoom in to see details).
Sensors 23 06525 g007
Figure 8. Visualization of facial geometry reconstruction errors. From left to right are the original facial scan, our reconstruction result, and the face recovered by Deep3D [17], MGCNet [58], and Face Rec [59]. The warmer the color visualization of a point, the higher the corresponding geometry reconstruction error. The score underneath each image is the root mean squared error between the reconstructed face and the original facial scan after rigid registration.
Figure 8. Visualization of facial geometry reconstruction errors. From left to right are the original facial scan, our reconstruction result, and the face recovered by Deep3D [17], MGCNet [58], and Face Rec [59]. The warmer the color visualization of a point, the higher the corresponding geometry reconstruction error. The score underneath each image is the root mean squared error between the reconstructed face and the original facial scan after rigid registration.
Sensors 23 06525 g008
Figure 9. Comparison of reconstructed faces using different method configurations of ours. The images from left to right within each group correspond to the input face, baseline result, result with sparse texture enhancement (STE, key design choice 1, Section 3.4), and result with both sparse texture enhancement and radiance residual learning (RRL, key design choice 2, Section 3.3). The VIF score for each reconstructed face is shown underneath (ref. Table 4).
Figure 9. Comparison of reconstructed faces using different method configurations of ours. The images from left to right within each group correspond to the input face, baseline result, result with sparse texture enhancement (STE, key design choice 1, Section 3.4), and result with both sparse texture enhancement and radiance residual learning (RRL, key design choice 2, Section 3.3). The VIF score for each reconstructed face is shown underneath (ref. Table 4).
Sensors 23 06525 g009
Table 1. Comparisons with state-of-the-art methods on the FaceWarehouse dataset [49]. Our method outperforms these methods on both VIF and SSIM metrics that measure visual fidelity of reconstructed facial images.
Table 1. Comparisons with state-of-the-art methods on the FaceWarehouse dataset [49]. Our method outperforms these methods on both VIF and SSIM metrics that measure visual fidelity of reconstructed facial images.
MethodVIF ↑SSIM ↑
MGCNet [58]0.41500.9359
Face Rec [59]0.44620.9544
Detailed 3D [57]0.44970.9561
Deep3D [17]0.45010.9540
TED-Face (ours)0.48020.9622
Table 2. Comparisons with state-of-the-art methods on the indoor-outdoor LS3D-300W dataset [25]. Our method outperforms these methods on both VIF and SSIM metrics that measure visual fidelity of reconstructed facial images.
Table 2. Comparisons with state-of-the-art methods on the indoor-outdoor LS3D-300W dataset [25]. Our method outperforms these methods on both VIF and SSIM metrics that measure visual fidelity of reconstructed facial images.
Method VIF SSIM
Indoor/OutdoorIndoor/Outdoor
MGCNet [58]0.3545/0.35810.8993/0.9060
Face Rec [59]0.3661/0.37990.8998/0.9120
Deep3D [17]0.3712/0.38710.8953/0.9052
TED-Face (ours)0.3922/0.41000.9079/0.9160
Table 3. Comparison of facial geometry reconstruction errors. We use the iterative closest point (ICP) algorithm to align original and reconstructed 3D faces for computing the mean and standard deviation of point-to-point distances (mm) between the two on the MICC Florence 3D face dataset [60] and the FaceWarehouse dataset [61].
Table 3. Comparison of facial geometry reconstruction errors. We use the iterative closest point (ICP) algorithm to align original and reconstructed 3D faces for computing the mean and standard deviation of point-to-point distances (mm) between the two on the MICC Florence 3D face dataset [60] and the FaceWarehouse dataset [61].
MethodFlorenceFaceWarehouse
Mean/Std ↓Mean/Std ↓
MGCNet [58]5.7375/8.10420.0313/0.332
Face Rec [59]5.0877/6.67750.0295/0.2999
Detailed 3D [57]2.4551/2.15020.0157/0.0126
Deep3D [17]1.7038/2.06800.0149/0.0123
TED-Face (ours)1.7065/2.06470.0140/0.0120
Table 4. Ablation study results of our method on the FaceWarehouse dataset [49]. The two key design choices we propose in this paper, namely sparse texture enhancement (STE) and radiance residual learning (RRL), improve the visual fidelity of reconstructed faces compared with the baseline that does not have those modules.
Table 4. Ablation study results of our method on the FaceWarehouse dataset [49]. The two key design choices we propose in this paper, namely sparse texture enhancement (STE) and radiance residual learning (RRL), improve the visual fidelity of reconstructed faces compared with the baseline that does not have those modules.
ConfigurationVIF ↑SSIM ↑
Baseline0.45990.9622
Baseline + STE 0.48280.9670
Baseline + STE + RRL (our full model)0.49030.9692
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Y.; Fang, L.; Hu, S. TED-Face: Texture-Enhanced Deep Face Reconstruction in the Wild. Sensors 2023, 23, 6525. https://doi.org/10.3390/s23146525

AMA Style

Huang Y, Fang L, Hu S. TED-Face: Texture-Enhanced Deep Face Reconstruction in the Wild. Sensors. 2023; 23(14):6525. https://doi.org/10.3390/s23146525

Chicago/Turabian Style

Huang, Ying, Lin Fang, and Shanfeng Hu. 2023. "TED-Face: Texture-Enhanced Deep Face Reconstruction in the Wild" Sensors 23, no. 14: 6525. https://doi.org/10.3390/s23146525

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop