Next Article in Journal
Structural Features of the Fragments from Cast Iron Cauldrons of the Medieval Golden Horde: Neutron Tomography Data
Previous Article in Journal
Mammography Datasets for Neural Networks—Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Face Aging by Explainable Conditional Adversarial Autoencoders

by
Christos Korgialas
,
Evangelia Pantraki
,
Angeliki Bolari
,
Martha Sotiroudi
and
Constantine Kotropoulos
*
Department of Informatics, Faculty of Sciences, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2023, 9(5), 96; https://doi.org/10.3390/jimaging9050096
Submission received: 30 March 2023 / Revised: 3 May 2023 / Accepted: 6 May 2023 / Published: 10 May 2023
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

:
This paper deals with Generative Adversarial Networks (GANs) applied to face aging. An explainable face aging framework is proposed that builds on a well-known face aging approach, namely the Conditional Adversarial Autoencoder (CAAE). The proposed framework, namely, xAI-CAAE, couples CAAE with explainable Artificial Intelligence (xAI) methods, such as Saliency maps or Shapley additive explanations, to provide corrective feedback from the discriminator to the generator. xAI-guided training aims to supplement this feedback with explanations that provide a “reason” for the discriminator’s decision. Moreover, Local Interpretable Model-agnostic Explanations (LIME) are leveraged to provide explanations for the face areas that most influence the decision of a pre-trained age classifier. To the best of our knowledge, xAI methods are utilized in the context of face aging for the first time. A thorough qualitative and quantitative evaluation demonstrates that the incorporation of the xAI systems contributed significantly to the generation of more realistic age-progressed and regressed images.

1. Introduction

Face aging attempts to synthesize a person’s future facial appearance as they age or their past appearance as they regress. As individuals age, their facial features gradually and cumulatively change, resulting in certain common patterns, such as fine lines around the eyes and mouth and changes in skin texture. Although the effects of aging on facial appearance may differ from person to person, these patterns can be learned and applied to produce accurate simulations of how a face may age or become rejuvenated. The topic of face aging has triggered the interest of the research community due to its diverse range of applications, such as age-invariant face recognition for security purposes [1,2,3], entertainment-related applications [4], and the cosmetics industry [5].
Generative Adversarial Networks (GANs) [6] have played a vital role in generating realistic synthetic images for various applications. In [7], GANs were applied to face aging, which involved generating face images across different age groups. In [8], GANs were used to create synthetic images of traffic signs. Although GANs are highly effective in generating new data, they typically require a large and balanced dataset for proper training. Methods for addressing the class imbalance in classification tasks through the employment of GANs are explored in [9,10].
This paper deals with GANs applied to facial images in the context of face aging. Figure 1 illustrates an example of face progression, where a young subject’s input face image is used to extract identity-related and age-related features fed into a face aging model. The model’s output is expected to preserve the identity-related features while incorporating aging patterns. In this paper, a state-of-the-art face aging GAN, namely, the Conditional Adversarial Autoencoder (CAAE) [7], is enriched with explanatory methods that provide insight into the discriminator’s decisions. The CAAE maps any face to a latent vector through a convolutional encoder. Then, the latent vector is projected to the face manifold conditional on age through a de-convolutional generator. More specifically, a facial image is faithfully reconstructed following the inversion process [11] by the generator. Simultaneous age progression and regression can be achieved by manipulating the age attribute.
The proposed framework integrates two explainable Artificial Intelligence (xAI) methods, namely, Saliency maps [12] and Shapley additive explanations (SHAP) [13], into the CAAE network. Saliency maps resort to the explanation matrix whose elements take values in the range [ 0 , 1 ] , referring to pixels. A value close to zero indicates that the pixel has no impact on the classification decision made by the discriminator. Similarly, a value close to one implies the pixel significantly contributes. Here, we are interested in a modified gradient descent to update the generator’s weights. That is, the gradient of the discriminator’s decision with respect to the generator’s output (i.e., generated image) is used to derive the explanation matrix by taking the absolute value element-wise and scaling it in the range [ 0 , 1 ] . The explanation matrix is multiplied by a weight and added to the generator’s gradients with respect to the loss. The origin of the SHAP method is traced back to cooperative game theory. The payoff is the discriminator’s output, while the pixels of the images generated by the generator are treated as actors. Assume that one pixel is hidden from the rest of the pixels. The Shapley value is defined as the average marginal contribution of the withheld pixel in the discriminator’s output, which is calculated by averaging over all the different subsets that emerge along the path to form the grand coalition of pixels (i.e., the fully generated image) from the empty coalition of pixels. Let us refer to the CAAE coupled with xAI methods as xAI-CAAE. We explore whether improved corrective feedback from discriminator to generator occurs with the xAI-CAAE, improving its performance against CAAE.
A third xAI technique called Local Interpretable Model-agnostic Explanations (LIME) [14] is employed to determine the significant facial features contributing to face aging. By perturbing a set of facial images and collecting age classification decisions, LIME trains a regression model that approximates the age classifier in the local area of the training image. Consequently, if the explanation is consistent locally, it can be used to identify the significance of features.
The proposed framework incorporates xAI explanations by Saliency or SHAP into its training to enhance age progression and regression performed by GANs on facial images. LIME posterior explanations are derived as well. To the best of our knowledge, explainable methods are utilized in face aging employing several benchmark datasets for the first time. The main contributions of the paper are as follows:
1.
Explainable gradient-based methods are added to CAAE: To add the reasoning of the discriminator’s decision into the generator’s training, the gradients of the generator are modified using either Saliency [12] or SHAP [13] explainable methods. The impact of xAI on the training process is explored by assigning different weights to the modified xAI gradients. The proposed xAI-CAAE is trained on a combination of images from the Cross-Age Celebrity Dataset (CACD) [15] and the UTKFace [7] datasets.
2.
Face aging assessment: Both qualitative and quantitative evaluations of the generated facial images are conducted. The qualitative evaluation resorts to a visual inspection of the GAN-generated facial images. The quantitative assessment is two-fold: (i) The Fréchet Inception Distance (FID) [16] is computed for images produced by CAAE and xAI-CAAE applied to the FGNET dataset [17] to assess their visual quality; (ii) To estimate the age range of the generated age progressed and regressed FGNET images, the age estimation algorithm described in [18] is used. A thorough performance evaluation demonstrates the strengths of the proposed xAI-CAAE.
3.
Interpretation of age classification results using explainable occlusion-based methods: To identify which facial areas are essential for age classification, experiments that utilize LIME’s posterior explanations are conducted on the Adience dataset [19].
The remainder of the paper is structured as follows. Related work and the proposed framework are surveyed in Section 2. Experimental evaluation is conducted in Section 3. Finally, Section 4 concludes the paper and recommends future work.

2. Materials and Methods

In Section 2.1, relevant research approaches on face aging are surveyed. Section 2.2 provides a brief overview of the xAI algorithms associated with the proposed method. Finally, the proposed xAI-CAAE is described in detail in Section 2.3.

2.1. Face Aging

Before the advent of deep learning, age progression and regression methods were primarily divided into physical model and prototype approaches. Physical model techniques [20,21] were focused on modeling the physical attributes of face aging, such as the alterations in hair, mouth, and skin texture, over time. These methods required a substantial amount of matched data and were time-consuming. Prototype-based approaches were focused on investigating the differences in facial images among different age classes to determine the aging patterns of each age class. Typically, these methods involved averaging the faces of individuals within the same age range to identify common features [22,23]. The face rejuvenation/progression process involved removing/adding textures that exhibited signs of aging, which was accomplished by applying a learned transformation across facial surfaces. Since this procedure produced smoothed facial images for each age class, maintaining identity information was challenging.
The success of GANs in image synthesis and translation tasks has led to remarkable progress in face aging techniques. The training process of GANs involves training two models simultaneously: the generator G, which learns to generate new samples that resemble the training samples and captures their distribution, and the discriminator D, which distinguishes the synthetically generated samples from the real training ones. Let the data distribution be p d a t a ( x ) , i.e., the training data x p d a t a ( x ) . If z is sampled from the distribution p z ( z ) , G and D engage in the min-max game [6]
{ G * , D * } = argmin G argmax D E x p d a t a ( x ) log D ( x ) + E z p z ( z ) log 1 D ( G ( z ) ) ,
where E [ · ] denotes the expectation operator. To circumvent the prohibitive cost of longitudinal collection of multiple face images for each subject, GAN-based methods resort to unpaired face aging data for training and primarily focus on modeling mappings between image contents. It is crucial to enforce identity consistency to prevent matching ambiguities when trying to simulate the aging process in an unpaired training scenario. This way, key semantic conditional information of the input, such as the unique facial features of each individual, are maintained. It should be noted that most GAN-based face aging algorithms do not enforce constraints in regions that are particularly relevant to age changes. Still, the generator re-estimates the pixel at each spatial location of the synthetical image.
In [24], a GAN-based framework for Attribute-Aware Attentive Face Aging (A3GAN) was proposed. By integrating facial attribute vectors into the generator and discriminator, semantic conditional information from the input was employed to train the model to create elderly face images with attributes faithful to the corresponding input. An attention mechanism that limited the alterations to age-related regions improved the visual quality of the synthesized face images. Since wrinkles, eye bags, and laugh lines are represented mainly by local textures, a wavelet packet transform extracted features at several scales in the frequency space, increasing aging details.
The effects of aging within a specific age class are related to the effects of aging in neighboring age classes, as aging is a gradual and continuous process. Moreover, aging transformations between distant age classes are likely more drastic and intense than those between nearby age classes. In [25], face aging was addressed as an unsupervised image-to-image translation problem. The Pyramid Face Aging-GAN (PFA-GAN) was suggested in particular, which contains a pyramid weight-sharing method. Face aging effects are therefore learned hierarchically, beginning with the subtle changes necessary between neighboring age classes and progressing to the more conspicuous and drastic changes required between distant age classes. No paired examples showing how the person looks at the target age class were needed, alleviating a severe limitation of many face-aging approaches.
In [26], a framework was developed to simulate aging in three dimensions. The framework consisted of three components—a 3D estimator for vertices and textures, a texture-aging GAN, and a module for rendering 2D and 3D faces. The 3D vertex and texture estimator determined the spatial vertices and textures of the face. The GAN applied aging effects to the estimated texture map. Finally, the rendering module produced 2D or 3D faces using the vertex map and the aged facial texture map.
A GAN network termed Age Gap Reducer-Generative Adversarial Network was introduced in [27] to reduce the age gap between face images using automatic age estimation. The network took into account both the gender of the individual in the input image and the desired age group to which the input face needed to be adjusted. This enabled the network to either regress the input image to a younger age group or progress it to an older age group, depending on the desired outcome.
Face aging depends on accurate age estimation. Age estimation is a challenging task because it is affected by gender, race, and various intrinsic or extrinsic attributes. An age estimation architecture was developed in [28], which included three convolutional neural networks (CNNs) and two extreme learning machine structures. There is a lack of large and reliably annotated datasets for training deep neural networks to estimate age. Knowledge distillation was exploited for accurate age estimation in [29] to address this problem. In a nutshell, class probability vectors were derived by a large model comprising multiple CNNs applied to a reference annotated dataset. The resulting predictions were then exploited as target labels to train a smaller model on a face dataset without age annotations.

2.2. Basic Elements of Explainable AI

The paper’s main contribution is to enrich CAAE, described in Section 2.3, with explainable artificial intelligence (xAI) techniques by adding an explanation system into its training procedure. In the following, the basic xAI techniques are briefly reviewed.
Pixel and feature attribution techniques attempt to explain individual predictions by crediting each input feature according to how much it alters the prediction. Pixel attribution techniques are known by various names, one of which is Saliency maps. Typical examples of feature attribution techniques include SHAP and LIME. Input pixels, tabular data, or text are used as features. There are two sorts of attribution techniques:
1.
Gradient-based: These techniques compute the gradient of the prediction (or classification score) concerning input features. The key distinction between various gradient-based techniques lies in the method they use to calculate the gradient.
2.
Occlusion-based: These techniques, such as SHAP and LIME, manipulate image regions to produce model-agnostic explanations.
Both techniques provide explanations in the context of a Saliency map that is the same size as the input image or at least projected onto the input image. Each pixel is given a value, which can be interpreted as its importance for the prediction or categorization task.

2.2.1. Saliency

Saliency maps [12] determine the significance of each feature in a given input for subsequent classification using a deep neural network inspired by how animals focus their attention. A nonlinear score function S c ( x ) is used to determine whether an image belongs to a particular class. This function is linearized using a first-order Taylor expansion around a specific image x 0
S c ( x ) S c ( x 0 ) + x S c ( x 0 ) x x 0 .
It is seen that the first term represents the classification score when the input is not perturbed, while the gradient term weighs the impact of perturbations. By reshaping the gradient to a two-dimensional matrix (i.e., an image), the Saliency map M is produced. The Saliency map is frequently normalized. Accordingly, M will refer to the normalized Saliency map hereafter. If more than one channels are present in the input image, the maximum Saliency map across all channels is considered. One can create a visual representation by taking the absolute values of the elements in the Saliency map or by distinguishing between the negative and positive contributions. The process of computing the Saliency map is not costly, as it only necessitates a single back-propagation step and does not assume the existence of any further annotations beyond the labels employed during the initial model training.

2.2.2. LIME

LIME [14] selectively activates or deactivates certain super-pixels in an image and then examines how these perturbations impact the predictions made by a classifier. To achieve this, LIME creates a synthetic neighborhood, denoted by N ( x ) , around the input instance to be explained, denoted by x R d , i.e.,
N ( x ) = x j | x j = x + p j , p j N 0 , Σ j = 1 , 2 , , H ,
where p j is a local perturbation, and N 0 , Σ denotes the zero-mean multivariate normal distribution with diagonal covariance matrix Σ estimated from the training set. Let x { 0 , 1 } d denote a binary vector for the interpretable representation of x , i.e., a binary vector indicating the “presence” or “absence” of a super-pixel. Let f : R d R be the model to be explained. f ( x ) can be either a probability value or a binary indicator that indicates whether x belongs to a specific class. To learn a potentially interpretable linear model g ( x ) = w g x , LIME samples instances x from N ( x ) by randomly selecting non-zero elements. Given a perturbed sample ζ { 0 , 1 } d for d < d that contains the fraction of non-zero elements in x , the ridge regression method is used to learn g ( x ) by defining a locally weighted square loss function as
L f , g , π x = x , ζ N ( x ) π x ( x ) f ( x ) g ( ζ ) 2 ,
where π x ( x ) = exp | | x ζ | | 2 2 ϱ 2 is an exponential kernel with width ϱ . Next, the function g * ( x ) is sought that minimizes
g * ( x ) = argmin g L f , g , π x + Ω ( g ) ,
where the minimization is performed over the set of linear models, and Ω ( g ) is a measure of interpretability (i.e., the number of non-zero weights for linear models). To solve the optimization problem (5), the least absolute shrinkage and selection operator (LASSO) is employed.
LIME utilizes a sequential feature selection [30,31] to fit multiple ridge regressors and select a subset of F features for the model g repeatedly. The kernel width used is ϱ = 3 4 F , which is chosen during the process.

2.2.3. SHAP

Shapley’s values, which came from cooperative game theory, are the cornerstone of SHAP [13]. In SHAP, features are considered actors that can potentially form coalitions to maximize future profits in a collaborative ecosystem. This approach has served as the foundation for several fields due to its adaptability. KernelSHAP (abbreviated as SHAP) is one of these branches.
SHAP approximates the original model with the surrogate model, chosen as a linear one. Let the original black-box model (i.e., GAN) be f and the surrogate model be denoted as g. Moreover, let z be a reduced vector of ones and zeros that enables or disables certain features of z , also known as the coalition vector. If m denotes the size of z and ξ is the number of ones in z , SHAP’s loss function is defined as
L f , g , π x = z Z f h x z g z 2 π x z ,
where h x z reshapes the reduced vector z R ξ to R m and Z denotes the set of all possible reduced vectors to be generated by taking subsets of features from z R m . The SHAP kernel π z in (6) is given by:
π x z = m ! ξ ! ( m ξ ) ! 1 .
The SHAP kernel is critical for giving tiny or big coalitions greater weight than coalitions that merely combine half of the traits (or close to it). The idea behind these behaviors is that we can learn more about individual features if we can analyze them separately (small coalitions) or if we have nearly all features except one (big coalitions).
The SHAP method is a reliable approach that can yield results equivalent to, if not better than, that of LIME. It also draws on well-established notions such as Shapley values, game theory, and LIME’s intuitive reasoning. However, similar to other permutation-based methods, SHAP has the issue of creating unrealistic data points by replacing missing attributes with random ones, which might lead to exaggerated interpretations.

2.2.4. Overview of xAI-Enhanced Approaches

This research parallels other initiatives to equip GANs with explainable techniques. In [32], a study was conducted to investigate the similarity of the inner structure of CNN-based generators employed in CycleGAN. The CycleGAN was previously used for face aging in [33]. The study introduced a cross-GAN filter similarity index to analyze the similarity of CNN filters across different GANs. Another explainable methodology called GAN-based Model EXplainability (GANMEX) was developed in [34] by incorporating the classifier into the GAN to generate one versus many explanations. By using Principal Component Analysis in the latent feature space, important latent directions were identified, which enabled a large number of interpretable controls through layer-wise perturbation [35]. In [36], an xAI-enhanced version of a baseline machine learning model is proposed that is proved to outperform the original model in terms of interpretability and classification accuracy. The SHAP technique was used to extract high-contributed features that led to more accurate identification of vegetation pixels in [37].

2.3. Proposed Workflow

The baseline CAAE network is depicted in Figure 2. The CAAE model includes the encoder E, the generator G, and two discriminators, namely, D i m g and D z . G is implemented as an autoencoder. Given an input face image, the encoder E generates an encoded z (also known as a latent vector), preserving the high-level personal feature of the input face. The generator G uses the encoded z and the target age information as a label to generate a facial image conditioned on the age. Two discriminator networks are imposed on the encoder E and the generator G, respectively. D z regularizes z to be uniformly distributed to smooth the age transformation. D i m g enforces G to generate photo-realistic and plausible faces for arbitrary z and age label. The objective function contains three terms: (1) the 2 norm of the reconstruction error between the input image and the generated image by the generator plus the total variation loss of the generated image to remove ghosting artifacts; (2) the min-max objective function to train the encoder and D z ; and (3) the min-max function to train D i m g . Let x denote the input face image, l refer to an age label, and z be the encoded variable in the output of the encoder E ( x ) = z . Moreover, let L ( · , · ) and T V ( · ) be the 2 error norm and the total variation, respectively. If p d a t a ( x ) is the distribution of the training face images, p ( z ) denotes the prior distribution, and z p ( z ) implies random sampling from the prior distribution, the objective function optimized by CAAE is defined in (8), i.e.,
{ E * , G * , D z * , D i m g * } = argmin E , G argmax D z , D i m g { λ L x , G ( E ( x ) , l ) + γ T V G ( E ( x ) , l ) + E z p z ( z ) [ log D z ( z ) ] + E x p d a t a ( x ) [ 1 log D z ( E ( x ) ) ] + E x , l p d a t a ( x , l ) [ log D i m g ( x , l ) ] + E x , l p d a t a ( x , l ) [ log 1 D i m g ( G ( E ( x ) , l ) ) ] } ,
where the coefficients λ and γ balance the smoothness and high-resolution terms.
CAAE is capable of generating highly realistic face images with both regressive and progressive effects. Unlike other methods, CAAE does not need paired examples of data for training or labeled faces in the testing data, making it more flexible. The separation of age and personality information in the latent space of z results in maintained individual personalities and the elimination of any ghosting artifacts. Finally, CAAE is not influenced by alterations in pose, emotions, or occlusion.
Another network termed xAI-GAN refers to the technique that aims to enhance GANs to provide better synthetic or reconstructed images through the incorporation of an xAI system [38]. In this way, corrective explanatory feedback is provided during training from the discriminator to the generator. Figure 3 depicts the system architecture of xAI-GAN.
Given a noise sample z from a noise distribution, the untrained generator G creates an image G ( z ) that is then fed to discriminator D. The output of the discriminator D ( G ( z ) ) , the generated image G ( z ) , and the discriminator network D are channeled to the xAI system seeking an explanation of the loss incurred by the synthetic image G ( z ) . The general idea behind the xAI-guided training process of a GAN is that the xAI system works as a guide. The xAI system acts by structuring the gradient descent in such a manner that generator training is focused on the most essential input features that the discriminator identifies.
The proposed framework follows the network structure in CAAE, which includes an encoder E that converts RGB images to latent vectors z , a generator G that converts z to RGB images, a discriminator D z that enforces a uniform distribution on the encoder’s output, and a discriminator D i m g that ensures the generator creates realistic images. CAAE incorporates two discriminators to improve the realistic properties of the generated facial images. The proposed xAI-CAAE aims to leverage xAI systems to strengthen and enrich the age progression and regression accomplished by CAAE. In CAAE, the discriminator offers feedback to the generator using a single loss value per generated image. The aim of xAI-guided training is to enhance this feedback by providing the xAI system’s “reasoning” for the discriminator’s decision. The architecture of xAI-CAAE is depicted in Figure 4.
In xAI-CAAE, a modified gradient descent generator training process is established so that generator training focuses on the most significant features for the discriminator’s prediction. An xAI system employs a score function S c to determine the explanation matrix M = S c ( G ( z ) ) after propagating the loss through the discriminator D i m g to find Δ G ( z ) . The matrix M , which refers to pixels, is made up of real numbers in the range [0, 1], with greater values indicating more important features for the discriminator’s prediction. More specifically, in Saliency maps, if the pixel is given a value of 0 or near 0 in M , the pixel under consideration does not influence the discriminator’s classification decision. On the contrary, the pixel is considered extremely essential if given a value of 1 or near 1. These values, when approaching 1, also show high-quality classification by the discriminator. In SHAP, the Shapley value can be described as the mean incremental impact of each pixel that is excluded in the discriminator’s output. This is computed by taking an average of all the different combinations of pixels that form the complete image, starting from an empty set of pixels. Accordingly, the explanation matrix M allows us to concentrate the learning process on the most important qualities, regardless of whether they were favorable or harmful to the classification.
The proposed framework utilizes M to update the generator’s weights in a modified gradient descent method. In CAAE, the adjustment of generator weights typically involves calculating the gradient of the generator’s output with respect to the loss and then applying the chain rule. This method is enhanced by first computing the explanation matrix M and then multiplying it by the gradient of the generator’s output with respect to the loss. The explanation matrix M is used to mask the latter gradient and the pixels that contributed to the discriminator’s classification. As described, the modified gradient Δ G ( z ) is obtained by taking the Hadamard product (element-wise multiplication) between Δ G ( z ) and M , denoted as Δ G ( z ) M , which serves as a mask for G ( z ) and restricts the gradient to the most significant elements. Finally, the generator’s gradients Δ G ( z ) are computed using the modified gradient as
Δ G ( z ) = Δ G ( z ) + θ Δ G ( z ) M ,
where θ is a parameter that determines the degree to which the xAI system affects the original gradients.

3. Results

Here, experimental findings for the proposed framework are disclosed. In Section 3.1, implementation details are provided for xAI-CAAE. Section 3.2 discusses the datasets utilized in the experimental evaluation. In Section 3.3, the qualitative evaluation for xAI-CAAE is described. A thorough quantitative evaluation for xAI-CAAE is discussed in Section 3.4, while important facial features for face aging are investigated in Section 3.5 using the LIME explanation system.

3.1. Implementation Details

The implementation of the proposed xAI-CAAE framework is based on the publicly available code for CAAE [7] (https://github.com/mattans/AgeProgression/tree/v1.0.0, accessed on 8 May 2023) and xAI-GAN [38] (https://github.com/explainable-gan/XAIGAN, accessed on 8 May 2023). xAI-CAAE is implemented using the Pytorch 1.2.0 [39] library. The Captum 0.4.0 [40] library is also used to implement the Saliency and SHAP explanations. For LIME, we used Lime 0.2.0.1.
The proposed xAI-CAAE, based on the setup described in [7], takes as input images of size 128 × 128 × 3 . Both the encoder E and the generator G use a kernel size of 5 × 5 . The encoder consists of five convolutional layers, each of which is followed by a ReLU activation function, and a fully connected layer, which is followed by the hyperbolic tangent activation function. The generator consists of a linear layer and seven de-convolution layers, each of which is followed by the ReLU activation function. The discriminator D z is composed of four fully connected layers. The discriminator D i m g uses a kernel size of 2 × 2 and comprises four convolutional layers, each of which is followed by batch normalization and a ReLU activation function, as well as two fully connected layers followed by the sigmoid activation function.
The input image intensities are normalized within the range [−1, 1], and then they are fed to the encoder E. The output of E is represented by the encoded vector z , whose elements are also limited within the range [−1, 1] due to the hyperbolic tangent activation function. Subsequently, the age and gender information is transformed into a one-hot vector, which is also constrained to the range [−1, 1] (instead of the usual range [0, 1]), and then concatenated with the encoded vector z . This concatenated vector is used as input for the generator G, which generates an image with intensities within the range [−1, 1] due to the hyperbolic tangent activation function. During the training process, the mini-batch size is set to 64, and the network’s four blocks (E, G, D z , and D i m g ) are updated accordingly. The Adaptive Moment Estimation (ADAM) optimizer [41] with a learning rate of 0.0002, β 1 = 0.9 , β 2 = 0.999 , and weight decay of 10 5 is employed. The network is trained for 200 epochs.
In order to integrate the xAI system with CAAE, a sigmoid activation layer is added to D i m g , which provides predictions within the range [0, 1]. By doing so, the prediction of D i m g for generated images is in the [0, 1] range, allowing images with high prediction values to be masked. The explanation matrix derived from any of the xAI systems undergoes two processing steps by taking the absolute value of the elements of M and then normalizing the resulting absolute values. The processed explanation matrix M is utilized as a mask for G. The autograd package [42] of Pytorch, which supports automatic tensor differentiation, is utilized in the xAI implementation. To modify the gradients of the generator using the explanation matrix M and adjust the backpropagation method, the register backward hook method is used, as described in Section 2.3. After half of the training epochs, xAI-guided gradient descent is utilized, as per [38].

3.2. Datasets

The proposed xAI-CAAE framework is trained on a set of images that were collected from the CACD [15] and the UTKFace [7] dataset. This set of images was collected and used to train the face aging approach in [25]. It includes 21,267 face images distributed to seven age classes: 0–10, 11–18, 19–29, 30–39, 40–49, 50–59, and 60+ years old (the oldest person is 80 years old). The same split to age classes has been considered in many facing approaches [25,43]. Approximately the same number of images belongs to each age class. Each gender is equally distributed in each class.
The FGNET aging dataset [17] is employed for testing xAI-CAAE. FGNET comprises 1002 face images of 82 different subjects, whose age varies from 0 to 69 years. FGNET is a dataset frequently used in facial aging research [7,44,45].
The Adience dataset [19] is a collection of images collected from the social network Flickr. The dataset comprises 26,580 images distributed to eight age classes: 0–2, 4–6, 8–13, 15–20, 25–32, 38–43, 48–53, and 60+. The dataset is very challenging for age and gender classification due to the unconstrained, real-life capturing conditions of its images. This dataset is used to investigate which facial regions are important for age classification by leveraging LIME explanations, as described in Section 3.5.

3.3. Qualitative Evaluation of xAI-CAAE

The proposed framework generated age progression and regression results on the FGNET dataset using two xAI systems, Saliency and SHAP. These results are depicted in Figure 5a and Figure 5b, respectively. Figure 5c depicts comparative results for the original CAAE. The images in the first column of each figure illustrate the sample FGNET images from each age class, while the rest of the columns illustrate the generated images in each age class. The ground truth age class of each image is indicated by a red box. In the experiments shown in Figure 5, the xAI-CAAE parameters were set to a z size of 100 and θ = 0.2 in (9). In Figure 6, results with a z size of 100 and θ = 0.5 in (9) are depicted.
Comparison between xAI-CAAE and CAAE. As can be observed in Figure 5, xAI-CAAE yields plausible and satisfying age progression and regression results. The images generated by xAI-CAAE are more realistic, with fewer distortions compared to the images generated by CAAE. For example, in the second row of Figure 5a–c, age-progressed and regressed images generated by xAI-CAAE using either Saliency or SHAP are notably more pleasing compared with the images generated by CAAE, which produced blurring in the image, especially around the area of the eyes. The same can be observed for the images in the fourth row of Figure 5a–c, where xAI-CAAE using either Saliency or SHAP generates more realistic facial images compared to the images generated by CAAE that include eye distortions, especially for age progression (columns 6–8). Similar observations can be made for the age progression and regression results in Figure 6a–c. For example, inspecting the images in the second, third, and fifth rows shows that the proposed framework produces more realistic images with fewer artifacts compared to the images generated by CAAE. The qualitative inspection reveals the advantages of using xAI techniques in the CAAE network.
Comparison between different xAI techniques in xAI-CAAE. Comparing the results achieved by xAI-CAAE in Figure 5a,b and Figure 6a,b, xAI-CAAE with Saliency achieves competitive performance to xAI-CAAE using SHAP. In some cases (e.g., the fifth and sixth rows in Figure 5a,b), the images generated by xAI-CAAE using Saliency are sharper and more detailed compared to the ones generated by xAI-CAAE using SHAP. However, in the same rows, the images generated by xAI-CAAE using SHAP more realistically represent the process of face aging, especially in the older age classes. More specifically, the SHAP method seems to render with greater success the characteristics of face age progression, i.e., faces with wrinkles around the eyes and mouth (see column 8 in the sixth row of Figure 5a,b).

3.4. Quantitative Evaluation of xAI-CAAE

In this Section, xAI-CAAE is evaluated using the quantitative evaluation metrics described in Section 3.4.1.

3.4.1. Evaluation Metrics

Fréchet Inception Distance. To evaluate the quality of images generated by CAAE and xAI-CAAE, the Fréchet Inception Distance (FID) is employed. This metric has been shown to align with human perception of image quality [16]. FID works by mapping a set of images to a feature space defined by a specific layer of the Inception model. The layer’s activation values are used to estimate statistics such as the mean vector and the covariance matrix, which are employed to create a multi-dimensional Gaussian distribution. Finally, the Fréchet Distance between the distributions estimated from the real and generated images is calculated and denoted by the FID score. When the FID score is lower, it signifies that the generated images closely resemble the real images, indicating that high-quality visual images have been produced.
Age Estimation. In order to assess the plausibility of the images generated for a specific age group, the DEX age estimation model [18] is used to estimate the age of both the age-progressed and regressed images. The classification accuracy and 1-off classification accuracy, which measures the accuracy when the estimated age class is off by one age group from the actual age class, are the evaluation criteria used. Additionally, the Adjusted-Mean Absolute Error (MAE) is also employed as an evaluation metric, following the methodology in [25]. The MAE considers the discrepancy between the estimated age and the age range of the targeted age group. Let a ˜ be the estimated age for a generated image that resembles the aging characteristics of an age class with range [ a l 0 , a l 1 ] . If a ˜ < a l 0 , Adjusted-MAE is calculated as | a l 0 a ˜ | . If a ˜ > a l 1 , Adjusted-MAE is calculated as | a l 1 a ˜ | . Finally, if a l 0 < a ˜ < a l 1 , Adjusted-MAE is zero.

3.4.2. Evaluation Results Using FID Score

FID scores for the original FGNET images. Table 1 summarizes the FID scores for the original FGNET images. The FID score is a measure of how closely the training images resemble the original FGNET images and is calculated by comparing the two sets of images. The results show that the FID scores for the images in the older age classes ( C 5 C 7 ) are higher than those in the younger age classes, suggesting that the training images in the older age classes are less similar to the original FGNET images in those classes compared to the training images selected from the younger age classes and the original FGNET images in those classes.
FID scores for images generated by CAAE. In Table 2 the FID scores of the images generated by CAAE in age classes C i , i = 1 , 7 are listed. Following the analysis in Section 3.4.1, the images with higher visual quality result in a lower FID score. The best value in each row is indicated in boldface. The lowest FID score is obtained in all age classes when the input age class is the same as the target age class ( C 5 ) or when the input and target age classes are adjacent ( C 1 C 4 , C 6 ). As can be seen, the input age class C 7 has the lowest FID score for translations to target class C 2 . Since a relatively limited amount of FGNET images are contained in this age class (see Table 1), the estimated distribution for the real FGNET images that belong to class C 7 may be unrepresentative.
FID scores for images generated by xAI-CAAE. Table 3 and Table 4 summarize the FID scores for the generated images by xAI-CAAE for latent vectors of size 100 and θ = 0.2 using Saliency and SHAP explanations, respectively. The best value in each row is indicated in boldface. Explanatory methods are used in order to more effectively assimilate the features of a facial image and provide images of higher quality than those generated by CAAE. In Table 3 and Table 4, the FID scores that outperform the original CAAE are marked with gray color to facilitate visual inspection. It is clear that when compared with the original CAAE, the xAI-CAAE using SHAP produces better FID scores for more translations than the xAI-CAAE utilizing Saliency explanations. Comparing the two xAI systems, it can be seen that the SHAP explanation method in Table 4 gives marginally improved images in some cases compared with the Saliency explanation method in Table 3. For example, all generated images by xAI-CAAE using SHAP to target age class C 3 (fourth column in Table 4) demonstrate lower FID scores compared with the images generated by xAI-CAAE using Saliency (fourth column in Table 3).
Comparison of FID scores for different θ values. To investigate the impact of parameter θ , another experiment is conducted by assigning a greater value to parameter θ , i.e., θ = 0.5 . This way, the modified gradient multiplied by the explanation matrix M is given greater weight, expecting to affect face aging more intensively. Table 5 and Table 6 summarize the FID scores for the images generated by xAI-CAAE with latent vectors of size 100 using Saliency and SHAP explanations, respectively. The best value in each row is indicated in boldface, while the FID scores that outperform those of the original CAAE are highlighted in gray. By inspecting Table 5 and Table 6, it can be seen that the proposed framework achieves lower FID scores when using SHAP than when using Saliency. More specifically, in 73.47% of the cells (36 cells out of 49) in Table 5 and Table 6, the xAI-CAAE with SHAP explanations achieved a better performance than the xAI-CAAE with Saliency explanations with respect to the FID score.
By comparing Table 3 and Table 5, it can be seen that the increase in parameter θ has lowered the FID scores in almost all cases of xAI-CAAE with Saliency explanations. Only six cells in Table 5 have greater FID scores than the corresponding cells in Table 3, i.e., cells [ C 1 , C 1 ] , [ C 7 , C 1 ] , [ C 5 , C 4 ] , [ C 7 , C 4 ] , [ C 7 , C 6 ] , and [ C 7 , C 7 ] . Comparing the FID scores for xAI-CAAE with SHAP explanations in Table 4 and Table 6, it is found that FID is reduced by increasing θ . FID scores for θ = 0.5 in Table 6 are smaller than the FID scores for θ = 0.2 in Table 4 in the vast majority of translations. Only six cells in Table 6 resulted in a larger FID than the same cells in Table 4. By inspecting the gray cells in Table 5 and Table 6, one finds that the increase in the value of θ has led to improved FID scores than those achieved by the original CAAE (see Table 2) as well as when θ = 0.2 . The quantitative evaluation using the FID score has demonstrated the improved performance of xAI-CAAE when explanations are incorporated from either Saliency or SHAP compared to the original CAAE without explanations. Notably, FID reduces when parameter θ increases, i.e., when a greater weight is assigned to the explanation system during training for both Saliency and SHAP. These quantitative results agree with the qualitative results depicted in Figure 5 and Figure 6 that demonstrate the visual quality of the age progressed and regressed images generated by xAI-CAAE.
Diagram comparison of FID scores for images generated by CAAE and xAI-CAAE. To simplify the comparison of xAI-CAAE with the original CAAE, the FID score for all generated images in each age group is presented in Table 7. The distribution of the original FGNET images belonging to each age class, such as C k , and all the generated images that resemble the aging features of this class, i.e., the images generated to age classes C l C k , l = 1 , , 7 , are used to calculate the FID score. Table 7 gathers FID scores for xAI-CAAE when parameter θ takes value either 0.2 or 0.5 . As can be seen in Table 7, xAI-CAAE with SHAP explanation marginally outperforms CAAE for θ = 0.2 . When θ = 0.5 , xAI-CAAE with both Saliency and SHAP explanations outperforms CAAE. Notably, xAI-CAAE using SHAP achieves a large percentage of improvement in FID score compared to the original CAAE. The FID scores obtained from the calculations in Table 7 are comparatively depicted with bar diagrams in Figure 7. This representation was chosen in order to provide a visual illustration of the impact of the parameter θ on the scores. For θ = 0.2 , the proposed framework achieves competitive performance compared to the original CAAE (Figure 7a). For θ = 0.5 , xAI-CAAE with either Saliency or SHAP explanations consistently outperforms CAAE (Figure 7b), while the greatest performance gain is achieved by xAI-CAAE that employs SHAP explanations. Figure 7a,b demonstrate that increasing the value of parameter θ , which enhances the impact of the incorporated xAI technique, results in improved FID scores for the proposed xAI-CAAE, thus enabling the xAI-CAAE to surpass the original CAAE.

3.4.3. Evaluation Results Using Age Estimation

Age estimation results for images generated by CAAE and xAI-CAAE. The pre-trained DEX age estimation model [18] was employed to determine the age estimation results of images produced by xAI-CAAE. The results of the age estimation are presented in Table 8, using the Adjusted-MAE evaluation metric described in Section 3.4.1. The fourth and fifth columns of Table 8 show the age estimation results for xAI-CAAE with latent vectors of size 100 and θ = 0.2 , while the last two columns display the results for the same size of latent vectors but with θ = 0.5 . In addition, age estimation results for images generated by CAAE are presented in the third column of Table 8, and the second column displays the evaluation metric computed for the original FGNET images. Table 8 includes results in the second column that serve as evidence of the accuracy of DEX age estimations on the 1002 FGNET images considered as ground truth. It should be noted that these results are not directly comparable to those computed for xAI-CAAE (columns 4–7) and CAAE (column 3), which were obtained for translations of the FGNET images to each of the 7 age classes resulting in a total of 7041, i.e., 1002 × 7 images per column. The evaluation of age classification accuracy and 1-off accuracy are summarized in Table 9 and Table 10, respectively.
Table 8 shows that xAI-CAAE achieves the best Adjusted-MAE for all age groups, using either Saliency or SHAP and either θ = 0.2 or θ = 0.5 . While the differences in Adjusted-MAE between xAI-CAAE and CAAE are negligible for some age groups (such as C 7 , where the top Adjusted-MAE for xAI-CAAE is 29.20 years compared to 29.64 years for CAAE), the differences are more significant for younger age groups, particularly C 1 and C 2 , where xAI-CAAE achieves Adjusted-MAE scores 1.96 and 1.53 years lower than CAAE, respectively. Furthermore, xAI-CAAE using SHAP with θ = 0.5 achieves the best Adjusted-MAE for all generated images (as indicated in the last row of Table 8). Hence, incorporating xAI systems in CAAE has made it easier to generate images that closely resemble the aging characteristics of the target age class.
Similar observations can be made for the age classification results in Table 9 and Table 10. The proposed framework (using either Saliency or SHAP, with either θ = 0.2 or θ = 0.5 ), achieves the top accuracy and the top 1-off accuracy for all age classes (all rows in Table 9 and Table 10). It should be noted that the most significant gain in accuracy (+5.69%) for xAI-CAAE compared to the original CAAE is achieved for age class C 2 when using SHAP with θ = 0.5 . As can be seen, the best results on age estimation accuracy are reported for age classes C 2 = [ 11 , 18 ] and C 3 = [ 19 , 29 ] for both CAAE and xAI-CAAE. From the results in column 2, which as mentioned are not directly comparable to the results in columns 3–7, it can be seen that the pre-trained age estimation model achieves better accuracy scores for the age classes C 2 = [ 11 , 18 ] , C 3 = [ 19 , 29 ] , C 6 = [ 50 , 59 ] , and C 7 = 60 + on the original FGNET images. The accuracies achieved by both CAAE and xAI-CAAE for classes C 2 and C 3 are also high, but lower scores are achieved for the elder age classes C 6 and C 7 . The results in column 2 for the age classes with a small number of original FGNET images (see Table 1) may not be fully representative due to the limited sample size.
Regarding 1-off accuracy, in Table 10, the most significant differences between xAI-CAAE and CAAE are noted for age classes C 1 , C 4 , and C 6 where the absolute differences compared to the results achieved by CAAE are 10.88%, 6.48%, and 5.59%, respectively. It can be seen that 1-off accuracy is significantly better than accuracy for all age classes, with the highest increases reported for age classes C 1 and C 4 ( 2752.11 % and 1999.09 % , respectively for xAI-CAAE with SHAP and θ = 0.5 ). The improved performance of the images generated by xAI-CAAE on age classification illustrates the strength of xAI explanations to guide the generator for realistic age progression and regression more effectively.
Comparison of age estimation results for different θ values. Regarding the impact of parameter θ on xAI-CAAE, it can be seen in Table 8 that both xAI-CAAE using Saliency and SHAP achieve better Adjusted-MAE scores for θ = 0.5 compared with the corresponding results for θ = 0.2 for the younger age classes ( C 1 C 4 ). The opposite is true for the older age classes ( C 5 C 7 ) where both xAI-CAAE using Saliency and xAI-CAAE using SHAP achieved better scores for θ = 0.2 . It is noted that, in general, the predictions of the pre-trained age estimation model for the images generated by both xAI-CAAE and CAAE to these age classes ( C 5 C 7 ) are less accurate, resulting in higher Adjusted-MAE scores compared with the ones achieved for age classes C 1 C 4 .

3.5. Interpreting Age Classification Using LIME xAI-System

This experiment uses the LIME explanation system to investigate which facial features are essential for face aging. As described in detail in Section 2.2.2, LIME provides an insight into image areas (known as super-pixels) that a model trained on age classification considers critical for the classification decision. The pre-trained age classification model in [46], which is trained on the Adience dataset (see Section 3.2), is assessed using LIME. More specifically, the predictions of the pre-trained age classification model are fed to the LIME explanation system. Subsequently, LIME is used to explain the areas of the image that mainly influence the classification decision. A good classification model is expected to highlight areas of the face and not noise from the background. Since this is a fine-grained classification problem, we expect the model to highlight areas in the face that indicate each age class.
LIME provides a local interpretation of the age classification model and can be used to explain the model’s behavior for each testing image, i.e., explain the individual predictions to find out which input features are essential for the particular prediction. Figure 8 shows a subset of the experimental results, which includes two examples of correctly classified testing images and two examples of misclassified testing images in each class. It can be seen that specific areas of the face light up according to age class, indicating the facial areas that contributed the most to the classification decision. For example, the area around the cheeks is highlighted in C 1 (age 0–2), a characteristic of humans at that age. In C 4 (age 15–20), the model identifies the area of the eyes as important, while the neck area is highlighted in C 7 (age 48–53). For the misclassified images, we notice that the same regions are highlighted, but the noise from the background is also considered important. This experiment demonstrates that by leveraging LIME in the training process of an age classifier, the classifier’s attention can be directed toward the facial areas that play a crucial role in accurate classification decisions. Additionally, considering the significant facial areas of misclassified images can help the classifier reduce its attention to these areas and improve the accuracy of age classification.

4. Discussion and Conclusions

In this paper, a novel explainable Conditional Adversarial Autoencoder, termed xAI-CAAE, aims to provide corrective feedback from the discriminator to the generator through an explanation matrix using Saliency maps and Shapley values as explanatory techniques. The proposed framework has been thoroughly evaluated both qualitatively and quantitatively. It has been demonstrated to have great potential as a competitive framework for generating more realistic face images. The xAI system has contributed significantly to face aging, as can be confirmed by quantitative evaluation metrics, such as the FID scores and the age estimation on the generated images. LIME has also been leveraged to investigate the facial areas important for age classification, yielding interesting results. Future work will focus on deepening the xAI methods for face aging.

Author Contributions

Conceptualization, C.K. (Christos Korgialas), E.P. and C.K. (Constantine Kotropoulos); methodology, C.K. (Christos Korgialas) and E.P.; software, C.K. (Christos Korgialas) and E.P.; validation, C.K. (Christos Korgialas), E.P., A.B. and M.S.; writing—original draft preparation, C.K. (Christos Korgialas) and E.P.; writing—review and editing, C.K. (Christos Korgialas), E.P. and C.K. (Constantine Kotropoulos); supervision, C.K. (Constantine Kotropoulos). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data showcased in this article are available for download and further analysis through their respective websites. Any additional information or clarifications on the data used in this study can be obtained by contacting the corresponding author.

Acknowledgments

Results presented in this work have been produced using the Aristotle University of Thessaloniki (AUTh) High-Performance Computing Infrastructure and Resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sethuram, A.; Patterson, E.; Ricanek, K.; Rawls, A. Improvements and performance evaluation concerning synthetic age progression and face recognition affected by adult aging. In Proceedings of the International Conference on Biometrics, Alghero, Italy, 2–5 June 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 62–71. [Google Scholar]
  2. Sawant, M.M.; Bhurchandi, K.M. Age invariant face recognition: A survey on facial aging databases, techniques and effect of aging. Artif. Intell. Rev. 2019, 52, 981–1008. [Google Scholar] [CrossRef]
  3. Moustafa, A.A.; Elnakib, A.; Areed, N.F.F. Age-invariant face recognition based on deep features analysis. Signal Image Video Process. 2020, 14, 1027–1034. [Google Scholar] [CrossRef]
  4. How Digital De-Aging Is Changing the Face of Movies. 2019. Available online: https://www.cbc.ca/news/entertainment/digital-de-aging-1.5397657 (accessed on 8 May 2023).
  5. Despois, J.; Flament, F.; Perrot, M. AgingMapGAN (AMGAN): High-Resolution Controllable Face Aging with Spatially-Aware Conditional GANs. In Proceedings of the Computer Vision–ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 613–628. [Google Scholar]
  6. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  7. Zhang, Z.; Song, Y.; Qi, H. Age progression/regression by conditional adversarial autoencoder. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5810–5818. [Google Scholar]
  8. Dewi, C.; Chen, R.C.; Liu, Y.T.; Yu, H. Various generative adversarial networks model for synthetic prohibitory sign image generation. Appl. Sci. 2021, 11, 2913. [Google Scholar] [CrossRef]
  9. Sharma, A.; Singh, P.K.; Chandra, R. SMOTified-GAN for class imbalanced pattern classification problems. IEEE Access 2022, 10, 30655–30665. [Google Scholar] [CrossRef]
  10. Huang, G.; Jafari, A.H. Enhanced balancing GAN: Minority-class image generation. Neural Comput. Appl. 2023, 35, 5145–5154. [Google Scholar] [CrossRef] [PubMed]
  11. Xia, W.; Zhang, Y.; Yang, Y.; Xue, J.H.; Zhou, B.; Yang, M.H. GAN inversion: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3121–3138. [Google Scholar] [CrossRef]
  12. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. In Proceedings of the International Conference on Learning Representations Workshop, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  13. Lundberg, S.M.; Lee, S. A Unified Approach to Interpreting Model Predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 4765–4774. [Google Scholar]
  14. Ribeiro, M.T.; Singh, S.; Guestrin, C. Why should I trust you? Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  15. Chen, B.C.; Chen, C.S.; Hsu, W.H. Cross-Age Reference Coding for Age-Invariant Face Recognition and Retrieval. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 768–783. [Google Scholar]
  16. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30, pp. 6626–6637. [Google Scholar]
  17. Lanitis, A.; Taylor, C.J.; Cootes, T.F. Toward automatic simulation of aging effects on face images. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 442–455. [Google Scholar] [CrossRef]
  18. Rothe, R.; Timofte, R.; Gool, L.V. DEX: Deep EXpectation of apparent age from a single image. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 10–15. [Google Scholar]
  19. Eidinger, E.; Enbar, R.; Hassner, T. Age and gender estimation of unfiltered faces. IEEE Trans. Inf. Forensics Secur. 2014, 9, 2170–2179. [Google Scholar] [CrossRef]
  20. Suo, J.; Zhu, S.C.; Shan, S.; Chen, X. A compositional and dynamic model for face aging. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 385–401. [Google Scholar]
  21. Shen, C.T.; Huang, F.; Lu, W.H.; Shih, S.W.; Liao, H.Y.M. 3D Age Progression Prediction in Children’s Faces with a Small Exemplar-Image Set. J. Inf. Sci. Eng. 2014, 30, 1131–1148. [Google Scholar]
  22. Shu, X.; Tang, J.; Lai, H.; Liu, L.; Yan, S. Personalized age progression with aging dictionary. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3970–3978. [Google Scholar]
  23. Kemelmacher-Shlizerman, I.; Suwajanakorn, S.; Seitz, S.M. Illumination-aware age progression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3334–3341. [Google Scholar]
  24. Liu, Y.; Li, Q.; Sun, Z.; Tan, T. A3GAN: An Attribute-Aware Attentive Generative Adversarial Network for Face Aging. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2776–2790. [Google Scholar] [CrossRef]
  25. Pantraki, E.; Kotropoulos, C. Face aging using global and pyramid generative adversarial networks. Mach. Vis. Appl. 2021, 32, 1–20. [Google Scholar] [CrossRef]
  26. Wu, Y.; Wang, R.; Gong, M.; Cheng, J.; Yu, Z.; Tao, D. Adversarial UV-Transformation Texture Estimation for 3D Face Aging. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4338–4350. [Google Scholar] [CrossRef]
  27. Yadav, D.; Kohli, N.; Vatsa, M.; Singh, R.; Noore, A. Age Gap Reducer-GAN for Recognizing Age-Separated Faces. In Proceedings of the 25th International Conference on Pattern Recognition, Milan, Italy, 10–15 January 2020; pp. 10090–10097. [Google Scholar]
  28. Duan, M.; Li, K.; Li, K. An Ensemble CNN2ELM for Age Estimation. IEEE Trans. Inf. Forensics Secur. 2018, 13, 758–772. [Google Scholar] [CrossRef]
  29. Greco, A.; Saggese, A.; Vento, M.; Vigilante, V. Effective training of convolutional neural networks for age estimation based on knowledge distillation. Neural Comput. Appl. 2021, 34, 21449–21464. [Google Scholar] [CrossRef]
  30. Pudil, P.; Novovičová, J.; Kittler, J. Floating search methods in feature selection. Pattern Recognit. Lett. 1994, 15, 1119–1125. [Google Scholar] [CrossRef]
  31. Ververidis, D.; Kotropoulos, C. Fast and accurate sequential floating forward selection with the Bayes classifier in speech emotion recognition. Signal Process. 2008, 88, 2956–2970. [Google Scholar] [CrossRef]
  32. Genovese, A.; Piuri, V.; Scotti, F. Towards Explainable Face Aging with Generative Adversarial Networks. In Proceedings of the 2019 IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; pp. 3806–3810. [Google Scholar]
  33. Zhou, S.; Zhao, W.; Feng, J.; Lai, H.; Pan, Y.; Yin, J.; Yan, S. Personalized and Occupational-aware Age Progression by Generative Adversarial Networks. arXiv 2017, arXiv:1711.09368. [Google Scholar]
  34. Shih, S.M.; Tien, P.J.; Karnin, Z. GANMEX: One-vs-One Attributions using GAN-based Model Explainability. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; Volume 139, pp. 9592–9602. [Google Scholar]
  35. Härkönen, E.; Hertzmann, A.; Lehtinen, J.; Paris, S. GANSpace: Discovering Interpretable GAN Controls. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 9841–9850. [Google Scholar]
  36. Aslam, N.; Khan, I.U.; Mirza, S.; AlOwayed, A.; Anis, F.M.; Aljuaid, R.M.; Baageel, R. Interpretable Machine Learning Models for Malicious Domains Detection Using Explainable Artificial Intelligence (XAI). Sustainability 2022, 14, 7375. [Google Scholar] [CrossRef]
  37. Abdollahi, A.; Pradhan, B. Urban vegetation mapping from aerial imagery using explainable AI (XAI). Sensors 2021, 21, 4738. [Google Scholar] [CrossRef]
  38. Nagisetty, V.; Graves, L.; Scott, J.; Ganesh, V. xAI-GAN: Enhancing Generative Adversarial Networks via Explainable AI Systems. arXiv 2020. [Google Scholar] [CrossRef]
  39. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32, pp. 8024–8035. [Google Scholar]
  40. Kokhlikyan, N.; Miglani, V.; Martin, M.; Wang, E.; Reynolds, J.; Melnikov, A.; Lunova, N.; Reblitz-Richardson, O. PyTorch Captum. 2019. Available online: https://github.com/pytorch/captum (accessed on 8 May 2023).
  41. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; Bengio, Y., LeCun, Y., Eds.; pp. 1–15. [Google Scholar]
  42. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in PyTorch. In Proceedings of the Advances in Neural Information Processing Systems Workshop on Autodiff, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  43. Liu, S.; Sun, Y.; Zhu, D.; Bao, R.; Wang, W.; Shu, X.; Yan, S. Face Aging with Contextual Generative Adversarial Nets. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 82–90. [Google Scholar]
  44. Zhao, J.; Cheng, Y.; Cheng, Y.; Yang, Y.; Lan, H.; Zhao, F.; Xiong, L.; Xu, Y.; Li, J.; Pranata, S.; et al. Look Across Elapse: Disentangled Representation Learning and Photorealistic Cross-Age Face Synthesis for Age-Invariant Face Recognition. In Proceedings of the 33rd Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  45. Yang, H.; Huang, D.; Wang, Y.; Jain, A.K. Learning Face Age Progression: A Pyramid Architecture of GANs. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 31–39. [Google Scholar]
  46. Levi, G.; Hassner, T. Age and gender classification using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 34–42. [Google Scholar]
Figure 1. The face aging task.
Figure 1. The face aging task.
Jimaging 09 00096 g001
Figure 2. The CAAE architecture.
Figure 2. The CAAE architecture.
Jimaging 09 00096 g002
Figure 3. xAI-GAN system architecture.
Figure 3. xAI-GAN system architecture.
Jimaging 09 00096 g003
Figure 4. Proposed xAI-CAAE architecture.
Figure 4. Proposed xAI-CAAE architecture.
Jimaging 09 00096 g004
Figure 5. Age progression and regression results obtained by xAI-CAAE for the size of latent vector z equal to 100 and θ = 0.2 , when using (a) Saliency and (b) SHAP xAI systems to sample images of FGNET. The results in (c) are obtained by the original CAAE without incorporating any xAI system for the same size of the latent vector z . The first column depicts input images, and the rest of the columns depict the age-progressed and regressed images. The red boxes indicate the generated images belonging to each input image’s ground truth age class.
Figure 5. Age progression and regression results obtained by xAI-CAAE for the size of latent vector z equal to 100 and θ = 0.2 , when using (a) Saliency and (b) SHAP xAI systems to sample images of FGNET. The results in (c) are obtained by the original CAAE without incorporating any xAI system for the same size of the latent vector z . The first column depicts input images, and the rest of the columns depict the age-progressed and regressed images. The red boxes indicate the generated images belonging to each input image’s ground truth age class.
Jimaging 09 00096 g005
Figure 6. Age progression and regression results obtained by xAI-CAAE for the size of latent vector z equal to 100 and θ = 0.5 , when using (a) Saliency and (b) SHAP. The results in (c) are obtained by the original CAAE without incorporating any xAI system for the same size of the latent vector z . The first column depicts input faces, and the rest of the columns depict the age-progressed and regressed generated images. The red boxes indicate the generated images belonging to each input image’s ground truth age class.
Figure 6. Age progression and regression results obtained by xAI-CAAE for the size of latent vector z equal to 100 and θ = 0.5 , when using (a) Saliency and (b) SHAP. The results in (c) are obtained by the original CAAE without incorporating any xAI system for the same size of the latent vector z . The first column depicts input faces, and the rest of the columns depict the age-progressed and regressed generated images. The red boxes indicate the generated images belonging to each input image’s ground truth age class.
Jimaging 09 00096 g006
Figure 7. Bar diagram comparison of the FID scores in Table 7 achieved by the xAI-CAAE with latent vectors z of size 100 and (a) θ = 0.2 , (b) θ = 0.5 against the FID scores of CAAE with latent vectors of the same size for all images generated to resemble the aging characteristics in each age class.
Figure 7. Bar diagram comparison of the FID scores in Table 7 achieved by the xAI-CAAE with latent vectors z of size 100 and (a) θ = 0.2 , (b) θ = 0.5 against the FID scores of CAAE with latent vectors of the same size for all images generated to resemble the aging characteristics in each age class.
Jimaging 09 00096 g007
Figure 8. Explaining age classification predictions on Adience dataset. Two examples of correctly classified images and two misclassified images in each class are shown. Super-pixels highlight important areas of the face in each age class.
Figure 8. Explaining age classification predictions on Adience dataset. Two examples of correctly classified images and two misclassified images in each class are shown. Super-pixels highlight important areas of the face in each age class.
Jimaging 09 00096 g008
Table 1. FID scores for the original FGNET images across age classes C i , i = 1 , , 7 . The number of images that belong to each age class is listed in the second column.
Table 1. FID scores for the original FGNET images across age classes C i , i = 1 , , 7 . The number of images that belong to each age class is listed in the second column.
Age ClassNumber of ImagesFID Score
C 1 = [ 0 , 10 ] 41176.81
C 2 = [ 11 , 18 ] 27661.69
C 3 = [ 19 , 29 ] 16768.03
C 4 = [ 30 , 39 ] 7987.53
C 5 = [ 40 , 49 ] 46104.21
C 6 = [ 50 , 59 ] 15141.61
C 7 = [ 60 , 80 ] 8162.13
Table 2. FID scores for the images generated by CAAE for the size of latent vectors z equal to 100 in age classes C i , i = 1 , 2 , , 7 . The best value in each row is indicated in boldface.
Table 2. FID scores for the images generated by CAAE for the size of latent vectors z equal to 100 in age classes C i , i = 1 , 2 , , 7 . The best value in each row is indicated in boldface.
InputTarget Age Class
Age Class C 1 C 2 C 3 C 4 C 5 C 6 C 7
C 1 93.5177.3078.5080.8378.6379.7180.49
C 2 79.9677.5076.5880.7776.9277.9379.04
C 3 83.5774.7178.5877.5476.0975.4277.43
C 4 97.5792.4091.8690.0689.4294.4191.21
C 5 115.48108.74105.38105.82103.31109.18108.22
C 6 137.41129.38132.62127.72126.99127.10124.24
C 7 145.50131.44143.80137.37134.87132.69137.51
Table 3. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.2 using Saliency explanation in age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform those of the original CAAE (Table 2) are highlighted in gray.
Table 3. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.2 using Saliency explanation in age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform those of the original CAAE (Table 2) are highlighted in gray.
InputTarget Age Class
Age Class C 1 C 2 C 3 C 4 C 5 C 6 C 7
C 1 88.4476.8179.7779.7377.6177.9379.62
C 2 83.9181.6881.4781.1279.8080.1678.72
C 3 83.1682.6483.0983.1581.4782.0283.61
C 4 103.9092.4694.31100.8294.2791.9596.96
C 5 119.68107.53112.02112.97113.55105.92113.72
C 6 153.28132.22142.72123.51136.57139.22139.33
C 7 134.93133.58136.56120.42128.27133.16136.79
Table 4. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.2 using SHAP explanation in age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform those of the original CAAE (Table 2) are highlighted in gray.
Table 4. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.2 using SHAP explanation in age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform those of the original CAAE (Table 2) are highlighted in gray.
InputTarget Age Class
Age Class C 1 C 2 C 3 C 4 C 5 C 6 C 7
C 1 91.7675.3172.0777.3375.0678.3876.11
C 2 85.2478.7479.1979.3879.1279.1679.17
C 3 82.5980.2577.8174.4376.8577.7377.74
C 4 101.4289.8791.1192.7692.8692.0194.62
C 5 119.23106.70108.25107.91109.82112.48111.47
C 6 145.47116.81121.69121.18126.64130.71127.28
C 7 137.68141.90134.83145.01133.76129.72146.28
Table 5. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.5 using Saliency explanation for age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform the original CAAE (in Table 2) are highlighted in gray.
Table 5. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.5 using Saliency explanation for age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform the original CAAE (in Table 2) are highlighted in gray.
InputTarget Age Class
Age Class C 1 C 2 C 3 C 4 C 5 C 6 C 7
C 1 89.1374.1873.1976.7377.1177.1075.58
C 2 81.3476.5379.0479.4076.6177.2375.80
C 3 80.0481.5278.9275.3777.5276.4477.20
C 4 94.7587.5589.6886.1889.4389.8988.95
C 5 114.29105.91101.32118.98112.86100.82110.00
C 6 141.69117.76120.19116.91126.94118.87125.98
C 7 137.04122.92135.70123.84127.32134.07149.07
Table 6. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.5 using SHAP explanation for age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform the original CAAE (in Table 2) are highlighted in gray.
Table 6. FID scores for images generated by the proposed xAI-CAAE for latent vectors z of size 100 and θ = 0.5 using SHAP explanation for age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface. The FID scores that outperform the original CAAE (in Table 2) are highlighted in gray.
InputTarget Age Class
Age Class C 1 C 2 C 3 C 4 C 5 C 6 C 7
C 1 89.4970.9069.9570.2269.2573.6572.88
C 2 76.0074.7773.0874.2672.8674.7974.50
C 3 79.1776.5676.5473.3674.7173.3372.67
C 4 93.8587.2286.0586.5889.3988.0590.12
C 5 112.37101.81109.32107.68109.64107.66106.52
C 6 135.24134.39117.76126.24152.73118.30142.22
C 7 131.26136.74134.48128.89144.58142.89135.19
Table 7. FID scores for the images generated by xAI-CAAE and CAAE for latent vectors z of size 100 in age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface.
Table 7. FID scores for the images generated by xAI-CAAE and CAAE for latent vectors z of size 100 in age classes C i , i = 1 , , 7 . The best value in each row is indicated in boldface.
Age ClassCAAExAI-CAAE ( θ = 0.2 )xAI-CAAE ( θ = 0.5 )
SaliencySHAPSaliencySHAP
C k C 1 , k = 1 , , 7 65.1665.3765.7662.7961.69
C k C 2 , k = 1 , , 7 58.1660.2059.4557.3954.67
C k C 3 , k = 1 , , 7 58.7861.2058.3157.7754.75
C k C 4 , k = 1 , , 7 61.1862.1859.1858.5154.97
C k C 5 , k = 1 , , 7 58.9760.3958.9658.9254.55
C k C 6 , k = 1 , , 7 59.7560.9959.6958.7155.79
C k C 7 , k = 1 , , 7 61.0461.7459.5458.2456.24
Table 8. Adjusted-MAE in age estimation for the proposed xAI-CAAE using Saliency and SHAP explanations. Age prediction is performed using the pre-trained model in [18]. Adjusted-MAE for the original FGNET images and the images generated by the original CAAE are also listed. The best value in each row (excluding column 2) is indicated in boldface. The results in the second column prove the correctness of the DEX age estimations on the ground truth FGNET images and cannot directly be compared to the results listed in the remaining columns.
Table 8. Adjusted-MAE in age estimation for the proposed xAI-CAAE using Saliency and SHAP explanations. Age prediction is performed using the pre-trained model in [18]. Adjusted-MAE for the original FGNET images and the images generated by the original CAAE are also listed. The best value in each row (excluding column 2) is indicated in boldface. The results in the second column prove the correctness of the DEX age estimations on the ground truth FGNET images and cannot directly be compared to the results listed in the remaining columns.
Adjusted-MAE
Age ClassFGNETCAAExAI-CAAE ( θ = 0.2 )xAI-CAAE ( θ = 0.5 )
SaliencySHAPSaliencySHAP
C 1 13.0312.9915.2911.4112.9311.03
C 2 5.835.655.984.565.314.12
C 3 5.813.083.943.092.302.74
C 4 10.949.3010.019.608.769.47
C 5 9.1516.0816.1915.9016.3416.91
C 6 0.2022.5621.5422.8322.4423.96
C 7 10.1329.6429.2030.2930.6431.01
C k , k = 1 , , 7 18.0517.4518.3717.4617.1516.90
Table 9. Accuracy of age classification for the proposed xAI-CAAE using Saliency and SHAP explanations. Age prediction is performed using the pre-trained model in [18]. Accuracy for the original FGNET images and the images generated by the original CAAE are also listed. The best value in each row (excluding column 2) is indicated in boldface. The results in column 2 prove the correctness of the DEX age estimations on the ground truth FGNET images and cannot directly be compared to the results in the remaining columns.
Table 9. Accuracy of age classification for the proposed xAI-CAAE using Saliency and SHAP explanations. Age prediction is performed using the pre-trained model in [18]. Accuracy for the original FGNET images and the images generated by the original CAAE are also listed. The best value in each row (excluding column 2) is indicated in boldface. The results in column 2 prove the correctness of the DEX age estimations on the ground truth FGNET images and cannot directly be compared to the results in the remaining columns.
Accuracy (%)
Age ClassFGNETCAAExAI-CAAE ( θ = 0 . 2 )xAI-CAAE ( θ = 0 . 5 )
SaliencySHAPSaliencySHAP
C 1 10.221.703.992.993.391.90
C 2 38.4135.0334.3339.0234.4340.72
C 3 52.6959.6855.9954.3961.4859.68
C 4 8.864.292.694.895.193.29
C 5 4.353.292.592.993.692.89
C 6 26.675.997.886.896.795.79
C 7 75.0010.0813.379.589.1810.88
C k , k = 1 , , 7 25.4517.1517.2717.2517.7417.88
Table 10. Results for 1-off accuracy of age classification for the proposed xAI-CAAE using Saliency and SHAP explanations. Age prediction is performed using the pre-trained model in [18]. The 1-off accuracy scores of age classification for the original FGNET images and the images generated by the original CAAE are also listed. The best value in each row (excluding column 2) is indicated in boldface. The results in column 2 provide proof of the correctness of the DEX age estimations on the ground truth FGNET images and cannot directly be compared to the results in the remaining columns.
Table 10. Results for 1-off accuracy of age classification for the proposed xAI-CAAE using Saliency and SHAP explanations. Age prediction is performed using the pre-trained model in [18]. The 1-off accuracy scores of age classification for the original FGNET images and the images generated by the original CAAE are also listed. The best value in each row (excluding column 2) is indicated in boldface. The results in column 2 provide proof of the correctness of the DEX age estimations on the ground truth FGNET images and cannot directly be compared to the results in the remaining columns.
1-Off Accuracy
Age ClassFGNETCAAExAI-CAAE ( θ = 0 . 2 )xAI-CAAE ( θ = 0 . 5 )
SaliencySHAPSaliencySHAP
C 1 52.8045.0143.7155.8951.0054.19
C 2 88.0489.0287.1391.0289.7292.71
C 3 80.8490.2287.8291.0292.8191.42
C 4 48.1066.3764.9764.3772.8569.06
C 5 54.3512.4815.5716.0713.0711.18
C 6 10018.2623.8519.2618.8615.97
C 7 75.0019.6622.0618.0616.7717.96
C k , k = 1 , , 7 67.7648.7249.3050.8150.7350.36
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Korgialas, C.; Pantraki, E.; Bolari, A.; Sotiroudi, M.; Kotropoulos, C. Face Aging by Explainable Conditional Adversarial Autoencoders. J. Imaging 2023, 9, 96. https://doi.org/10.3390/jimaging9050096

AMA Style

Korgialas C, Pantraki E, Bolari A, Sotiroudi M, Kotropoulos C. Face Aging by Explainable Conditional Adversarial Autoencoders. Journal of Imaging. 2023; 9(5):96. https://doi.org/10.3390/jimaging9050096

Chicago/Turabian Style

Korgialas, Christos, Evangelia Pantraki, Angeliki Bolari, Martha Sotiroudi, and Constantine Kotropoulos. 2023. "Face Aging by Explainable Conditional Adversarial Autoencoders" Journal of Imaging 9, no. 5: 96. https://doi.org/10.3390/jimaging9050096

APA Style

Korgialas, C., Pantraki, E., Bolari, A., Sotiroudi, M., & Kotropoulos, C. (2023). Face Aging by Explainable Conditional Adversarial Autoencoders. Journal of Imaging, 9(5), 96. https://doi.org/10.3390/jimaging9050096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop