Next Article in Journal
Apple Tree Leaves (Malus domestica Borkh) as a Valuable Source of Polyphenolic Compounds with a High Antioxidant Capacity
Previous Article in Journal
Influence of Monolith Length on Temperature Field of Concrete Gravity Dams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SupCon-MPL-DP: Supervised Contrastive Learning with Meta Pseudo Labels for Deepfake Image Detection

1
Department of Computer Engineering, Dong-A University, Busan 49315, Republic of Korea
2
Department of Information Convergence Engineering, Pusan National University, Busan 46241, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(8), 3249; https://doi.org/10.3390/app14083249
Submission received: 8 March 2024 / Revised: 6 April 2024 / Accepted: 6 April 2024 / Published: 12 April 2024
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Recently, there has been considerable research on deepfake detection. However, most existing methods face challenges in adapting to the advancements in new generative models within unknown domains. In addition, the emergence of new generative models capable of producing and editing high-quality images, such as diffusion, consistency, and LCM, poses a challenge for traditional deepfake training models. These advancements highlight the need for adapting and evolving existing deepfake detection techniques to effectively counter the threats posed by sophisticated image manipulation technologies. In this paper, our objective is to detect deepfake videos in unknown domains using unlabeled data. Specifically, our proposed approach employs Meta Pseudo Labels (MPL) with supervised contrastive learning, so-called SupCon-MPL, allowing the model to be trained on unlabeled images. MPL involves the simultaneous training of both a teacher model and a student model, where the teacher model generates pseudo labels utilized to train the student model. This method aims to enhance the adaptability and robustness of deepfake detection systems against emerging unknown domains. Supervised contrastive learning utilizes labels to compare samples within similar classes more intensively, while encouraging greater distinction from samples in dissimilar classes. This facilitates the learning of features in a diverse set of deepfake images by the model, consequently contributing to the performance of deepfake detection in unknown domains. When utilizing the ResNet50 model as the backbone, SupCon-MPL exhibited an improvement of 1.58% in accuracy compared with traditional MPL in known domain detection. Moreover, in the same generation of unknown domain detection, there was a 1.32% accuracy enhancement, while in the detection of post-generation unknown domains, there was an 8.74% increase in accuracy.

1. Introduction

Recently, with the advancement of generative artificial intelligence models [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15], deepfakes have become increasingly similar to real images/videos, making them difficult to distinguish. Deepfakes can be broadly categorized into three generations based on the evolution of generative models. First-generation deepfake generative models [1,2,3,4,5] typically attempt to generate simple and low-resolution images/videos based on probability distributions or synthesize multiple images/videos by exploiting features in tasks such as Face2Face and FaceSwap. In particular, generative adversarial network (GAN)-based models such as CGAN [2], WGAN [4], and WGAN-GP [5], as well as autoencoder-based models like VAE [6] and conditional VAE [7] have enabled the generation of various deepfake images/videos. However, first-generation deepfakes often exhibit noticeable artifacts that can be discerned by the human eye. With the transition to second-generation deepfake generative models [6,7,8], there has been progress in generating high-resolution deepfake images that are more difficult to distinguish compared with first-generation ones, along with performance improvements in various tasks. In particular, second-generation deepfake generative models like StyleGAN, proposed by T. Karras et al. [8], produce deepfake images that are difficult for the human eye to distinguish, excluding some flaws such as artifacts in hair. Second-generation deepfakes can be generated using deepfake generation tools such as DeepfaceLab [10], DeepSwap [11], Synthesia [12], and others. Finally, third-generation deepfake generative models [9,13,14,15] produce images/videos that are even more flexible and difficult to distinguish than those generated by second-generation models, across various tasks. In particular, Stable Diffusion, proposed by R. Rombach et al. [13], is currently being used for the generation of various human and artwork images, raising concerns related to copyright and human rights issues. Furthermore, the consistency model proposed by Y. Song et al. [15] has enabled state-of-the-art deepfake generative model training at a lower cost by reducing the extensive iteration process required by previous diffusion models for restoring original images from noise. As deepfakes increasingly become difficult to distinguish from real images/videos, they are being utilized in various criminal activities.
To address issues caused by deepfakes, methods have been proposed to identify flaws in landmarks that occur when deepfake generative models create images, aiding in the detection of deepfakes [16,17,18,19]. Meanwhile, recent advancements in deepfake detection for single models have shown improvement in detecting deepfake videos and images. D.A. Coccomini et al. [20] enhanced the detection performance of deepfake videos and images by combining EfficientNet [21] and Vision Transformer [22] when training a single model. In other words, existing deepfake detection models [16,17,18,19,20,21,22,23] verify flaws in facial landmarks during the preprocessing stage and construct large models for flexible predictions.
Previous studies have primarily focused on the detection performance of labeled known domain (known domain) tasks in deepfake detection. However, deepfake generation models are rapidly evolving, and similar generations of deepfake generation models are also being developed diversely. Therefore, detecting deepfake images in unknown domains (unknown domain) is also crucial. A few studies have proposed generalized deepfake detection models using techniques such as contrastive learning, meta learning, and others [24,25,26,27,28,29,30,31,32,33,34,35,36].
In this paper, we propose SupCon-MPL, a combination of the Meta Pseudo Labels (MPL) [37] with supervised contrastive learning (SupCon) [38] to further train the model with unlabeled images/videos, simultaneously enhancing the model’s generalization ability to distinguish deepfakes in unknown domains. The proposed SupCon-MPL utilizes the basic structure of MPL, where two models, namely teacher and student, are simultaneously trained. Each model influences the other during training. The teacher model constructs pseudo labels for unlabeled images and transfers them to the student model. Through this approach, the student model learns from unlabeled data, providing the potential to train effectively with limited labeled data. Furthermore, during the training process, we apply the supervised contrastive loss (SupConLoss) [38] to the encoder of each model, enabling contrastive representation learning, thereby inducing generalized model training.
The performance evaluation experiments were conducted in two parts: model validation experiments and deepfake detection experiment based on scenario. In the model validation experiments, we utilized the data from five domains within FaceForensics++ [39]. We evaluated the detection performance in labeled known domains by combining the data in various ways and assessed the generalized detection performance in unknown domains. The deepfake detection experiment based on scenario involves training the model with first-generation deepfake datasets (FaceForensics++ [39], DFDC [40], Celeb-DF [41]) and evaluating the detection performance on first- and second-generation unknown deepfake datasets (StyleGAN [8], NeuralTextures [39]). The experimental results showed that SupCon-MPL achieved performance improvements of 1.58%, 1.32%, and 8.74% over the baseline MPL model in the proposed evaluation scenario, respectively. The main contributions of this paper are as follows:
(1)
The proposed method enables additional training through unlabeled data. Especially, while two models are trained simultaneously, the student model infers information about unlabeled data from the other model and provides feedback, allowing for additional training to be conducted with less bias towards a specific model. Ultimately, the proposed method enhances the performance of deepfake detection by enabling additional training with a large amount of unlabeled data.
(2)
Our model enables generalized deepfake detection model training through contrastive learning. We improved the generalized deepfake detection performance on unknown data, which was previously low in the Meta Pseudo Labels-based deepfake detection model [24], through contrastive learning.
(3)
Our model exhibited higher deepfake detection performance compared with all other models in the comparison with various generalized deepfake detection models [24,31,34,35]. The experimental results demonstrate that our model outperforms existing deepfake detection models, showing robust detection capability across diverse labeled datasets and even unknown generational deepfakes.

2. Related Works

As deepfake generation models advance, various detection methods have also been researched. A common approach in deepfake detection is to explore flaws in facial images [16,17,18,19]. However, the continual development of new generative models has led to the problem of being unable to train detection models using data from all generative models. To address this problem, a few studies have explored training generalized deepfake detection models [24,25,26,27,28,29,30,31,32,33,34,35,36].

2.1. Generalized Deepfake Detection

The generalization of deepfake detection implies the ability to detect deepfake videos generated not only by the models used during training but also by unseen or new generative models. In other word, as generative models progress from GANs and VAEs to diffusion and consistency models, achieving the ability to detect deepfakes generated by various and new generative models simultaneously is the main goal of generalized deepfake detection techniques. Recently, research has been conducted on detecting deepfake videos that are unknown from both the data and training perspectives.
On the data perspective, SBL [29] and OST [30] enhanced the generalization of deepfake detection by synthesizing additional training data by combining original images from each generative model with various other images and selectively using them. On the training perspective, A. Jain et al. [25] utilized datasets from Google, Jigsaw, FaceForensics++ [39], Celeb-DF [41], Deepfake-TIMIT [42], and their own database DF-Mobio to train a generalized deepfake detection model using contrastive representation learning across various domains. A. Nadimpalli et al. [26] proposed a hybrid learning technique combining supervised learning and reinforcement learning. In particular, during the training process, the reinforcement learning agent selects the top k augmentations that have the most significant impact on performance improvement when training through convolutional neural networks (CNN) and uses them for testing, enabling the training of a generalized deepfake detection model. We employed the meta-learning technique, Meta Pseudo Labels, in the deepfake training process, applying it after domain splitting for each data, resulting in training a model with higher performance in the same model training [24], and conducted experiments using various CNN-based models specialized in extracting features from images, including EfficientNet [21], ResNet [43], ResNext [44], and WideResNet [45], which are image classification models.

2.2. Contrastive Representation Learning in Deepfake Detection

Currently, research applying contrastive representation learning (CRL) for training generalized deepfake detection models is conducted. CRL enables learning similar features in the feature space between a specific image and from the same domain (positive images), while also learning features that differentiate from different domains (negative images). H. Chih-Chung et al. [27] utilized contrastive loss [43] to train the encoder, following which they trained the classifier to generalize the discriminative performance on deepfake images generated by various GAN-based models. S. Fung et al. [36] trained the encoder using unsupervised CRL with image pairs that include random augmentation applied to the same image during the training process. Following this, they trained the classifier using labeled images to develop a generalized deepfake detection model. X. Ying et al. [35] addressed the issue of conventional CRL techniques not utilizing label information of deepfake images by applying supervised contrastive learning (SupCon) [38], which uses label information. However, CRL requires a large amount of data, especially a significant number of negative samples. In this paper, we performed CRL using SupCon [38], while simultaneously combining the SupCon model with meta-learning method, MPL [37]. This approach also allows for additional CRL training on unlabeled data, even when using the same labeled data, enabling the training of a more generalized deepfake detection model compared with conventional Meta Pseudo Labels.

2.3. Meta Pseudo Labels

Meta Pseudo Labels (MPL) [37] trains the model using unlabeled images, and when the same model is trained on an image classification task, it has shown improved performance compared with conventional models. MPL gained significant attention by achieving over 90% Top 1 score on the ImageNet [44] classification task, marking a significant milestone. Figure 1a shows the MPL facilitates the learning of the teacher model through feedback from the student model, thus enhancing the conventional learning techniques such as knowledge distillation [45] or noisy student [46], where the teacher model passes on information to the student model. This improvement addresses the issue of inadequate learning of the student model when the performance of the teacher model is subpar. The training process of the MPL is shown in Figure 1b. The student model in MPL learns through the Pseudo Labels inferred by the teacher model. Subsequently, it imparts the feedback value regarding the learning to the teacher model. The teacher model learns through the labeled loss from the labeled data, UDA loss [47], feedback from the student model, and MPL loss from the unlabeled data. However, it consumes substantial computing resources due to the simultaneous training of the two models.
From the perspective of training a generalized deepfake detection model, MPL can enhance the performance of generalized deepfake detection by enabling additional training through unlabeled data, compared with models trained solely with labeled data. In this paper, we experiment with the enhancement of detection capabilities for unknown domains and post-generation deepfakes, using both MPL [37] and SupCon [38].

3. Proposed SupCon-MPL-Based Deepfake Detection

To detect deepfake videos in the deepfake unknown domain, the proposed method introduces SupCon-MPL, a meta-learning model based on contrastive learning, utilizing unlabeled images from the deepfake known domain. You can find notations used in the rest of paper summarized in Table A1.

3.1. Proposed Training Strategy

3.1.1. Known Domain and Unknown Domain in Deepfake

A deepfake domain can be defined as a collection of images and their features, generated from a “specific deepfake generative model”. In this paper, we distinguish deepfake domains into the known domain ( K ) and unknown domain ( U ). The known domain ( K ) refers to a collection of deepfake images that are labeled when training models. The data in K is labeled and therefore can be directly used for training. Meanwhile, the unknown domain ( U ) refers to data created by unknown deepfake generative models. The data in U is not labeled, hence it is not possible to determine whether the image is real or fake. Also, as they are created from various generative models, they can involve various features. The known domain K can be defined as K = { K 1 , K 2 , } where K i is i-th known deepfake generative model, and the deepfake dataset D K = { D K 1 , D K 2 , } consists of a dataset D K i = { x i , y i } composed with a set of deepfake images x i and labels y i generated by K i . On the other hand, the unknown domain U can be defined as U = { U 1 , U 2 , } where U i is i-th unknown deepfake generative model, and the deepfake dataset D U = { D U 1 , D U 2 , } consists of a dataset D U i = { x i } composed with a set of deepfake images x i generated by U i .
In this paper, to address U , we first experiment by distinguishing D K into a labeled dataset ( D L ) and an unlabeled dataset ( D U L ), as shown in Figure 2a. Subsequently, to verify the influence of D U on the training process, we assume D U as D U L and perform experiments, as shown in Figure 2b. In the deepfake training scenario, from the perspective of generative models by generation, both D L and D U L constitute with first-generation D K , and evaluation is conducted using the first- and second-generation D U .

3.1.2. Training Strategy for Deepfake Unknown Domain Detection

The training process is employed based on a comparison between the base model and the student model of the MPL (SupCon-MPL). Upon completion of training the base model with the entire dataset D K , the model is subsequently employed as the teacher model to train the student model. In other words, we aim to verify performance improvement when training the model under the same conditions. If performance enhancement is validated at this method, it suggests that superior performing models can be trained under identical training conditions, even when employing larger or state-of-the-art (SOTA) models.
The training images are constructed considering the problems of existing deepfake detection. While deepfakes by known generative models exist in K , deepfake images by unknown generative models also exist in U . Therefore, during the training phase, we enhance the deepfake detection performance in K using labeled data and contribute to the generalization of the learning model by using data D K and D U from K and U as unlabeled data, respectively. Consistent with this approach, the data D L and D U L are structured into D K , with images from dataset D U serving as D U L .
In the proposed method, we combine data in three strategies to detect the unknown domain dataset D U . The first strategy is to use the data from D L and D U L as the same domain, aiming to verify whether unlabeled data from a specific K contributes to the improvement of model performance. Figure 2a illustrates the training strategy of using D K as unlabeled data. The second strategy aims to solve the realistic deepfake problem by experimenting with the impact of unlabeled data on the detection performance of the corresponding domain. Figure 2b illustrates the feasibility of improving model performance by employing dataset D K as labeled data D L and dataset D U as unlabeled data D U L . Finally, in the deepfake scenario experiment, after training the model using the first-generation deepfake dataset as D L and D U L , the generalized deepfake detection model learning is assessed through the first-generation D K and the first- and second-generation D U .

3.2. SupCon-MPL: Supervised Contrastive Learning with Meta Pseudo Labels

In the proposed method, following the strategy in Figure 2b, the MPL model is trained for the detection of deepfakes in the unknown domain U . SupCon-MPL allows supplementary training utilizing unlabeled videos, and with the aid of CRL, it enhances the deepfake detection in feature space. Furthermore, it affords the flexibility to employ diverse encoder models during the training phase and enables the fine tuning of the SupCon-MPL-trained model.
In particular, the limitations of deepfake detection with limited labeled data can be mitigated by using unlabeled data, and a generalized detection model can be trained through CRL. Another notable advantage lies in the capability to conduct concurrent learning via feedback from the student model, even if the performance of the T model is low. The details of the proposed method are elucidated in Figure 3. The most significant distinction from the conventional MPL and SupCon model training is that learning through unlabeled data not only resolves the training issue of CRL due to limited data but also enhances detection capabilities in both K and U . Ultimately, the final goal is to enhance the detection capabilities of deepfakes in domains that are not targeted, especially in a situation where new deepfake models in U continue to be developed.
Supcon-MPL consists of a teacher model T and a student model S . Each model has the same structure but different parameter values. The teacher model ( T ) utilizes a pre-trained model, while the student model ( S ) starts its training from the initial state before being trained. Each model consists of an identical structure of an encoder and a classifier. The encoder is modified to extract a 128-dimensional feature by removing the classifier layer of a specific model, allowing it to primarily learn the representations of labeled and unlabeled images. The classifier is composed of a single linear layer, which performs classification based on the representation values received from the encoder.
The training of the SupCon-MPL is conducted by first having the student model T use unlabeled data to perform CRL, followed by fine tuning with labeled data. In this process, the teacher model T ’s classifier learns through the feedback from S , while S learns dependently on T .

3.3. SupCon-MPL Loss Function

SupCon-MPL, as shown in Figure 3, is composed of a teacher model ( T ) and a student model ( S ), each of which consists of an encoder and a linear classifier. SupCon-MPL has two loss functions in order to sequentially train each model. One involves the teacher model T distilling knowledge to the student model S , while the other entails the teacher model T training from the feedback factor provided by S on the labeled data. The knowledge distilled by T includes previously learned content about deepfakes.
In SupCon-MPL, let the parameters of T classifier and S classifier be θ T , θ S , respectively, and denote the batch of images and labels on the labeled data as x l , y l D K , and the batch of images on the unlabeled data as x u D U . The goal of SupCon-MPL is to minimize the parameters θ S P L of the generalized deepfake detection model S :
θ S P L = argmin θ S E x u C E T x u ; θ T , S x u ; θ S L u θ T , θ S
Hence, the objective function of SupCon-MPL is defined as follows.
L l w i t h r e s p e c t i v e t o θ T : min θ T L l θ S P L θ T , w h e r e θ S P L θ T = argmin θ S L u θ T , θ S .
For optimization, SupCon-MPL approximates θ S P L θ T by the learning rate η s , and then,
θ S P L θ T θ S η s θ S L u θ T , θ S
defines the final objective function as follows:
L l w i t h r e s p e c t i v e t o θ T : min θ T L l θ S η s θ S L u θ T , θ S , w h e r e θ S P L θ T = argmin θ S L u θ T , θ S .
Both T and S consist of an encoder and a classifier and are trained according to their respective loss functions. The loss function of E N C T , the encoder of T , is composed of SupConLoss [38], and the loss function of C L F T , the classifier, is composed of labeled loss for the labeled data x l and MPL loss, reflecting the feedback from S . First and foremost, g T , c o n t r a s t t the loss function of E N C T , receives image pairs R a n d A u g m e n t a x l , R a n d A u g m e n t b x l , y l as inputs that reflect different random augmentations on the same image x l and the label y l . Subsequently, the loss value is obtained by passing image pairs through SupConLoss [38]. At this juncture, given the similarity between the current training process and that of the original MPL’s UDA loss, the utilization of the UDA loss is no longer used:
g T , c o n t r a s t t = θ T S u p C o n L o s s R a n d A u g m e n t a x l , R a n d A u g m e n t b x l , y l | θ T = θ T t
E N C T is promptly updated following the computation of the g T , c o n t r a s t t :
θ T , E N C t + 1 = θ T , E N C t η S g T , c o n t r a s t t
The labeled loss of C l F T , g T , s u p e r v i s e d t , measures the difference between y l and the label predicted by T through cross-entropy loss (CE Loss). Here, e m b l T denotes the embedding value derived by passing the labeled data x l through E N C T :
g T , s u p e r v i s e d t = θ T C E y l , C L F T e m b l T ; θ T | θ T = θ T t
The MPL loss g T t calculates the difference between the hard pseudo label y u , which is the maximum value extracted from the pseudo labels generated by T through x u , and the logit. Here, e m b u T denotes the embedding value derived by passing the labeled data x u through E N C T :
g T t = h θ T C E y ^ u , C L F T e m b u T ; θ T | θ T = θ T t
The feedback factor h of S was calculated in the same way as the original Meta Pseudo Labels [37], using Taylor expansion to calculate the difference before and after the training of S . In the proposed method, we approximated h using the difference from the CE loss value for the labeled data after S was trained to the value before training. This allows the final loss value to converge as the training progresses.
h = C E y l , S x l ; θ S t + 1 C E y l , S x l ; θ S t
The final loss function of C l F T is composed of the sum of each loss function value:
θ T t + 1 = θ T t η s g T t + g T , s u p e r v i s e d t
S is trained through unlabeled data. The loss function of the student model’s encoder E N C S , denoted as g S , c o n t r a s t t , is trained utilizing SupConLoss [38], akin to E N C T . It leverages x u , y ^ u , comprising an unlabeled image x u paired with pseudo labels y ^ u , generated by T :
g S , c o n t r a s t ( t ) = θ T S u p C o n L o s s R a n d A u g m e n t a x u , R a n d A u g m e n t b x u , y ^ u | θ S = θ S ( t )
E N C S is also promptly updated following the computation of the g S , c o n t r a s t t :
θ S , E N C t + 1 = θ S , E N C t η S g S , c o n t r a s t t
The loss function of C L F S is calculated using CE loss for the hard pseudo label y ^ u of T for x u and the prediction of S . Here, e m b u S denotes the embedding value derived by passing the labeled data x u through E N C S :
θ S t + 1 = θ S t η s θ S C E y ^ u , C L F S e m b u S ; θ S | θ S = θ S t
The SupConLoss in g T , c o n t r a s t t , g S , c o n t r a s t t on the teacher model T and student model S are as follows:
S u p C o n L o s s = i I S u p C o n L o s s i = i I 1 P i p P i l o g e x p z i z p / τ a A i e x p z i z a / τ
Here, i I { 1 , , 2 N } is the index of the randomly augmented data, and P i { p A i : y p = y i } is the set of indices for all positives in the batch (since y p and y i are labels of images that have been randomly augmented from y l , they are the same as y l ). z i and z p are the embedding values of each randomly augmented image in R a n d A u g m e n t a x l , R a n d A u g m e n t b x l passed through the encoder E N C , and A i I \ { i } , and τ is the temperature parameter. In other words, the inner product between positive pairs ( i and p are the same class but different samples) is maximized through e x p z i z p / τ , and the inner product between negative pairs is minimized through e x p z i z a / τ , so that the SupConLoss is minimized.
The training process of the proposed SupCon-MPL model for deepfake detection is shown in Algorithm 1.
Algorithm 1 The Deepfake detection method based on SupCon-MPL (Pseudo code)
Set Labeled data, Unlabeled data with domain splitting [37].
Input: Labeled data x l , y l and unlabeled data x u .
Initialize θ T ( 0 ) and θ S ( 0 ) .
Pretrain Teacher model with x l , y l .
For  t = 0 to N 1  do
Sample an unlabeled example x u and a labeled example x l , y l .
Sample a pseudo label y ^ u P x u ; θ T .
Compute contrastive loss of student encoder E N C S using the pseudo label y ^ u :
g S , c o n t r a s t t = θ T S u p C o n L o s s R a n d A u g m e n t a x u , R a n d A u g m e n t b x u , y ^ u | θ S = θ S t
Update the student encoder E N C S using the pseudo label y ^ :
θ S , E N C ( t + 1 ) = θ S , E N C ( t ) η S g S ,   contrast ( t )
Update the student classifier C L F S using the pseudo label y ˆ u :
θ S ( t + 1 ) = θ S ( t ) η s θ S C E y ˆ u , C L F S e m b u S ; θ S θ S = θ S ( t )
Compute contrastive loss of teacher encoder E N C T using the labeled data x l , y l :
g T , c o n t r a s t t = θ T S u p C o n L o s s R a n d A u g m e n t a x l , R a n d A u g m e n t b x l , y l | θ T = θ T t
Update the teacher encoder E N C T using the labeled data x l , y l :
θ T , E N C ( t + 1 ) = θ T , E N C ( t ) η S g T ,   contrast   ( t )
Compute gradient on labeled data x l , y l :
g T ,   supervised   ( t ) = θ T C E y l , C L F T e m b l T ; θ T θ T = θ T ( t )
Compute feedback factor h from student:
h = C E y l , S x l ; θ S ( t + 1 ) C E y l , S x l ; θ S ( t )
Compute MPL loss from unlabeled data x l :
g T ( t ) = h θ T C E y ˆ u , C L F T e m b u T ; θ T θ T = θ T ( t )
Update the teacher classifier C L F T :
θ T ( t + 1 ) = θ T ( t ) η s g T ( t ) + g T ,   supervised   ( t )
end for
return  θ S , E N C ( N ) , θ S ( N )          Only the student encoder and classifier are returned for evaluations.

4. Experiment

In this chapter, we first describe the experimental setup. Subsequently, we present the experimental results for Figure 2a,b in Section 4.2 and Section 4.3, respectively. Our main results consist of comparisons with the pretrained model and SupCon model in Section 4.4 and comparisons with state-of-the-art models in Section 4.5.

4.1. Experiment Setup

The experiments are conducted using NVIDIA Tesla V100 32 and NVIDIA RTX-3090 (NVIDIA, Santa Clara, CA, USA) with ubuntu 20.04 environment for reproducibility and stability. The single-domain experiment and the multi-domain experiment are existing outputs of the Meta Learning-based Deepfake Detection Project [24]. The pretrained model, Meta Pseudo Labels model (MPL model), SupCon model, and SupCon-MPL model are experimented for their training performance under the same conditions and hyperparameters. The training dataset uses the videos of the deepfakes (DF), Face2Face (F2F), FaceSwap (FS), NeuralTextures (NT), and real videos in FaceForensics++ [39], with DFDC [40], and Celeb-DF [41]. In scenario evaluation, deepfake videos of first-generation’s unknown domain are NeuralTextures(NT) [39] with real videos, and for post-generation’s unknown domain, we selected StyleGAN [8] images with CelebA [48] videos. We used MTCNN [49] to extract face images frame by frame of each video.
In the single-domain experiment, 260,000 real and 340,000 fake data from each domain in the FaceForensics++ [39] are used. During training, the amount of validation data used is 20% of the training data, and the evaluation dataset uses 150,000 per each data domain. In the multi-domain experiment, 200,000 labeled data are randomly extracted from four domains, and 180,000 unlabeled data are extracted from a single domain for use.
The generational deepfake scenario trains using 170,000 each of the first-generation known domain’s FaceForensics++ (DF, F2F, FS, Real) [39], DFDC [40], Celeb-DF [41] data, and then evaluates using 51,200 each of first- and second-generation unknown domain data. In the backbone model in scenario evaluation, we used ResNet50 [50] due to lack of computational resources.
The hyperparameters used in the experiment are a learning rate of 10−4, an image size of 64, and a batch size of 512. In the MPL, SupCon-MPL models, the batch sizes of labeled and unlabeled images are 64 and 448, respectively. Finally, the threshold is set at 0.95. The training models used were ResNet50 [50], ResNet101 [50], ResNext50 [51], EfficientNet-b5 [21], and WideResNet50 [52].
Experiment data and evaluation data use a mix of fake and real data. In the experiment in Section 4.2, video data from one domain is used as labeled and unlabeled data, and in the experiment in Section 4.3, videos from multiple domains are used as labeled data, and one domain is used as unlabeled data.

4.2. Single-Domain Experiment

In this section, single-domain experiments evaluate the detection performance for a single deepfake generation model, while, in Section 4.3, the performance on new deepfakes is evaluated through training on various deepfake generation models. Section 4.4 and Section 4.5 assess the performance of deepfake detection across diverse deepfake datasets and scenario-based deepfake detection, respectively.
In the experiment using only one domain, as shown in Table 1, the performance of the known domain increased in most of the situation. Furthermore, it was confirmed that the performance in unknown domains also increased in most of the situation. Based on EfficientNet-b5 [21] in Table 1, ACC and AUC in K improved by an average of 4.47% and 4.53%, respectively, and ACC in U improved by an average of 0.20%, but AUC decreased by an average of 0.13%. However, overall, out of a total of 64 ACC and AUC validations recorded in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8, 44 and 41 cases improved, respectively. Through this, it was confirmed that when using the same data, the performance of the MPL model is higher than the pretrained model.

4.3. Multi-Domain Experiment

In the multi-domain experiment, the combination of data is configured considering the actual deepfake situation. The situation is assumed to have K deepfake data from multiple domains and D U L data from U . Afterwards, the evaluation is conducted through the U data used as D U L . Therefore, MPL trains with deepfake videos from multiple domains, and, after training, MPL experiments the ability to detect unlabeled data using unlabeled data. As a result of the experiment, ACC and AUC increased by an average of 1.59% and 1.26% in two models in Table 9 and Table 10. Through this, it was confirmed that the MPL model improved the deepfake detection ability of the unknown model by learning with unlabeled data.

4.4. SupCon-MPL Experiment

The experiment uses the labeled data identically to the multi-domain experiment and evaluates according to each combination of known domain and unknown domain. At this time, Celeb-DF [41] is used as unlabeled data in the training of the SupCon-MPL model. As a result of the experiment, shown in Table 11, when evaluating FS data as an unknown domain compared with the SupCon model [35], ACC and AUC decreased, but in other validations, the performance of the SupCon-MPL model was similar or higher than the performance of the two models being compared. This shows that the SupCon-MPL model has been trained as a generalized detection model compared with the existing deepfake detection model.

4.5. Deepfake Scenario Experiment

In this section, we construct a training scenario for a deepfake detection model in the real world and train the model. The scenario involves training a deepfake model with first-generation deepfake data, then experimenting with the detection of first-generation deepfakes (NT) that were not participated while training, and post-generation (second-generation) deepfakes (StyleGAN) that are newly developed and unknown. Table 12 shows the training results of various models according to the scenario. As a result of the experiment, among various models, the SupCon-MPL model achieved highest performance in all scenario evaluations compared with other models.

4.6. Limitations

The main goal of the SupCon-MPL is to enhance the deepfake detection performance in unknown domains using meta-learning. Therefore, in this paper, we conducted experiments by reconfiguring a limited deepfake dataset into scenarios.
The main limitation is related to computing resources. As the training in Section 4.3. and Section 4.4. was conducted using NVIDIA RTX-3090, only ResNet50 [50] could be used as the backbone model in SupCon-MPL, which uses two models. The ResNet50 [50] model currently stands out among various image classification models for its stability and consistently decent performance scores. Additionally, within the given resources, it can be utilized for training the SupCon-MPL model. Subsequent experiments are needed with various backbone models and larger image sizes through more computing resources.
The next limitation is that we could not find a verified deepfake dataset for the third generation and higher. Further experiments are needed through the corresponding dataset in the future.

5. Conclusions

With the development of various deepfake generative models, it has become important to develop a generalized deepfake detection model that guarantees the detection performance of unknown domain deepfakes, not just the data of the domain used for training. The proposed SupCon-MPL allows for training from unlabeled data and performs contrastive learning to enhance the detection performance of unknown deepfakes. This enables contrastive learning with a diverse range of data, and, during training, the teacher model and the student model infer information and provide feedback to each other, facilitating the student model to surpass the performance of the teacher model.
Our model improved all detection performances over other generalized deepfake detection models’ known and unknown domains in deepfake scenario evaluation. Indeed, one of the significant features of SupCon-MPL is its ability to train models using a large amount of unlabeled data. This provides a method to enhance the model’s performance utilizing the countless images and videos available on the internet and so on. Through this, our model enables the training of a generalized deepfake detection model and provides robust detection capability for real-world deepfakes. SupCon-MPL can contribute to the detection of newly developed deepfake generative models as they become increasingly difficult to distinguish from real images, especially with the development of various deepfake generative models.
Future research will focus on improving the detection performance of higher generation deepfake images/videos using these methods. Additionally, studies on reducing the training cost of SupCon-MPL will be conducted. The goal is to develop a more efficient and economical deepfake detection model through these efforts.

Author Contributions

Conceptualization, K.-H.M. and S.-H.L.; Methodology, K.-H.M., S.-Y.O. and S.-H.L.; Software, K.-H.M.; Validation, K.-H.M.; Resources, S.-Y.O.; Writing—original draft, K.-H.M.; Writing—review & editing, S.-H.L.; Supervision, S.-Y.O. and S.-H.L.; Project administration, K.-H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Dong-A University research fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Summary of notations.
Table A1. Summary of notations.
SymbolDefinition
CRLContrastive Representation Learning.
MPLMeta Pseudo Labels.
GANGenerative Adversarial Networks.
VAEVariational Auto Encoder.
K Known domain which involves a set of known deepfake generative models.
U Unknown domain which involves a set of unknown deepfake generative models.
D K Deepfake datasets of the known domain which are labeled.
D U Deepfake datasets of the unknown domain which are unlabeled.
K i i -th known deepfake generative model in known domain.
U i i -th unknown deepfake generative model in unknown domain.
x i Images of real and deepfake.
y i Labels of real and deepfake.
D L Labeled dataset used while training.
D U L Unlabeled dataset used while training.
T Teacher model of SupCon-MPL.
S Student model of SupCon-MPL.
θ T Parameters of the teacher model’s classifier.
θ S Parameters of the student model’s classifier.
θ S P L Minimized parameters of model in SupCon-MPL.
θ T , E N C t Parameters of teacher encoder.
θ S , E N C t Parameters of student encoder.
η s Learning rate used while training.
C l F T Classifier in the teacher model.
E N C T Encoder in the teacher model.
C L F S Classifier in the student model.
E N C S Encoder in the student model.
x l Labeled images used while training.
x u Unlabeled images used while training.
y l Labels used while training.
y ^ u Pseudo label generated by the teacher model.
e m b l T Embedding value derived by passing the labeled data through the teacher encoder.
e m b u S Embedding value derived by passing the unlabeled data through the student encoder.
h Student model’s feedback factor.
I Index of the randomly augmented data.
P i Set of indices for all positives in the batch.
z i ,   z p Embedding values of each randomly augmented image.

References

  1. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 1–9. [Google Scholar] [CrossRef]
  2. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  3. Choi, Y.; Choi, M.; Kim, M.; Ha, J.W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar] [CrossRef]
  4. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  5. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of wasserstein gans. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  6. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  7. Sohn, K.; Lee, H.; Yan, X. Learning structured output representation using deep conditional generative models. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar]
  8. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4401–4410. [Google Scholar] [CrossRef]
  9. Ramesh, A.; Pavlov, M.; Goh, G.; Gray, S.; Voss, C.; Radford, A.; Sutskever, I. Zero-shot text-to-image generation. In Proceedings of the ICML 2021 Workshop on Unsupervised Reinforcement Learning, Virtual, 18–24 July 2021; pp. 8821–8831. [Google Scholar]
  10. DeepFaceLab. Available online: https://github.com/iperov/DeepFaceLab (accessed on 5 March 2024).
  11. Deepswap. Available online: https://deepfaceswap.ai/ (accessed on 5 March 2024).
  12. Synthesia. Available online: https://www.synthesia.io (accessed on 5 March 2024).
  13. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10684–10695. [Google Scholar] [CrossRef]
  14. Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.L.; Norouzi, M. Photorealistic text-to-image diffusion models with deep language understanding. Adv. Neural Inf. Process. Syst. 2022, 35, 36479–36494. [Google Scholar]
  15. Song, Y.; Dhariwal, P.; Chen, M.; Sutskever, I. Consistency models. arXiv 2023, arXiv:2303.01469. [Google Scholar]
  16. Li, Y.; Lyu, S. Exposing deepfake videos by detecting face warping artifacts. arXiv 2018, arXiv:1811.00656. [Google Scholar]
  17. Matern, F.; Riess, C.; Stamminger, M. Exploiting visual artifacts to expose deepfakes and face manipulations. In Proceedings of the 2019 IEEE Winter Applications of Computer Vision Workshops, Waikoloa, HI, USA, 7–11 June 2019; pp. 83–92. [Google Scholar] [CrossRef]
  18. Li, Y.; Chang, M.C.; Lyu, S. In ictu oculi: Exposing ai generated fake face videos by detecting eye blinking. arXiv 2018, arXiv:1806.02877. [Google Scholar]
  19. Ciftci, U.A.; Demir, I.; Yin, L. Fakecatcher: Detection of synthetic portrait videos using biological signals. IEEE Trans. Pattern Anal. Mach. Intell. 2020, early access. [CrossRef]
  20. Coccomini, D.A.; Messina, N.; Gennaro, C.; Falchi, F. Combining efficientnet and vision transformers for video deepfake detection. In Proceedings of the International Conference on Image Analysis and Processing, Lecce, Italy, 23–27 May 2022; pp. 219–229. [Google Scholar] [CrossRef]
  21. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Los Angeles, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  22. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Houlsby, N. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  23. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar] [CrossRef]
  24. Moon, K.-H.; Ok, S.-Y.; Seo, J.; Lee, S.-H. Meta Pseudo Labels Based Deepfake Video Detection. J. Korea Multimed. Soc. 2024, 27, 9–21. [Google Scholar] [CrossRef]
  25. Jain, A.; Korshunov, P.; Marcel, S. Improving generalization of deepfake detection by training for attribution. In Proceedings of the 2021 IEEE 23rd International Workshop on Multimedia Signal Processing, Tampere, Finland, 6–8 October 2021; pp. 1–6. [Google Scholar] [CrossRef]
  26. Nadimpalli, A.V.; Rattani, A. On improving cross-dataset generalization of deepfake detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 91–99. [Google Scholar] [CrossRef]
  27. Hsu, C.C.; Lee, C.Y.; Zhuang, Y.X. Learning to detect fake face images in the wild. In Proceedings of the 2018 International Symposium on Computer, Consumer and Control, Taichung, Taiwan, 6–8 December 2018; pp. 388–391. [Google Scholar] [CrossRef]
  28. Dong, F.; Zou, X.; Wang, J.; Liu, X. Contrastive learning-based general Deepfake detection with multi-scale RGB frequency clues. J. King Saud Univ.-Comput. Inf. Sci. 2023, 35, 90–99. [Google Scholar] [CrossRef]
  29. Shiohara, K.; Yamasaki, T. Detecting deepfakes with self-blended images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18720–18729. [Google Scholar] [CrossRef]
  30. Chen, L.; Zhang, Y.; Song, Y.; Wang, J.; Liu, L. Ost: Improving generalization of deepfake detection via one-shot test-time training. Adv. Neural Inf. Process. Syst. 2022, 35, 24597–24610. [Google Scholar]
  31. Aneja, S.; Nießner, M. Generalized zero and few-shot transfer for facial forgery detection. arXiv 2020, arXiv:2006.11863. [Google Scholar]
  32. Kim, M.; Tariq, S.; Woo, S.S. Fretal: Generalizing deepfake detection using knowledge distillation and representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1001–1012. [Google Scholar] [CrossRef]
  33. Qi, H.; Guo, Q.; Juefei-Xu, F.; Xie, X.; Ma, L.; Feng, W.; Zhao, J. Deeprhythm: Exposing deepfakes with attentional visual heartbeat rhythms. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 4318–4327. [Google Scholar] [CrossRef]
  34. Lee, S.; Tariq, S.; Kim, J.; Woo, S.S. Tar: Generalized forensic framework to detect deepfakes using weakly supervised learning. In Proceedings of the IFIP International Conference on ICT Systems Security and Privacy Protection, Oslo, Norway, 22–24 June 2021; pp. 351–366. [Google Scholar] [CrossRef]
  35. Xu, Y.; Raja, K.; Pedersen, M. Supervised contrastive learning for generalizable and explainable deepfakes detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 379–389. [Google Scholar]
  36. Fung, S.; Lu, X.; Zhang, C.; Li, C.T. DeepfakeUCL: Deepfake Detection via Unsupervised Contrastive Learning. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar] [CrossRef]
  37. Pham, H.; Dai, Z.; Xie, Q.; Le, Q.V. Meta pseudo labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11557–11568. [Google Scholar] [CrossRef]
  38. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Krishnan, D. Supervised contrastive learning. Adv. Neural Inf. Process. Syst. 2020, 33, 18661–18673. [Google Scholar]
  39. Rossler, A.; Cozzolino, D.; Verdoliva, L.; Riess, C.; Thies, J.; Niessner, M. Faceforensics++: Learning to Detect Manipulated Facial Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1–11. [Google Scholar] [CrossRef]
  40. Dolhansky, B.; Bitton, J.; Pflaum, B.; Lu, J.; Howes, R.; Wang, M.; Ferrer, C.C. The deepfake detection challenge (dfdc) dataset. arXiv 2020, arXiv:2006.07397. [Google Scholar]
  41. Li, Y.Z.; Yang, X.; Sun, P.; Qi, H.G.; Lyu, S. Celeb-DF: A Large-scale Challenging Dataset for DeepFake Forensics. In Proceedings of the Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3207–3216. [Google Scholar] [CrossRef]
  42. Korshunov, P.; Marcel, S. Deepfakes: A new threat to face recognition? assessment and detection. arXiv 2018, arXiv:1812.08685. [Google Scholar]
  43. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; pp. 1735–1742. [Google Scholar] [CrossRef]
  44. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  45. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  46. Xie, Q.; Luong, M.T.; Hovy, E.; Le, Q.V. Self-training with noisy student improves imagenet classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10687–10698. [Google Scholar] [CrossRef]
  47. Xie, Q.; Dai, Z.; Hovy, E.; Luong, T.; Le, Q. Unsupervised data augmentation for consistency training. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; pp. 6256–6268. [Google Scholar]
  48. Liu, Z.; Luo, P.; Wang, X.; Tang, X. Deep learning face attributes in the wild. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3730–3738. [Google Scholar] [CrossRef]
  49. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint Face Detection and Alignment Using Multitask Cascaded Convolutional Networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  51. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar] [CrossRef]
  52. Zagoruyko, S.; Komodakis, N. Wide residual networks. arXiv 2016, arXiv:1605.07146. [Google Scholar]
Figure 1. (a) Training overview and (b) process with label/unlabeled datasets of Meta Pseudo Labels [37].
Figure 1. (a) Training overview and (b) process with label/unlabeled datasets of Meta Pseudo Labels [37].
Applsci 14 03249 g001aApplsci 14 03249 g001b
Figure 2. Deepfake image discrimination strategy targeting for the (a) known domain and (b) unknown domain.
Figure 2. Deepfake image discrimination strategy targeting for the (a) known domain and (b) unknown domain.
Applsci 14 03249 g002
Figure 3. Modified Meta Pseudo Labels and loss functions for deepfake detection.
Figure 3. Modified Meta Pseudo Labels and loss functions for deepfake detection.
Applsci 14 03249 g003
Table 1. The performance of pretrained and MPL model on the known domain (EfficientNetb5 [21]).
Table 1. The performance of pretrained and MPL model on the known domain (EfficientNetb5 [21]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
EfficientNetb5 [21]DFDF89.3589.3589.1990.0890.0890.36
F2FF2F77.2177.2177.6180.3580.3579.45
FSFS84.5284.3183.2087.9087.4385.98
NTNT64.1363.3358.1874.7974.4771.36
Table 2. The performance of pretrained and MPL model on the unknown domain (EfficientNetb5 [21]).
Table 2. The performance of pretrained and MPL model on the unknown domain (EfficientNetb5 [21]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
EfficientNetb5 [21]DFF2F51.5951.6022.0951.7051.7023.01
FS55.9752.0623.2356.4552.6525.01
NT55.3751.2620.0655.4151.3721.28
F2FDF57.5657.5547.7853.8753.8432.11
FS56.4754.0539.6655.4851.9226.57
NT54.6351.9133.9654.7650.9923.26
FSDF57.2557.2239.8457.6157.8836.54
F2F51.7651.7725.7851.9851.9920.16
NT53.7949.8920.8054.3249.8012.35
NTDF58.7258.7152.7361.0361.0052.78
F2F54.5254.5345.2355.8855.8943.68
FS51.6549.4234.0353.2949.3427.25
Table 3. The performance of pretrained and MPL model on the known domain (ResNet50 [50]).
Table 3. The performance of pretrained and MPL model on the known domain (ResNet50 [50]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
ResNet50 [50]DFDF91.1691.1690.9191.2991.3090.97
F2FF2F82.4782.4881.8483.9883.9782.75
FSFS87.5587.2186.0188.3188.0987.06
NTNT74.9674.2871.2076.2275.8473.15
Table 4. The performance of pretrained and MPL model on the unknown domain (ResNet50 [50]).
Table 4. The performance of pretrained and MPL model on the unknown domain (ResNet50 [50]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
ResNet50 [50]DFF2F52.9152.8922.4151.9251.8620.36
FS56.8052.6821.4157.0253.0022.63
NT55.9151.7718.5455.4551.3318.53
F2FDF53.9453.9031.7253.7953.7528.28
FS54.6250.9724.4755.3251.3521.84
NT54.7651.2024.0754.9651.0421.00
FSDF56.3456.3033.8959.5759.5341.50
F2F51.1651.1020.0851.5751.5221.01
NT53.6749.4414.3253.8149.6115.49
NTDF60.4560.4349.9360.1760.1550.85
F2F54.7454.7038.8853.5653.5136.77
FS51.5948.1522.3051.8448.5125.20
Table 5. The performance of pretrained and MPL model on the known domain (ResNet101 [50]).
Table 5. The performance of pretrained and MPL model on the known domain (ResNet101 [50]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
ResNet101 [50]DFDF91.1691.1690.8491.1391.1390.85
F2FF2F81.4181.4180.2783.5083.4982.73
FSFS87.5987.3786.1387.7587.6686.39
NTNT74.1774.5673.1676.0375.5373.22
Table 6. The performance of pretrained and MPL model on the unknown domain (ResNet101 [50]).
Table 6. The performance of pretrained and MPL model on the unknown domain (ResNet101 [50]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
ResNet101 [50]DFF2F52.6952.6321.2251.2651.2018.01
FS56.9052.7320.4256.0251.8418.71
NT55.9351.7418.3354.9350.7316.53
F2FDF53.6353.5930.4453.5053.4627.42
FS54.3350.0822.8254.8650.8519.67
NT54.6251.0123.3055.1751.2820.98
FSDF57.0457.7037.0059.5559.5142.55
F2F51.4351.3720.6351.8451.7924.60
NT53.6649.5515.0453.6149.5917.59
NTDF62.6662.6559.0860.1660.1449.99
F2F52.6952.6347.5551.2651.2037.97
FS56.9052.7329.9856.0251.8421.29
Table 7. The performance of pretrained and MPL model on the known domain (ResNext50 [51]).
Table 7. The performance of pretrained and MPL model on the known domain (ResNext50 [51]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Sore
ResNext50 [51]DFDF90.2990.2990.0191.1491.1491.02
F2FF2F81.1981.1780.0782.8782.8582.11
FSFS87.0486.7785.4487.3687.3686.36
NTNT74.2674.4373.2176.3275.7873.30
Table 8. The performance of pretrained and MPL model on the unknown domain (ResNext50 [51]).
Table 8. The performance of pretrained and MPL model on the unknown domain (ResNext50 [51]).
Baseline ModelTrain DatasetTest DatasetPretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
ResNext50 [51]DFF2F51.8351.7320.1751.6651.5618.92
FS56.5752.5321.4857.2453.2123.08
NT55.2551.5718.5054.7250.5015.47
F2FDF54.5754.5831.5654.4654.4731.25
FS54.3150.3818.2155.2851.4522.28
NT54.6750.8921.1254.8051.0221.58
FSDF59.2359.2341.2560.3360.3346.22
F2F52.0251.9323.4352.9852.8927.91
NT53.3149.2215.0553.4149.5517.30
NTDF60.5860.5843.4658.5058.5144.82
F2F54.1654.1130.6952.4652.3831.00
FS49.1746.7819.3651.2647.8021.71
Table 9. The performance of pretrained and MPL model on the targeted unknown domain (ResNext50 [51]).
Table 9. The performance of pretrained and MPL model on the targeted unknown domain (ResNext50 [51]).
Baseline ModelTrain DatasetUnlabeled
Dataset
Pretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
ResNext50 [51]F2F, FS, NTDF63.3363.3270.0266.2366.2270.35
DF, FS, NTF2F57.3557.3960.6958.0458.0861.64
DF, F2F, NTFS50.6950.5346.6951.8151.5447.22
DF, F2F, FSNT53.6353.0547.6553.8953.0646.28
Table 10. The performance of pretrained and MPL model on the targeted unknown domain (WideResNet50 [52]).
Table 10. The performance of pretrained and MPL model on the targeted unknown domain (WideResNet50 [52]).
Baseline ModelTrain DatasetUnlabeled
Dataset
Pretrained ModelMPL Model
ACCAUCF1 ScoreACCAUCF1 Score
WideResNet50 [52]F2F, FS, NTDF63.1763.1469.1966.0366.0070.15
DF, FS, NTF2F56.8656.8760.7657.6057.6058.91
DF, F2F, NTFS48.3747.7439.3252.8651.1341.60
DF, F2F, FSNT54.2253.4347.3753.9251.9639.87
Table 11. The performance of SupCon-MPL compared with the baseline model and SupCon model [35].
Table 11. The performance of SupCon-MPL compared with the baseline model and SupCon model [35].
Baseline ModelTrain
Dataset
Unlabeled
Dataset
Pretrained
Model
SupCon
Model [35]
SupCon-MPL(Ours)
Model
ACCAUCF1 ScoreACCAUCF1 ScoreACCAUCF1 Score
ResNet50 [50]FF (without DF)DF (unknown)64.2464.2765.4962.8862.8458.5564.6064.5561.60
F2F + FS + NT (known)70.5671.3683.3675.4475.5283.0175.8476.0083.50
FF (without F2F)F2F (unknown)55.7655.6440.6356.6156.6449.4758.7458.7751.52
DF + FS + NT (known)77.2676.8981.6675.5475.5484.1378.1178.2885.84
FF (without FS)FS (unknown)54.4752.0779.1855.7553.6877.2055.7253.4179.29
DF + F2F + NT (known)75.9975.8781.6676.0275.8884.8477.2277.1285.84
FF (without NT)NT (unknown)56.7154.9562.6656.5854.3166.2756.7654.0269.43
DF + F2F + FS (known)77.3977.4181.6679.0979.2684.1381.2281.3285.84
Table 12. The performance of SupCon-MPL compared with other deepfake detection methods.
Table 12. The performance of SupCon-MPL compared with other deepfake detection methods.
ModelScenario Deepfakes
(Known)
Current-Generation
Deepfakes (Unknown)
Post-Generation
Deepfakes (Unknown)
Tar [34]52.4044.6249.96
DDT [31]80.4144.6249.49
MPL [24]79.8256.5343.16
SupCon [35]79.0156.6647.77
SupCon-MPL (ours)81.4057.8551.90
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moon, K.-H.; Ok, S.-Y.; Lee, S.-H. SupCon-MPL-DP: Supervised Contrastive Learning with Meta Pseudo Labels for Deepfake Image Detection. Appl. Sci. 2024, 14, 3249. https://doi.org/10.3390/app14083249

AMA Style

Moon K-H, Ok S-Y, Lee S-H. SupCon-MPL-DP: Supervised Contrastive Learning with Meta Pseudo Labels for Deepfake Image Detection. Applied Sciences. 2024; 14(8):3249. https://doi.org/10.3390/app14083249

Chicago/Turabian Style

Moon, Kyeong-Hwan, Soo-Yol Ok, and Suk-Hwan Lee. 2024. "SupCon-MPL-DP: Supervised Contrastive Learning with Meta Pseudo Labels for Deepfake Image Detection" Applied Sciences 14, no. 8: 3249. https://doi.org/10.3390/app14083249

APA Style

Moon, K. -H., Ok, S. -Y., & Lee, S. -H. (2024). SupCon-MPL-DP: Supervised Contrastive Learning with Meta Pseudo Labels for Deepfake Image Detection. Applied Sciences, 14(8), 3249. https://doi.org/10.3390/app14083249

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop