A Novel Data Sanitization Method Based on Dynamic Dataset Partition and Inspection Against Data Poisoning Attacks
Abstract
:1. Introduction
- We proposed a novel data sanitization method based on dynamic partitioning and inspection to defend against data poisoning attacks on deep learning (DL) models. Considering two data poisoning attack strategies (concentrated poisoning attacks and random poisoning attacks) [13,35,36], we design and implement two partition and inspection algorithms such as the Sequential Partitioning and Inspection Algorithm (SPIA) and the Randomized Partitioning and Inspection Algorithm (RPIA).
- We conducted two kinds of experiments in the Python environment (Experiment 1) and DL environment (Experiment 2) to validate and evaluate the defensive performance of our two proposed methods under concentrated poisoning attacks and random poisoning attacks. According to our experimental results, the SPIA completely removed all poisoned data under concentrated poisoning attacks in both Python and DL environments. In addition, the RPIA removed up to 91.1% and 99.1% of poisoned data under random poisoning attacks in Python and DL environments, respectively.
2. Background and Related Works
2.1. Adversarial Attacks
2.2. Poisoning Attacks on Deep Learning Models
2.3. Defense Approach Against Poisoning Attacks
2.4. Existing Methods Using Data Sanitization Approach
3. Proposed Defense Method
3.1. Basic Idea and Working Steps
3.2. Design
3.2.1. Motivation of n-Way Partition and Inspection
3.2.2. Two Partition and Inspection Algorithms: SPIA and RPIA
- Sequential partition and inspection algorithm (SPIA)
Algorithm 1: Sequential Partition and Inspection Algorithm (SPIA) | |||
Input: Mt1: a model pre-trained on a clean dataset Dc Dt: a new dataset of size m from an outer source δ: detection threshold n: the number of partitions | |||
Output: Du: a cleaned dataset after inspection | |||
1 | load model Mt1, Dt | ||
2 | # the base size of each dataset with floor function | ||
3 | r = m mod n | ||
4 | Du = [ ] | ||
5 | Partitions ← [ ] # List to store the split datasets | ||
6 | index ← 0 # Starting index for splitting | ||
7 | for i = 1 to r do | ||
8 | Partition ← D[index: index + q + 1] # Create a partition of size (q + 1) | ||
9 | Partitions.APPEND(Partition) | ||
10 | index ← index + q + 1 | ||
11 | end for | ||
12 | for i = r + 1 to n do | ||
13 | Partition ← D[index: index + q ] # Create a partition of size q | ||
14 | Partitions.APPEND(Partition) | ||
15 | index ← index + q | ||
16 | end for | ||
17 | for each Partition in Partitions do # Evaluate each partition | ||
18 | if accuracy(partition, Mt1) (Mt1.accuracy δ) then | ||
19 | Du ← Du + Partition | ||
20 | end if | ||
21 | end for | ||
22 | Return Du |
- 2.
- Random partition and inspection algorithm (RPIA)
Algorithm 2: Randomized Partition and Inspection Algorithm (RPIA) | ||||
Input: Mt1: a model pre-trained on a clean dataset Dc Dt: a new dataset of size m from an outer source δ: detection threshold n: the number of partitions k: the number of inspection iterations | ||||
Output: Du: a cleaned dataset after inspection | ||||
1 | load model Mt1, Dt | |||
2 | for iteration = 1 to k do | |||
3 | Partitions ← [ ] # List to store the split datasets | |||
4 | m ← size of Dt | |||
5 | # the base size of each dataset with floor function | |||
6 | r = m mod n | |||
7 | for i = 1 to r do | |||
8 | Partition ← randomly select a subset of size (q + 1) from Dt | |||
9 | Partitions.APPEND(Partition) | |||
10 | end for | |||
11 | for i = r + 1 to n do | |||
12 | Partition ← randomly select a subset of size q from Dt | |||
13 | Partitions.APPEND(Partition) | |||
14 | end for | |||
15 | for each Partiton in Partitions do | |||
16 | if accuracy(partition, Mt1) < (Mt1.accuracy δ) then | |||
17 | Dt ← Dt Partiton | |||
18 | end if | |||
19 | end for | |||
20 | end for | |||
21 | Du ← Dt | |||
21 | Return Du |
4. Experiments
- Experiment 1: Performance evaluation using Python simulation
- Experiment 2: Performance evaluation using DL model training
4.1. Experiment 1: Performance Evaluation Using Python Simulation
4.1.1. Experimental Setup and Procedure
- Removed poison rate (RPR): RPR represents the ratio (%) of successfully removed poisoned data to all poisoned data after applying the defense method. A higher RPR indicates more effective removal of poisoned data. RPR is the most important metric for evaluating the performance of the defense method.
- Attack success rate (ASR): ASR measures the success rate of the poisoning attack and is defined as the ratio (%) of non-removed poisoned data to all poisoned data after applying the defense method. Thus, ASR is equal to 100 − RPR. A lower ASR indicates better defense performance.
- Removed benign data rate (RBR): RBR indicates the proportion of benign data falsely removed during the partition and inspection and is defined as the ratio (%) of falsely removed benign data to all benign data. A lower RBR reflects more accurate detection performance.
- Accuracy: Accuracy measures how a DL model correctly predicts given input. Accuracy is calculated as (TP + TN)/(TP + TN + FP + FN) where TP denotes positive samples correctly classified as positive, TN denotes negative samples correctly classified as negative, FP denotes negative samples misclassified as positive, and FN denotes positive samples misclassified as negative. Thus, a higher accuracy indicates a better performance of the DL model. We note that since Experiment 1 is a Python simulation experiment to briefly analyze the partition and inspection performance of our proposed methods and thus DL model training is not conducted, and measuring accuracy is not used. Meanwhile, we use the accuracy metric in Experiment 2.
- Data structure and experimental design: To evaluate the feasibility of the proposed methods, we implemented a Python-based simulation and structured a data pool with integers ranging from 1 to 10,000. For each integer, we determined whether it was correctly detected or misclassified by the detection accuracy (81.34%) of a pre-trained model used in the DL experiment described in Section 4.2.
- Poisoning attack methods: To vary the distribution of poisoned data in a new dataset, we considered two poisoning attacks (concentrated poisoning attacks and random poisoning attacks). For concentrated poisoning attacks, the first 20% of the dataset was designated as poisoned data. In contrast, for random poisoning attacks, we selected and poisoned 20% of the dataset randomly.
- Partition methods for defense: For partition methods, two partition and inspection algorithms (SPIA and RPIA) were used and implemented based on Algorithm 1 and Algorithm 2, respectively. The detection threshold δ was set to 0.2. For the number of partitions n, we used 2, 4, 8, 10, 50, 100, 200, 500, 1000, 2000, 5000, and 10,000. For the RPIA, the number of iterations k was set to 100. The total number of data samples was 10,000.
- Inspection methods for defense: Partitioned datasets were inspected by using a pre-trained DL model Mt1 as a detector. To implement this process in the Python experiment, for data samples of each sub-dataset, we classified them according to the accuracy (81.34%) of a pre-trained model used in the DL experiment in Section 4.2. If the evaluated accuracy of a sub-dataset was lower than a detection threshold of 61.34%, then it was removed. After the inspection and removal processes were completed, the performance metrics were calculated for the refined dataset combining all remaining sub-datasets.
4.1.2. Results and Analysis
4.2. Experiment 2: Performance Evaluation Using DL Model Training
4.2.1. Setup and Procedure
- Target DL model: The target deep learning model used in this experiment is ResNet18, which is pre-trained on the CIFAR-10 dataset [45]. ResNet18 is a convolutional neural network (CNN) model based on residual blocks. A DL model trained with ResNet18 and CIFAR-10 can be used for developing applications based on computer vision tasks such as object recognition and image classification. It consists of 18 layers and approximately 11.1 million parameters. The CIFAR-10 dataset comprises 60,000 image samples (size: 32 × 32 pixels, the number of classes: 10), representing various objects such as birds, frogs, and airplanes (see Figure 9). For the experiments, an initial clean dataset Dc was created using 30,000 training images and 6000 test images from CIFAR-10. During transfer learning, an additional dataset Dt was used, which consists of 10,000 training images and 2000 test images.
- Poisoning attack methods: With a poison rate of 20%, we poisoned 2000 out of 10,000 samples in the dataset such that the labels of the poisoned data were flipped (dirty-label attack) and the poisoned data were distributed by using concentrated poisoning attacks and random poisoning attacks. For concentrated poisoning attacks, the goal was to target specific classes (class 0 and class 1); to this end, the entire dataset was sorted by class order, and then we manipulated the labels of the first 2000 samples randomly. Random poisoning attacks are used to degrade the model’s reliability. In this case, 2000 poisoned samples were randomly distributed across all classes such that 200 poisoned samples were allocated to each class. The resulting poisoned dataset Dp was incorporated into the transfer learning dataset Dt.
- Training and test dataset for performance evaluation: To evaluate the performance of the proposed method, the transfer learning dataset Dt containing 10,000 samples was divided into multiple sub-datasets with varying numbers of partitions based on a logarithmic scale. Each subset was inspected using the pre-trained model Mt1. and removed according to the inspection result. After completing the inspection process, the remaining sub-datasets were used for transfer learning to create the updated model Mt. Finally, the performance of Mt2 was evaluated based on a test dataset consisting of 2000 images and four evaluation metrics.
4.2.2. Experimental Results
4.3. Discussion
- Lowering high false-positive rate: While our methods demonstrated effective defense capabilities by successfully removing most of the poisoned data, they also falsely identified a substantial portion of benign data as poisoned. This limitation could adversely impact the utility of new datasets, particularly in transfer learning applications. In general, there exists an inherent trade-off between the preservation of benign data (RBR) and the removal of poisoned data (RPR), making the simultaneous achievement of low RBR and high RPR a particularly challenging problem. A potential solution to address this issue involves integrating a clustering algorithm with a sanitization method. Specifically, poisoned samples could be grouped into distinct clusters, which are subsequently inspected and removed using a reliable sanitization technique.
- Lowering high computation cost: In the random partition and inspection algorithm (RPIA), as the iteration k grows, the computation cost also increases; the time complexity of the RPIA is O(kn) where n is the number of partitions, while the time complexity of the SPIA is O(n). Since a larger k improves the possibility of capturing poisoned data samples in a randomly constructed sub-dataset, it is necessary to optimize the RPIA or devise a mechanism that reduces its computational cost. One potential solution to address the latter is to redesign the RPIA by leveraging the parallel processing approach, which could significantly mitigate the computational burden.
- Finding optimal parameters and mechanisms: This study primarily evaluated the performance of our two algorithms using a fixed detection threshold. The detection threshold was carefully set to 0.2, as preliminary experiments—considering various thresholds such as 0.1 and 0.3 on small datasets—indicated that it provided the best defense performance. However, we did not extensively investigate the identification of optimal thresholds under diverse attack scenarios and conditions. Given the numerous possible combinations of attack scenarios and conditions, determining the optimal detection thresholds in such settings remains a complex and challenging task. Furthermore, exploring alternative partitioning strategies could be a valuable direction for future research. For instance, datasets could be partitioned into multiple heterogeneous sub-datasets of varying sizes or designed to allow overlapping data samples among partitions.
- Defending against various sophisticated attacks and scenarios: This study focuses on two types of data poisoning attack models—concentrated poisoning attacks and random poisoning attacks—employing the dirty-label flipping approach. However, numerous other sophisticated data poisoning attack methods exist. Furthermore, it is expected that novel and increasingly complex attack models will continue to emerge, designed to circumvent the existing defense mechanisms known to adversaries. Consequently, it is essential to assess the performance and limitations of our proposed methods against such advanced attacks under diverse adversarial scenarios. Examples of these include clean-label attacks [42], dirty-label attacks incorporating complex patterns [46], and backdoor attacks leveraging adversarial examples [47].
- Conducting various, comparative empirical studies: This study demonstrates the effectiveness of our proposed methods within an experimental framework utilizing the ResNet18 model and the CIFAR-10 dataset as the target deep learning model. However, numerous deep learning models and architectures exist, such as transformers and recurrent neural networks (RNNs), alongside diverse data types including time series, tabular, text, and video. To comprehensively evaluate the scalability and performance of our methods, it is necessary to experiment with significantly larger datasets. Furthermore, given the variety of existing defense approaches, conducting extensive comparative studies under fair and standardized conditions is essential. Such empirical investigations would not only enable a thorough evaluation of the defense models’ performance but also provide valuable insights to further optimize and advance their development.
5. Conclusions and Future Works
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Hinton, G.; Deng, L.; Yu, D.; Dahl, G.E.; Mohamed, A.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.N.; et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
- Andor, D.; Alberti, C.; Weiss, D.; Severyn, A.; Presta, A.; Ganchev, K.; Petrov, S.; Collins, M. Globally Normalized Transition-Based Neural Networks. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016—Long Papers, Berlin, Germany, 7–12 August 2016; Volume 4. [Google Scholar]
- Bono, F.M.; Cinquemani, S.; Chatterton, S.; Pennacchi, P. A Deep Learning Approach for Fault Detection and RUL Estimation in Bearings. In Proceedings of the NDE 4.0, Predictive Maintenance, and Communication and Energy Systems in a Globally Networked World, Long Beach, CA, USA, 6 March–11 April 2022. [Google Scholar]
- Wang, Y.; Mianjy, P.; Arora, R. Robust Learning for Data Poisoning Attacks. In Proceedings of the Machine Learning Research, Online, 18–24 July 2021; Volume 139. [Google Scholar]
- Shayegani, E.; Al Mamun, A.; Fu, Y.; Zaree, P.; Dong, Y.; Abu-Ghazaleh, N. Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks. arXiv 2023, arXiv:2310.10844. [Google Scholar]
- Jiang, W.; Li, H.; Liu, S.; Luo, X.; Lu, R. Poisoning and Evasion Attacks against Deep Learning Algorithms in Autonomous Vehicles. IEEE Trans. Veh. Technol. 2020, 69, 4439–4449. [Google Scholar] [CrossRef]
- Ren, K.; Zheng, T.; Qin, Z.; Liu, X. Adversarial Attacks and Defenses in Deep Learning. Engineering 2020, 6, 346–360. [Google Scholar] [CrossRef]
- Akhtar, N.; Mian, A. Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey. IEEE Access 2018, 6, 14410–14430. [Google Scholar] [CrossRef]
- Qiu, S.; Liu, Q.; Zhou, S.; Wu, C. Review of Artificial Intelligence Adversarial Attack and Defense Technologies. Appl. Sci. 2019, 9, 909. [Google Scholar] [CrossRef]
- Goldblum, M.; Tsipras, D.; Xie, C.; Chen, X.; Schwarzschild, A.; Song, D.; Madry, A.; Li, B.; Goldstein, T. Dataset Security for Machine Learning: Data Poisoning, Backdoor Attacks, and Defenses. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1563–1580. [Google Scholar] [CrossRef] [PubMed]
- Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Fan, J.; Yan, Q.; Li, M.; Qu, G.; Xiao, Y. A Survey on Data Poisoning Attacks and Defenses. In Proceedings of the Proceedings—2022 7th IEEE International Conference on Data Science in Cyberspace, DSC 2022, Guilin, China, 11–13 July 2022. [Google Scholar]
- Yuan, D.; Li, G.; Li, Q.; Zheng, Y. Sybil Defense in Crowdsourcing Platforms. In Proceedings of the International Conference on Information and Knowledge Management, Proceedings, Singapore, 6–10 November 2017. Volume Part F131841. [Google Scholar]
- Miao, C.; Li, Q.; Su, L.; Huai, M.; Jiang, W.; Gao, J. Attack under Disguise: An Intelligent Data Poisoning Attack Mechanism in Crowdsourcing. In Proceedings of the Web Conference 2018—World Wide Web Conference, WWW 2018, Lyon, France, 23–27 April 2018. [Google Scholar]
- Li, Y.; Gao, J.; Lee, P.P.C.; Su, L.; He, C.; He, C.; Yang, F.; Fan, W. A Weighted Crowdsourcing Approach for Network Quality Measurement in Cellular Data Networks. IEEE Trans. Mob. Comput. 2017, 16, 300–313. [Google Scholar] [CrossRef]
- Levine, A.; Feizi, S. Deep partition aggregation: Provable defenses against general poisoning attacks. In Proceedings of the ICLR 2021—9th International Conference on Learning Representations, Online, 3–7 May 2021. [Google Scholar]
- Li, Y.; Gao, J.; Meng, C.; Li, Q.; Su, L.; Zhao, B.; Fan, W.; Han, J. A Survey on Truth Discovery. ACM Sigkdd Explor. Newsl. 2016, 17, 1–16. [Google Scholar] [CrossRef]
- Borgnia, E.; Cherepanova, V.; Fowl, L.; Ghiasi, A.; Geiping, J.; Goldblum, M.; Goldstein, T.; Gupta, A. Strong Data Augmentation Sanitizes Poisoning and Backdoor Attacks without an Accuracy Tradeoff. In Proceedings of the ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing—Proceedings, Toronto, ON, Canada, 6–11 June 2021; Volume 2021. [Google Scholar]
- Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. MixUp: Beyond Empirical Risk Minimization. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018—Conference Track Proceedings, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Yun, S.; Han, D.; Chun, S.; Oh, S.J.; Choe, J.; Yoo, Y. CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; Volume 2019. [Google Scholar]
- Park, H.; Cho, Y. A Dilution-Based Defense Method against Poisoning Attacks on Deep Learning Systems. Int. J. Electr. Comput. Eng. 2024, 14, 645–652. [Google Scholar] [CrossRef]
- Koh, P.W.; Steinhardt, J.; Liang, P. Stronger Data Poisoning Attacks Break Data Sanitization Defenses. Mach. Learn. 2022, 111, 1–47. [Google Scholar] [CrossRef]
- Steinhardt, J.; Koh, P.W.; Liang, P. Certified Defenses for Data Poisoning Attacks. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 2017. [Google Scholar]
- Li, M.; Lian, Y.; Zhu, J.; Lin, J.; Wan, J.; Sun, Y. A Sampling-Based Method for Detecting Data Poisoning Attacks in Recommendation Systems. Mathematics 2024, 12, 247. [Google Scholar] [CrossRef]
- Kearns, M.; Li, M. Learning in the Presence of Malicious Errors. SIAM J. Comput. 1993, 22, 807–837. [Google Scholar] [CrossRef]
- Yang, Y.; Liu, T.Y.; Mirzasoleiman, B. Not All Poisons Are Created Equal: Robust Training against Data Poisoning. In Proceedings of the Machine Learning Research, Baltimore, MD, USA, 17–23 July 2022; Volume 162. [Google Scholar]
- Poudel, S. Improving Collaborative Filtering Recommendation Systems via Optimal Sub-Sampling and Aspect-Based Interpretability. Ph.D. Thesis, North Carolina Agricultural and Technical State University, Greensboro, NC, USA, 2022. [Google Scholar]
- Barreno, M.; Nelson, B.; Joseph, A.D.; Tygar, J.D. The Security of Machine Learning. Mach. Learn. 2010, 81, 121–148. [Google Scholar] [CrossRef]
- Chan, P.P.K.; He, Z.; Hu, X.; Tsang, E.C.C.; Yeung, D.S.; Ng, W.W.Y. Causative Label Flip Attack Detection with Data Complexity Measures. Int. J. Mach. Learn. Cybern. 2021, 12, 103–116. [Google Scholar] [CrossRef]
- Chan, P.P.K.; He, Z.M.; Li, H.; Hsu, C.C. Data Sanitization against Adversarial Label Contamination Based on Data Complexity. Int. J. Mach. Learn. Cybern. 2018, 9, 1039–1052. [Google Scholar] [CrossRef]
- Ho, S.; Reddy, A.; Venkatesan, S.; Izmailov, R.; Chadha, R.; Oprea, A. Data Sanitization Approach to Mitigate Clean-Label Attacks Against Malware Detection Systems. In Proceedings of the Proceedings—IEEE Military Communications Conference MILCOM, Rockville, MD, USA, 28 November–2 December 2022; Volume 2022. [Google Scholar]
- Seetharaman, S.; Malaviya, S.; Vasu, R.; Shukla, M.; Lodha, S. Influence Based Defense Against Data Poisoning Attacks in Online Learning. In Proceedings of the 2022 14th International Conference on Communication Systems and Networks, COMSNETS 2022, Bangalore, India, 4–8 January 2022. [Google Scholar]
- Biggio, B.; Nelson, B.; Laskov, P. Poisoning Attacks against Support Vector Machines. In Proceedings of the 29th International Conference on Machine Learning, ICML 2012, Edinburgh, Scotland, 26 June–1 July 2012. [Google Scholar]
- Barreno, M.; Nelson, B.; Sears, R.; Joseph, A.D.; Tygar, J.D. Can Machine Learning Be Secure? In Proceedings of the 2006 ACM Symposium on Information, Computer and Communications Security, ASIACCS ’06, Taipei, Taiwan, 21–24 March 2006; Volume 2006. [Google Scholar]
- Jagielski, M.; Oprea, A.; Biggio, B.; Liu, C.; Nita-Rotaru, C.; Li, B. Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning. In Proceedings of the Proceedings—IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 20–24 May 2018; Volume 2018. [Google Scholar]
- Vorobeychik, Y.; Kantarcioglu, M. Adversarial Machine Learning; Synthesis Lectures on Artificial Intelligence and Machine Learning; Springer: Cham, Switzerland, 2018; Volume 12. [Google Scholar] [CrossRef]
- Tian, Z.; Cui, L.; Liang, J.; Yu, S. A Comprehensive Survey on Poisoning Attacks and Countermeasures in Machine Learning. ACM Comput. Surv. 2022, 55, 1–35. [Google Scholar] [CrossRef]
- Liang, H.; He, E.; Zhao, Y.; Jia, Z.; Li, H. Adversarial Attack and Defense: A Survey. Electronics 2022, 11, 1283. [Google Scholar] [CrossRef]
- Demontis, A.; Melis, M.; Pintor, M.; Jagielski, M.; Biggio, B.; Oprea, A.; Nita-Rotaru, C.; Roli, F. Why Do Adversarial Attacks Transfer? Explaining Transferability of Evasion and Poisoning Attacks. In Proceedings of the 28th USENIX Security Symposium, Santa Clara, CA, USA, 14–16 August 2019. [Google Scholar]
- Yerlikaya, F.A.; Bahtiyar, Ş. Data Poisoning Attacks against Machine Learning Algorithms. Expert Syst. Appl. 2022, 208, 118101. [Google Scholar] [CrossRef]
- Shafahi, A.; Ronny Huang, W.; Najibi, M.; Suciu, O.; Studer, C.; Dumitras, T.; Goldstein, T. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Volume 2018. [Google Scholar]
- Wang, S.; Nepal, S.; Rudolph, C.; Grobler, M.; Chen, S.; Chen, T. Backdoor Attacks Against Transfer Learning with Pre-Trained Deep Learning Models. IEEE Trans. Serv. Comput. 2022, 15, 1526–1539. [Google Scholar] [CrossRef]
- Cho, Y. Intelligent On-off Web Defacement Attacks and Random Monitoring-Based Detection Algorithms. Electronics 2019, 8, 1338. [Google Scholar] [CrossRef]
- Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images; Science Department, University of Toronto: Toronto, ON, Canada, 2009; Technical Report. [Google Scholar]
- Paudice, A.; Muñoz-González, L.; Lupu, E.C. Label Sanitization Against Label Flipping Poisoning Attacks. In Proceedings of the Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Dublin, Ireland, 10–14 September 2018; Springer: Cham, Switzerland, 2019; Volume 11329 LNAI. [Google Scholar]
- Li, Y.; Lyu, X.; Koren, N.; Lyu, L.; Li, B.; Ma, X. Anti-Backdoor Learning: Training Clean Models on Poisoned Data. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–14 December 2021; Volume 18. [Google Scholar]
Attack Characteristic | Attack Type |
---|---|
Attack Timing | Decision time (evasion attack) vs. training time (poisoning attack) |
Attacker Information | White-box attacks vs. black-box attacks |
Attack Goals | Targeted attacks vs. reliability attacks |
Method | Metric | No Defense (n = 1) | n | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2 | 4 | 8 | 10 | 50 | 100 | 200 | 500 | 1000 | 2000 | 4000 | 8000 | 10,000 | |||
SPIA | ASR(%) | 100 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.5 | 8.2 | 3.4 | 18.7 |
RPR(%) | 0 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 99.5 | 91.8 | 96.6 | 81.3 | |
RBR(%) | 0 | 37.5 | 6.3 | 6.3 | 0 | 0 | 0 | 0 | 1.3 | 9.3 | 23.1 | 22.1 | 22.5 | 18.7 | |
RPIA | ASR(%) | 100 | 100 | 100 | 100 | 100 | 61.8 | 32.6 | 19.2 | 10.8 | 9.2 | 10.4 | 13.2 | 16.1 | 18.7 |
RPR(%) | 0 | 0 | 0 | 0 | 0 | 38.2 | 67.4 | 80.8 | 89.2 | 90.8 | 89.6 | 86.8 | 83.9 | 81.3 | |
RBR(%) | 0 | 0 | 0 | 0 | 0 | 30.4 | 50.3 | 60.7 | 62.6 | 60.2 | 55.9 | 40.5 | 28.5 | 18.7 |
Method | Metric | No Defense (n = 1) | n | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2 | 4 | 8 | 10 | 50 | 100 | 200 | 500 | 1000 | 2000 | 4000 | 8000 | 10,000 | |||
SPIA | ASR | 100 | 100 | 100 | 100 | 100 | 97.8 | 91.9 | 86.5 | 65.5 | 46.6 | 28.2 | 39.2 | 16.9 | 18.7 |
RPR | 0 | 0 | 0 | 0 | 0 | 2.2 | 8.1 | 13.5 | 34.5 | 53.4 | 71.8 | 60.8 | 83.1 | 81.3 | |
RBR | 0 | 0 | 0 | 0 | 0 | 2 | 6.7 | 9.8 | 24.9 | 35.4 | 44.9 | 28.2 | 29.3 | 18.7 | |
RPIA | ASR | 100 | 100 | 100 | 100 | 100 | 61.1 | 32.2 | 17.9 | 11 | 8.9 | 11 | 13.2 | 16.5 | 18.7 |
RPR | 0 | 0 | 0 | 0 | 0 | 38.9 | 67.8 | 82.1 | 89 | 91.1 | 89 | 86.8 | 83.5 | 81.3 | |
RBR | 0 | 0 | 0 | 0 | 0 | 31.9 | 51.5 | 61.9 | 61.9 | 59 | 54.7 | 40.8 | 29.3 | 18.7 |
Method | Metric | No Defense (n = 1) | n | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2 | 4 | 8 | 10 | 50 | 100 | 200 | 500 | 1000 | 2000 | 4000 | 8000 | 10,000 | |||
SPIA | ACC(%) | 67.1 | 80.7 | 80.8 | 80.8 | 80.7 | 81.1 | 81.2 | 80.7 | 80.8 | 80.4 | 80.8 | 80.7 | 80.8 | 80.5 |
ASR(%) | 100 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | |
RPR(%) | 0 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 99 | |
RBR(%) | 0 | 37.5 | 6.3 | 6.3 | 12.5 | 10 | 7.5 | 8.1 | 12 | 17.5 | 28.1 | 22.8 | 25.3 | 20.4 | |
RPIA | ACC(%) | 67.1 | 69.1 | 66.8 | 68.2 | 67.3 | 76.3 | 71.4 | 78.9 | 79.8 | 80.3 | 79.6 | 80.4 | 80.7 | 79.8 |
ASR(%) | 100 | 100 | 100 | 73.6 | 69.5 | 19.4 | 8.2 | 3.2 | 1.1 | 0.6 | 0.8 | 0.9 | 1.3 | 1.5 | |
RPR(%) | 0 | 0 | 0 | 26.4 | 30.5 | 80.6 | 91.8 | 96.8 | 98.9 | 99.4 | 99.2 | 99.1 | 98.7 | 98.5 | |
RBR(%) | 0 | 0 | 0 | 22.7 | 26.2 | 68.1 | 74.8 | 76 | 71.5 | 67.4 | 63.3 | 44.2 | 31.6 | 20.3 |
Method | Metric | No Defense (n = 1) | n | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2 | 4 | 8 | 10 | 50 | 100 | 200 | 500 | 1000 | 2000 | 4000 | 8000 | 10,000 | |||
SPIA | ACC(%) | 67.1 | 66.8 | 67.6 | 67.4 | 66.2 | 63.4 | 64.9 | 67.6 | 69 | 68.3 | 75.2 | 75 | 80.3 | 80.2 |
ASR(%) | 100 | 100 | 100 | 100 | 100 | 90.6 | 82.2 | 68.5 | 49.1 | 36.2 | 20.2 | 28.3 | 2.4 | 2.8 | |
RPR(%) | 0 | 0 | 0 | 0 | 0 | 9.4 | 17.8 | 31.5 | 50.9 | 63.8 | 79.8 | 71.7 | 97.6 | 97.2 | |
RBR(%) | 0 | 0 | 0 | 0 | 0 | 7.7 | 14.3 | 22.7 | 36.5 | 42.9 | 50 | 29.9 | 29.9 | 18.4 | |
RPIA | ACC(%) | 67.1 | 68.4 | 67.6 | 69.6 | 69.5 | 70.5 | 76.1 | 76.8 | 79.5 | 79.1 | 76.6 | 79.9 | 79.9 | 79.8 |
ASR(%) | 100 | 100 | 100 | 100 | 100 | 35.9 | 17.2 | 7.1 | 1.6 | 1.2 | 0.9 | 1.2 | 2.4 | 2.8 | |
RPR(%) | 0 | 0 | 0 | 0 | 0 | 64.1 | 82.8 | 92.9 | 98.4 | 98.8 | 99.1 | 98.8 | 97.6 | 97.2 | |
RBR(%) | 0 | 0 | 0 | 0 | 0 | 51.1 | 63.6 | 68.5 | 67.4 | 62.1 | 59.5 | 41.9 | 29.7 | 18.4 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lee, J.; Cho, Y.; Lee, R.; Yuk, S.; Youn, J.; Park, H.; Shin, D. A Novel Data Sanitization Method Based on Dynamic Dataset Partition and Inspection Against Data Poisoning Attacks. Electronics 2025, 14, 374. https://doi.org/10.3390/electronics14020374
Lee J, Cho Y, Lee R, Yuk S, Youn J, Park H, Shin D. A Novel Data Sanitization Method Based on Dynamic Dataset Partition and Inspection Against Data Poisoning Attacks. Electronics. 2025; 14(2):374. https://doi.org/10.3390/electronics14020374
Chicago/Turabian StyleLee, Jaehyun, Youngho Cho, Ryungeon Lee, Simon Yuk, Jaepil Youn, Hansol Park, and Dongkyoo Shin. 2025. "A Novel Data Sanitization Method Based on Dynamic Dataset Partition and Inspection Against Data Poisoning Attacks" Electronics 14, no. 2: 374. https://doi.org/10.3390/electronics14020374
APA StyleLee, J., Cho, Y., Lee, R., Yuk, S., Youn, J., Park, H., & Shin, D. (2025). A Novel Data Sanitization Method Based on Dynamic Dataset Partition and Inspection Against Data Poisoning Attacks. Electronics, 14(2), 374. https://doi.org/10.3390/electronics14020374