Adaptive Oversampling via Density Estimation for Online Imbalanced Classification
Abstract
:1. Introduction
2. Related Works
2.1. Online Learning Framework
2.2. Concept Drift in Online Learning
2.3. Class Imbalance in Online Learning
2.4. Online Learning Methods for Class Imbalance
3. The Proposed Method
3.1. Motivation
3.2. Adaptive Weighted Kernel Density Estimator
Algorithm 1: Adaptive weighted kernel density estimation (awKDE) |
Input: : Kernel function, we choose the Gaussian function Output: Estimated density
|
3.3. Online Biased Learning with awKDE Oversampling
Algorithm 2: Online biased learning with awKDE oversampling |
input: f: classifier W: max queue size, : biased ratio, : time-decayed factor for class size metrics, output: Updated classifier |
4. Comparative Experiments
4.1. Data
4.1.1. Synthetic Dataset
- Sine [30]: This contains two relevant variables; ; . Before the drift, all points under are labeled as minority(positive) class. After the drift, the true label is reversed. Both x and y are normalized before classification, so that .
- Circle [30]: This contains two relevant variables; . We construct a circle as a true boundary and all points inside of the circle are labeled as minority(positive) class. Before the drift, we set and after the drift it varies to .
- SEA [31]: This contains two relevant variables; . All points under are labeled as minority (positive) class. The true decision boundary changes to after the drift. Both are normalized before classification, so that .
4.1.2. Real-World Dataset
- Gesture phase segmentation [33,34]: This dataset consists of seven videos recorded using Microsoft Kinect sensors, capturing gestures performed by three participants narrating comic strips. Each video is annotated with five gesture phases: ‘Rest’, ‘Preparation’, ‘Stroke’, ‘Hold’, and ‘Retraction’. For our experiments, we focused on distinguishing between the ‘Rest’ and ‘Hold’ phases, which are challenging to classify because both involve stationary hand positions. Furthermore, the imbalance ratio varies between individuals, with some having ‘Hold’ as the majority class and others having it as the minority class.
- Fraud [6]: This dataset contains transactions made with credit cards in September 2013 by European cardholders. This dataset presents transactions that occurred over two days, where we have 492 frauds out of 284,807 transactions. The dataset is highly unbalanced, and the positive class (frauds) account for 0.172% of all transactions. The exact imbalance types and drift types are unknown; however, they are expected to be reflected given the massive dynamics in financial environments.
4.2. Classifier
4.3. Competing Methods
- Baseline model: An online incremental learning algorithm with no specific treatments. This method uses only a single instance at to predict the label at time t.
- Sliding queue: This method trains the classifier using the most recent instances stored in a queue of size 100. While it leverages more data than the proposed method, it does so without considering the quality or adaptability of the stored instances.
- Oversampling-based Online Bagging (OOB) [12]: This method employs a bagging ensemble to oversample the minority class. The ensemble size is set to 30, requiring significantly more memory to store multiple models and their corresponding data.
- Adaptive REBalancing method (AREBA) [13]: This method adaptively adjusts the queue size to balance each class, with a maximum queue size of 20 per class, as specified by the authors as the optimal configuration. Since AREBA stores data separately for each class (positive and negative), the total memory usage effectively amounts to 40.
- OB-awKDE: The proposed method. To achieve memory optimization, the maximum queue size for each class (positive, negative, and KDE-generated samples) was set to 20, resulting in a total memory usage of 60. We set the bias ratio to 0.5 .
4.4. Performance Measure for Online Learning Methods
4.5. Results
4.5.1. Synthetic Dataset
4.5.2. Real-World Dataset
5. Discussion and Concluding Remarks
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Elnahrawy, E.; Nath, B. Online data cleaning in wireless sensor networks. In Proceedings of the 1st International Conference on Embedded Networked Sensor Systems, Los Angeles, CA, USA, 5–7 November 2003; pp. 294–295. [Google Scholar]
- Wang, S.; Minku, L.L.; Yao, X. Online class imbalance learning and its applications in fault detection. Int. J. Comput. Intell. Appl. 2013, 12, 134001. [Google Scholar] [CrossRef]
- Meseguer, J.; Puig, V.; Escobet, T. Fault diagnosis using a timed discrete-event approach based on interval observers: Application to sewer networks. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2010, 40, 900–916. [Google Scholar] [CrossRef]
- Sun, Y.; Tang, K.; Minku, L.L.; Wang, S.; Yao, X. Online ensemble learning of data streams with gradually evolved classes. IEEE Trans. Knowl. Data Eng. 2016, 28, 1532–1545. [Google Scholar] [CrossRef]
- Sousa, M.R.; Gama, J.; Brandão, E. A new dynamic modeling framework for credit risk assessment. Expert Syst. Appl. 2016, 45, 341–351. [Google Scholar] [CrossRef]
- greenDal Pozzolo, A.; Caelen, O.; Johnson, R.A.; Bontempi, G. Calibrating probability with undersampling for unbalanced classification. In Proceedings of the IEEE Symposium Series on Computational Intelligence, Cape Town, South Africa, 7–10 December 2015; pp. 159–166. [Google Scholar]
- Ditzler, G.; Roveri, M.; Alippi, C.; Polikar, R. Learning in nonstationary environments: A survey. IEEE Comput. Intell. Mag. 2015, 10, 12–25. [Google Scholar] [CrossRef]
- Minku, L.L. Online Ensemble Learning in the Presence of Concept Drift. Ph.D. Dissertation, School of Computer Science, The University of Birmingham, Birmingham, UK, 2011. [Google Scholar]
- Mahdi, O.A.; Pardede, E.; Ali, N.; Cao, J. Fast reaction to sudden concept drift in the absence of class labels. Appl. Sci. 2020, 10, 606. [Google Scholar] [CrossRef]
- Feng, Y.; Zhou, M.; Tong, X. Imbalanced classification: A paradigm-based review. Stat. Anal. Data Min. ASA Data Sci. J. 2021, 14, 383–406. [Google Scholar] [CrossRef]
- Wang, S.; Minku, L.L.; Yao, X. A learning framework for online class imbalance learning. In Proceedings of the IEEE Symposium on Computational Intelligence and Ensemble Learning (CIEL), Singapore, 16–19 April 2013; pp. 36–45. [Google Scholar]
- Wang, S.; Minku, L.L.; Yao, X. Resampling-based ensemble methods for online class imbalance learning. IEEE Trans. Knowl. Data Eng. 2015, 27, 1356–1368. [Google Scholar] [CrossRef]
- Malialis, K.; Panayiotou, C.G.; Polycarpou, M.M. Online learning with adaptive rebalancing in nonstationary environments. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4445–4459. [Google Scholar] [CrossRef]
- Oza, N.; Russell, S. Experimental comparisons of online and batch versions of bagging and boosting. In Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 26–29 August 2001. [Google Scholar]
- Wang, S.; Minku, L.L.; Yao, X. A systemic study of online class imbalance learning with concept drift. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4802–4821. [Google Scholar] [CrossRef]
- Polikar, R.; Upda, L.; Upda, S.; Honavar, V. Learn++: An incremental learning algorithm for supervised neural networks. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2001, 31, 497–508. [Google Scholar] [CrossRef]
- Gama, J.; Žliobaitė, I.; Bifet, A.; Pechenizkiy, M.; Bouchachia, A. A survey on concept drift adaptation. ACM Comput. Surv. 2014, 46, 1–37. [Google Scholar] [CrossRef]
- Pesaranghader, A.; Viktor, H.L.; Paquet, E. McDiarmid Drift Detection Methods for Evolving Data Streams. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–9. [Google Scholar]
- Brzezinski, D. Combining Block-Based and Online Methods in Learning Ensembles from Concept Drifting Data Streams. Ph.D. Dissertation, Poznan University of Technology, Poznan, Poland, 2015. [Google Scholar]
- Wadewale, K.; Desai, S. Survey on method of drift detection and classification for time varying data set. Int. Res. J. Eng. Technol. 2015, 2, 709–713. [Google Scholar]
- Iwashita, A.S.; Papa, J.P. An overview on concept drift learning. IEEE Access 2019, 7, 1532–1547. [Google Scholar] [CrossRef]
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Mao, W.; Wang, J.; Wang, L. Online sequential classification of imbalanced data by combining extreme learning machine and improved smote algorithm. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–8. [Google Scholar]
- Wang, B.; Pineau, J. Online bagging and boosting for imbalanced data streams. IEEE Trans. Knowl. Data Eng. 2016, 28, 3353–3366. [Google Scholar] [CrossRef]
- Wang, J.; Zhao, P.; Hoi, S.C. Cost-sensitive online classification. IEEE Trans. Knowl. Data Eng. 2013, 26, 2425–2438. [Google Scholar] [CrossRef]
- Ghazikhani, A.; Monsefi, R.; Yazdi, H.S. Online neural network model for non-stationary and imbalanced data stream classification. Int. J. Mach. Learn. Cybern. 2014, 5, 51–62. [Google Scholar] [CrossRef]
- Parzen, E. On estimation of a probability density function and mode. Ann. Math. Stat. 1962, 33, 1065–1076. [Google Scholar] [CrossRef]
- Chen, Y.-C. A tutorial on kernel density estimation and recent advances. Biostat. Epidemiol. 2017, 1, 161–187. [Google Scholar] [CrossRef]
- Wang, B.; Wang, X. Bandwidth selection for weighted kernel density estimation. arXiv 2007, arXiv:0709.1616. [Google Scholar]
- Gama, J.; Medas, P.; Castillo, G.; Rodrigues, P. Learning with drift detection. In Proceedings of the SBIA Brazilian Symposium on Artificial Intelligence, São Luís, Brazil, 29 September–1 October 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 286–295. [Google Scholar]
- Street, W.N.; Kim, Y. A streaming ensemble algorithm (SEA) for large-scale classification. In Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’01), New York, NY, USA, 26–29 August 2001; Association for Computing Machinery: New York, NY, USA, 2001; pp. 377–382. [Google Scholar]
- Napierala, K.; Stefanowski, J. Identification of different types of minority class examples in imbalanced data. In Hybrid Artificial Intelligent Systems; Corchado, E., Snášel, V., Abraham, A., Woźniak, M., Graña, M., Cho, S.-B., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 139–150. [Google Scholar]
- Madeo, R.C.B.; Peres, S.M.; de Moraes Lima, C.A. Gesture phase segmentation using support vector machines. Expert Syst. Appl. 2016, 56, 100–115. [Google Scholar] [CrossRef]
- Madeo, R.C.B.; Wagner, P.K.; Peres, S.M. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 25 November 2024).
- He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; IEEE: New York, NY, USA, 2015; pp. 1026–1034. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; Citeseer: Princeton, NJ, USA, 2013; Volume 30, p. 3. [Google Scholar]
- Dawid, A.P.; Vovk, V.G. Prequential Probability: Principles and Properties. Bernoulli 1999, 5, 125–162. [Google Scholar] [CrossRef]
Predicted Positive | Predicted Negative | |
---|---|---|
Actual Positive | TP (True Positive) | FN (False Negative) |
Actual Negative | FP (False Positive) | TN (True Negative) |
Dataset | IR | Baseline | Sliding Queue | OOB | AREBA | OB-awKDE |
---|---|---|---|---|---|---|
10% | 0.207 ± 0.136 (0.000) | 0.592 ± 0.086 (0.000) | 0.694 ± 0.003 (0.001) | 0.618 ± 0.171 (0.000) | 0.752 ± 0.034 | |
Sine | 1% | 0.0 ± 0.0 (0.000) | 0.114 ± 0.114 (0.000) | 0.395 ± 0.135 (0.000) | 0.329 ± 0.118 (0.000) | 0.712 ± 0.079 |
0.1% | 0.0 ± 0.0 (0.000) | 0.0 ± 0.0 (0.000) | 0.023 ± 0.107 (0.000) | 0.074 ± 0.196 (0.000) | 0.261 ± 0.316 | |
10% | 0.028 ± 0.058 (0.000) | 0.703 ± 0.027 (0.000) | 0.701 ± 0.033 (0.000) | 0.666 ± 0.155 (0.000) | 0.749 ± 0.033 | |
Circle | 1% | 0.0 ± 0.0 (0.000) | 0.103 ± 0.146 (0.000) | 0.405 ± 0.109 (0.000) | 0.386 ± 0.155 (0.000) | 0.734 ± 0.058 |
0.1% | 0.0 ± 0.0 (0.000) | 0.0 ± 0.0 (0.000) | 0.029 ± 0.120 (0.000) | 0.060 ± 0.176 (0.000) | 0.285 ± 0.335 | |
10% | 0.302 ± 0.059 (0.000) | 0.575 ± 0.044 (0.000) | 0.813 ± 0.026 (0.842) | 0.680 ± 0.197 (0.000) | 0.799 ± 0.030 | |
SEA | 1% | 0.0 ± 0.0 (0.000) | 0.088 ± 0.116 (0.000) | 0.568 ± 0.097 (0.000) | 0.333 ± 0.201 (0.000) | 0.758 ± 0.064 |
0.1% | 0.0 ± 0.0 (0.000) | 0.0 ± 0.0 (0.000) | 0.038 ± 0.134 (0.000) | 0.056 ± 0.172 (0.000) | 0.194 ± 0.290 |
Dataset | Baseline | Sliding Queue | OOB | AREBA | OB-awKDE |
---|---|---|---|---|---|
Gesture of A | 0.382 ± 0.259 | 0.106 ± 0.256 | 0.565 ± 0.314 | 0.378 ± 0.246 | 0.626 ± 0.251 |
Gesture of B | 0.767 ± 0.234 | 0.748 ± 0.253 | 0.829 ± 0.213 | 0.697 ± 0.293 | 0.854 ± 0.262 |
Gesture of C | 0.136 ± 0.162 | 0.597 ± 0.280 | 0.189 ± 0.172 | 0.724 ± 0.224 | 0.732 ± 0.213 |
Fraud | 0.424 ± 0.159 | 0.064 ± 0.202 | 0.066 ± 0.205 | 0.770 ± 0.121 | 0.877 ± 0.128 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lee, D.; Kim, H. Adaptive Oversampling via Density Estimation for Online Imbalanced Classification. Information 2025, 16, 23. https://doi.org/10.3390/info16010023
Lee D, Kim H. Adaptive Oversampling via Density Estimation for Online Imbalanced Classification. Information. 2025; 16(1):23. https://doi.org/10.3390/info16010023
Chicago/Turabian StyleLee, Daeun, and Hyunjoong Kim. 2025. "Adaptive Oversampling via Density Estimation for Online Imbalanced Classification" Information 16, no. 1: 23. https://doi.org/10.3390/info16010023
APA StyleLee, D., & Kim, H. (2025). Adaptive Oversampling via Density Estimation for Online Imbalanced Classification. Information, 16(1), 23. https://doi.org/10.3390/info16010023