Next Article in Journal / Special Issue
Prediction of Drivers’ Red-Light Running Behaviour in Connected Vehicle Environments Using Deep Recurrent Neural Networks
Previous Article in Journal
Digital Assistance Systems to Implement Machine Learning in Manufacturing: A Systematic Review
Previous Article in Special Issue
Node-Centric Pruning: A Novel Graph Reduction Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Continual Semi-Supervised Malware Detection

by
Matthew Chin
and
Roberto Corizzo
*,†
Department of Computer Science, American University, Washington, DC 20016, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mach. Learn. Knowl. Extr. 2024, 6(4), 2829-2854; https://doi.org/10.3390/make6040135
Submission received: 30 October 2024 / Revised: 4 December 2024 / Accepted: 5 December 2024 / Published: 10 December 2024

Abstract

:
Detecting malware has become extremely important with the increasing exposure of computational systems and mobile devices to online services. However, the rapidly evolving nature of malicious software makes this task particularly challenging. Despite the significant number of machine learning works for malware detection proposed in the last few years, limited interest has been devoted to continual learning approaches, which could allow models to showcase effective performance in challenging and dynamic scenarios while being computationally efficient. Moreover, most of the research works proposed thus far adopt a fully supervised setting, which relies on fully labelled data and appears to be impractical in a rapidly evolving malware landscape. In this paper, we address malware detection from a continual semi-supervised one-class learning perspective, which only requires normal/benign data and empowers models with a greater degree of flexibility, allowing them to detect multiple malware types with different morphology. Specifically, we assess the effectiveness of two replay strategies on anomaly detection models and analyze their performance in continual learning scenarios with three popular malware detection datasets (CIC-AndMal2017, CIC-MalMem-2022, and CIC-Evasive-PDFMal2022). Our evaluation shows that replay-based strategies can achieve competitive performance in terms of continual ROC-AUC with respect to the considered baselines and bring new perspectives and insights on this topic.

1. Introduction

Malware infections are on the rise, especially with the pervasiveness of IoT and mobile devices exposed to online services. A malware intends to alter, disrupt, or destroy software elements that are essential to the proper functioning of a computer system or device. It can steal sensitive information, gain persistent access to victim networks, as well as encrypt and send system information to the threat actor’s command and control (C&C) centers.
Modern malware is coded using sophisticated techniques, which may make conventional security mechanisms (such as firewalls and anti-viruses) ineffective. For this reason, machine learning-based approaches for malware detection appear particularly attractive. However, one central issue pertains to the adequacy of existing solutions in real-world scenarios. While machine learning and deep learning have shown a promising detection performance [1,2,3], they may provide a limited robustness in a continuous scenario where malware morphology evolves over time, due to their reliance on labelled malware data, as well as their demanding computational requirements.
In this context, continual learning (CL) could provide a significant advance to machine learning-based malware detection, allowing models to effectively perform in challenging and evolving scenarios with limited computational requirements. The goal is to continuously adapt models to maintain and improve performance across all tasks encountered in a lifelong learning setup. Success is measured by the model’s ability to balance plasticity (learning new tasks) and stability (retaining old tasks) using metrics such as accuracy on a sequence of tasks, forgetting rate, and forward/backward transfer. Learning strategies are task-driven and can be broadly categorized as regularization-based, replay-based, and architectural-based.
One of the most popular strategies is experience replay (ER), which maintains and revisits a subset of past data during training on new tasks. ER is also commonly adopted in reinforcement learning (RL) [4] and Q-learning [5] to support agents in learning from a mix of past experiences and approximate the value function or policy better [6,7]. In the context of CL, however, ER focuses on mitigating catastrophic forgetting, i.e., the tendency of neural networks to overwrite knowledge of previously learned tasks when learning new ones [8]. While RL and CL share the overlapping goal of adapting to new information over time, they are structurally different. In fact, RL focuses on training an agent that interacts with an environment to maximize cumulative rewards over time [9]. The agent learns policies for sequential decision-making tasks, often in stochastic or dynamic settings. On the other hand, CL focuses on training models on a sequence of tasks without forgetting knowledge of previously learned tasks. A structured breakdown of the differences for ER in CL and RL learning settings is presented in Table 1.
Despite the high potential of CL for malware detection, this has been scarcely explored so far. In this paper, our focus is investigating malware detection from a continual, semi-supervised, one-class learning perspective. One-class learning allows us to train models using exclusively normal/benign data. This learning setting can empower models with a greater degree of flexibility and robustness when compared to a fully supervised setting. Specifically, (i) it overcomes the limitation of requiring labelled data for different malware classes, and (ii) it allows models to detect novel malware types, possibly unknown at training time. Moreover, in our work, a continual learning perspective is added to this setting, where models are challenged to incorporate emerging normal tasks/concepts while retaining knowledge of previously observed concepts and while being exposed to new malware instances. We assess the detection performance and the forgetting of a diverse set of anomaly detection models in two devised continual learning scenarios, adopting continual learning strategies. Our contribution is three-fold. Specifically, we do the following:
(i)
Formulate a semi-supervised one-class continual malware detection workflow, which is essential for real-world applications but has received limited attention in previous research.
(ii)
Devise two model-agnostic experience replay strategies that support anomaly detection models popular in conventional learning settings.
(iii)
Conduct a comprehensive empirical evaluation and analysis involving seven diverse models, three real-world malware detection datasets, and two continual learning scenarios, assessing the effectiveness of experience replay strategies compared to proposed lower and upper-bound baselines.

2. Background

2.1. Malware Detection

Malware detection has become an urgent need in recent years. Operating systems are subject to run on a wide variety of devices, including desktops, servers, routers, security cameras, drones, etc., which exacerbates the difficulty of this task [10].
Machine learning-based approaches for malware detection have become popular in recent years. An LSTM-based approach to detect five malware families, i.e., Trojan, rootkit, backdoor, virus, and worm, was proposed by [11]. Deep recurrent neural networks have been explored in the context of Android malware detection [12]. A multi-head squeeze-and-excitation residual network to detect malware in Android apps has been devised [3], analyzing manifest file permissions, API calls, and hardware features. Specularly, a Chameleon–Hunter algorithm has been proposed [13] to address mobile malware in iOS devices and detect UI-based illicit activity threats.
Some approaches specialize in ransomware detection, spanning from conventional machine learning to deep learning models [14]. Recently, multi-classifier network-based [1], concept-drift-aware approaches [15], and multi-view feature intelligence [2] have been proposed. However, a significant pitfall of existing approaches stands in their fully supervised nature. Semi-supervised learning in this context can provide more flexibility, allowing models to generalize to unforeseeable classes of attacks that are not described by known patterns or profiles.
The authors in [16] propose a semi-supervised approach to identify malicious traffic by leveraging multimodal traffic characteristics, where two independent neural networks are adopted to learn sequence and topological features from the traffic. The model is trained using a joint strategy that minimizes both the reconstruction error from the autoencoder and the classification loss, allowing it to effectively utilize limited labeled data alongside a large amount of unlabeled data. A semi-supervised technique to detect Android malware from Android permissions and Application Programmer Interface (API) call logs is proposed in [17]. The ML technique is incorporated into an Android application to scan the installed applications and detect the corresponding levels of maliciousness with success. The work in [18] introduces a Semi-Supervised Vulnerability Detection (SSVD) mtehod that leverages the information gain of model parameters as the certainty of the correctness of pseudo-labels and prioritizes high-certainty pseudo-labeled code snippets as training data. The proposed approach incorporates a triplet loss to maximize the separation between vulnerable and non-vulnerable code snippets to better propagate labels from labeled code snippets to nearby unlabeled snippets.
An ensemble semi-supervised classification algorithm named Random Forest of Tensors (RFoT) is proposed in [19]. RFoT leverages tensor decomposition to extract intricate latent patterns from the data and combines multidimensional analysis with clustering to capture sample groupings within latent components, aiding in distinguishing between malware and benign-ware. On a similar thread, the authors in [20] propose a hierarchical semi-supervised algorithm (HNMFk) based on non-negative matrix factorization with automatic model selection. The method exploits the hierarchical structure of the malware data together with a semi-supervised setup, which enables the classification malware families under conditions of extreme class imbalance.
Despite the advantages provided by semi-supervised learning over fully supervised approaches, continual learning has rarely been considered for malware detection, resulting in a substantial gap. In fact, the unknown signatures of modern malware require models to be proactive rather than reactive. As a result, investigating models that can evolve and deal with dynamic scenarios where adaptation and knowledge retention are simultaneously considered is of paramount importance for future research. With this present paper, we attempt to fill this gap at the intersection of continual learning and semi-supervised malware detection, providing a new perspective on this problem.

2.2. Continual Learning

Also known as lifelong learning, it focuses on a continuous process in which a series of different tasks (or concepts) are presented to a machine learning method over time [8], challenging its adaptation and knowledge retention capabilities. In some cases, continual learning strategies draw inspiration from diverse disciplines, including biology and neuroscience [8]. Most common continual learning settings tackle image classification problems in task-incremental, class-incremental, and domain-incremental scenarios [21]. Emerging types of scenarios include online learning [22] and recurring tasks [23]. In anomaly detection, an emerging benchmark for continual learning with naturally occurring changes over time is proposed in [24].
Continual learning strategies in the literature can be loosely grouped into three main categories:
Regularization-based: These strategies introduce constraints on weight updates during the training process of neural network models. This goal could be realized by preventing or limiting updates to weights learned in previously observed tasks [25] or by freezing early layers in the model architecture to mitigate forgetting previous tasks while updating late layers to incorporate new tasks [26]. Another option is to use specialized losses to prevent drastic changes in already learned weights, as in EWC [27] and LWF [28].
Dynamic architectures: They adaptively adjust the model architecture during the learning process by expanding the network with new neurons or layers as they encounter new tasks [29,30]. Another option is to perform dynamic adaptation with pruning capabilities to keep model capacity under control by removing insignificant weights, as in PackNet [31] and WSN [32], as well as by introducing quantization capabilities [33]. Dynamic architecture strategies consider a neural network model as built upon independent sub-networks, each specialized in addressing a different task. Some approaches are also referred to as forget-free [33], i.e., not subject to forgetting. However, this behavior holds only under certain assumptions, such as the availability of task labels and unbounded capacity.
Replay-based: They focus on collecting a memory that represents a summarized version of data from previous tasks, which can be used for model updates. This goal can be achieved by selecting and storing salient samples in a buffer and replaying them during model updates [8].
Several experience replay strategies have been recently proposed in the literature. Examples include Retrospective Adversarial Replay (RAR) [34], which perturbs a buffered sample towards its nearest neighbors drawn from the current task in a latent representation space. Replaying such samples was shown to refine the boundary between previous and current tasks.
A brain-inspired variant of replay is proposed in [35], where hidden representations are replayed based on the neural network’s own, context-modulated feedback connections.
Adaptive-experience replay (AdaER) [36] introduces a contextually-cued memory recall (C-CMR) strategy, which selectively replays memories that are most conflicting with the current input data in terms of both data and task. It also incorporates an entropy-balanced reservoir sampling (E-BRS) strategy to enhance the performance of the memory buffer by maximizing information entropy.
Uncertainty-Aware Sampling (UAS) [37] employs model and data uncertainties to select samples that are stable to the model and have low noise for rehearsal. The method leverages a dual Convolutional Neural Network (CNN) and Bayesian Neural Network (BNN) to continuously learn and consolidate knowledge.
The authors in [38] propose a mixup-based training approach to mitigate representation shifts by incorporating asymmetric mixup training into the replay method. The method selectively targets the old data stored in the memory buffer, deliberately excluding classes from the newly incoming data and enabling the model to learn new data while preserving the representation of the old data.
In general, experience replay approaches adopt selection strategies, which are devised to identify the most relevant samples for every task, with the intention to keep a compact replay buffer size and limit resource utilization [39]. Variants of these approaches have also been explored in distributed and collaborative learning settings [40].
Alternatively to collecting samples, artificial samples can be obtained through generative models [41], resulting in a lower memory footprint.
Despite the growing interest in continual learning, proposed strategies are mostly focused on training neural-based models in a supervised image classification setting. A few works explored continual learning from an anomaly detection perspective [42], although not strictly focused on malware detection. An exploration on continual learning for malware detection was recently conducted [43]. Albeit limited to supervised approaches, important takeaways of this study include an emphasis on continual learning methods as a more memory-efficient alternative to conventional machine learning and a thorough analysis that shows that the detection performance of available methods is still unsatisfactory.
In our work, we argue that continual semi-supervised learning could provide a significant advantage in malware detection settings. Table 2 compares different learning approaches considering the key properties considered in our study and highlighting the lack of a continual learning perspective in popular supervised and semi-supervised malware detection approaches.
An overview of our semi-supervised one-class continual malware detection workflow is graphically presented in Figure 1. Selected results in Figure 2 allow us to highlight this potential: learning new concepts while mitigating forgetting of previously learned concepts can lead to more robust models that exploit general knowledge to improve their performance on all concepts. A more detailed discussion on the adopted continual learning strategies for malware detection as well as the evaluation approaches used in our work is provided in the following sections.

3. Methodology

3.1. Experience Replay Strategies

Our proposed continual malware detection workflow consists of alternate model training and evaluation phases. In the experimental scenario, training concepts (In this work, the terms task and concept are interchangeable. We refer to a concept as a self-consistent behavior of the normal class in an anomaly detection setting, as opposed to the conventional term task used in a classification setting.) C i = 1 , 2 , , N are presented to the model, which is updated accordingly using a replay strategy. In the evaluation phase, one-class models perform inference (anomaly detection) for all evaluation concepts E i = 1 , 2 , , k , and their performance is evaluated.
In our work, we adopt commonly used anomaly detection models in a semi-supervised one-class learning setting that does not require knowledge of anomalous patterns for the model training phase. Therefore, models are trained on background data (normal) and extract anomaly scores on newly observed data. A known drawback of model updates in sequential learning is that models become prone to forgetting, which implies a decrease in the model performance for previously learned concepts as new knowledge is acquired. As a result, the challenge presented by a continual learning scenario is to incorporate new concepts in the model while preserving knowledge acquired from previously observed concepts. In our work, we leverage experience replay to consider stability and plasticity in our base models simultaneously. The replay buffer allows storing a summarized version of all concepts observed so far and leverages this knowledge to prevent forgetting while updating the model. Using a subset of data points allows for small storage requirements for the replay data. Notably, replay is model-agnostic, supporting any machine learning-based anomaly detection model.
For each training concept C i , we store its summarized version R i as part of the replay buffer. The complete experience replay buffer R can be defined as
R = { R i ; i { 1 , 2 , , k } } ,
where k is the number of training concepts observed in the scenario. As a result, the replay buffer contains a subset of samples from each concept and allows us to maintain their representation throughout the whole scenario. Thus, replay-based model training takes into account all concepts observed so far. It is worth noting that data samples in the replay buffer are selected from the original data in each concept without replacement. As a result, each sample can be only included once in the replay buffer.
The replay buffer has a limited size controlled by budget B, which depends on the available storage resources. This budget is typically relatively small in comparison to observed data. In our work, the budget for each concept C i is equal for all concepts in the scenario based on the total budget B. We devise two variants of experience replay given a concept C i :
Random: selects data points at random from C i according to budget B. This strategy gives equal weight to all experiences, regardless of their distance from the centroid of each concept. It favors the diversity of selected data samples, but it may increase the risk of concept overlap and noise as more concepts are observed by anomaly detection models, potentially impacting their ability to discriminate between normal data and anomalies.
Selective: selects data points closest to the centroid of C i according to budget B. It is based on the assumption that experiences that lie closest to the centroid in each concept are the most relevant and should be preserved. It can potentially aid anomaly detection models in defining a restricted concept boundary that does not overlap with other concepts but may be subject to underfitting and limited exploitation of diversity.
Once the replay buffer is formed, the entire buffer (all samples) is used for model training independently from the chosen strategy. Intuitively, the proposed strategies allow models to adapt during the learning scenario by processing new data and adjusting the replay buffer to include a summary of past and new data. Performing training and model updates using data collected in this buffer has the effect of incorporating new relevant information without forgetting past knowledge, which is one of the most desired behaviors in continual learning.

3.2. Scenarios

To define continual learning scenarios, we exploit the characterization defined in [44] for continual/lifelong anomaly detection. The procedure leverages clustering functions to create normal and anomaly concepts (multiple self-consistent sets of data points) based on normal and anomaly data, respectively. In general, a concept corresponds to a new distribution, change in a performed activity, or a new state of the environment. For instance, in network traffic data, the entire normal/benign class can be thought of as a set of concepts describing legitimate traffic: web browsing, file transfer, and video streaming. Similarly, in our malware detection context, a normal concept can be regarded as a set of benign files with similar characteristics, where each file is described by a feature vector, and each concept is characterized by similar vectors, i.e., vectors with numeric feature values in a specific range.
Scenarios are built based upon the selection of the number of desired concepts k for normal (N) and anomaly (A) data from a given dataset and functions ϕ and γ to create multiple self-consistent sets of data points (normal and anomaly concepts, respectively) having common characteristics based on the entire classes.
The scenario creation procedure is described in more detail in Algorithm 1 [44]. First, we create concepts for the normal class through a concept creation function ϕ (Line 1). The function can consider any aspect (or feature values) that can delineate the boundaries of one concept. Second, we create concepts for the anomaly class through the anomalous concepts creation function γ (Line 2). Third, for each normal concept C N i , we select a corresponding anomaly concept C A j using a function λ (Line 3–5). The combination of C N i and C A j is a concept added to the continual learning scenario (Line 6). Each time a concept is built, the selected anomaly concept C A j is removed from the set of available anomaly concepts C A (Line 7). The algorithm returns the resulting scenario as a sequence of concepts (Line 9), each of which may need to be separated into training and evaluation data depending on the learning settings, e.g., unsupervised or semi-supervised.
In this work, we leverage the k-Means clustering algorithm to implement ϕ and γ . The assignment function λ matches each normal concept with an anomaly concept, leading to a combined concept, which allows for model training (normal data) and evaluation (normal and anomaly data). We leverage two continual learning scenario types based on different choices of λ :
  • A: clustered anomaly concepts assigned to the closest normal concept.
  • C: clustered anomaly concepts assigned randomly to normal concepts.
In regards to experiments and the relationship between datasets, scenarios, and corresponding model performance, it is worth noting that the ordering of data samples in the replay buffer does not affect experiments, since samples used for model updates have all equal weight, and models have no notion of time to consider them differently. A different ordering of concepts in the learning scenario may, however, lead to a different model performance on single concepts. Emerging studies on curriculum learning in continual learning scenarios recently studied this phenomenon [45,46,47].
Algorithm 1: Scenario creation protocol [44]
Make 06 00135 i001

4. Experiments

All experiments are run on a machine with an Intel Core i7-9750H CPU and 32 GB of RAM. The code and processed dataset of this paper are available at https://github.com/rcorizzo/cl-malware (accessed on 29 October 2024).

4.1. Datasets

The datasets adopted in our study were selected to satisfy goals of realism, scale, and diversity. First, datasets are representative and significant examples in real-world malware detection, as they are curated by the Canadian Institute for Cybersecurity (CIC), a comprehensive multidisciplinary training, research and development, and entrepreneurial unit that draws on the expertise of researchers in social sciences, business, computer science, engineering, law, and science. Second, they are large-scale, as they contain a large number of samples and malware types. Third, they cover diverse malware types spanning from Android applications to malware memory analysis and infections in PDF files.
CIC-MalMem-2022 [48]: This dataset is designed to test obfuscated malware detection methods through memory. The dataset was created to represent real-world malware types, including Spyware, Ransomware, and Trojan Horse. Debug mode was used for the memory dump process to reproduce typical use cases of users with applications in execution when malware attacks occur. The dataset contains a total of 58,596 instances (29,298 benign, 29,298 malicious).
CIC-Evasive-PDFMal2022 [49]: PDF is the most widely used document format due to its portability and reliability. Unfortunately, its popularity and advanced features have allowed attackers to misuse them to deliver a malicious payload. Authors collected a large number of malicious files from Contagio and VirusTotal and extracted 32 representative features, including 12 general and 25 structural. After processing, this dataset consists of 10,025 instances (4468 benign, 5557 malicious).
CIC-AndMal-2017 [50]: Android malware detection is becoming increasingly challenging for cybersecurity experts, due to the large number of variants released every day. This dataset includes 10,854 samples (4354 malware and 6500 benign) from several sources. Benign data are collected from the Google Play market. A total of 5000 of the collected apps (426 malware and 5065 benign) were installed on real devices. Malware samples are classified into four categories: Adware, Ransomware, Scareware, and SMS Malware.

4.2. Strategies

In addition to the proposed replay-based strategies, we adopt the following baselines in our experiments:
Naive: Models are updated each time a new training concept is available, without exploiting any smart continual learning strategy to provide both adaptation and knowledge retention. This strategy has the goal of simulating lower-bound performance provided by incremental learning, which is expected to gradually or catastrophically forget knowledge of previously presented concepts.
Cumulative: It accumulates all data from concepts observed so far, and it updates the model accordingly. It can be thought of as an upper bound adoptable when unlimited memory resources are available. Despite being unrealistic, it is useful to compare performances against resource-aware learning strategies.
In our experiments, we aim to compare the performance of two variants of experience replay (ER) with naive and cumulative.

4.3. Anomaly Detection Models

We adopt standard anomaly detection learning methods that are often used in unsupervised and semi-supervised settings. We selected seven methods, spanning a large diversity of modeling approaches:
LOF (Local Outlier Factor) [51]: simple but effective anomaly detection method based on a nearest-neighbor-based approach. LOF measures the deviation in the local density of data points with respect to their neighbors and compares the local density of data points with the average local density of nearest neighbors, considering it as an anomaly score.
Considering an object O 1 , local reachability densities are compared with those of its neighbors as
LOF k ( O 1 ) : = Σ O 2 N k ( O 1 ) lr k ( O 2 ) rr k ( O 1 ) N k ( O 1 ) = Σ O 2 N k ( O 1 ) lr k ( O 2 ) N k ( O 1 ) · lr k ( O 1 ) ,
which denotes the average local reachability density of neighbors divided by the local reachability density of the object. If the score is less than 1, the object has a higher density than its neighbors, indicating it is an inlier. A score greater than 1 suggests a lower density than its neighbors, classifying the object as an outlier.
LOF measures the deviation in the local density of data objects with respect to their neighbors. The prediction score returned by LOF for a particular data object is obtained as the ratio between its local density and the average local density of the nearest neighbors.
Considering the k-distance k D as the distance of object O 1 from its k-th nearest neighbors, the notion of reachability distance can be defined as
R D k ( O 1 , O 2 ) = max { k D ( O 2 ) , d ( O 1 , O 2 ) } .
Based on this definition, objects that belong to the k nearest neighbors of O 2 are considered to be equally distant. The local reachability density of an object O 1 defined as lr k ( O 2 ) is the inverse of the average reachability distance of the object O 1 from its neighbors.
A value of approximately 1 indicates that the object is similar to its neighbors (and thus not an outlier). A value below 1 indicates a denser region (i.e., an inlier), whereas values significantly larger than 1 indicate outliers. A value lower than 1 expresses a higher density than neighbors (inlier), while a value greater than 1 denotes a lower density than neighbors (outlier).
IF (Isolation Forests) [52]: The tree and ensemble-based modeling capabilities of IF showed great potential in anomaly detection works. The method computes an isolation score for every data object. The average distance from the tree’s root to the leaf associated with the data object (corresponding to the number of splits required to reach the object) is used to predict the anomaly score. Considering that more pronounced variations in values equal shorter paths in the tree, Isolation Forest uses this information to distinguish an abnormal data object from the rest. The anomaly score is defined as
s ( x , n ) = 2 E ( h ( x ) ) c ( n ) ,
where h ( x ) is the path length of the object x, c ( n ) is the average path length of an unsuccessful search in the Binary Search Tree, and n is the number of external nodes. An anomaly score close to 1 indicates that an object has a high chance of being an anomaly, while scores that are smaller than 0.5 are indicative of a regular (or non-anomaly) data object. OCSVM (One-Class Support Vector Machines) [53]: A well-recognized hyperplane-based method in anomaly detection is due to its ability to detect out-of-distribution data points considering their distance from the decision boundary. It conceptually operates in a similar way to Support Vector Machines, which identify a hyperplane to separate data instances from two classes. However, the one-class learning counterpart uses a hyperplane to encompass all of the background data instances (human essays). Solving the OCSVM optimization problem corresponds to solving the dual quadratic programming problem:
min α 1 2 i j α i α j K x i , x j
subject to the constraints 0 α i 1 ν l and i α i = 1 , where α i is the weight for the instance i, vectors with non-zero weights are defined as support vectors and determine the optimal hyperplane, ν is a parameter that represents a trade-off between the distance of the hyperplane from the origin and the number of instances covered by the hyperplane, l is the number of instances in training data, and K x i , x j is the kernel function. Leveraging the kernel function to project input vectors into a feature space allows for nonlinear decision boundaries. Specifically, a feature map can be defined as
ϕ : X R N ,
where ϕ maps training vectors from the input space X to a dimensional feature space, and the kernel function is defined as
K ( x , y ) = ϕ ( x ) , ϕ ( y )
The adoption of kernel values K ( x , y ) allows one to avoid the explicit computation of feature vectors, with a great improvement in computational efficiency. Common kernels include Linear, Radial Basis Function (RBF), Polynomial, and Sigmoid.
Following the training phase, the OCSVM-learned hyperplane can categorize a new data instance (essay) as regular/normal (human) or different/anomaly (AI-generated) with regard to the training data distribution, based on its geometric location within the decision boundary.
COPOD (Copula-based Outlier-Detection) [54]: Copulas are defined as multi-variate cumulative distribution functions with a uniform marginal probability distribution for each feature. Copula-based anomaly detection predicts the degree of “extremeness” of data samples based on tail probabilities. COPOD is known as a robust parameterless one-class learning method.
HBOS (Histogram-based Outlier Score) [55]: a statistical approach that assumes independence across features. It generates a histogram for each feature and multiplies the inverse height of the bins, assessing the density of all features. Even though feature relationships are ignored (i.e., the method assumes feature independence), this simplification allows the method to converge quickly. HBOS builds histograms in two different modalities: (i) static bin sizes and a preset bin width, and (ii) dynamic bins with a close-to-equal number of bins. This approach, similar in nature to the Naive Bayes algorithm, provides a quick and effective way to identify anomalies.
ABOD (Angle–Base Outlier Detection) [56] computes the variance of weighted cosine scores between data points and their neighbors and considers them as the anomaly score. It is known for efficiently identifying outliers in high-dimensional datasets and for being robust to false positives.
Auto-Encoder (AE) [57]: Auto-encoders compress data into a simplified latent representation and then reconstruct the original data from it. Auto-encoders present a mirrored structure, i.e., the number of layers in the decoding phase matches the encoding phase. The final output layer has the same size as the input layer and reconstructs the input data. A common approach is to train an auto-encoder on normal data and then detect anomalies by identifying data that the model cannot reconstruct well.
Auto-encoders consist of a model architecture with multiple hidden layers, where the output of the i-th hidden layer is the i-th encoding level of the input data. Typically, the model presents a mirrored architecture, where the number of layers for the decoding stage corresponds to the number of layers defined for the encoding stage. Finally, the last layer of the auto-encoder has the same size as the input layer and returns the reconstructed input representation by means of the decoded stage.
After training, the model is used to reconstruct unseen data objects. When a high reconstruction error is observed for a new data object, this is assumed to belong to a different class than that observed during training (normal/benign), i.e., it is classified as malicious. Otherwise, it is recognized to belong to the training data distribution, i.e., it is classified as benign.

4.4. Model Evaluation

We devise an evaluation protocol to assess model performance for any one-class learning model across all concepts in our continual learning scenarios. We initialize a matrix R to store anomaly detection results for specific tasks/concepts. The protocol iterates over training concepts and trains/updates the model accordingly. Following up, models are evaluated on all testing concepts, i.e., previous, current, and future concepts. In the resulting matrix R, entries R i , j contain the ROC–AUC metric of the model on concept j after learning concept i. The resulting matrix R can then be used to compute continual learning metrics directly, such as backward and forward transfer, which allow us to assess model behavior more precisely than standard performance metrics, taking into account model performance on previous, current, and future concepts:
Continual ROC–AUC ( ROC AUC ): Inspired by [58], we propose a variant of ROC-AUC that assesses each model’s performance on all concepts after learning every new concept:
ROC AUC = i j N R i , j N ( N + 1 ) 2
This metric is calculated considering previously learned concepts, including the current concept (corresponding to averaging over N ( N + 1 ) 2 lower triangular entries). One major advantage of ROC–AUC over threshold-dependent metrics such as F1–Score is that it evaluates the model’s performance in a more detailed manner, considering all possible detection thresholds. However, ROC–AUC may be swapped with other metrics of choice in our protocol.
Backward transfer for ROC–AUC ( BWT ): It measures the impact of learning new concepts on the model’s performance of all previously learned concepts. Negative BWT indicates that the model presents a degree of forgetting (a strongly negative BWT value is also known as catastrophic forgetting). Positive BWT signals that learning new concepts benefits models’ performance on previously observed concepts:
BWT = i = 2 N j = 1 i 1 R i , j R j , j N ( N 1 ) 2
Forward transfer for ROC–AUC ( FWT ): It measures the impact of learning each concept on the model’s performance on future concepts, i.e., zero-shot model performance on future concepts:
FWT = i < j N R i , j N ( N 1 ) 2

4.5. Results and Discussion

In this subsection, we analyze the extracted results. The main research question we aim to address is whether the adoption of experience replay is beneficial for the performance of one-class learning models in continual malware detection scenarios. We recall that a replay strategy is in charge of selecting and maintaining a buffer of data samples from previously observed concepts, and it is limited in size by a budget, which represents a subset of available data points for concepts. As the model is provided with a new task (concept), the buffer is updated in order to capture knowledge of all concepts presented in the scenario so far. This behavior should, in principle, equip models with both adaptation and knowledge retention capabilities [39]. As a result, the model should be able to provide a satisfactory performance on all concepts, without a significant degree of forgetting for any of them.
Results in Table 3, Table 4 and Table 5 show the performance achieved by all models on all datasets (the best results for each strategy in both scenarios are marked in bold—all reported ER configurations are with a budget of 0.15 , i.e., 15 % of data for each concept is stored in the replay buffer). We compare replay strategies in terms of anomaly detection performance (ROC-AUC), backward transfer (BWT), and forward transfer (FWT) with respect to the naive (lower bound) and cumulative (upper bound) strategies. Results show that the Replay strategy (ER) brings various degrees of improvement. In some cases, the improvement in ROC-AUC over naive is non-existent, as seen in Table 3 (Scenario: A) for ER (Random) with IF as the base model, i.e., the ROC-AUC is the same (0.511). In many other cases, the improvement is significant. For example, focusing on ABOD, we can observe in Table 5 that ER (Random) increases ROC-AUC from 0.613 to 0.674 (Scenario: A) and from 0.669 to 0.73 (Scenario: C). There are also cases where the improvement is remarkable. This is the case for LOF in Table 3 (from 0.621 to 0.662 in Scenario: A and from 0.586 to 0.654 in Scenario: C) and for ABOD in Table 4 (from 0.591 to 0.864 in Scenario: A and from 0.559 to 0.821 in Scenario: C).
Our results emphasize that some anomaly detection models (LOF, ABOD) are more robust than others in terms of detection performance and resistance to forgetting. One explanation for this phenomenon is that LOF and ABOD are more suitable for datasets with complex and irregular distributions, as they can capture local density variations effectively. In contrast, IF and OCSVM tend to struggle when the data distribution is highly irregular or presents overlapping clusters and are more sensitive to hyperparameter tuning. ABOD and LOF also tend to be more effective at detecting local anomalies or clusters of anomalies, compared to IF and OCSVM.
Overall, we observe that at least one of the replay strategies presents a performance that is equal to or higher than naive in 34 out of 42 configurations (7 models, 3 datasets, 2 scenarios). Ablation results (see Appendix A) reveal the performance achieved with different replay budgets ( 0.05 , 0.1 , 0.15 , 0.2 ) . One interesting result is that model performance does not increase dramatically with larger budgets, which represents a promising direction for lightweight replay strategies in settings with limited memory requirements. It is also worth noting that, in our experiments, model performance did not increase significantly with budget values higher than 0.2 .
Another important consideration is that ROC-AUC generally improves symmetrically with BWT, which means that replay strategies are beneficial for both the anomaly detection performance and the ability for one-class models to reuse knowledge across different tasks. It should be noted that in many of the cases in which replay does not provide an improvement over naive, we observe that all strategies perform poorly for the specific base model, as noted by their poor scores (close to random performance). This phenomenon suggests that the poor performance may not depend on the replay strategy but on predictive models being fundamentally inadequate to deal with the complexity presented by specific datasets or continual learning scenarios.
We note that there is still a gap between results obtained with the replay strategies and the simulated cumulative strategy considered as the upper-bound. We can observe that cumulative presents the best results in 21 out of 42 cases. This result demonstrates that it is possible to achieve a higher performance on the considered malware detection datasets and scenarios and signals the need for more sophisticated continual learning strategies to address this gap.
The heatmaps shown in Figure 3, Figure 4 and Figure 5 allow us to zoom into the performance of different replay strategies on specific tasks/concepts. Considering the CIC-MalMem-2022 dataset presented in Figure 3, we observe that, for the A scenario, the ER strategy manages to preserve knowledge of multiple concepts (3, 10–12), especially in the second half of the scenario, as visible by the darker colored areas. Training the model on concept 12 shows a significant spike in model performance on all concepts, whereas the performance on concept 7 is preserved throughout the entire scenario. However, its performance is still far from the upper bound achieved by cumulative. In the C scenario, ER appears less robust on concept 7 across the entire scenario, although training the model on task 7 appears to boost the performance on all concepts. The performance on concept 13 is preserved throughout the entire scenario. Shifting the focus to the CIC-Evasive-PDFMal2022 dataset in Figure 4, we observe a robust performance for ER in the A scenario, as shown by the dark-colored areas in the lower diagonal part of the heatmap. This performance is quite close to that achieved by the cumulative strategy. However, the performance for ER appears weaker on concept 2 in the C scenario. On the CIC-AndMal2017 dataset, results in Figure 5 show a generally satisfactory performance for ER across the scenario. The model appears relatively robust to forgetting, and its performance is significantly better than naive. One interesting case is that of concept 3, where ER achieves 0.93 after the model is trained on concept 0, possibly due to a high concept similarity. On this dataset, it can also be noted that the difference in performance between the ER and cumulative strategies is minimal, which represents a successful case for ER.

5. Conclusions

In this work, we investigate a semi-supervised one-class continual malware detection workflow, leveraging two model-agnostic experience replay strategies. In addition to adopting a conventional replay strategy that selects data samples at random from each concept, we proposed a selective replay strategy, which attempts to select the most relevant data by prioritizing samples that are closer to the centroid of each concept. Our extensive evaluation involves seven diverse anomaly detection models, three real-world malware detection datasets, and two continual learning scenarios. Our results show that anomaly detection models combined with replay strategies are in some cases effective in providing a proper stability–plasticity trade-off in the presence of complex malware detection scenarios. In our experiments, replay strategies are capable of improving the performance of a naive incremental learning strategy in 34 out of 42 of the considered settings, both in terms of continual ROC-AUC and backward transfer. Our results suggest that the selective replay strategy achieves a similar performance to conventional replay, with slight improvements in some configurations. Moreover, the experiments emphasize that some models (LOF, ABOD) are more robust than others in terms of detection performance and limited forgetting.
Overall, we advocate continual learning as a suitable learning framework for malware detection, with the potential to provide reactive model capabilities in evolving scenarios with limited computational requirements. However, we underscore that further research is required to design more sophisticated strategies in order to fill the existing gap between replay strategies and simulated upper-bound performance, as emphasized by our results with the cumulative strategy. In future work, we aim to explore continual malware detection in task-free/task-agnostic scenarios where task boundaries are not known. Moreover, we aim to investigate new continual learning strategies specifically tailored for malware detection.

Author Contributions

Conceptualization, R.C.; methodology, R.C. and M.C.; software, R.C. and M.C.; validation, R.C.; investigation, R.C. and M.C.; resources, R.C.; data curation, M.C.; writing—original draft preparation, R.C.; writing—review and editing, M.C.; visualization, R.C. and M.C.; supervision, R.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The code and the processed datasets of this paper are available at https://github.com/rcorizzo/cl-malware accessed on 29 October 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Model Hyperparameters

For experience replay (ER), the budget B adopted for results reported in the main paper is 20% of the data samples available in each concept.
For reproducibility, Table A1 shows hyperparameter configurations for all models considered in our experiments. Using ROC-AUC as an evaluation metric allows us to overcome the burden of optimizing hyperparameters that influence different scales for the anomaly score since no decision function based on a threshold is used for the predictions. For all methods, we follow the recommended guidelines or default values provided in the PyOD [59] documentation or recommended in reference papers of the methods, where suitable.
Table A1. Hyperparameters for all the anomaly detection models considered in our experiments. All models are trained and evaluated five times using all hyperparameter values in the sets shown in the table, and the final results are averaged.
Table A1. Hyperparameters for all the anomaly detection models considered in our experiments. All models are trained and evaluated five times using all hyperparameter values in the sets shown in the table, and the final results are averaged.
LOFleaf_size = 20, n_neighbors = 5
IFn_estimators = 85
OCSVMkernel = rbf, gamma = 8, shrinking = True
COPODparameterless

Appendix A.1. Qualitative Analysis

Figure A1. Visualization of extracted concepts via t-SNE: CIC-MalMem-2022 dataset. Normal class (left) and anomaly class (right).
Figure A1. Visualization of extracted concepts via t-SNE: CIC-MalMem-2022 dataset. Normal class (left) and anomaly class (right).
Make 06 00135 g0a1
Figure A2. Visualization of extracted concepts via t-SNE: CIC-Evasive-PDFMal2022 dataset. Normal class (left) and anomaly class (right).
Figure A2. Visualization of extracted concepts via t-SNE: CIC-Evasive-PDFMal2022 dataset. Normal class (left) and anomaly class (right).
Make 06 00135 g0a2
Figure A3. Visualization of extracted concepts via t-SNE: CIC-AndMal-2017 dataset. Normal class (left) and anomaly class (right).
Figure A3. Visualization of extracted concepts via t-SNE: CIC-AndMal-2017 dataset. Normal class (left) and anomaly class (right).
Make 06 00135 g0a3

Appendix A.2. Ablation Experiments

Table A2. Results: CIC-MalMem-2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies including two experience replay (ER) variants in two scenarios (A,C) with different budget rates (B = 0.05, 0.1, 0.15, 0.2).
Table A2. Results: CIC-MalMem-2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies including two experience replay (ER) variants in two scenarios (A,C) with different budget rates (B = 0.05, 0.1, 0.15, 0.2).
Scenario: A
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.3950.0020.4540.4620.0060.5120.5750.0540.6390.3650.0070.442
ER (Random) 0.10.3980.0010.4560.461−0.0010.510.5840.0590.6470.3660.0080.444
ER (Random) 0.150.40.0010.4590.4610.0010.5110.5940.0590.6590.3670.010.447
ER (Random) 0.20.3970.0010.4560.462−0.0060.5110.6010.0640.6690.3680.0110.448
ER (Selective) 0.050.3890.0010.4530.4520.0010.4990.5660.0580.6330.3610.0050.436
ER (Selective) 0.10.387−0.0030.4580.4460.0010.4970.5730.0590.6430.3590.0070.437
ER (Selective) 0.150.39−0.0050.4590.44−0.0050.4850.5860.0720.6620.3590.0080.438
ER (Selective) 0.20.393−0.0070.4580.434−0.0070.4790.6020.0760.6730.360.0060.439
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.4540.0010.5110.612−0.0980.5360.5870.0010.569
ER (Random) 0.10.4550.00.5110.587−0.1560.4760.5880.0010.57
ER (Random) 0.150.4540.0010.5110.549−0.20.4190.5880.00.571
ER (Random) 0.20.4540.0010.5110.511−0.2230.3720.588−0.0010.57
ER (Selective) 0.050.449−0.0010.5040.616−0.0420.5760.5880.0010.569
ER (Selective) 0.10.445−0.0020.5010.604−0.0490.5580.588−0.00.57
ER (Selective) 0.150.444−0.0040.4980.591−0.0530.5460.588−0.00.571
ER (Selective) 0.20.445−0.0080.4930.577−0.0490.5340.5880.0010.571
Scenario: C
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.3840.0020.4760.4970.0010.4860.6240.0420.6160.370.0010.446
ER (Random) 0.10.380.0010.4720.4950.0040.4850.630.050.6340.3740.0030.448
ER (Random) 0.150.3810.0010.4740.4960.0010.4830.6320.0610.6480.3820.0020.45
ER (Random) 0.20.3820.0010.4760.4950.0020.4830.6390.0720.6690.3870.0040.451
ER (Selective) 0.050.403−0.0130.4530.485−0.0090.4660.6150.0450.6070.367−0.0040.437
ER (Selective) 0.10.406−0.0090.4480.472−0.0110.460.6210.050.6230.374−0.0040.436
ER (Selective) 0.150.409−0.0080.4440.465−0.0110.450.630.0640.6540.384−0.0050.435
ER (Selective) 0.20.411−0.010.4410.464−0.010.4470.6410.0720.6670.386−0.0060.434
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.495−0.0010.4790.59−0.1450.4550.5620.0010.591
ER (Random) 0.10.496−0.0010.4790.538−0.250.3210.5730.0010.593
ER (Random) 0.150.495−0.0010.4790.512−0.2650.2860.5640.0010.595
ER (Random) 0.20.495−0.0010.4790.498−0.2470.2940.5640.00.593
ER (Selective) 0.050.489−0.0060.4690.617−0.0140.5820.5660.0020.593
ER (Selective) 0.10.481−0.010.4620.605−0.0210.5680.5640.0010.595
ER (Selective) 0.150.474−0.0120.4580.59−0.0290.5530.5660.00.594
ER (Selective) 0.20.473−0.0140.4550.574−0.0290.5420.5640.0010.593
Table A3. Results: CIC-Evasive-PDFMal2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies including two experience replay (ER) variants in two scenarios (A,C) with different budget rates (B = 0.05, 0.1, 0.15, 0.2).
Table A3. Results: CIC-Evasive-PDFMal2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies including two experience replay (ER) variants in two scenarios (A,C) with different budget rates (B = 0.05, 0.1, 0.15, 0.2).
Scenario: A
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.187−0.2790.3990.555−0.0860.6980.2−0.2630.5570.507−0.0890.613
ER (Random) 0.10.191−0.2820.3790.558−0.130.70.186−0.0790.6740.508−0.0510.628
ER (Random) 0.150.186−0.2660.3830.564−0.1110.7040.184−0.0220.7120.51−0.0220.642
ER (Random) 0.20.187−0.2570.3880.563−0.1060.6890.184−0.0330.7020.51−0.0080.652
ER (Selective) 0.050.187−0.1960.5180.544−0.0950.7110.199−0.3030.5370.508−0.090.61
ER (Selective) 0.10.186−0.1790.5130.557−0.0840.7210.196−0.1630.6240.51−0.0370.63
ER (Selective) 0.150.187−0.1780.5070.565−0.1070.7210.184−0.1230.660.511−0.0130.644
ER (Selective) 0.20.185−0.180.5030.564−0.0520.7310.184−0.1150.6680.5110.0020.653
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.561−0.0260.7780.239−0.1220.8420.427−0.0040.617
ER (Random) 0.10.56−0.010.7760.216−0.1020.8530.346−0.0170.615
ER (Random) 0.150.557−0.0020.7760.221−0.0840.8640.376−0.0490.54
ER (Random) 0.20.553−0.0020.7730.221−0.0710.8750.371−0.0160.551
ER (Selective) 0.050.557−0.0220.7820.203−0.3060.7180.432−0.1650.513
ER (Selective) 0.10.559−0.0170.7780.237−0.20.7940.382−0.1440.517
ER (Selective) 0.150.558−0.0080.7790.232−0.1720.8140.315−0.0280.578
ER (Selective) 0.20.555−0.0090.7750.204−0.1450.8320.451−0.0690.569
Scenario: C
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.383−0.4710.5610.626−0.2450.7220.552−0.0250.6660.579−0.1090.611
ER (Random) 0.10.378−0.4290.5510.657−0.180.7340.5510.0750.7540.582−0.0830.625
ER (Random) 0.150.373−0.4190.5380.652−0.1380.7280.480.0670.7810.594−0.070.631
ER (Random) 0.20.372−0.4110.5250.629−0.090.7150.4820.040.7610.596−0.0590.637
ER (Selective) 0.050.372−0.4640.5820.635−0.2390.7170.502−0.1550.6360.574−0.1080.611
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Selective) 0.10.374−0.3980.5740.641−0.1610.7090.414−0.050.7540.58−0.0820.624
ER (Selective) 0.150.375−0.3830.5650.664−0.1380.7330.405−0.0680.7610.592−0.0670.633
ER (Selective) 0.20.376−0.3680.5520.663−0.1320.7190.3890.0220.7670.595−0.0550.638
ER (Random) 0.050.7−0.1210.7720.483−0.1440.7920.396−0.2540.516
ER (Random) 0.10.704−0.0740.7750.489−0.1030.8180.419−0.1760.535
ER (Random) 0.150.708−0.0630.7730.488−0.090.8210.471−0.270.541
ER (Random) 0.20.708−0.0510.770.487−0.0770.8210.454−0.1820.536
ER (Selective) 0.050.698−0.1250.7550.38−0.2640.7280.352−0.0810.577
ER (Selective) 0.10.703−0.0960.7460.411−0.1620.7910.4280.1410.507
ER (Selective) 0.150.706−0.0750.7530.422−0.1550.7940.481−0.1690.566
ER (Selective) 0.20.706−0.0580.7580.417−0.1550.7940.495−0.1860.535
Table A4. Results: CIC-AndMal2017 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies including two experience replay (ER) variants in two scenarios (A,C) with different budget rates (B = 0.05, 0.1, 0.15, 0.2).
Table A4. Results: CIC-AndMal2017 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies including two experience replay (ER) variants in two scenarios (A,C) with different budget rates (B = 0.05, 0.1, 0.15, 0.2).
Scenario: A
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.62−0.1030.6440.619−0.0550.6440.616−0.0910.6420.518−0.0950.484
ER (Random) 0.10.606−0.1020.6480.599−0.0540.6560.596−0.1290.6220.522−0.0980.5
ER (Random) 0.150.608−0.1050.6480.594−0.0360.6730.589−0.1020.6360.524−0.10.51
ER (Random) 0.20.604−0.0990.6460.609−0.0270.6650.581−0.0890.6430.526−0.0980.52
ER (Selective) 0.050.666−0.1230.6350.596−0.0630.6390.654−0.1220.620.52−0.0970.488
ER (Selective) 0.10.664−0.120.6380.597−0.0420.6620.648−0.1350.610.524−0.0980.506
ER (Selective) 0.150.664−0.1140.6410.602−0.0630.6450.637−0.10.6290.526−0.0960.518
ER (Selective) 0.20.662−0.1070.6450.594−0.0150.6780.631−0.0890.6350.528−0.0910.528
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.535−0.040.6220.616−0.0030.6730.584−0.0050.642
ER (Random) 0.10.538−0.0310.6260.5910.0050.6750.612−0.0060.649
ER (Random) 0.150.539−0.0260.6280.60.0030.6740.594−0.0130.65
ER (Random) 0.20.542−0.0180.6310.6020.0060.6820.588−0.0090.641
ER (Selective) 0.050.532−0.040.6230.675−0.0570.6520.6030.0120.651
ER (Selective) 0.10.532−0.0320.6280.68−0.030.6680.5980.0010.654
ER (Selective) 0.150.533−0.0250.6310.68−0.0290.6680.622−0.0020.652
ER (Selective) 0.20.533−0.020.6330.68−0.0310.6660.6670.0010.653
Scenario: C
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.6−0.2620.6830.587−0.2360.6730.6−0.2140.6770.431−0.1590.545
ER (Random) 0.10.591−0.2420.6860.579−0.1670.680.585−0.2010.7070.438−0.1430.553
ER (Random) 0.150.591−0.2240.6920.592−0.1510.6780.573−0.1840.7290.444−0.130.56
ER (Random) 0.20.595−0.2110.690.575−0.1440.6720.569−0.1580.7510.449−0.1190.565
ER (Selective) 0.050.602−0.2750.6870.572−0.2140.6880.562−0.2620.6940.431−0.1570.546
ER (Selective) 0.10.601−0.2540.6960.604−0.1850.6720.552−0.2270.7140.438−0.1380.556
ER (Selective) 0.150.602−0.2340.7010.595−0.1540.6780.553−0.1970.7260.444−0.1220.563
ER (Selective) 0.20.601−0.220.7030.593−0.1380.6710.552−0.1760.7390.449−0.1110.568
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
ER (Random) 0.050.555−0.1080.630.593−0.180.7070.626−0.0360.635
ER (Random) 0.10.555−0.0920.630.594−0.150.7210.613−0.0350.629
ER (Random) 0.150.557−0.0830.6280.586−0.1350.7260.615−0.0360.63
ER (Random) 0.20.559−0.0760.6290.59−0.1310.7270.619−0.0360.628
ER (Selective) 0.050.552−0.1070.6210.576−0.2110.7170.613−0.0550.646
ER (Selective) 0.10.553−0.0950.6260.581−0.1790.7290.614−0.0610.641
ER (Selective) 0.150.554−0.0860.6240.581−0.1690.730.616−0.0580.647
ER (Selective) 0.20.557−0.0760.6270.583−0.1560.7370.625−0.0450.644

References

  1. Almashhadani, A.O.; Carlin, D.; Kaiiali, M.; Sezer, S. MFMCNS: A multi-feature and multi-classifier network-based system for ransomworm detection. Comput. Secur. 2022, 121, 102860. [Google Scholar] [CrossRef]
  2. Qiu, J.; Han, Q.L.; Luo, W.; Pan, L.; Nepal, S.; Zhang, J.; Xiang, Y. Cyber code intelligence for android malware detection. IEEE Trans. Cybern. 2022, 53, 617–627. [Google Scholar] [CrossRef] [PubMed]
  3. Zhu, H.j.; Gu, W.; Wang, L.m.; Xu, Z.c.; Sheng, V.S. Android malware detection based on multi-head squeeze-and-excitation residual network. Expert Syst. Appl. 2023, 212, 118705. [Google Scholar] [CrossRef]
  4. Hessel, M.; Modayil, J.; Van Hasselt, H.; Schaul, T.; Ostrovski, G.; Dabney, W.; Horgan, D.; Piot, B.; Azar, M.; Silver, D. Rainbow: Combining improvements in deep reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  5. Gu, S.; Lillicrap, T.; Sutskever, I.; Levine, S. Continuous deep q-learning with model-based acceleration. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 20–22 June 2016; pp. 2829–2838. [Google Scholar]
  6. Saglam, B.; Mutlu, F.B.; Cicek, D.C.; Kozat, S.S. Actor prioritized experience replay. J. Artif. Intell. Res. 2023, 78, 639–672. [Google Scholar] [CrossRef]
  7. Andrychowicz, M.; Wolski, F.; Ray, A.; Schneider, J.; Fong, R.; Welinder, P.; McGrew, B.; Tobin, J.; Pieter Abbeel, O.; Zaremba, W. Hindsight experience replay. arXiv 2017, arXiv:1707.01495. [Google Scholar]
  8. Parisi, G.I.; Kemker, R.; Part, J.L.; Kanan, C.; Wermter, S. Continual lifelong learning with neural networks: A review. Neural Netw. 2019, 113, 54–71. [Google Scholar] [CrossRef]
  9. Shakya, A.K.; Pillai, G.; Chakrabarty, S. Reinforcement learning algorithms: A brief survey. Expert Syst. Appl. 2023, 231, 120495. [Google Scholar] [CrossRef]
  10. Bensaoud, A.; Kalita, J.; Bensaoud, M. A survey of malware detection using deep learning. Mach. Learn. Appl. 2024, 16, 100546. [Google Scholar] [CrossRef]
  11. Zhu, D.; Jin, H.; Yang, Y.; Wu, D.; Chen, W. DeepFlow: Deep learning-based malware detection by mining Android application for abnormal usage of sensitive data. In Proceedings of the 2017 IEEE Symposium on Computers and Communications (ISCC), Heraklion, Greece, 3–6 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 438–443. [Google Scholar]
  12. HaddadPajouh, H.; Dehghantanha, A.; Khayami, R.; Choo, K.K.R. A deep recurrent neural network based approach for internet of things malware threat hunting. Future Gener. Comput. Syst. 2018, 85, 88–96. [Google Scholar] [CrossRef]
  13. Lee, Y.; Wang, X.; Liao, X.; Wang, X. Understanding illicit UI in iOS apps through hidden UI analysis. IEEE Trans. Dependable Secur. Comput. 2019, 18, 2390–2402. [Google Scholar] [CrossRef]
  14. Beaman, C.; Barkworth, A.; Akande, T.D.; Hakak, S.; Khan, M.K. Ransomware: Recent advances, analysis, challenges and future research directions. Comput. Secur. 2021, 111, 102490. [Google Scholar] [CrossRef] [PubMed]
  15. Fernando, D.W.; Komninos, N. FeSA: Feature selection architecture for ransomware detection under concept drift. Comput. Secur. 2022, 116, 102659. [Google Scholar] [CrossRef]
  16. Liu, M.; Yang, Q.; Wang, W.; Liu, S. Semi-Supervised Encrypted Malicious Traffic Detection Based on Multimodal Traffic Characteristics. Sensors 2024, 24, 6507. [Google Scholar] [CrossRef] [PubMed]
  17. Memon, M.; Unar, A.A.; Ahmed, S.S.; Daudpoto, G.H.; Jaffari, R. Feature-based semi-supervised learning approach to android malware detection. Eng. Proc. 2023, 32, 6. [Google Scholar] [CrossRef]
  18. Yu, X.; Lin, G.; Hu, X.; Keung, J.W.; Xia, X. Less is More: Unlocking Semi-Supervised Deep Learning for Vulnerability Detection. Acm Trans. Softw. Eng. Methodol. 2024. [Google Scholar] [CrossRef]
  19. Eren, M.E.; Alexandrov, B.S.; Nicholas, C. Classifying Malware Using Tensor Decomposition. In Malware: Handbook of Prevention and Detection; Springer: Berlin/Heidelberg, Germany, 2024; pp. 3–36. [Google Scholar]
  20. Eren, M.E.; Bhattarai, M.; Joyce, R.J.; Raff, E.; Nicholas, C.; Alexandrov, B.S. Semi-supervised classification of malware families under extreme class imbalance via hierarchical non-negative matrix factorization with automatic model selection. ACM Trans. Priv. Secur. 2023, 26, 1–27. [Google Scholar] [CrossRef]
  21. Van de Ven, G.M.; Tolias, A.S. Three scenarios for continual learning. arXiv 2019, arXiv:1904.07734. [Google Scholar]
  22. De Lange, M.; Tuytelaars, T. Continual Prototype Evolution: Learning Online From Non-Stationary Data Streams. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 8250–8259. [Google Scholar]
  23. Cossu, A.; Graffieti, G.; Pellegrini, L.; Maltoni, D.; Bacciu, D.; Carta, A.; Lomonaco, V. Is Class-Incremental Enough for Continual Learning? arXiv 2022, arXiv:2112.02925. [Google Scholar] [CrossRef]
  24. Dragoi, M.; Burceanu, E.; Haller, E.; Manolache, A.; Brad, F. AnoShift: A distribution shift benchmark for unsupervised anomaly detection. Adv. Neural Inf. Process. Syst. 2022, 35, 32854–32867. [Google Scholar]
  25. Sharif Razavian, A.; Azizpour, H.; Sullivan, J.; Carlsson, S. CNN features off-the-shelf: An astounding baseline for recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, 23–28 June 2014; pp. 806–813. [Google Scholar]
  26. Lomonaco, V.; Maltoni, D. Core50: A new dataset and benchmark for continuous object recognition. In Proceedings of the Conference on Robot Learning—PMLR, Mountain View, CA, USA, 13–15 November 2017; pp. 17–26. [Google Scholar]
  27. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
  28. Li, Z.; Hoiem, D. Learning without Forgetting. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2935–2947. [Google Scholar] [CrossRef] [PubMed]
  29. Diethe, T.; Borchert, T.; Thereska, E.; Balle, B.; Lawrence, N. Continual learning in practice. arXiv 2018, arXiv:1903.05202. [Google Scholar]
  30. Mignone, P.; Corizzo, R.; Ceci, M. Distributed and explainable GHSOM for anomaly detection in sensor networks. Mach. Learn. 2024, 113, 4445–4486. [Google Scholar] [CrossRef]
  31. Mallya, A.; Lazebnik, S. PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning. arXiv 2017, arXiv:1711.05769. [Google Scholar]
  32. Kang, H.; Mina, R.J.L.; Rizky, S.; Madjid, H.; Yoon, J.; Hasegawa-Johnson, M.; Ju-Hwang, S.; Yoo, C.D. Forget-free Continual Learning with Winning Subnetworks. ICML 2022, 162, 10734–10750. [Google Scholar]
  33. Pietroń, M.; Żurek, D.; Faber, K.; Corizzo, R. Ada-QPacknet–adaptive pruning with bit width reduction as an efficient continual learning method without forgetting. In Proceedings of the European Conference on Artificial Intelligence (ECAI), Krakow, Poland, 30 September–4 October 2023; pp. 1882–1889. [Google Scholar]
  34. Kumari, L.; Wang, S.; Zhou, T.; Bilmes, J.A. Retrospective adversarial replay for continual learning. Adv. Neural Inf. Process. Syst. 2022, 35, 28530–28544. [Google Scholar]
  35. Van de Ven, G.M.; Siegelmann, H.T.; Tolias, A.S. Brain-inspired replay for continual learning with artificial neural networks. Nat. Commun. 2020, 11, 4069. [Google Scholar] [CrossRef]
  36. Li, X.; Tang, B.; Li, H. AdaER: An adaptive experience replay approach for continual lifelong learning. Neurocomputing 2024, 572, 127204. [Google Scholar] [CrossRef]
  37. Wang, Q.; Ji, Z.; Pang, Y.; Zhang, Z. Uncertainty-aware enhanced dark experience replay for continual learning. Appl. Intell. 2024, 54, 7135–7150. [Google Scholar] [CrossRef]
  38. Lim, W.S.; Zhou, Y.; Kim, D.W.; Lee, J. MixER: Mixup-Based Experience Replay for Online Class-Incremental Learning. IEEE Access 2024, 12, 41801–41814. [Google Scholar] [CrossRef]
  39. Buzzega, P.; Boschini, M.; Porrello, A.; Calderara, S. Rethinking experience replay: A bag of tricks for continual learning. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 2180–2187. [Google Scholar]
  40. Faber, K.; Sniezynski, B.; Corizzo, R. Distributed Continual Intrusion Detection: A Collaborative Replay Framework. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 3255–3263. [Google Scholar]
  41. Shin, H.; Lee, J.K.; Kim, J.; Kim, J. Continual Learning with Deep Generative Replay. In Proceedings of the NeurIPS; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  42. Faber, K.; Corizzo, R.; Sniezynski, B.; Japkowicz, N. VLAD: Task-agnostic VAE-based lifelong anomaly detection. Neural Netw. 2023, 165, 248–273. [Google Scholar] [CrossRef] [PubMed]
  43. Rahman, M.S.; Coull, S.; Wright, M. On the Limitations of Continual Learning for Malware Classification. arXiv 2022, arXiv:2208.06568. [Google Scholar]
  44. Faber, K.; Corizzo, R.; Sniezynski, B.; Japkowicz, N. Lifelong Learning for Anomaly Detection: New Challenges, Perspectives, and Insights. arXiv 2023, arXiv:2303.07557. [Google Scholar] [CrossRef]
  45. Fayek, H.M.; Cavedon, L.; Wu, H.R. Progressive learning: A deep learning framework for continual learning. Neural Netw. 2020, 128, 345–357. [Google Scholar] [CrossRef]
  46. Faber, K.; Zurek, D.; Pietron, M.; Japkowicz, N.; Vergari, A.; Corizzo, R. From MNIST to ImageNet and back: Benchmarking continual curriculum learning. Mach. Learn. 2024, 113, 8137–8164. [Google Scholar] [CrossRef]
  47. Kesgin, H.T.; Amasyali, M.F. Cyclical curriculum learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 12864–12872. [Google Scholar] [CrossRef]
  48. Carrier, T. Detecting obfuscated malware using memory feature engineering. In Proceedings of the ICISSP, Virtual, 11–13 February 2022; pp. 177–188. [Google Scholar]
  49. Issakhani, M.; Victor, P.; Tekeoglu, A.; Lashkari, A.H. PDF Malware Detection based on Stacking Learning. In Proceedings of the ICISSP, Virtual, 9–11 February 2022; pp. 562–570. [Google Scholar]
  50. Lashkari, A.H.; Kadir, A.F.A.; Taheri, L.; Ghorbani, A.A. Toward developing a systematic approach to generate benchmark android malware datasets and classification. In Proceedings of the 2018 International Carnahan Conference on Security Technology (ICCST), Madrid, Spain, 22–25 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–7. [Google Scholar]
  51. Breunig, M.M.; Kriegel, H.P.; Ng, R.T.; Sander, J. LOF: Identifying density-based local outliers. In Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, TX, USA, 15–18 May 2000; pp. 93–104. [Google Scholar]
  52. Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; IEEE: Piscataway, NJ, USA; pp. 413–422. [Google Scholar]
  53. Schölkopf, B.; Williamson, R.C.; Smola, A.J.; Shawe-Taylor, J.; Platt, J.C. Support vector method for novelty detection. In Advances in Neural Information Processing Systems; MIT Press: Denver, CO, USA, 2000; pp. 582–588. [Google Scholar]
  54. Li, Z.; Zhao, Y.; Botta, N.; Ionescu, C.; Hu, X. COPOD: Copula-Based Outlier Detection. In Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy, 17–20 November 2020; pp. 1118–1123. [Google Scholar] [CrossRef]
  55. Goldstein, M.; Score, A.D.H.b.O. A fast Unsupervised Anomaly Detection Algorithm. In Proceedings of the KI-2012: Poster and Demo Track, Saarbrücken, Germany, 24–27 September 2012; pp. 59–63. [Google Scholar]
  56. Kriegel, H.; Schubert, M.; Zimek, A. Angle-based outlier detection in high-dimensional data. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 444–452. [Google Scholar] [CrossRef]
  57. Gehring, J.; Miao, Y.; Metze, F.; Waibel, A. Extracting deep bottleneck features using stacked auto-encoders. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 3377–3381. [Google Scholar]
  58. Díaz-Rodríguez, N.; Lomonaco, V.; Filliat, D.; Maltoni, D. Don’t forget, there is more than forgetting: New metrics for Continual Learning. arXiv 2018, arXiv:1810.13166. [Google Scholar]
  59. Zhao, Y.; Nasrullah, Z.; Li, Z. PyOD: A Python Toolbox for Scalable Outlier Detection. J. Mach. Learn. Res. 2019, 20, 1–7. [Google Scholar]
Figure 1. Semi-supervised one-class continual malware detection workflow consisting of model training and evaluation phases. Training concepts C i contain exclusively normal data, whereas evaluation concepts E i contain normal and anomalous data points. The experience replay component updates the replay buffer R each time a new training concept C i is presented, according to the budget B. Two strategies are used: Random and Selective.
Figure 1. Semi-supervised one-class continual malware detection workflow consisting of model training and evaluation phases. Training concepts C i contain exclusively normal data, whereas evaluation concepts E i contain normal and anomalous data points. The experience replay component updates the replay buffer R each time a new training concept C i is presented, according to the budget B. Two strategies are used: Random and Selective.
Make 06 00135 g001
Figure 2. Example of model evaluation in continual malware detection: concept-level ROC-AUC as a heatmap (a) and as a line plot showing performance over time (b)—CIC-MalMem-2022 dataset—Strategy: Cumulative—Scenario: A—Model: LOF. The results show that learning a new concept without forgetting previous concepts leads to a more comprehensive and robust model, which results in a performance improvement on all other concepts.
Figure 2. Example of model evaluation in continual malware detection: concept-level ROC-AUC as a heatmap (a) and as a line plot showing performance over time (b)—CIC-MalMem-2022 dataset—Strategy: Cumulative—Scenario: A—Model: LOF. The results show that learning a new concept without forgetting previous concepts leads to a more comprehensive and robust model, which results in a performance improvement on all other concepts.
Make 06 00135 g002
Figure 3. Continual ROC-AUC performance (CIC-MalMem-2022 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class model (LOF).
Figure 3. Continual ROC-AUC performance (CIC-MalMem-2022 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class model (LOF).
Make 06 00135 g003
Figure 4. Continual ROC-AUC performance (CIC-Evasive-PDFMal2022 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class model (ABOD).
Figure 4. Continual ROC-AUC performance (CIC-Evasive-PDFMal2022 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class model (ABOD).
Make 06 00135 g004
Figure 5. Continual ROC-AUC performance (CIC-AndMal2017 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class models (ABOD and LOF, respectively).
Figure 5. Continual ROC-AUC performance (CIC-AndMal2017 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class models (ABOD and LOF, respectively).
Make 06 00135 g005
Table 1. Comparison of replay in reinforcement learning and continual learning settings.
Table 1. Comparison of replay in reinforcement learning and continual learning settings.
Reinforcement Learning (RL)Continual Learning (CL)
GoalImprove convergenceMitigate catastrophic forgetting
Stored Data Transitions or episodesTask data or summaries
Sampling Random or prioritizedTask or diversity-aware
Constraints Less constrained by memoryHighly memory-constrained
Focus Single task, dynamic explorationMulti-task knowledge retention
Table 2. Comparison of surveyed research works on supervised, semi-supervised, and continual learning, based on three key aspects considered in our study. The first two groups are strictly focused on malware detection. The third group includes continual learning works focused on different downstream tasks such as image classification, reinforcement learning, and anomaly detection in different domains.
Table 2. Comparison of surveyed research works on supervised, semi-supervised, and continual learning, based on three key aspects considered in our study. The first two groups are strictly focused on malware detection. The third group includes continual learning works focused on different downstream tasks such as image classification, reinforcement learning, and anomaly detection in different domains.
ApproachesLearn from
Known Malware
Patterns
Generalize to
Unseen Malware
types
Provide Adaptation
and Knowledge
Retention
MFMCNS [1]X
Qiu et al. [2]X
Zhu et al. [3]X
Deepflow [11]X
HaddadPajouh et al. [12]X
Lee et al. [13]X
Beaman et al. [14]X
FeSA [15]X
Liu et al. [16]XX
Memon et al. [17]XX
Yu et al. [18]XX
Eren et al. [19]XX
Eren et al. [20]XX
Razavian et al. [25] X
EWC [27] X
LWF [28] X
Diethe et al. [29] X
Mignone et al. [30] X
Packnet [31] X
WSN [32] X
Ada-Q-Packnet [33] X
RAR [34] X
Van De Ven et al. [35] X
AdaER [36] X
UAS [37] X
MixER [38] X
Buzzega et al. [39] X
Faber et al. [40] X
Shin et al. [41] X
Rahman et al. [43]X X
ProposedXXX
Table 3. Results: CIC-MalMem-2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC (AUC) with all models and strategies, including two experience replay (ER) variants in two scenarios (A,C). For all metrics, higher values are indicative of a better performance: zero-shot model capabilities on future concepts (FWT), models’ ability to avoid forgetting and improve on previously learned concepts (BWT), and anomaly detection performance (AUC). Values in bold highlight the best model for each strategy according to AUC.
Table 3. Results: CIC-MalMem-2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC (AUC) with all models and strategies, including two experience replay (ER) variants in two scenarios (A,C). For all metrics, higher values are indicative of a better performance: zero-shot model capabilities on future concepts (FWT), models’ ability to avoid forgetting and improve on previously learned concepts (BWT), and anomaly detection performance (AUC). Values in bold highlight the best model for each strategy according to AUC.
Scenario: A
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.3980.0040.4580.460.0030.5110.5610.0490.6210.3640.0060.44
Cumulative0.3950.00.4550.4620.00.510.6690.1220.7830.3830.0060.469
ER (Random)0.40.0010.4590.4610.0010.5110.5940.0590.6590.3670.010.447
ER (Selective)0.39−0.0050.4590.44−0.0050.4850.5860.0720.6620.3590.0080.438
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.4550.0010.5110.624−0.040.5860.5860.00.569
Cumulative0.455−0.0010.5110.315−0.0490.2890.587−0.0010.569
ER (Random)0.4540.0010.5110.549−0.20.4190.5880.00.571
ER (Selective)0.444−0.0040.4980.591−0.0530.5460.588−0.00.571
Scenario: C
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.3830.0010.4720.4980.0030.4850.610.0250.5860.3680.0010.444
Cumulative0.3810.0010.4750.4950.0020.4820.6980.160.8220.4050.0040.455
ER (Random)0.3810.0010.4740.4960.0010.4830.6320.0610.6480.3820.0020.45
ER (Selective)0.409−0.0080.4440.465−0.0110.450.630.0640.6540.384−0.0050.435
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.495−0.0010.4780.629−0.0090.5910.5620.00.592
Cumulative0.495−0.00.4780.323−0.0710.2830.562−0.00.591
ER (Random)0.495−0.0010.4790.512−0.2650.2860.5640.0010.595
ER (Selective)0.474−0.0120.4580.59−0.0290.5530.5660.00.594
Table 4. Results: CIC-Evasive-PDFMal2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies, including two experience replay (ER) variants in two scenarios (A,C). For all metrics, higher values are indicative of a better performance: zero-shot model capabilities on future concepts (FWT), models’ ability to avoid forgetting and improve on previously learned concepts (BWT), and anomaly detection performance (AUC). Values in bold highlight the best model for each strategy according to AUC.
Table 4. Results: CIC-Evasive-PDFMal2022 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies, including two experience replay (ER) variants in two scenarios (A,C). For all metrics, higher values are indicative of a better performance: zero-shot model capabilities on future concepts (FWT), models’ ability to avoid forgetting and improve on previously learned concepts (BWT), and anomaly detection performance (AUC). Values in bold highlight the best model for each strategy according to AUC.
Scenario: A
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.187−0.3560.5080.566−0.1380.7390.205−0.4010.4840.505−0.2040.559
Cumulative0.18−0.1540.410.549−0.0610.6990.173−0.0140.7330.5150.0540.698
ER (Random)0.186−0.2660.3830.564−0.1110.7040.184−0.0220.7120.51−0.0220.642
ER (Selective)0.187−0.1780.5070.565−0.1070.7210.184−0.1230.660.511−0.0130.644
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.56−0.0770.7840.249−0.5120.5910.27−0.170.507
Cumulative0.5580.0060.7710.214−0.0160.9090.3150.0430.573
ER (Random)0.557−0.0020.7760.221−0.0840.8640.376−0.0490.54
ER (Selective)0.558−0.0080.7790.232−0.1720.8140.315−0.0280.578
Scenario: C
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.37−0.6490.5340.575−0.3740.7030.439−0.5410.4920.574−0.1390.596
Cumulative0.301−0.1990.3990.651−0.0240.70.3920.0020.7490.606−0.0030.671
ER (Random)0.373−0.4190.5380.652−0.1380.7280.480.0670.7810.594−0.070.631
ER (Selective)0.375−0.3830.5650.664−0.1380.7330.405−0.0680.7610.592−0.0670.633
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.656−0.2730.7590.395−0.6260.5590.568−0.3960.528
Cumulative0.728−0.0080.7750.411−0.0310.850.397−0.0130.581
ER (Random)0.708−0.0630.7730.488−0.090.8210.471−0.270.541
ER (Selective)0.706−0.0750.7530.422−0.1550.7940.481−0.1690.566
Table 5. Results: CIC-AndMal2017 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies, including two experience replay (ER) variants in two scenarios (A,C). For all metrics, higher values are indicative of a better performance: zero-shot model capabilities on future concepts (FWT), models’ ability to avoid forgetting and improve on previously learned concepts (BWT), and anomaly detection performance (AUC). Values in bold highlight the best model for each strategy according to AUC.
Table 5. Results: CIC-AndMal2017 dataset: forward transfer (FWT), backward transfer (BWT), and ROC-AUC with all models and strategies, including two experience replay (ER) variants in two scenarios (A,C). For all metrics, higher values are indicative of a better performance: zero-shot model capabilities on future concepts (FWT), models’ ability to avoid forgetting and improve on previously learned concepts (BWT), and anomaly detection performance (AUC). Values in bold highlight the best model for each strategy according to AUC.
Scenario: A
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.668−0.1070.6350.616−0.090.6150.652−0.0760.650.508−0.0840.462
Cumulative0.606−0.0820.6550.604−0.0240.6580.579−0.0030.6730.526−0.0370.586
ER (Random)0.608−0.1050.6480.594−0.0360.6730.589−0.1020.6360.524−0.10.51
ER (Selective)0.664−0.1140.6410.602−0.0630.6450.637−0.10.6290.526−0.0960.518
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.532−0.0630.6120.663−0.1180.6130.66−0.0020.66
Cumulative0.55−0.0050.6440.5950.0010.6790.61−0.0080.645
ER (Random)0.539−0.0260.6280.60.0030.6740.594−0.0130.65
ER (Selective)0.533−0.0250.6310.68−0.0290.6680.622−0.0020.652
Scenario: C
StrategyOCSVMIFLOFCOPOD
FWTBWTAUCFWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.595−0.3430.670.583−0.3060.6540.584−0.3170.6710.42−0.180.536
Cumulative0.606−0.1090.690.612−0.0510.6710.535−0.0710.7580.471−0.0440.58
ER (Random)0.591−0.2240.6920.592−0.1510.6780.573−0.1840.7290.444−0.130.56
ER (Selective)0.602−0.2340.7010.595−0.1540.6780.553−0.1970.7260.444−0.1220.563
StrategyHBOSABODAE
FWTBWTAUCFWTBWTAUCFWTBWTAUC
Naive0.55−0.1410.6220.571−0.3470.6690.622−0.0920.64
Cumulative0.569−0.0290.6320.566−0.0820.7270.624−0.0370.659
ER (Random)0.557−0.0830.6280.586−0.1350.7260.615−0.0360.63
ER (Selective)0.554−0.0860.6240.581−0.1690.730.616−0.0580.647
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chin, M.; Corizzo, R. Continual Semi-Supervised Malware Detection. Mach. Learn. Knowl. Extr. 2024, 6, 2829-2854. https://doi.org/10.3390/make6040135

AMA Style

Chin M, Corizzo R. Continual Semi-Supervised Malware Detection. Machine Learning and Knowledge Extraction. 2024; 6(4):2829-2854. https://doi.org/10.3390/make6040135

Chicago/Turabian Style

Chin, Matthew, and Roberto Corizzo. 2024. "Continual Semi-Supervised Malware Detection" Machine Learning and Knowledge Extraction 6, no. 4: 2829-2854. https://doi.org/10.3390/make6040135

APA Style

Chin, M., & Corizzo, R. (2024). Continual Semi-Supervised Malware Detection. Machine Learning and Knowledge Extraction, 6(4), 2829-2854. https://doi.org/10.3390/make6040135

Article Metrics

Back to TopTop