Next Article in Journal
Comparison of Wearable and Clinical Devices for Acquisition of Peripheral Nervous System Signals
Previous Article in Journal
New Service Virtualisation Approach to Generate the Categorical Fields in the Service Response
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Continual Learning Strategy in One-Stage Object Detection Framework Based on Experience Replay for Autonomous Driving Vehicle

1
Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan
2
Information and Communications Research Laboratories, Embedded Vision and Graphics Technology Department, Division for Embedded System and SoC Technology, Industrial Technology Research Institute, Hsinchu 31057, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(23), 6777; https://doi.org/10.3390/s20236777
Submission received: 12 November 2020 / Revised: 24 November 2020 / Accepted: 24 November 2020 / Published: 27 November 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Object detection is an important aspect for autonomous driving vehicles (ADV), which may comprise of a machine learning model that detects a range of classes. As the deployment of ADV widens globally, the variety of objects to be detected may increase beyond the designated range of classes. Continual learning for object detection essentially ensure a robust adaptation of a model to detect additional classes on the fly. This study proposes a novel continual learning method for object detection that learns new object class(es) along with cumulative memory of classes from prior learning rounds to avoid any catastrophic forgetting. The results of PASCAL VOC 2007 have suggested that the proposed ER method obtains 4.3% of mAP drop compared against the all-classes learning, which is the lowest amongst other prior arts.

1. Introduction

A deep learning algorithm is expected to learn from data that are always changing. Consequently, a deep learning model is required to be trained repeatedly from the ground-up to stay relevant. In nature, humans are capable of learning new things continuously while retaining their previous knowledge. This ability to learn new things continuously is also known as continual or life-long learning. In contrast, for a machine, especially a deep learning model, learning new things naïvely means overwriting its previous knowledge. Instead of learning additional knowledge, the model will lose its ability to detect previous class(es). However, naïvely forcing a model that is already trained to learn additional classes without any strategy can lead to catastrophic forgetting. Catastrophic forgetting is a phenomenon in which a model forgets the previously learned knowledge. This forgetting of the previous knowledge is a major shortcoming of the convolutional neural networks (CNN). In order to preserve the performance on previous knowledge, a training strategy is needed to overcome the shortcoming of the naive training while learning additional classes [1]. Continual learning challenge has persisted for decades in the deep learning field. Moreover, as new data with more classes have become easier to obtain in recent years, continual learning has gained more attention from the research community [2,3].
The phenomenon of catastrophic forgetting is illustrated in Figure 1. Suppose that a model l [ 0 ] is capable of detecting n number of classes in l [ 0 ] dataset, then given a certain condition, the model is required to detect t additional classes in l [ 1 ] dataset, as illustrated in Figure 1. If in the model l [ 1 ] training, for the purpose of detecting these t number of additional classes, the dataset that is provided only contains the (n + 1)th to tth classes without providing previously learned (1st to nth) classes’ dataset, model l [ 1 ] will forget these previously learned classes. As a result the model l [ 1 ] loses its ability to detect any object from dataset l [ 0 ] . This phenomenon happens because weights and biases of previously learned classes get completely overwritten when training for additional classes. Any research attempt at dealing with catastrophic forgetting is also known as continual learning. Generally, continual learning is represented as incremental learning and active learning in the research literature.
In this era of increasing data, the autonomous driving vehicle (ADV) datasets are also developing. Thus, the complexity of deep learning application for ADV has increased with the introduction of novel vehicles and traffic signs. Therefore, it is expected that existing models could be updated to detect new classes [4,5,6,7,8]. Furthermore, the public dataset available on the Internet may not share the same classes and characteristics due to the distinctive driving environment and vehicles in a different country. For instance, in mainstream ADV datasets that are publicly accessible like KITTI [9] and Cityscape [10], some unique classes available in Taiwan, such as Scooter, can not be found in those datasets. For generic classes presented in public datasets, the developer may be able to obtain a pre-trained model with high accuracy trained on these public datasets. Nevertheless, this high-performance model still needs to be trained again to detect certain classes that are not present in other public datasets. Thus, it becomes increasingly urgent to increase the number of classes that our model can detect without risking the drop in accuracy on the previously learned classes by utilizing continual learning in ADV.
Numerous researches have been proposed to deal with catastrophic forgetting. However, most of those solutions only focus on the classification problem while the problem of catastrophic forgetting on object detection is still largely untouched [11,12,13,14]. Even though one can attach an object proposal algorithm before the classification network to create an object detection framework, the resulting object detection framework would be considered as a two-stage detection framework. However, this framework is not suitable for real-time application because of its high inference time [1]. In gradient episodic memory (GEM) [15], previous data are stored in an episodic memory to avoid forgetting previous knowledge in the current continual training. They offer forward and backward transfer of knowledge. However, the backward transfer is essential for the previous task that increases the computation time. Moreover, they used a gradient constraint approach. This approach limits the gradient to prevent the weight from being drastically changed. While such a strategy is able to preserve the weights to some degree, it also limits the network’s ability to learn new classes effectively. Consequently, this strategy often faces a dilemma where it should choose between preserving the previously learned classes’ weights or radically changes its weights to learn new classes effectively. In recent years, a new strategy which employs knowledge distillation (KD) [16] that is originally utilized for model compression, has been proposed as an alternative [1,17]. While this strategy is generally more stable and yields better results than gradient constraint, it demands huge additional data that should be similar to previous classes. This additional data requirement highly affects the performance of the model. Thus, it defeats the original purpose of continual learning to add scarce and unique classes incrementally.
To overcome the shortcoming of prior works, a method that mimics the biological system’s ability to replay past experiences is proposed. A similar strategy has been implemented successfully in image classification [15] and reinforcement learning [18]. Nevertheless, object detection is a completely different domain. Several modifications and adjustments have to be made for the proposed method to work properly, mainly on the issues concerning memory integration during l [ 1 ] training phase. Our research has demonstrated that replay can be used as an effective strategy in continual learning. In this paper, the proposed method is demonstrated in You-Only-Look-Once (YOLO) since it is one of the most popular one-stage detection framework [19]. However, the proposed method is flexible and can be implemented in other object detection frameworks without modifying the network architecture.

2. Related Work

Even though there has been plenty of work addressing continual learning, they typically only focused on the classification problem. This research focuses on the more practical setting in the context of object detection. There has been less research on continual learning in object detection scenarios. Amidst those few is the continual learning scheme applied to faster region convolution neural networks (Faster R-CNN) [1,20] object detection framework. They avoid catastrophic forgetting of previously learned knowledge by distilling knowledge from the previous model. The external distilled proposals of the prior network are saved and utilized as pseudo-data in continual learning. They freeze previous network layers while training the continual one. The whole network layers are utilized at the inference stage. However, Faster R-CNN is a two-stage object detector that consists of an external network for extracting proposals that result in higher computational complexity. Another recent approach to overcome the catastrophic forgetting is proposed in [21], that improved the learning procedure of EWC for object detection. To remember the previous knowledge, they proposed a pseudo-annotation of previously learned classes. A Laplace approximation [22] is proposed for the likelihood of each task to be diagonal.
Other research that focuses on continual learning on object detection is deep model consolidation (DMC) [17]. In the DMC, a double distillation loss has been proposed to combine two models that specialize in different classes into one compact model that can detect all the preceding models’ classes. First, they trained two networks on different data. Then, they consolidated both models into one single model using double distillation loss through training on unlabeled auxiliary data. Although DMC is fast at the inference stage, the training time is extremely long compared to baseline object detectors and demands higher computational power since the consolidation phase requires three models to run simultaneously. Moreover, additional data needed for the consolidation phase are massive since they are not labeled. Furthermore, auxiliary data needed for the consolidation phase highly affect the performance of the consolidated model. Thus, results will vary according to the number of images and domain similarity between the original and auxiliary data. Therefore, in a case where many unique and rare classes are involved, DMC may not perform very well.
Other prior works regarding continual learning methods for object detection are inspired by KD [16], where the previous knowledge is saved in the frozen copy of the previous model. Moreover, the object detection frameworks considered for knowledge distillation are proposal generation-based methods and utilizing stored proposals from previous tasks [1]. Some models are based on auxiliary data and multi-models training as proposed in DMC [17] that consumes more computational power and resources. In contrast, the proposed approach is more general and can be used in any conventional object detection models. The proposed approach does not require any auxiliary data while learning the current data. A comprehensive comparison between the advantages and disadvantages of the proposed method and previous methods are described in Section 4.

3. Proposed Methodology

Suppose that a model l [ 0 ] is trained normally on n number of classes from the first task dataset that will be referred to as l [ 0 ] dataset, then the model l [ 0 ] is needed to detect additional t number of classes without using the whole data from the l [ 0 ] dataset. To avoid catastrophic forgetting where the model forgets features learned in classes belonging to l [ 0 ] dataset, a method which utilizes memory is proposed. The proposed method, which will be referred to as Experience Replay (ER), works by saving a portion of the l [ 0 ] dataset into the memory. Then, the images in the memory will be concatenated with the second task dataset, denoted as l [ 1 ] dataset in every iteration during l [ 1 ] training phase as shown in Figure 2. Furthermore, the dynamic omission is implemented in order to ensure that the memory and l [ 1 ] dataset can be adequately combined during l [ 1 ] training phase.

3.1. YOLO Architecture

The proposed ER method utilizing YOLO is suitable for real-time applications. Nevertheless, the proposed ER method is a flexible strategy and easy to implement in other frameworks without modifying the architecture. The architecture of YOLO that is utilized can be seen in Figure 3. To predict objects’ location, YOLO utilizes grid cells with a size of N × N with three different scales located at layer 89, 101, and 113. Each grid cell has a fixed number of predictions depending on the number of anchor boxes specified. To update YOLO’s weight, the loss between the target and model prediction needs to be calculated. First, in each grid cell, the bounding box regression loss ( L b b ) is calculated using generalized intersection over union (GIOU) [23]. Suppose that a target bounding box, which is denoted as T x , and a predicted bounding box denoted as T y , both of which contain the coordinate and size of bounding boxes. Meanwhile, the algorithm can calculate the value of C by measuring the minimum rectangular area that surrounds both bounding boxes. Thus, the algorithm can get the GIOU by using Equation (1). Then, the resulting value is used as the input for Equations (3) and (5).
Aside from coordinates and sizes, each bounding box also contains an objectness score, which is denoted as X o and Y o for predicted and target bounding boxes, respectively. The objectness score of a predicted bounding box can measure between 0 and 1, whereas for target bounding boxes, the score is only 0 or 1 depending on the target object’s presence in the corresponding grid cell. Meanwhile, the predicted and target bounding boxes’ classification scores are denoted as X c and Y c , respectively. However, unlike the objectness score with a single value for each bounding box, the confidence score is represented as a one-hot vector with a length equal to the number of the model’s classes. The confident score of each bounding box is the multiplication of objectness and classification scores, as shown in Equations (4) and (5), respectively. Both objectness and classification losses are calculated using binary cross-entropy (BCE) as written in Equation (2), whereas C x and C y represent the input and output of the function, respectively. It is important to note that the input of BCE should be normalized using the sigmoid function ( σ ) to avoid an exploding gradient. In Equation (2), w denotes the positive weight which should be used when there is a drastic imbalance between precision and recall. If there is no problem with the imbalance, the algorithm uses the default value of w, which is 1. Lastly, the total loss function, as shown in Equation (7) is calculated by adding α × L c o n f and β × L b b , with α and β denote the weight for each of their respective loss.
G I O U ( T x , T y ) = | T x T y | | T x T y | | C \ ( T x T y ) | | C | ,
B C E ( C x , C y ) = w C y × log σ ( C x ) + ( 1 C y ) × log ( 1 σ ( C x ) ) ,
L b b ( X t , Y t ) = 1 N 2 i , j = 1 N 1 G I O U ( X t i j , Y t i j ) ,
L c l s ( X c , Y c ) = 1 N 2 i , j = 1 N B C E ( X c i j , Y c i j ) ,
L o b j ( X o , Y o ) = 1 N 2 i , j = 1 N B C E ( X o i j , G I O U ( X o i j , Y o i j ) ) ,
L c o n f = L o b j × L c l s ,
L o s s = ( α × L c o n f ) + ( β × L b b ) .

3.2. Task Distribution and Data Augmentation

The Pascal VOC 2007 [24] dataset is used for training the proposed ER method. For experiment purposes, the dataset is divided into the l [ 0 ] and l [ 1 ] tasks. The l [ 0 ] dataset is used to train YOLO normally, whereas the l [ 1 ] dataset is used for continual learning. The purpose of training on the l [ 1 ] dataset is to simulate a real-world condition where the model is needed to learn additional classes. The l [ 1 ] dataset has entirely different classes from the l [ 0 ] dataset. After the model l [ 0 ] is trained normally using the l [ 0 ] dataset, it will be used as the pre-trained weight for the continual learning on the l [ 1 ] . It is also important to increase the variance of training data. By utilizing data augmentation correctly, the accuracy of the model can be further improved as elaborated in Section 4.2.
The most recent YOLO utilizes Cut-Mix [25] for data augmentation. In ER implementation for the task l [ 1 ] , Cut-Mix should be applied after images are loaded and concatenated with the new dataset. Saving augmented images into memory instead of non-augmented ones can slightly decrease the accuracy as shown in our experiment in Section 4. Another important thing to note is the effect of the imbalance number of data between the l [ 0 ] and l [ 1 ] dataset, which is shown in Table 1. In the 19 + 1 scheme, the purpose of switching between TV and person classes as the l [ 1 ] task is to observe this particular issue. The imbalance data between l [ 0 ] and l [ 1 ] tasks may provide a better understanding of how to implement ER correctly.

3.3. Memory Replay on Continual Learning

Given a training with b-size batch, each iteration will contain c-number of training images from the l [ 1 ] dataset and ( b c ) images from the l [ 0 ] dataset. Then, images from the l [ 0 ] dataset and l [ 1 ] dataset are concatenated. By mixing images from the l [ 0 ] dataset, the model would be able to avoid catastrophic forgetting. However, this strategy causes the model to train longer since each iteration only contains c-number of images from the l [ 1 ] task dataset. Therefore, one training epoch will have the same number of iterations as the number of images in the new dataset divided by c, regardless of the batch size. Details of the proposed continual learning algorithm is described in Algorithm 1, specifically in line 12–23. When training the model l [ 1 ] in the continual learning scheme, the loss function is the same as the one used in normal training. The L c o n f is the multiplication result of L c l s and L o b j which are calculated using BCE in Equation (6), whereas for L b b the loss function used is the GIOU loss as shown in Equation (3). During the l [ 1 ] training phase, the L o b j is affected by dynamic omission, which is described as in Algorithm 1 from line 21 to 27. Suppose a bounding box is predicted as a class that belongs to l [ 0 ] dataset, then the L o b j of the corresponding bounding box is invalidated as it may happen because of an unlabeled target object. This rule is to prevent reckless training, which is illustrated in Figure 4. Reckless training occurs since the pre-trained model that is used in l [ 1 ] training phase is already capable of predicting objects which belong to l [ 0 ] dataset. Thus, when training l [ 1 ] model, it may detect an object which belongs to l [ 0 ] dataset in the l [ 1 ] . Conversely, the opposite occurrence, where the model detects objects that belong l [ 1 ] dataset in the memory, can also occur. However, since labels of l [ 0 ] objects are not present in l [ 1 ] dataset and contrariwise, calculating losses from these predictions is inadvisable. It is important to note that this strategy is also highly affected by the number of images from l [ 0 ] dataset denoted as m that is stored in the memory. It is not compulsory to store all images from l [ 0 ] dataset into the memory. However, having more images in the memory can lead to better accuracy for l [ 0 ] classes.
Algorithm 1 The proposed continual learning strategy
Sensors 20 06777 i001

4. Results

In this section, the proposed ER method is extensively evaluated on Pascal VOC 2007 dataset. Results are provided for three different modifications to the proposed ER method. ER for 2500 memory size, ER for 2500 memory size with data augmentation, and ER for 1000 memory size with data augmentation. The proposed ER method is implemented on NVIDIA 1080 GPU. The proposed ER method is trained using stochastic gradient descent (SGD) [26] for 100 iterations. All experiments are written in Python using Pytorch [27] as the machine learning framework. Parameters for training the model is described in Table 2.
The distribution of the trainval and the test set for training and testing, respectively, is unchanged from the Pascal VOC 2007 official release. Following experiments presented in [1,17], the proposed continual learning method is evaluated on two tasks such that ten classes for the l [ 0 ] task and ten classes for l [ 1 ] or the continual task. The other experiment is performed on 19 classes for l [ 0 ] task and 1 class for l [ 1 ] task. The distribution of objects and images for 10 + 10 classes and 19 + 1 classes are presented in Table 1 in Section 3. A few images of the Pascal VOC 2007 are visualized in Figure 5. The graphs of loss, classification, recall, precision, and mean average precision ( m A P ) for both training such that the previous task and the current task are shown in Figure 6.

4.1. Performance Evaluation Parameters

The m A P is calculated independently for the l [ 0 ] and l [ 1 ] tasks. To calculate the m A P , the precision and recall score have to be obtained first from the discrepancy between ground-truth and prediction. Precision, also known as sensitivity, is the ratio of true positives to the sum of true positives and true negatives. Recall, also known as specificity, is the ratio of true positives to the total number of true positives and false negatives. Formulas for the precision and recall is given in Equations (8) and (9) respectively. The A P and m A P are calculated in Equations (10) and (11). A Jaccard index ( J ) is used to calculate the overlap area between predicted and ground-truth bounding box A and B, as shown in Equation (12).
P r e c i s i o n = T P T P + T N ,
R e c a l l = T P T P + F N ,
A P = 0 1 p ( r ) d r ,
m A P = 1 / N i = 1 N A P i ,
J ( A , B ) = | A B | | A B | ,
where A P , T P , and F P denotes average precision, true positive, and false positive, respectively, whereas p ( r ) represents the probability of the event under the precision and recall curve. The m A P is the mean of A P for N number of classes. The proposed ER method is compared with prior state-of-the-art continual learning methods [1,15,17] for object detection. These methods have been implemented on different object detection framework by respective authors. However, for a fair comparison, these methods are adopted to the YOLO framework. Results of these adopted methods are then compared with the proposed ER method.

4.2. Addition of Classes Incrementally

In the first experiment, consider the first 19 classes that appear in alphabetical order in Pascal VOC 2007 dataset as the l [ 0 ] task, and the remaining one as the l [ 1 ] task. The model l [ 0 ] is trained normally on the first (1–19) classes on trainval subset, and the model l [ 1 ] is trained using continual strategies on the remaining one class, which is the TV class. A summary of the comparison between the proposed ER method and state-of-the-art methods are shown in Table 3. The baseline approach for continual learning is to store some of the images from l [ 0 ] task in memory and replay that memory repeatedly while training on l [ 1 ] task. By replaying those memories, ER is able to preserve features of classes in the l [ 0 ] dataset. Furthermore, it maintains the same accuracy on l [ 1 ] classes. In the l [ 0 ] training phase, 19 classes are trained first. Then during the l [ 1 ] training phase, only one class is incrementally trained. This scheme is performed to observe the effect of class distribution during continual learning. In comparison to previous methods [1,15,17], the proposed ER method has significantly higher m A P for all classes. The proposed ER method have increased the m A P to 8.9%, 8.8%, and 30.5% in comparison to GEM [15], DMC [17], and KD [1] respectively. Specifically, the proposed ER method with data augmentation has achieved 68.9% m A P for all 20 classes.
The second experiment is performed on the 10 + 10 class scheme. The first ten classes (1–10) are trained as the l [ 0 ] task, and the remaining ten (11–20) are trained as l [ 1 ] task for continual learning. As presented in Table 4, GEM is the only method that achieves a higher result than the proposed ER method, yet only for the chair class. The reason behind this is the imbalanced data in the Pascal VOC 2007 dataset. The proposed ER method with 1000 memory size has slightly higher results than the 5000 memory size for l [ 0 ] classes (11–20). However, the proposed ER method for 5000 memory size with augmentation has improved the m A P for all classes up to 65.5%. In the 10 + 10 class scenario the proposed ER method has increased the m A P up to 54.3%, 30%, and 14.6% as compared to GEM [15], KD [1], and DMC [17] respectively. To better investigate the proposed ER method’s continual learning behaviour, an experiment is performed on the 19 + 1 scheme with the person class as the incremental class instead of TV. Table 5 presents experiments performed on normal training (1–19), normal training (20), GEM, and the proposed ER method. As shown in Table 5, the m A P for the normal training is 68.8% while the proposed ER method achieved 67.1%. In comparison to GEM, the proposed ER method has 15.2% increased the m A P for all classes. Results presented in Table 5 indicates that the proposed ER method performed better in preserving features of l [ 0 ] task.

4.3. Visualization and Effect of Different Memory Size

In order to provide an insight of the proposed ER method’s performance compared with prior methods, several prediction results are visualized from the Pascal VOC 2007 dataset. Those images are obtained from the test set that contain both the l [ 0 ] and l [ 1 ] classes. The performance of each method is visualized in Figure 7. Images from Figure 7a–d presents the object detection performance on KD [1], DMC [17], GEM [15], and the proposed ER method respectively. In Figure 7a KD [1] and Figure 7c GEM [15] are shown to have many false negatives. Generally, KD [1] has comparatively better performance on the l [ 1 ] task in comparison to l [ 0 ] task. Note that, the person class belongs to the l [ 1 ] task whereas the car class belong to the l [ 0 ] task. The same also happens with GEM [15], since it fails to recognise the bicycle class which belongs to l [ 0 ] task. Meanwhile, the performance of DMC [17] as shown in Figure 7b is considerably better than KD [1] and GEM [15]. However, DMC [17] has more false positive than other methods which are visualized as black bounding boxes such as train and dog in Figure 7a first row. This false positives occur because the mean of DMC’s confidence score is higher than other methods. In comparison to all these prior methods, the proposed ER method which utilizes 2500 memory as shown in Figure 7d has correctly localized and classified all objects. However, as shown in Figure 7d first row, far objects are not detected by the proposed ER method. The false negative of far objects is the limitation of this work.
In Figure 8, four identical images as the one presented in Figure 7 have been considered for prediction using ER in various memory size to provide better comprehension of the effect of different memory size. In order to evaluate the effect of memory size on results, different experiments are performed on 500, 1000, 2500, and 5000 memory size in the 10 + 10 class scenario. As shown in Table 6, the difference between the m A P for the memory size 5000 and 2500 are 0.1%. However, there is 4.3% accuracy drop for the 1000 memory size. The proposed ER method has comparable results in all the memory size. As the memory size is increased, the capability of the model to predict many objects also increases. However, the results of memory size 2500 and 5000 are relatively similar with only 0.1% difference in the average m A P . The fewer memory size requires fewer training time and hardware’s memory requirement. However, it will reduce the performance on l [ 0 ] classes.

4.4. Performance Evaluation on ITRI-DrvieNet60 Dataset

The proposed ER model is trained on the ITRI-DriveNet-60 private dataset. ITRI-DriveNet-60 private dataset is taken on the highways of Taiwan. It has similar characteristics with famous autonomous driving object detection dataset such as KITTI [9] and Cityscape [10]. The number of images and objects for train and test set are shown in Table 7. Four classes (four-wheel vehicle, rider, two-wheel vehicle, and person) are introduced in this dataset. The task distribution strategy for the ITRI-DriveNet-60 dataset’s continual learning is presented in Table 8. We trained the model on the 1 + 3 classes scheme. Therefore, the four-wheel vehicle class is considered in l [ 0 ] training, and the other classes are not considered in l [ 0 ] training those are represented by dashes in Table 8. and the other three classes are considered in l [ 1 ] training. While training the l [ 1 ] classes, the l [ 0 ] class labels are ignored and not utilized for training. The AP for the normal training, l [ 0 ] training, and l [ 1 ] training are illustrated in Table 8. Notably, the single class accuracy is higher because it is easier for the model to classify a single class. However, in the l [ 1 ] training, the four-wheel vehicle class accuracy is dropped to 4.5% still the four-wheel vehicle accuracy is higher among other classes. The total accuracy of four-wheel vehicle class in the l [ 1 ] training is 85.2%. The m A P of the proposed method is almost similar to the mAP of the normal training. Specifically, the proposed method obtains 77.1% m A P for all four classes. The m A P drops only 0.1% in comparison to normal training on the YOLOv3 object detection framework. This indicates the effectiveness of the proposed ER method for continual learning. Conclusively, the proposed ER method achieved comparably the same accuracy as the normal training.
The proposed ER method has excellent detection results when using the higher memory size (2500 and 5000 frames) as shown in Figure 9c,d respectively. On the other hand, the fewer memory size (500 and 1000 frames) has a tendency to detect one of the tasks better than the other. As illustrated in Figure 9a the model with memory size 500 frames can detect classes which belong to l [ 1 ] dataset better than classes which belong to l [ 0 ] dataset, whereas the opposite is true for the model with memory size 1000 frames. This condition occurs because employing a memory size of 1000 frames will preserve more features from the l [ 0 ] dataset compared with only utilizing a memory size of 500 frames. However, using higher memory will results in better performance from both l [ 0 ] and l [ 1 ] dataset since more images in memory can be shuffled, resulting in more variance while training. It is notable from Figure 9d that the memory size 5000 frames have better results among others. However, similar to prediction results in Figure 8, the detection result of the memory size 2500 and 5000 frames are comparable. This result indicates that even utilizing ER with the fewer memory size can achieve acceptable performance regardless of the dataset, as proven in Table 6 and Table 8 which are performed in two different datasets.
Continual learning strategies tend to have higher training time compared with normal training. In normal training, after the model is run in inference for obtaining the prediction, backpropagation is performed to update the weights based on the loss between prediction and ground-truth. However, the prior works typically run the l [ 0 ] or previous model alongside the l [ 1 ] model and use the l [ 0 ] model prediction to learn previous features (KD [1] and DMC [17]) or compare the gradients of both models (GEM [15]). These approaches are rather cumbersome and take a longer time compared with normal training. Therefore, in the proposed ER method, a straight-forward approach has been considered, which is able to reduce the training time. In the proposed ER method, the noticeable addition in each iteration’s training time is caused by the integration and augmentation of the memory in the l [ 1 ] training phase, which has an average of 200 ms for each iteration as described in Table 9. Meanwhile, the dynamic omission added to the algorithm is very simple. Thus, its time consumption is negligible. Table 9 also presents the training time of each process in the prior works. Based on these processes, the training time for a single iteration is calculated as the sum of all these processes. The average training time for the proposed ER methods is shown to be the lowest among other continual strategies. Table 10 provides the advantages and disadvantages of the proposed ER method in comparison to the previous methods. The proposed ER method achieves better detection results in comparison to other continual learning methods while having the lowest time complexity.

5. Conclusions

In this research, a novel method to address the problem of catastrophic forgetting is proposed. Using YOLO architecture as the benchmark framework, ER can preserve the features from l [ 0 ] task while training on the l [ 1 ] task. A guide to implementing the data augmentation technique is added to the proposed ER method to improve the learning process in the current task with varying memory sizes. Experimental results on Pascal VOC show that the proposed ER method provides acceptable results for continual learning. Specifically, the proposed ER method has achieved m A P of 65.5% and 68.9% in 10 + 10 and 19 + 1 classes scenario, respectively, higher than the state-of-the-art method. Nevertheless, continual learning for object detection still desires more improvement. A further evaluation that integrates with thoroughly extensive experiments on improving the continual learning process is expected to be performed for future study. Furthermore, the proposed ER method has alleviated the problem of catastrophic forgetting. However, enough improvements are needed to reduce the use of memory size to the minimum level while maintaining classes accuracy.

Author Contributions

Conceptualization, J.-L.S.; methodology, J.-L.S., M.A.H., Q.M.u.H. and S.K.; software, J.-L.S., M.A.H. and Q.M.u.H.; validation, M.A.H. and Q.M.u.H.; formal analysis, Q.M.u.H.; investigation, J.-L.S.; resources, D.-Q.G.; data curation, J.-L.S.; writing—original draft preparation, M.A.H., Q.M.u.H. and S.K.; writing—review and editing, P.C. and S.-J.R.; visualization, M.A.H.; supervision, P.C., D.-Q.G. and S.-J.R.; project administration, J.-L.S.; funding acquisition, P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Ministry of Economic Affairs, Republic of China under project “Real-time Situation Awareness Subsystem Technologies for Automatic Driving (3/4)” (Grant No. K352BL1210) and Ministry of Science and Technology, Republic of China under project “A deep learning-based overexposure suppression codec for OLED power saving” (Grant No. MOST 109-2221-E-011-122-MY3).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADVAutonomous Driving Vehicles
APAverage Precision
BCEBinary Cross Entropy
DMCDeep Model Consolidation
ERExperience Replay
Faster R-CNNFaster Region Convolution Neural Networks
GEMGradient Episodic Memory
GIOUGeneralized Intersection Over Union
ITRIIndustrial Technology Research Institute
KDKnowledge Distillation
YOLOYou Only Look Once

References

  1. Shmelkov, K.; Schmid, C.; Alahari, K. Incremental Learning of Object Detectors without Catastrophic Forgetting. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 3420–3429. [Google Scholar] [CrossRef] [Green Version]
  2. Ren, B.; Wang, H.; Li, J.; Gao, H. Life-long learning based on dynamic combination model. Appl. Soft Comput. 2017, 56, 398–404. [Google Scholar] [CrossRef]
  3. Parisi, G.I.; Kemker, R.; Part, J.L.; Kanan, C.; Wermter, S. Continual lifelong learning with neural networks: A review. Neural Netw. 2019, 113, 54–71. [Google Scholar] [CrossRef] [PubMed]
  4. Lin, H.Y.; Dai, J.M.; Wu, L.T.; Chen, L.Q. A Vision-Based Driver Assistance System with Forward Collision and Overtaking Detection. Sensors 2020, 20, 5139. [Google Scholar] [CrossRef] [PubMed]
  5. Fayyad, J.; Jaradat, M.A.; Gruyer, D.; Najjaran, H. Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review. Sensors 2020, 20, 4220. [Google Scholar] [CrossRef] [PubMed]
  6. Dominguez-Sanchez, A.; Cazorla, M.; Orts-Escolano, S. A new dataset and performance evaluation of a region-based cnn for urban object detection. Electronics 2018, 7, 301. [Google Scholar] [CrossRef] [Green Version]
  7. Feng, D.; Haase-Schütz, C.; Rosenbaum, L.; Hertlein, H.; Gläser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges. IEEE Trans. Intell. Transp. Syst. 2020, 1–20. [Google Scholar] [CrossRef] [Green Version]
  8. Zhao, Z.; Zheng, P.; Xu, S.; Wu, X. Object Detection With Deep Learning: A Review. IEEE Trans. Neural Networks Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar] [CrossRef]
  10. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  11. Yoon, J.; Yang, E.; Lee, J.; Hwang, S.J. Lifelong learning with dynamically expandable networks. arXiv 2017, arXiv:1708.01547. [Google Scholar]
  12. Chen, C.P.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Networks Learn. Syst. 2017, 29, 10–24. [Google Scholar] [CrossRef] [PubMed]
  13. Chen, C.P.; Liu, Z.; Feng, S. Universal approximation capability of broad learning system and its structural variations. IEEE Trans. Neural Networks Learn. Syst. 2018, 30, 1191–1204. [Google Scholar] [CrossRef] [PubMed]
  14. Xing, Y.; Shen, F.; Zhao, J. Perception evolution network based on cognition deepening model—Adapting to the emergence of new sensory receptor. IEEE Trans. Neural Networks Learn. Syst. 2015, 27, 607–620. [Google Scholar] [CrossRef] [PubMed]
  15. Lopez-Paz, D.; Ranzato, M.A. Gradient Episodic Memory for Continual Learning. In Advances in Neural Information Processing Systems 30; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Long Beach, CA, USA, 2017; pp. 6467–6476. [Google Scholar]
  16. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  17. Zhang, J.; Zhang, J.; Ghosh, S.; Li, D.; Tasci, S.; Heck, L.; Zhang, H.; Jay Kuo, C.C. Class-incremental Learning via Deep Model Consolidation. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020; pp. 1120–1129. [Google Scholar] [CrossRef]
  18. Rolnick, D.; Ahuja, A.; Schwarz, J.; Lillicrap, T.; Wayne, G. Experience Replay for Continual Learning. In Advances in Neural Information Processing Systems 32; Wallach, H., Larochelle, H., Beygelzimer, A., Alché-Buc, F.d., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; pp. 350–360. [Google Scholar]
  19. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  20. Ren, S.; He, K.; Girshick, R.B.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015. [Google Scholar] [CrossRef] [Green Version]
  21. Liu, L.; Kuang, Z.; Chen, Y.; Xue, J.H.; Yang, W.; Zhang, W. Incdet: In defense of elastic weight consolidation for incremental object detection. IEEE Trans. Neural Networks Learn. Syst. 2020. [Google Scholar] [CrossRef] [PubMed]
  22. MacKay, D.J. A practical Bayesian framework for backpropagation networks. Neural Comput. 1992, 4, 448–472. [Google Scholar] [CrossRef]
  23. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16 June–20 June 2019. [Google Scholar]
  24. Everingham, M.; Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vision 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  25. Yun, S.; Han, D.; Chun, S.; Oh, S.J.; Yoo, Y.; Choe, J. CutMix: Regularization Strategy to Train Strong Classifiers With Localizable Features. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 6022–6031. [Google Scholar] [CrossRef] [Green Version]
  26. Zinkevich, M.; Weimer, M.; Li, L.; Smola, A.J. Parallelized stochastic gradient descent. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Vancouver, BC, Canada, 2010; pp. 2595–2603. [Google Scholar]
  27. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Wallach, H., Larochelle, H., Beygelzimer, A., dAlché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; pp. 8024–8035. [Google Scholar]
Figure 1. Continual learning purpose is to increase the number of classes a model can detect while prevents overwriting of previous knowledge.
Figure 1. Continual learning purpose is to increase the number of classes a model can detect while prevents overwriting of previous knowledge.
Sensors 20 06777 g001
Figure 2. Block diagram of the proposed continual learning strategy method.
Figure 2. Block diagram of the proposed continual learning strategy method.
Sensors 20 06777 g002
Figure 3. Architecture of YOLOv3 with DarkNet53 as target network for the proposed continual learning strategy.
Figure 3. Architecture of YOLOv3 with DarkNet53 as target network for the proposed continual learning strategy.
Sensors 20 06777 g003
Figure 4. Dynamic omission prevents models from training recklessly from predictions that have no corresponding labels.
Figure 4. Dynamic omission prevents models from training recklessly from predictions that have no corresponding labels.
Sensors 20 06777 g004
Figure 5. Sample images from the pascal VOC 2007 dataset.
Figure 5. Sample images from the pascal VOC 2007 dataset.
Sensors 20 06777 g005
Figure 6. Training details for previous and current task. The x-axis represents epochs in which epoch 0–99 is the training on l [ 0 ] task, whereas epoch 100–199 is the training on l [ 1 ] task.
Figure 6. Training details for previous and current task. The x-axis represents epochs in which epoch 0–99 is the training on l [ 0 ] task, whereas epoch 100–199 is the training on l [ 1 ] task.
Sensors 20 06777 g006
Figure 7. Object detection performance of the proposed ER method with KD [1], DMC [17], GEM [15], and the proposed ER method on test dataset as shown above from (ad) respectively.
Figure 7. Object detection performance of the proposed ER method with KD [1], DMC [17], GEM [15], and the proposed ER method on test dataset as shown above from (ad) respectively.
Sensors 20 06777 g007
Figure 8. Prediction of the proposed ER method with different memory size, (a) memory size of 500 frames, (b) memory size of 1000 frames, (c) memory size of 2500 frames, and (d) memory size of 5000 frames.
Figure 8. Prediction of the proposed ER method with different memory size, (a) memory size of 500 frames, (b) memory size of 1000 frames, (c) memory size of 2500 frames, and (d) memory size of 5000 frames.
Sensors 20 06777 g008
Figure 9. The object detection result on ITRI-DriveNet-60 private dataset on various memory size, (a) memory size of 500 frames, (b) memory size of 1000 frames, (c) memory size of 2500 frames, and (d) memory size of 5000 frames.
Figure 9. The object detection result on ITRI-DriveNet-60 private dataset on various memory size, (a) memory size of 500 frames, (b) memory size of 1000 frames, (c) memory size of 2500 frames, and (d) memory size of 5000 frames.
Sensors 20 06777 g009
Table 1. The task distribution between three different training scheme.
Table 1. The task distribution between three different training scheme.
Task DistributionTask l [ 0 ] Task l [ 1 ]
10 + 102868 images (4852 objects)3252 images (7756 objects)
19 + TV4931 images (7918 objects)256 images (324 objects)
19 + Person4112 images (3552 objects)2008 images (4690 objects)
Table 2. Parameters for training the model.
Table 2. Parameters for training the model.
ParametersValue
Momentum0.9
Batch size32
Epoch100
Weight decay0.0005
Initial learning rate0.002
Final learning rate0.0005
Image size416 × 416
Table 3. Comparison of the proposed ER method with state-of-the-art methods for 19 + 1 classes scenario. Best results are written in bold number.
Table 3. Comparison of the proposed ER method with state-of-the-art methods for 19 + 1 classes scenario. Best results are written in bold number.
MethodsAeroBikeBirdBoatBottleBusCarCatChairCowDin. TableDogHorseM BikePersonPlantSheepSofaTrainTVmAP at 0.5
All data76.584.468.959.355.682.185.979.955.473.767.879.985.284.483.146.17170.382.171.773.2
KD [1]50.253.029.827.513.355.467.245.513.336.934.238.760.756.338.510.533.833.055.615.838.4
DMC [17]77.267.863.546.550.27079.380.343.565.50.271.174.671.578.239.666.634.879.242.460.1
GEM [15]66.672.654.948.146.772.580.568.737.966.348.462.58172.567.333.56845.876.928.460
ER*73.28362.956.353.781.184.977.748.469.163.678.585.782.680.943.367.760.879.639.768.6
ER**76.783.162.556.853.283.88578.150.168.662.177.485.679.780.742.870.666.281.133.368.9
ER***74.379.559.555.349.578.983.577.145.26657.673.985.580.378.340.16858.178.744.766.7
ER* method without augmentation before uploading to memory (2500), ER** method with augmentation after uploading to memory (2500), ER*** method with augmentation after uploading to memory (1000).
Table 4. Comparison of the proposed ER method with state-of-the-art methods for 10 + 10 classes scenario. Best results are written in bold number.
Table 4. Comparison of the proposed ER method with state-of-the-art methods for 10 + 10 classes scenario. Best results are written in bold number.
MethodsAeroBikeBirdBoatBottleBusCarCatChairCowDin. TableDogHorseM BikePersonPlantSheepSofaTrainTVmAP at 0.5
All data76.584.468.959.355.682.185.979.955.473.767.879.985.284.483.146.17170.382.171.773.2
KD [1]10.810.236.19.66.35.414.341.73.819.442.553.576.566.857.927.860.450.472.24635.5
DMC [17]56.646.548.725.745.562.971.565.736.852.40.660.171.258.27340.758.912.66466.650.9
GEM [15]3.13.43.66.119.76.158.69.555.714.20.77.64.22.923.50.21.22.63.40.211.2
ER*67.274.147.344.444.77180.164.536.957.142.254.978.566.763.526.351.44568.952.156.9
ER**75.58258.951.549.578.784.874.147.265.16165.478.776.674.634.660.658.477.861.965.5
ER***66.768.53845.740.769.881.460.330.952.367.460.481.176.679.643.961.161.377.464.761.4
ER* method without augmentation before uploading to memory (2500), ER** method with augmentation after uploading to memory (2500), ER*** method with augmentation after uploading to memory (1000).
Table 5. Comparison of the proposed ER method with the state-of-the-art methods for 19 + 1 classes scenario when the incremental class person. Best results are written in bold number.
Table 5. Comparison of the proposed ER method with the state-of-the-art methods for 19 + 1 classes scenario when the incremental class person. Best results are written in bold number.
MethodsAeroBikeBirdBoatBottleBusCarCatChairCowDin. tableDogHorseM bikeTVPlantSheepSofaTrainPersonmAP at 0.5
Classes (1–19)75.883.966.758.053.181.885.478.353.77566.578.584.68382.743.571.469.481.2-68.8
Class (20)-------------------64.164.1
GEM [15]61.765.942.637.234.765.268.459.630.950.345.154.965.665.252.723.852.550.162.150.351.9
ER76.7815853.250.78184.4754767.662.471.882.379.767.738.867.464.980.851.267.1
Table 6. Comparison of 10 + 10 classes with different memory sizes.
Table 6. Comparison of 10 + 10 classes with different memory sizes.
MethodsmAP(1–10)mAP(10–20)mAP
ER** (500)42.564.453.5
ER** (1000)55.467.461.4
ER** (2500)66.365.265.7
ER** (5000)66.765.065.8
Table 7. The number of images and objects for training and testing the proposed method on the ITRI-DriveNet-60 dataset.
Table 7. The number of images and objects for training and testing the proposed method on the ITRI-DriveNet-60 dataset.
DatasetFour-Wheel VehicleRiderTwo-Wheel VehiclePerson
Train3099 images (17,879 objects)652 images (1911 objects)1146 images (5697 objects)726 images (1053 objects).
Test644 images (3116 objects)117 images (311 objects)195 images (881 objects)128 images (202 objects).
Total3643 images (20,995 objects)769 images (2222 objects)1341 images (6578 objects)854 images (1255 objects).
Table 8. Test results of ER trained on ITRI DriveNet-60 dataset in 1 + 3 class continual scheme.
Table 8. Test results of ER trained on ITRI DriveNet-60 dataset in 1 + 3 class continual scheme.
MethodsFour-wheel vehicleRiderTwo-wheel vehiclePersonmAP
Normal Training89.77280.266.977.2
l [ 0 ] Training91.3---91.3
l [ 1 ] Training85.277.380.265.677.1
Table 9. Training time for single iteration of the proposed ER method and prior works.
Table 9. Training time for single iteration of the proposed ER method and prior works.
MethodsProccessTime (ms)
Normal Training 1 × inference + 1 × backpropagation2380
KD [1] 2 × inference + 1 × backpropagation2660
DMC [17] 3 × inference + 1 × backpropagation2940
GEM [15] 2 × (inference + backpropagation)4760
ER 1 × inference + 1 × backpropagation + 1 × augmentation2580
Inference time = 280 ms, backpropagation time = 2100 ms, augmentation time = 200 ms.
Table 10. Advantages and disadvantages of the proposed ER method with state-of-the-art methods.
Table 10. Advantages and disadvantages of the proposed ER method with state-of-the-art methods.
MethodsAdvantagesDisadvantages
KD [1]No memory required and faster training timeAuxiliary data required.
DMC [17]No memory requiredThree models required to train and auxiliary data required
GEM [15]No memory and no auxiliary data requiredHigher training time and lower performance
ERNo auxiliary data and faster training timeMemory required
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shieh, J.-L.; Haq, Q.M.u.; Haq, M.A.; Karam, S.; Chondro, P.; Gao, D.-Q.; Ruan, S.-J. Continual Learning Strategy in One-Stage Object Detection Framework Based on Experience Replay for Autonomous Driving Vehicle. Sensors 2020, 20, 6777. https://doi.org/10.3390/s20236777

AMA Style

Shieh J-L, Haq QMu, Haq MA, Karam S, Chondro P, Gao D-Q, Ruan S-J. Continual Learning Strategy in One-Stage Object Detection Framework Based on Experience Replay for Autonomous Driving Vehicle. Sensors. 2020; 20(23):6777. https://doi.org/10.3390/s20236777

Chicago/Turabian Style

Shieh, Jeng-Lun, Qazi Mazhar ul Haq, Muhamad Amirul Haq, Said Karam, Peter Chondro, De-Qin Gao, and Shanq-Jang Ruan. 2020. "Continual Learning Strategy in One-Stage Object Detection Framework Based on Experience Replay for Autonomous Driving Vehicle" Sensors 20, no. 23: 6777. https://doi.org/10.3390/s20236777

APA Style

Shieh, J. -L., Haq, Q. M. u., Haq, M. A., Karam, S., Chondro, P., Gao, D. -Q., & Ruan, S. -J. (2020). Continual Learning Strategy in One-Stage Object Detection Framework Based on Experience Replay for Autonomous Driving Vehicle. Sensors, 20(23), 6777. https://doi.org/10.3390/s20236777

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop