Next Article in Journal
Dynamic Detection of Forest Change in Hunan Province Based on Sentinel-2 Images and Deep Learning
Next Article in Special Issue
Meta-Knowledge Guided Weakly Supervised Instance Segmentation for Optical and SAR Image Interpretation
Previous Article in Journal
D-MFPN: A Doppler Feature Matrix Fused with a Multilayer Feature Pyramid Network for SAR Ship Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Leveraging Self-Paced Semi-Supervised Learning with Prior Knowledge for 3D Object Detection on a LiDAR-Camera System

1
School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
2
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430072, China
3
School of Electronic and Control Engineering, Changan University, Xi’an 710064, China
4
Institute of Computer Application, China Academy of Engineer Physics, Mianyang 621900, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(3), 627; https://doi.org/10.3390/rs15030627
Submission received: 8 December 2022 / Revised: 9 January 2023 / Accepted: 18 January 2023 / Published: 20 January 2023
(This article belongs to the Special Issue Weakly Supervised Deep Learning in Exploiting Remote Sensing Big Data)

Abstract

:
Three dimensional (3D) object detection with an optical camera and light detection and ranging (LiDAR) is an essential task in the field of mobile robot and autonomous driving. The current 3D object detection method is based on deep learning and is data-hungry. Recently, semi-supervised 3D object detection (SSOD-3D) has emerged as a technique to alleviate the shortage of labeled samples. However, it is still a challenging problem for SSOD-3D to learn 3D object detection from noisy pseudo labels. In this paper, to dynamically filter the unreliable pseudo labels, we first introduce a self-paced SSOD-3D method SPSL-3D. It exploits self-paced learning to automatically adjust the reliability weight of the pseudo label based on its 3D object detection loss. To evaluate the reliability of the pseudo label in accuracy, we present prior knowledge based SPSL-3D (named as PSPSL-3D) to enhance the SPSL-3D with the semantic and structure information provided by a LiDAR-camera system. Extensive experimental results in the public KITTI dataset demonstrate the efficiency of the proposed SPSL-3D and PSPSL-3D.

1. Introduction

Three dimensional (3D) environment perception has an important role in the field of autonomous driving [1]. It analyzes the real-time information of the surroundings to ensure traffic safety. To avoid vehicle collision [2], 3D object detection is an important approach among the techniques of 3D environment perception. Its task is to identify the classification and predict the 3D bounding box of a targeted object a the traffic scenario. In a word, 3D object detection performance affects the traffic safety of intelligent driving [3]. As 3D object detection requires spatial information from the environment, light detection and ranging (LiDAR) is a suitable sensor because it can generate a 3D point cloud in real-time [4]. Thanks to its ranging accuracy and stability, multi-beam mechanical LiDAR is the mainstream LiDAR sensor for environment perception [1,5]. It is referred to as LiDAR henceforth for discussion simplicity. Due to the limited rotation frequency and beam number, the vertical and horizontal resolution angles are limited, causing sparsity of the LiDAR point cloud, thus increasing the difficulty of 3D object detection [1].
The current 3D object detection method exploits the technique of deep learning and takes LiDAR points as the main input to identify and localize 3D objects [6]. To decrease the negative impact of the sparse LiDAR point cloud, researchers have done lots of work in several areas such as (i) detector architecture [7], (ii) supervised loss function [4], and (iii) data augmentation [8], which have made progress in fully supervised 3D object detection (FSOD-3D). By training with the sufficient labeled data, FSOD-3D can achieve performance in 3D environment perception close to that of humans.
However, there is a contradiction between the demand for 3D object detection performance and the cost of human annotation on the LiDAR point cloud. Due to the sparsity of the LiDAR point cloud and occlusion of the 3D object, the annotation cost of the 3D object is high, so the labeled dataset is insufficient. Therefore, it is essential to utilize unlabeled data to train the 3D object detector.
Semi-supervised 3D object detection (SSOD-3D) [9,10,11] has attracted a lot of attention, for it improves the generalization ability of the 3D object detector with both labeled and lots of unlabeled samples recorded in various traffic scenarios. From the viewpoint of optimization, SSOD-3D is regarded as a problem that alternatively optimizes the weights of 3D objects detector w and pseudo labels from the unlabeled dataset. For one unlabeled sample x (i.e., the LiDAR point cloud), its pseudo label l consists of the 3D bounding boxes of the targeted objects (i.e., car, pedestrian, cyclist), predicted from x . This means that the capacity of w and quality of x are coupled. To obtain the optimal w * , it is essential to decrease the false-positives (FP) and true-negatives (TN) in l . To improve the quality of l , one common approach is to utilize the label filter to remove the incorrect objects in l . Sohn et al. [12] employed a confidence-based filter to remove pseudo labels of which the classification confidence score is below threshold τ cls . Wang et al. [11] extended this filter [12] in their SSOD-3D architecture, with both the τ cls and the 3D intersection-over-union (IoU) threshold. However, in practical application, the optimal thresholds are different with detector architecture, the training dataset and even the object category. It takes a lot of time to search the optimal thresholds in the label filter, which is inefficient in the actual application. Thus, it is a challenging problem to design a more effective and convenient SSOD-3D method.
In the background of intelligent driving, most self-driving cars are equipped with LiDAR and an optical camera. A sensor system with LiDAR and a camera is called a LiDAR-camera system. To remedy the sparsity of the LiDAR point cloud, researchers have studied 3D object detection methods on a LiDAR-camera system [13,14,15,16,17]; the LiDAR-camera system provides a dense texture feature from the RGB image, improving the classification accuracy and confidence of the 3D detection result. Thus, it is wise for the SSOD-3D to consider the prior knowledge provided by LiDAR-camera systems.
Motivated by this, we present a novel SSOD-3D method on a LiDAR-camera system. First, in order to train a 3D object detector with reliable pseudo labels, we introduce a self-paced, semi-supervised and learning-based 3D object detection (SPSL-3D) framework. It exploits the theory of self-paced learning (SPL) [18] to adaptively estimate the reliability weight of pseudo label with its 3D object detection loss. After that, we notice that the prior knowledge in the LiDAR point cloud and RGB image benefits the evaluation of the reliability of pseudo label, and propose a prior knowledge-based SPSL-3D (named PSPSL-3D) framework. Experiments are conducted in the autonomous driving dataset KITTI [19]. With the different labeled training samples, both comparison results and ablation studies demonstrate the efficiency of the SPSL-3D and PSPSL-3D frameworks. Therefore, SPSL-3D and PSPSL-3D benefit SSOD-3D on a LiDAR-camera system. The remainder of this paper is organized as follows. At first, the related works of FSOD-3D and SSOD-3D are illustrated in Section 2. In the next, the proposed SPSL-3D and PSPSL-3D methods are discussed in Section 3. After that, experimental configuration and results are analyzed in Section 4. Finally, this work is concluded in Section 5.

2. Related Works

2.1. Fully Supervised 3D Object Detection

To achieve high performance in environment perception in autonomous driving, FSOD-3D on LiDAR has been widely studied in recent years. Its architecture commonly has three modules [6]: (i) data representation, (ii) backbone network, (iii) detection head. LiDAR data mainly have three representations: point-based [4], pillar-based [20], and voxel-based [21]. Selection of a backbone network is dependent on data representation. Point-based features are extracted with PointNet [22], PointNet++ [23], or a graph neural network (GNN) [24]. As the pillar feature is regarded as the pseudo image, a 2D convolutional neural network (CNN) can be used. To deal with the sparsity of the LiDAR voxel, a 3D sparse convolutional neural network (Spconv) [25] is exploited for feature extraction. The detection head can be classified as anchor-based [26] and anchor-free [27]. The anchor-based 3D detector first generates the 3D bounding boxes with the pre-defined size of the different categories (called anchors) that are placed uniformly in the ground, then predicts the size, position shift, and confidence score of each anchor, and removes the incorrect anchors with lower confidence scores. After that, to remove the redundant 3D bounding boxes, 3D detection results are obtained by using non maximum suppression (NMS) on the remaining shifted anchors. The anchor-free 3D detector first usually predicts the foreground point cloud from the raw points [28], and then predicts the 3D bounding box from each foreground LiDAR point with a fully connected (FC) layer. Then, NMS is exploited to remove the bounding boxes with high overlap.
Recently, many researchers have produced lots of work; Zheng et al. [29] trained a baseline 3D detector with knowledge distillation. The teacher detector generates the pseudo label, and supervises the student detector with shape-aware data augmentation. Schinagl et al. [30] analyzed the importance of each LiDAR point for 3D object detection by means of Monte Carlo sampling. Man et al. [31] noticed that LiDAR can obtain multiple return signals with a single laser pulse and use this mechanism to encode a meaningful feature for the localization and classification of 3D proposals. Yin et al. [27] proposed a light anchor-free 3D detector to regress the heat maps of 3D bounding box parameters. In the training stage, it did not need target object alignment, thus saving lots of time. Based on 3D Spconv [25], Chen et al. [32] presented a focal sparse convolution to dynamically select the receptive field of voxel features for convolution computation. This can be extended for the multi-sensor feature fusion. As for FSOD-3D on a LiDAR-camera system, Wu et al. [33] used multi-model-based depth completion to generate a dense colored point cloud of 3D proposals for accurate proposal refinement. Li et al. [34] presented the multi-sensor-based cross attention module to utilize the LiDAR point as query, and its neighbored projected pixel coordinates and RGB values as keys and values for the fusion feature computation. Piergiovanni et al. [35] studied a general 4D detection framework for both RGB images and LiDAR point clouds in a time series. To deal with the sparsity of the LiDAR point cloud, Yin et al. [36] attempted to generate 3D virtual points of a targeted object with the guidance of the instance segmentation result predicted from the RGB image.

2.2. Semi-Supervised 3D Object Detection

Semi-supervised learning (SSL) is a classical problem in machine learning and deep learning [37]. Compared with the booming development of FSOD-3D and the rapid development of SSOD-2D, relatively fewer works on SSOD-3D have been published in academia. However, it is a challenging and meaningful problem for both industry and academia. First, unlike FSOD-3D, SSOD-3D needs to both consider how to generate reliable pseudo or weak labels from unlabeled point clouds, and how to exploit pseudo labels with uncertain quality for 3D detector training. Second, a classical SSL framework is difficult to directly use in SSOD-3D. For the labeled data x i and unlabeled data x j , traditional SSL theory [37] emphasises their similarity w ( x i , x j ) [ 0 , 1 ] , and constructs a manifold regularization term for SSL optimization. However, in SSOD-3D, point clouds x i and x j are collected from different places and have different data distributions; thus, it is difficult to measure their similarity. Third, as the sparse and unstructured LiDAR point cloud contains fewer texture features than the dense and structured RGB image, it is more difficult for SSOD-3D to extract salient prior knowledge of the targeted object than it is for SSOD-2D.
Some insightful works of SSOD-3D are discussed hereafter. Tang and Lee [9] exploited the weak label (i.e., the 2D bounding box in RGB image) of an unlabeled 3D point cloud to train a 3D detector. The weak label is generated via the 2D object detector. To compute 3D detection loss with the weak label, the predicted 3D bounding box is projected onto the image plane. After that, 3D detection loss is converted into 2D detection loss. This method requires both RGB imaging and point cloud, and it works for both RGB-D camera and LiDAR-camera systems. Xu et al. [38] adaptively filtered the incorrect 3D object in the unlabeled data with a statistical and adaptive confidence threshold, and added the remaining predicted 3D objects into the 3D object database for 3D object detector training in the next iteration.
Mean teacher [39] is a common SSL paradigm in SSOD-3D. It consists of teacher and student detectors. For one unlabeled data, its pseudo label is generated from the teacher detector and used to supervise the student detector. Zhao et al. [10] were the first to utilize the mean teacher framework [39] for SSOD-3D. For the unlabeled data x , they generated its pseudo label from the teacher detector and constructed consistency loss to minimize the difference between the pseudo label and result predicted by the student detector with data augmentation on x . After that, parameters in the teacher detector were updated with the trained student detector via exponential moving average (EMA). Some current literature [11,40,41] has attempted to improve the previous work [10]. Wang et al. [11] focused on how to remove incorrect annotations from the predicted label l with multi-thresholds of object confidence, class, and 3D IoU. Wang et al. [40] attempted to generate accurate predicted labels with temporal smoothing. The teacher 3D detector predicted multi-frame labels from the multi-frame LiDAR data. After that, temporal GNN was used to generate the accurate labels at the current frame from these multi-frame labels. Park et al. [41] exploited a multi-task (i.e., 3D and 2D object detection) teacher detector to establish multi-task guided consistency loss for supervision. It works on a LiDAR-camera system. Sautier et al. [42] presented a self-supervised distillation method to pre-train the backbone network in a 3D object detector, with the guidance of super-pixel segmentation results, on an RGB image.
Some researchers consider that weak label is convenient and time-efficient for annotation, and study weak-supervised 3D object detection (WSOD-3D). Meng et al. [43] proposed a weak and fast 3D annotation procedure to generate a 3D cylindrical bounding box by clicking the object center in an RGB image. With the cylindrical label, they converted SSOD-3D as WSOD-3D and provided a two-stage training scheme for the 3D object detector. Qin et al. [44] designed an unsupervised 3D proposal generation method to obtain the 3D bounding box with anchor size, using the normalized point cloud density. Peng et al. [45] presented a WSOD-3D method for a monocular camera, which utilizes the alignment constraint of predicted 3D proposal and LiDAR points for the weak supervision. Xu et al. [46] dealt with WSOD-3D under the condition that the position-level annotations are known. A virtual scene with GT annotation is constructed with the known object centers. Then, samples in the real scene with weak labels and in the virtual scene with GT labels are both used for detector training.

2.3. Discussions

From the above analysis, most of the current study of SSOD-3D emphasises improving the quality of pseudo labels. Two common schemes are utilized: (i) label filter [10,11,38,41] and (ii) temporal smoothing [40]. However, both of these have space for improvement. The label filter scheme is not time-efficient enough to search for the optimal filter thresholds. The temporal smoothing scheme needs multi-frame LiDAR point clouds with the accurate sensor pose information; this need is difficult to satisfy in some actual situations. Although the weak label (i.e., 2D bounding box [9,45], 3D cylindrical bounding box [43] and 3D center position [46]) is easier for annotation than the standard label (i.e., 3D bounding box), it still costs time and human resources for  the amount of unlabeled data that needs annotation in the context of autonomous driving. Therefore, an effective SSOD-3D method is still required.

3. Proposed Semi-Supervised 3D Object Detection

3.1. Problem Statement

SSOD-3D is a training framework to learn baseline detection from the labeled and unlabeled datasets X l and X u . The baseline detector is the arbitrary 3D object detector based on the LiDAR point cloud. Let w be the parameter set of baseline detector. SSOD-3D aims to learn w * with higher generalization ability.
Some symbols are discussed here. Let x i = { P i } be the i-th training sample where x i X l or x i X u means that it is ground truth (GT), labeled or not. P i is [ N i , 4 ] LiDAR point cloud where N i is number of LiDAR points. It contains the 3D position and reflected intensity of the LiDAR point cloud. Let l i = { l i j } j = 1 n i be the 3D object label of x i . n i is the object number. l i j represents the 3D bounding box of the j-th object using the parameter vector of the 3D bounding box [28]. l i is the pseudo or GT label if x i X u or x i X l . Let l i p = f ( x i ; w ) be the output of the 3D object detector with the input of x i and weight of w . l i p = { l i k p } k = 1 n i is the pseudo label of x i .

3.2. Previous Semi-Supervised 3D Object Detection

Before illustrating the proposed SPSL-3D, we briefly revisit the previous SSOD-3D approach [10]. The pipeline of the previous SSOD-3D is presented in Figure 1a. For the 3D object detector with high generalization ability, its prediction results from the unlabeled sample x i and its augmented sample A ( x i ) are both consistent and closed to the GT labels. Based on this analysis, as the unlabeled sample does not have annotation, L cons was proposed to minimize the difference in labels predicted from x i and A ( x i ) . A ( x i ) is the affine transformation on P i of x i , which contains scaling, X/Y-axis flipping, and Z-axis rotating operations. L cons is the core in this scheme [10], for this loss can update the weights in the 3D object detector via back-propagation. The current SSOD-3D [10,11,38,41,47] optimizes w * by minimizing the function as:
L SSOD - 3 D ( w ) = i = 1 N l L 3 d ( l i p , l i ) 1 + L cons , x i X l
L cons ( w ) = j = 1 N u L 3 d ( l j p , A 1 ( l j , Aug p ) ) 1 , x j X u
l j p = f ( x j ; w ) , l j , Aug p = f ( A ( x j ) ; w )
where L 3 d ( l i p , l i ) is the common 3D object detection loss of each detected object [20,28]. It is represented as [ n i , 1 ] vector to describe the detection loss of each object. With the inverse affine transformation A 1 ( · ) , A 1 ( l i , Aug p ) is obtained with the same reference coordinate system as in l i p . In the end, we also provide discussion of the relation of a previous SSOD-3D, and traditional SSL theory is further discussed in Appendix A.1.

3.3. Self-Paced Semi-Supervised Learning-Based 3D Object Detection

The main challenge of consistency loss in Equation (2) is that the quality of the pseudo label l j p is uncertain. As  N u > N l , if  l j p is noisy or even incorrect, the baseline detector with the optimized parameter set w * tends to detect 3D objects with low localization accuracy. To deal with this problem, we needs to evaluate the reliability weight v j of l j p , where v j = ( v j 1 , , v j k , , v j n j ) T is a vector to reflect the reliability score of objects in l j p ( v j k [ 0 , 1 ] ). In the training stage, unreliable pseudo labels are filtered out with v j . However, determining v j is a crucial problem.
One naive idea is to adjust the reliability weight v j with the guidance of the consistency loss of l j p . If the consistency loss of l j p enlarges, the pseudo labels l j p are unreliable. Based on this idea, we exploit the theory of SPL [18] to construct the mathematical relation of v j to L 3 d ( l j p , A 1 ( l j , Aug p ) ) , and propose a novel SSOD-3D framework, SPSL-3D, in this paper. Its pipeline is presented in Figure 1b. SPSL-3D optimizes w * by minimizing the function as:
L raw ( w , v , λ l , λ u ) = i = 1 N l v i T L 3 d ( l i p , l i ) + j = 1 N u v j T L 3 d ( l j p , A 1 ( l j , Aug p ) ) + f u ( { v j } j = 1 N u , λ u ) , v = { v j } j = 1 N u , x i X i , x j X j
λ u = e E max ( L 3 d ( l j p , A 1 ( l j , Aug p ) ) ) + 1 e E mean ( L 3 d ( l j p , A 1 ( l j , Aug p ) ) )
where λ u is age parameter to control the learning pace [48]. Let the current epoch and maximum training epoch be e and E. Furthermore, f u ( { v j } j = 1 N u , λ u ) is a self-paced regularization term [48] for the unlabeled sample:
f u ( { v j } i = 1 N u , λ u ) = λ u j = 1 N u k = 1 n j 1 2 v j k 2 + v j k
However, in deep learning, w contains lots of parameters, so it is difficult to directly optimize Equation (3). As the modern deep neural network (DNN) is trained with a batch of data the size of B l + B u [49], the loss of SPSL-3D is simplified as:
L SPSL - 3 D ( w , v , λ l , λ u ) = i = 1 B l v i T L 3 d ( l i p , l i ) + j = 1 B u v j T L 3 d ( l j p , A 1 ( l j , Aug p ) ) λ u k = 1 n j 1 2 v j k 2 + v j k , v = { v j } j = 1 B u , x i X i , x j X j
An alternative optimization scheme is used to optimize w and v . With the fixed w , v needs to be optimized. The closed-form solution is obtained as Equation (7) via L SPSL - 3 D / v = 0 . For a vector L, [ L ] k is its k-th element. With the fixed v , w is optimized in Equation (6) with the deep learning-based optimizer (i.e., Adam and SGD).
v j k = 1 [ L 3 d ( l j p , A 1 ( l j , Aug p ) ) ] k λ u , [ L 3 d ( l j p , A 1 ( l j , Aug p ) ) ] k < λ u 0 , [ L 3 d ( l j p , A 1 ( l j , Aug p ) ) ] k λ u
Intuitive explanation of Equation (7) is discussed here. For the k-th object in the sample with the pseudo label, if its loss is larger than λ u , it is regarded as an unreliable label and cannot be used. If its loss is smaller than λ u , SPSL-3D evaluates its reliability score with its consistency loss. SPSL-3D emphasizes the most reliable pseudo label in the training stage to enhance the robustness of the baseline detector. When epoch e grows, λ u increases (seen Equation (4)), meaning that SPSL-3D enlarges the size of the unlabeled samples for training, thus improving the generalization ability of baseline detector. The improvement can be found in Figure 1c.

3.4. Improving SPSL-3D with Prior Knowledge

From Equation (7), SPSL-3D can adaptively adjust the reliability weight of pseudo label using its 3D object detection loss. In fact, the reliability weight of pseudo label is not only dependent on consistency loss, but also dependent on the prior knowledge in the LiDAR point cloud and RGB image provided by the LiDAR-camera system. If the LiDAR point cloud or image feature of one predicted object is not salient, its pseudo label is not reliable. Based on this analysis, to further enhance the performance of PSPL-3D with information from the LiDAR point cloud and RGB image, we propose a prior knowledge-based SPSL-3D named PSPSL-3D, which is presented in Figure 2.
In PSPSL-3D, we attempt to represent the reliability of the pseudo label with the LiDAR point cloud and RGB image. For the k-th object o j k in l j p , its prior reliability coefficient is modeled as ρ det , j k = ρ diff , j k · ρ anno , j k . It consists of the detection difficulty coefficient ρ diff , j k [ 0 , 1 ] and the label accuracy coefficient ρ anno , j k [ 0 , 1 ] . The motivation of designing ρ det , j k is to constrain v j k with both 3D detection loss and prior knowledge from the RGB image and LiDAR point cloud.
Due to the LiDAR mechanism, ρ diff , j k is mainly dependent on the occlusion and resolution of o j k in the LiDAR point cloud. However, due to the complex situation of 3D object in the real traffic situation, it is difficult to model the relationship between the occlusion, resolution, and  detection difficulty of o j k . An approximate solution is provided here. Following the thought in the literature [50], a statistical variable m j k is used as the LiDAR point number inside the 3D bounding box of o j k to describe ρ diff , j k :
ρ diff , j k = min ( m j k , m t h ( c j k ) ) m t h ( c j k )
where c j k is the category (i.e., car, pedestrian, cyclist) of o j k , and  m t h ( c j k ) is the minimal threshold of the LiDAR point number of the corresponding category. For the actual implementation, m t h ( c i k ) is a statistic variable from X l , discussed in Section 4.2. If  o j k has higher resolution and less occlusion, m j k is closer and even higher than m t h ( c j k ) , so that ρ diff , j k is closer to 1. The 3D detection difficulty of o j k is largely decreased.
Then, ρ anno , j k is discussed. As GT is unknown, we attempt to evaluate the annotation accuracy indirectly. On the one hand, as for the current 3D detection method, a confidence score s j k p [ 0 , 1 ] of o j k is supervised with the 3D IoU of the predicted and GT 3D bounding box [21,26]. s j k p can be used to describe ρ anno , j k . On the other hand, a semantic segmentation map predicted from RGB image also contains annotation information. As the RGB image is more dense than the LiDAR point cloud, semantic segmentation on the RGB image is more accurate than the semantic segmentation on the LiDAR point cloud. Projecting the 3D bounding box of o j k onto the image plane generates a 2D bounding box B j k . Its pixel area is A j k . The pixel area of the semantic map of c j k inside B j k is S j k . If the predicted 3D bounding box of o j k is accurate, s j k p and ratio of S j k and A j k are closer to 1. Due to the occlusion of object, S j k is not accurate enough. Thus, the arithmetic mean of s j k p and the pixel area ratio is used to describe ρ anno , j k :
ρ anno , j k = s i k p · S i k A i k
From the above discussion, prior knowledge is not directly extracted from the RGB image, for the RGB feature of the targeted object is affected by shadow and blur in the complex traffic scenario. Compared with the RGB image, the semantic segmentation map is more stable to reflect the location information of a targeted object. Thus, the semantic feature is used to describe ρ anno , j k .
After obtaining ρ det , j k , a scheme designed to constrain v j k with ρ det , j k is required. Referring to the thought in self-paced curriculum learning [51], the interval of v j k can be constrained from [ 0 , 1 ] to [ 0 , ρ det , j k ] . This means that the interval of v j k is dependent on its prior detection coefficient ρ det , j k . To achieve this scheme, v j k in Equation (6) is replaced with v j k / ρ det , j k , and the loss function of PSPSL-3D is presented as:
L PSPSL - 3 D ( w , v , λ l , λ u ) = i = 1 B l v i T L 3 d ( l i p , l i ) + j = 1 B u v j T L 3 d ( l j p , A 1 ( l j , Aug p ) ) λ u k = 1 n j 1 2 v j k ρ det , j k 2 + v j k ρ det , j k , v = { v j } j = 1 B u , x i X i , x j X j
As the same in Section 3.2, the alternative optimization scheme is used to find the optimal w and v . The close-formed solution of v is shown in Equation (11) via L PSPSL - 3 D / v = 0 . The procedure of PSPSL-3D is summarized in Algorithm 1.
v j k = ρ det , j k · 1 [ L 3 d ( l j p , A 1 ( l j , Aug p ) ) ] k λ u , [ L 3 d ( l j p , A 1 ( l j , Aug p ) ) ] k < λ u 0 , [ L 3 d ( l j p , A 1 ( l j , Aug p ) ) ] k λ u
Algorithm 1 Proposed SPSL-3D and PSPSL-3D framework for SSOD-3D.
Inputs: Baseline detector, maximum epoch E, datasets X l and X u , batch sizes B l and B u ;
Parameters: Baseline detector weight w , current epoch e, sample weight v , age parameters λ l and λ u ;
Output: Optimal 3D detector weight w *
      1:
Pre-training baseline 3D detector in X l obtains w 0
      2:
Let e = 1 and w = w 0
      3:
while  e E   do
      4:
   Let k = 1
      5:
   while  k N l  do
      6:
      { x i , l i } i = 1 B l = DataLoader ( X l , B l ) , { x j } j = 1 B u = DataLoader ( X u , B u )
      7:
     Computing λ u with e and E via Equation (4)
      8:
     if SPSL-3D is exploited then
      9:
          Optimizing v using w , λ u via Equation (7)
    10:
        Optimizing w with v via Equations (6)
    11:
     end if
    12:
     if PSPSL-3D is exploited then
    13:
        Computing ρ det , j k via Equations (8) and (9)
    14:
        Optimizing v using w , ρ det , i k , λ u via Equation (11)
    15:
        Optimizing w with v via Equation (10)
    16:
     end if
    17:
      k = k + B l
    18:
   end while
    19:
    e = e + 1
    20:
end while
    21:
Return w * as w .

4. Experiments

4.1. Dataset and Configuration

The classical outdoor KITTI dataset [19] is exploited to evaluate 3D detection performance in the outdoor traffic scenario. It contains (i) training dataset D train with 3712 samples, (ii) validation dataset D val with 3769 samples, and (iii) testing dataset D test with 7518 samples. All of them have GT annotation. The LiDAR point cloud and RGB image are provided in each sample. As the raw KITTI dataset does not contain semantic segmentation images, we generate semantic maps with four categories (i.e., car, pedestrian, bicycle, background) using the pre-trained deeplab v3 [52]. Three categories (i.e., car, pedestrian, cyclist) of targeted object are considered in the following experiments. To verify the performance of SSOD-3D methods, a semi-supervised condition is established in the experiments. D train is divided as X l and X u where X l X u = D train and X l X u = . GT labels in X u are disabled in the training stage. X u is regarded as unlabeled dataset. Let R anno = N l / ( N l + N u ) be the labeled ratio. In the following experiments, to evaluate the performance of SSOD-3D comprehensively, we set various training situations with the different labeled ratios, from 4% (hard SSL case ) to 64% (easy SSL case). Specifically, R anno is set as 4%, 8%, 16%, 32%, 64%, respectively. Furthermore, we mainly focus on the comparison results in the hard SSL case ( R anno 16 % ).
To measure the results of SSOD-3D methods, SSOD-3D methods are trained on D train , and then evaluated on D val . 3D average precision (AP) is the main metric for comparison. In order to further evaluate the different SSOD-3D methods, a bird’s eye view (BEV) AP and 3D recall rate are also provided. The IoU thresholds of 3D and BEV objects are 0.7 (car) and 0.5 (pedestrian and cyclist). As object label in KITTI dataset has three levels (i.e., easy, moderate, hard), these metrics of all level objects are provided for the comprehensive comparison.
The proposed SPSL-3D, PSPSL-3D needs a baseline detector. Voxel RCNN [21] is selected as the baseline 3D detector as it has simple detector architecture and fast and accurate inference performance. THe optimizer, learning rate policy, and hyper-parameters in A ( · ) are default [21]. As the proposed method is implemented on a single Nvidia GTX 3070, the batch size is set as 2, where B l = 1 and B u = 1 . w 0 is needed for SPSL-3D and PSPSL-3D. This is obtained by per-training Voxel R-CNN on X l with 80 epochs. The maximum epoch E is related to R anno . Fine-tuning experience shows that better results are achieved when E = 30 if R anno 16 % and E = 50 if R anno > 16 % . In the actual training stage, data augmentation contains not only 3D affine transformation on the LiDAR point cloud, but also the cut-and-paste operation [4]. This operation aims to increase the object number in x i by putting extra 3D objects with GT annotation into the point cloud P i . 3D objects with GT labels are stored in the object bank before training. To prevent data leakage, the object bank should be built only on X l instead of D train .
Current SSOD-3D methods are selected for comparison. SESS [10] is the first to utilize the mean-teacher SSL framework [39] in SSOD-3D. SESS is enhanced with the multi-threshold label filter (LF) proposed in work [11], and the improved method is named as 3DIoUMatch. UDA [53] is a classical SSL framework. It exploits the consistency loss L cons for supervision. We consider that it can work for SSOD-3D. As curriculum learning (CL) [54] is a fundamental part of SPL theory [18], CL can also be used in SSOD-3D, so that method UDA+CL is designed. As unlabeled data increase learning uncertainty, UDA+CL tries to learn w with increasing unlabeled data. For L cons in Equation (3), N l is replaced with α ( e ) N l where α ( e ) = e / E . These methods also require a baseline 3D detector. For a fair comparison, Voxel RCNN [21] is marked as Baseline and used as the baseline 3D detector for all SSOD-3D methods. The above mentioned methods are trained in the same condition. As for the SSOD-3D method on the LiDAR point cloud, only the open-source code of 3DIoUMatch [11] is provided (https://github.com/THU17cyz/3DIoUMatch-PVRCNN, accessed on 1 July 2022). Other methods are implemented by authors on the open-source FSOD-3D framework OpenPCDet (https://github.com/open-mmlab/OpenPCDet, accessed on 1 March 2022).

4.2. Comparison with Semi-Supervised Methods

This experiment investigates the comparison results of the proposed PSPSL-3D methods with the current SSOD-3D methods. The results of all SSOD-3D methods at R anno = 4 % , 8 % , 16 % , 32 % , 64 % are presented in Table 1, Table 2, Table 3, Table 4 and Table 5. In these tables, gain from baseline means the improvement of the proposed PSPSL-3D over the baseline method. For the baseline, its 3D mAPs of all categories are dramatically increased from R anno = 4 % to 8 % , suggesting the large potential improvement of SSOD-3D. The 3D mAPs of baseline increased slowly when R anno 16 % , and the improvements of SSOD-3D methods are also relatively small. The 3D mAP of 3DIoUMatch [11] is higher than SESS [10], as the multi-threshold-based label filter in 3DIoUMatch can remove some incorrectly predicted labels. The 3D mAP of UDA+CL [54] is higher than UDA [53], as the curriculum can reduce certain instances of overfitting of easily detected objects in the training stage. In most cases, the proposed SPSL-3D is superior to SESS [10] and UDA [53], because it exploits SPL theory [18] to filter incorrect and too difficult labeled and unlabeled training samples adaptively. It is noticed that the proposed PSPSL-3D has higher 3D mAPs than other current methods because it adds prior knowledge from the RGB image and LiDAR point cloud as self-paced regularization terms in SPSL-3D to achieve robust and accurate learning results. It is found that 3DIouMatch [11] has better performance than other previous methods. Compared with 3DIoUMatch, the proposed PSPSL-3D makes a significant improvement in 3D cyclist detection and also has a certain improvement in 3D pedestrian detection, because the prior knowledge from the LiDAR point cloud and RGB image is beneficial to modeling the reliability of objects with a relatively small size. Therefore, it is concluded that the proposed SPSL-3D and PSPSL-3D benefit SSOD-3D in a LiDAR-camera system.

4.3. Comparison with Fully Supervised Methods

This experiment investigates the comparison results of the proposed PSPSL-3D methods with current FSOD-3D methods. FSOD-3D methods are all trained on the entire D train . For the proposed method PSPSL-3D at R anno = 100 % , the unlabeled testing samples D test in the KITTI dataset are used in the training procedure. Results in the KITTI validation dataset are provided in Table 6. The 3D APs of almost all categories of SPSL-3D and PSPSL-3D are smaller than fully supervised Voxel RCNN [21], but the 3D APs difference between car and pedestrian is not large. Compared with other FSOD-3D methods, it is found that 3D APs of pedestrians of PSPSL-3D are larger than some of classical methods [4,20,26], while 3D APs of the car category are smaller than the state-of-the-art FSOD-3D methods [14,29,55,56]. Additionally, the BEV AP results of SPSL-3D, PSPSL-3D, and fully supervised Voxel RCNN [21] are also presented in Table 7. As BEV object detection is easier than 3D object detection, it is found that most of the BEV APs of the proposed SSOD-3D methods and fully supervised methods are fairly close. In conclusion, the proposed PSPSL-3D method, which needs only 64% labeled data, can achieve a performance that is close to the BEV and 3D object detection performance of current FSOD-3D methods. By exploiting more unlabeled samples ( R anno = 100 % ), the proposed PSPSL-3D outperforms than other FSOD-3D methods, which means that SSOD-3D has huge research potential in the field of autonomous driving.

4.4. Visualizations

To better show the learning efficiency of PSPSL-3D, visualizations of 3D object detection are provided and discussed in this experiment. Comparisons with a fully supervised baseline 3D detector and PSPSL-3D trained in R anno = 64 % -labeled data are shown in Figure 3. In the different outdoor scenarios, the proposed PSPSL-3D has nearly the same performance as the fully supervised baseline detector. If the object is far from the LiDAR, only a few false-positives are generated in PSPSL-3D. To further compare 3D object detection results in the LiDAR point cloud, more visualizations of the proposed method in the condition of R anno = 8 % are presented in Figure 4 and Figure 5. As 3DIoUMatch [11] has a performance close to that of PSPSL-3D, it is used for the main comparison. In the complex street scenarios with lots of cars and cyclists, PSPSL-3D has a higher 3D recall rate than 3DIoUMatch, because the proposed PSPSL-3D emphasizes the reliability of the pseudo label, thus achieving stable learning results. PSPSL-3D also has accurate results in simple outdoor scenarios. It is concluded that the proposed PSPSL-3D has stable and accurate 3D object detection performance.

4.5. Ablation Study

This experiment evaluates the effectiveness of the proposed PSPSL-3D framework. From Table 6, Table 7 and Table 8, it is also found that the 3D AP, BEV AP, and 3D recall rates of PSPSL-3D are higher than those of SPSL-3D in most of cases. Specifically, in the task of BEV object detection, PSPSL-3D has significantly improved pedestrian and cyclist detection under the same annotation conditions. In the task of 3D object detection, under the condition that R anno 16 % , PSPSL-3D has a larger improvement than SPSL-3D in detecting 3D objects of all categories. The reason for this is provided in the following. Based on an SPSL-3D framework, PSPSL-3D exploits prior knowledge from the RGB image and LiDAR point cloud and then establishes extra regularization terms to prevent incorrectly predicted labels, thus achieving a higher 3D object detection performance. Therefore, ablation studies demonstrate the effectiveness of the proposed PSPSL-3D.

5. Discussion

The proposed SSOD-3D frameworks, SPSL-3D and PSPSL-3D, have several advantages. At first, we consider the reliability of pseudo label in the SSOD-3D training stage. As the GT annotation of pseudo label is unknown, we attempt to use the consistency loss to represent the weight of the pseudo label. If the pseudo label is incorrect, its consistency loss is larger than other pseudo labels. To reduce the negative effect of the pseudo label with large noise, we need to decrease the reliability weight of this pseudo label in the training stage. To adaptively and dynamically adjust the reliability weight of all pseudo labels, we exploit the theory of SPL [18] and then propose SPSL-3D as a novel and efficient framework. Second, we utilize the multi-model sensor data in the semi-supervised learning stage, thus further enhancing the capacity of the baseline 3D object detector based on LiDAR. The reason for the usage of multi-model data is that we notice that the LiDAR-camera system is widely equipped in the autonomous driving system. Thus, both the RGB image and LiDAR point cloud can be used in the training stage of SSOD-3D. For one object, there are generally abundant structural and textural features in the RGB image and LiDAR point cloud. However, as shown in Figure 6, the RGB image has shadow, occlusion, and blur, so it is difficult to extract prior knowledge of an object in the RGB image. Compared with RGB images, semantic segmentation images can directly reflect the object category information. Based on this analysis, we used the area of semantic segmentation of the object to describe the annotation accuracy. However, the semantic segmentation image also has two main limitations: it cannot identify occluded objects, and finds it hard to determine objects that are far from the sensor. Thus, the proposed PSPSL-3D exploits Equation (9) to approximately represent the annotation accuracy of the pseudo label.
Extensive experiments in Section 4 demonstrate the effectiveness of the proposed SPSL-3D and PSPSL-3D. Firstly, compared with the state-of-the-art SSOD-3D methods, the proposed SPSL-3D and PSPSL-3D frameworks have achieved the better results than other methods because the proposed frameworks consider the reliability of the pseudo labels, thus decreasing the negative effect of incorrect pseudo labels in the training stage. Second, the proposed frameworks are also suitable to train baseline 3D object in a fully supervising way. Compared with the current FSOD-3D methods, the baseline 3D detector trained with PSPSL-3D outperforms other FSOD-3D methods. This means that the training scheme which emphasizes the weight of label is beneficial to baseline 3D object detector training.
In the future, we will study SSOD-3D in the several ways. Firstly, computer graphic (CG) techniques can be used to generate a huge number of labeled simulated samples. Exploiting the theory of SPL [57] in SSOD-3D with unlabeled samples and labeled simulated samples is an ongoing problem. Secondly, in actual application, the data distribution of the labeled LiDAR point cloud might be different from that of the unlabeled LiDAR point cloud because the dataset is collected with a different type of LiDAR sensor at a different place. Utilizing domain adaptation in a SSOD-3D is a challenging problem. We will deal with the above problems in subsequent studies.

6. Conclusions

The main challenge of learning-based 3D object detection is the shortage of labeled samples. To make full use of unlabeled samples, SSOD-3D is an important technique. In this paper, we propose a novel and efficient SSOD-3D framework for 3D object detection on a LiDAR-camera system. Firstly, to avoid the negative effect of unreliable pseudo labels, we propose SPSL-3D to adaptively evaluate the reliability weight of pseudo labels. Secondly, to better evaluate the reliability weight of pseudo labels, we utilize prior knowledge from the LiDAR-camera system and present the PSPSL-3D framework. Extensive experiments show the effectiveness of the proposed SPSL-3D and PSPSL-3D on the public dataset. Hence, we believe that the proposed framework benefits 3D environmental perception in autonomous driving.

Author Contributions

Methodology and writing—original draft preparation, P.A.; software, J.L.; investigation, X.H.; formal analysis, T.M.; validation, S.Q.; supervision, Y.C. and L.W.; Resources, and funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (U1913602, 61991412, 62201536). Equipment Pre-Research Project (41415020202, 41415020404, 305050203).

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: http://www.cvlibs.net/datasets/kitti/ (accessed on 1 March 2022).

Acknowledgments

The authors thank Siying Ke, Bin Fang, Junfeng Ding, Zaipeng Duan, and Zhenbiao Tan from Huazhong University of Science and Technology, and anonymous reviewers for providing many valuable suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LiDARLight detection and ranging
3DThree dimensional
SSLSemi-supervised learning
SSOD-3DSemi-supervised 3D object detection
FSOD-3DFully supervised 3D object detection
WSOD-3DWeakly supervised 3D object detection
SPLSelf-paced learning
SPSL-3DSelf-paced semi-supervised learning based 3D object detection
PSPSL-3DSelf-paced semi-supervised learning based 3D object detection with prior knowledge
APAverage precision
BEVBird’s eye view
UDAUnsupervised data augmentation
CLCurriculum learning
IoUIntersection over union
GTGround truth
EMAExponential moving average
NMSNon maximum suppression
FCFully connected
CNNConvolutional neural network
GNNGraph neural network
SpconvSparse convolutional neural network

Appendix A

Figure A1. Relation of classical SSL, current SSOD-3D, and proposed SPSL-3D, PSPSL-3D. As weight function w ( x i , x j ) is difficult to design in complex traffic scenarios, the regularization (reg.) term w I in traditional SSL theory is simplified as the consistency loss L cons . In this paper, emphasising the impact of the reliability of pseudo labels on SSOD-3D training, we present SPL-based consistency loss L SPSL - 3 D and SPSL-3D. Prior information (info.) is exploited to enhance the effect of SPL on SPSL-3D.
Figure A1. Relation of classical SSL, current SSOD-3D, and proposed SPSL-3D, PSPSL-3D. As weight function w ( x i , x j ) is difficult to design in complex traffic scenarios, the regularization (reg.) term w I in traditional SSL theory is simplified as the consistency loss L cons . In this paper, emphasising the impact of the reliability of pseudo labels on SSOD-3D training, we present SPL-based consistency loss L SPSL - 3 D and SPSL-3D. Prior information (info.) is exploited to enhance the effect of SPL on SPSL-3D.
Remotesensing 15 00627 g0a1

Appendix A.1. Relation of Traditional SSL Theory and Previous SSOD-3D Method

We briefly revisit the previous SSOD-3D approach [10] and discuss the relation between the classical SSL theory and existing SSOD-3D method (seen in Figure A1). Classical SSL [37] aims to find w * by minimizing the following cost function:
L SSL ( w ) = i = 1 N l L 3 d ( l i p , l i ) 1 + α w 2 2 + γ w I 2 , x i X l
where α and γ are coefficients of regularization terms. Suppose that x lies in a compact manifold M [37]. P x is the marginal distribution of x . w I reflects the intrinsic structure of P x on M [37]:
w I 2 = x M M w 2 d P x ( x ) 1 N l + N u i j w ( x i , x j ) L 3 d ( l i p , l j p ) , x i , x j X l X u
In the above equation, data weight w ( x i , x j ) [ 0 , 1 ] describes the similarity of x i and x j . However, as presented in Figure 6, x i and x j collected in the different scenarios have different data distribution. Their object number and category are also not the same. Therefore, it is hard to design a suitable w ( x i , x j ) . There is a gap between traditional SSL theory and SSOD-3D.
To solve this problem, researchers attempted to relax w I and present the consistency loss L cons as an approximation of SSOD-3D. They exploit data augmentation operator A ( · ) to create datum A ( x i ) similar to x i [10,39]. A ( x i ) is used to replace x j in Equation (A2). A ( x i ) is the affine transformation on P i of x i , which contains scaling, X/Y-axis flipping and Z-axis rotating operations [26]. As affine transformation does not change the structure of point cloud, it is safe to assume that w ( x i , A ( x i ) ) 1 . l i , Aug p is one of the results in f ( A ( x i ) ; w ) . With the inverse affine transformation A 1 ( · ) , A 1 ( l i , Aug p ) is obtained, which has the same reference coordinate system as l i p . After that, w I 2 in Equation (A1) is simplified and replaced as one consistency loss L cons [10] as Equation (2). It is the core loss function in the pseudo label based SSOD-3D methods [10,11,38,41,47]. In the other literature, L cons is called unsupervised data augmentation (UDA) [53] because it does not utilize any GT information. The current SSOD-3D [10] optimizes w * by minimizing the function in Equation (1).

References

  1. Li, X.; Zhou, Y.; Hua, B. Study of a Multi-Beam LiDAR Perception Assessment Model for Real-Time Autonomous Driving. IEEE Trans. Instrum. Meas. 2021, 70, 1–15. [Google Scholar] [CrossRef]
  2. Yuan, Z.; Song, X.; Bai, L.; Wang, Z.; Ouyang, W. Temporal-Channel Transformer for 3D Lidar-Based Video Object Detection for Autonomous Driving. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2068–2078. [Google Scholar] [CrossRef]
  3. Zhang, J.; Lou, Y.; Wang, J.; Wu, K.; Lu, K.; Jia, X. Evaluating Adversarial Attacks on Driving Safety in Vision-Based Autonomous Vehicles. IEEE Internet Things J. 2022, 9, 3443–3456. [Google Scholar] [CrossRef]
  4. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud. In Proceedings of the CVPR, Long Beach, CA, USA, 16–20 June 2019; pp. 770–779. [Google Scholar]
  5. Zhu, H.; Yuen, K.; Mihaylova, L.; Leung, H. Overview of Environment Perception for Intelligent Vehicles. IEEE Trans. Intell. Transp. Syst. 2017, 18, 2584–2601. [Google Scholar] [CrossRef] [Green Version]
  6. Zamanakos, G.; Tsochatzidis, L.T.; Amanatiadis, A.; Pratikakis, I. A comprehensive survey of LIDAR-based 3D object detection methods with deep learning for autonomous driving. Comput. Graph. 2021, 99, 153–181. [Google Scholar] [CrossRef]
  7. He, C.; Zeng, H.; Huang, J.; Hua, X.; Zhang, L. Structure Aware Single-stage 3D Object Detection from Point Cloud. In Proceedings of the CVPR, Seattle, WA, USA, 14–19 June 2020; pp. 11870–11879. [Google Scholar]
  8. Fang, J.; Zuo, X.; Zhou, D.; Jin, S.; Wang, S.; Zhang, L. LiDAR-Aug: A General Rendering-Based Augmentation Framework for 3D Object Detection. In Proceedings of the CVPR, Virtual, 19–25 June 2021; pp. 4710–4720. [Google Scholar]
  9. Tang, Y.S.; Lee, G.H. Transferable Semi-Supervised 3D Object Detection From RGB-D Data. In Proceedings of the IEEE ICCV, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1931–1940. [Google Scholar]
  10. Zhao, N.; Chua, T.; Lee, G.H. SESS: Self-Ensembling Semi-Supervised 3D Object Detection. In Proceedings of the CVPR, Seattle, WA, USA, 13–19 June 2020; pp. 11076–11084. [Google Scholar]
  11. Wang, H.; Cong, Y.; Litany, O.; Gao, Y.; Guibas, L.J. 3DIoUMatch: Leveraging IoU Prediction for Semi-Supervised 3D Object Detection. In Proceedings of the CVPR, Virtual, 19–25 June 2021; pp. 14615–14624. [Google Scholar]
  12. Sohn, K.; Berthelot, D.; Carlini, N.; Zhang, Z.; Zhang, H.; Raffel, C.; Cubuk, E.D.; Kurakin, A.; Li, C. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence. In Proceedings of the NeurIPS, Virtual, 6–12 December 2020; pp. 1–13. [Google Scholar]
  13. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D Object Detection Network for Autonomous Driving. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 6526–6534. [Google Scholar]
  14. Yoo, J.H.; Kim, Y.; Kim, J.S.; Choi, J.W. 3D-CVF: Generating Joint Camera and LiDAR Features Using Cross-View Spatial Feature Fusion for 3D Object Detection. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; pp. 1–16. [Google Scholar]
  15. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum PointNets for 3D Object Detection from RGB-D Data. In Proceedings of the CVPR, Salt Lake City, UT, USA, 18–23 June 2018; pp. 918–927. [Google Scholar]
  16. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3D Proposal Generation and Object Detection from View Aggregation. In Proceedings of the IROS, Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar]
  17. An, P.; Liang, J.; Yu, K.; Fang, B.; Ma, J. Deep structural information fusion for 3D object detection on LiDAR-camera system. Comput. Vis. Image Underst. 2022, 214, 103295. [Google Scholar] [CrossRef]
  18. Kumar, M.P.; Packer, B.; Koller, D. Self-Paced Learning for Latent Variable Models. In Proceedings of the NeurIPS, Vancouver, BC, Canada, 6–9 December 2010; pp. 1189–1197. [Google Scholar]
  19. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In In Proceedings of the CVPR, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  20. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection From Point Clouds. In Proceedings of the CVPR, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  21. Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; Li, H. Voxel R-CNN: Towards High Performance Voxel-based 3D Object Detection. In Proceedings of the AAAI, Virtually, 2–9 February 2021; pp. 1201–1209. [Google Scholar]
  22. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  23. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the NIPS, Long Beach, CA, USA, 4–9 December 2017; pp. 5099–5108. [Google Scholar]
  24. Shi, W.; Rajkumar, R. Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud. In Proceedings of the CVPR, Seattle, WA, US, 13–19 June 2020; pp. 1708–1716. [Google Scholar]
  25. Graham, B.; Engelcke, M.; van der Maaten, L. 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks. In Proceedings of the CVPR, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9224–9232. [Google Scholar]
  26. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection. In Proceedings of the CVPR, Seattle, WA, USA, 13–19 June 2020; pp. 10526–10535. [Google Scholar]
  27. Yin, T.; Zhou, X.; Krähenbühl, P. Center-Based 3D Object Detection and Tracking. In Proceedings of the CVPR, Virtual, 19–25 June 2021; pp. 11784–11793. [Google Scholar]
  28. Shi, S.; Wang, Z.; Shi, J.; Wang, X.; Li, H. From Points to Parts: 3D Object Detection From Point Cloud With Part-Aware and Part-Aggregation Network. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2647–2664. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Zheng, W.; Tang, W.; Jiang, L.; Fu, C. SE-SSD: Self-Ensembling Single-Stage Object Detector From Point Cloud. In Proceedings of the CVPR, Virtual, 19–25 June 2021; pp. 14494–14503. [Google Scholar]
  30. Schinagl, D.; Krispel, G.; Possegger, H.; Roth, P.M.; Bischof, H. OccAM’s Laser: Occlusion-based Attribution Maps for 3D Object Detectors on LiDAR Data. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 1131–1140. [Google Scholar]
  31. Man, Y.; Weng, X.; Sivakumar, P.K.; O’Toole, M.; Kitani, K. Multi-Echo LiDAR for 3D Object Detection. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2021; pp. 3743–3752. [Google Scholar]
  32. Chen, Y.; Li, Y.; Zhang, X.; Sun, J.; Jia, J. Focal Sparse Convolutional Networks for 3D Object Detection. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 5418–5427. [Google Scholar]
  33. Wu, X.; Peng, L.; Yang, H.; Xie, L.; Huang, C.; Deng, C.; Liu, H.; Cai, D. Sparse Fuse Dense: Towards High Quality 3D Detection with Depth Completion. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 5408–5417. [Google Scholar]
  34. Li, Y.; Yu, A.W.; Meng, T.; Caine, B.; Ngiam, J.; Peng, D.; Shen, J.; Wu, B.; Lu, Y.; Zhou, D.; et al. DeepFusion: Lidar-Camera Deep Fusion for Multi-Modal 3D Object Detection. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 17161–17170. [Google Scholar]
  35. Piergiovanni, A.J.; Casser, V.; Ryoo, M.S.; Angelova, A. 4D-Net for Learned Multi-Modal Alignment. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2021; pp. 15415–15425. [Google Scholar]
  36. Yin, T.; Zhou, X.; Krähenbühl, P. Multimodal Virtual Point 3D Detection. In Proceedings of the NeurIPS, Virtual, 6–14 December 2021; pp. 16494–16507. [Google Scholar]
  37. Belkin, M.; Niyogi, P.; Sindhwani, V. Manifold Regularization: A Geometric Framework for Learning from Labeled and Unlabeled Examples. J. Mach. Learn. Res. 2006, 7, 2399–2434. [Google Scholar]
  38. Xu, H.; Liu, F.; Zhou, Q.; Hao, J.; Cao, Z.; Feng, Z.; Ma, L. Semi-Supervised 3d Object Detection Via Adaptive Pseudo-Labeling. In Proceedings of the ICIP, Anchorage, AK, USA, 19–22 September 2021; pp. 3183–3187. [Google Scholar]
  39. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the NeurIPS, Long Beach, CA, USA, 4–9 December 2017; pp. 1195–1204. [Google Scholar]
  40. Wang, J.; Gang, H.; Ancha, S.; Chen, Y.; Held, D. Semi-supervised 3D Object Detection via Temporal Graph Neural Networks. In Proceedings of the 3DV, Virtual, 1–3 December 2021; pp. 413–422. [Google Scholar]
  41. Park, J.; Xu, C.; Zhou, Y.; Tomizuka, M.; Zhan, W. DetMatch: Two Teachers are Better Than One for Joint 2D and 3D Semi-Supervised Object Detection. arXiv 2022, arXiv:2203.09510. [Google Scholar] [CrossRef]
  42. Sautier, C.; Puy, G.; Gidaris, S.; Boulch, A.; Bursuc, A.; Marlet, R. Image-to-Lidar Self-Supervised Distillation for Autonomous Driving Data. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 9881–9891. [Google Scholar]
  43. Meng, Q.; Wang, W.; Zhou, T.; Shen, J.; Gool, L.V.; Dai, D. Weakly Supervised 3D Object Detection from Lidar Point Cloud. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; Volume 12358, pp. 515–531. [Google Scholar]
  44. Qin, Z.; Wang, J.; Lu, Y. Weakly Supervised 3D Object Detection from Point Clouds. In Proceedings of the ACM MM, Istanbul, Turkey, 8–11 June 2020; pp. 4144–4152. [Google Scholar]
  45. Peng, L.; Yan, S.; Wu, B.; Yang, Z.; He, X.; Cai, D. Weakly Supervised 3D Object Detection from Point Clouds. In Proceedings of the ICLR, Virtual, 25–29 April 2022; pp. 1–20. [Google Scholar]
  46. Xu, X.; Wang, Y.; Zheng, Y.; Rao, Y.; Zhou, J.; Lu, J. Back to Reality: Weakly-supervised 3D Object Detection with Shape-guided Label Enhancement. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 8428–8437. [Google Scholar]
  47. Zhang, J.; Liu, H.; Lu, J. A semi-supervised 3D object detection method for autonomous driving. Displays 2022, 71, 102117. [Google Scholar] [CrossRef]
  48. Meng, D.; Zhao, Q.; Jiang, L. A theoretical understanding of self-paced learning. Inf. Sci. 2017, 414, 319–328. [Google Scholar] [CrossRef]
  49. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the ICML, Lille, France, 7–9 July 2015; Volume 37, pp. 448–456. [Google Scholar]
  50. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. In Proceedings of the CVPR, Seattle, WA, USA, 13–19 June 2020; pp. 2443–2451. [Google Scholar]
  51. Zhang, D.; Meng, D.; Zhao, L.; Han, J. Bridging Saliency Detection to Weakly Supervised Object Detection Based on Self-Paced Curriculum Learning. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016; pp. 3538–3544. [Google Scholar]
  52. Chen, L.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  53. Xie, Q.; Dai, Z.; Hovy, E.H.; Luong, T.; Le, Q. Unsupervised Data Augmentation for Consistency Training. In Proceedings of the NeurIPS, Virtual, 6–12 December 2020; pp. 1–21. [Google Scholar]
  54. Bengio, Y.; Louradour, J.; Collobert, R.; Weston, J. Curriculum learning. In Proceedings of the ICML, Montreal, QC, Canada, 14–18 June 2009; pp. 41–48. [Google Scholar]
  55. Huang, T.; Liu, Z.; Chen, X.; Bai, X. EPNet: Enhancing Point Features with Image Semantics for 3D Object Detection. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; pp. 1–16. [Google Scholar]
  56. Liu, Z.; Zhao, X.; Huang, T.; Hu, R.; Zhou, Y.; Bai, X. TANet: Robust 3D Object Detection from Point Clouds with Triple Attention. In Proceedings of the AAAI, New York, NY, USA, 7–12 February 2020; pp. 1–9. [Google Scholar]
  57. Jiang, L.; Meng, D.; Yu, S.; Lan, Z.; Shan, S.; Hauptmann, A.G. Self-Paced Learning with Diversity. In Proceedings of the NeurIPS, Montreal, QC, Canada, 8–13 December 2014; pp. 2078–2086. [Google Scholar]
Figure 1. Difference between the previous SSOD-3D [10] and the proposed SPSL-3D. (a) Consistency loss [10]. (b) Loss in SPSL-3D. This emphasizes the quality of pseudo label l j p , adjusting the reliability weight of object in l j p and thus enhancing the generalization ability of the baseline detector. (c) Improvement of SPSL-3D in 3D object detection.
Figure 1. Difference between the previous SSOD-3D [10] and the proposed SPSL-3D. (a) Consistency loss [10]. (b) Loss in SPSL-3D. This emphasizes the quality of pseudo label l j p , adjusting the reliability weight of object in l j p and thus enhancing the generalization ability of the baseline detector. (c) Improvement of SPSL-3D in 3D object detection.
Remotesensing 15 00627 g001
Figure 2. Framework of the proposed PSPSL-3D. It can evaluate the reliability weight of a pseudo label from prior knowledge extracted from the LiDAR point cloud and RGB image.
Figure 2. Framework of the proposed PSPSL-3D. It can evaluate the reliability weight of a pseudo label from prior knowledge extracted from the LiDAR point cloud and RGB image.
Remotesensing 15 00627 g002
Figure 3. 3D object detection results of the fully supervised (FS) baseline method and the proposed PSPSL-3D trained in R anno = 64 % -labeled data in the KITTI validation dataset.
Figure 3. 3D object detection results of the fully supervised (FS) baseline method and the proposed PSPSL-3D trained in R anno = 64 % -labeled data in the KITTI validation dataset.
Remotesensing 15 00627 g003
Figure 4. 3D object detection results in the complex scenarios. Baseline, 3DIoUMatch, and PSPSL-3D are trained with R anno = 8 % -labeled data. The recall rate of PSPSL-3D is higher than in other methods.
Figure 4. 3D object detection results in the complex scenarios. Baseline, 3DIoUMatch, and PSPSL-3D are trained with R anno = 8 % -labeled data. The recall rate of PSPSL-3D is higher than in other methods.
Remotesensing 15 00627 g004
Figure 5. 3D object detection results of PSPSL-3D trained with R anno = 8 % -labeled data in simple traffic scenarios. Few false-positives and true-negatives were obtained.
Figure 5. 3D object detection results of PSPSL-3D trained with R anno = 8 % -labeled data in simple traffic scenarios. Few false-positives and true-negatives were obtained.
Remotesensing 15 00627 g005
Figure 6. Multi-sensor data collected by the LiDAR-camera system in the different outdoor scenarios. Occlusion, shadow, blur, and sparsity in the LiDAR point cloud and RGB image have impacts on the detection difficulty and annotation accuracy (only for pseudo label) of the 3D object, limiting the efficiency of SSOD-3D. Blur in RGB images causes ambiguity in the semantic image. The image and 3D point cloud show significant differences in the various scenes. 3D objects are detected using the proposed PSPSL-3D framework with R anno = 100 % .
Figure 6. Multi-sensor data collected by the LiDAR-camera system in the different outdoor scenarios. Occlusion, shadow, blur, and sparsity in the LiDAR point cloud and RGB image have impacts on the detection difficulty and annotation accuracy (only for pseudo label) of the 3D object, limiting the efficiency of SSOD-3D. Blur in RGB images causes ambiguity in the semantic image. The image and 3D point cloud show significant differences in the various scenes. 3D objects are detected using the proposed PSPSL-3D framework with R anno = 100 % .
Remotesensing 15 00627 g006
Table 1. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 4 % .
Table 1. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 4 % .
4% Labeled3D AP of Car3D AP of Pedestrian3D AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
Baseline [21]36.88%32.05%30.03%19.71%16.76%16.40%13.84%13.45%13.50%
SESS [10]39.07%33.81%32.20%15.02%12.35%12.37%13.98%12.80%12.75%
UDA [53]38.43%35.94%31.57%20.25%19.81%17.06%15.42%7.67%7.79%
3DIoUMatch [11]47.73%41.15%39.33%21.53%22.01%18.04%18.33%14.57%13.32%
UDA+CL [54]40.21%36.17%32.19%22.52%19.34%18.11%15.97%10.12%9.09%
SPSL-3D46.98%40.87%38.16%25.58%21.71%18.05%19.27%13.70%13.73%
PSPSL-3D52.24%42.52%40.30%24.83%21.22%21.17%20.57%15.15%13.79%
Gain from baseline +15.36% +10.47% +10.27% +5.12% +4.46% +4.77% +6.73% +1.70% +0.29%
Table 2. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 8 % .
Table 2. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 8 % .
8% Labeled3D AP of Car3D AP of Pedestrian3D AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
Baseline [21]74.64%63.57%57.25%37.79%31.96%30.53%50.73%34.51%30.18%
SESS [10]77.70%64.14%59.66%41.11%33.79%33.17%50.92%33.24%32.67%
UDA [53]76.02%64.88%61.47%37.90%32.91%31.92%49.87%35.20%30.07%
3DIoUMatch [11]76.12%65.16%58.19%43.28%36.56%32.17%48.82%34.85%34.12%
UDA+CL [54]76.14%65.15%62.59%38.08%33.47%31.99%50.35%35.19%30.25%
SPSL-3D76.40%65.79%63.70%39.54%35.01%30.78%52.15%32.23%31.18%
PSPSL-3D77.03%66.03%63.28%42.24%37.14%32.14%55.14%38.76%32.69%
Gain from baseline +2.39% +2.46% +6.03% +4.45% +5.18% +1.61% +4.41% +4.25% +2.51%
Table 3. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 16 % .
Table 3. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 16 % .
16% Labeled3D AP of Car3D AP of Pedestrian3D AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
Baseline [21]86.52%75.06%68.98%50.17%45.96%43.09%64.72%47.99%47.05%
SESS [10]85.74%75.25%68.52%51.16%47.36%45.36%63.22%48.03%46.54%
UDA [53]86.21%75.73%73.90%50.42%48.33%43.34%66.12%47.39%46.88%
3DIoUMatch [11]87.41%76.41%74.57%52.39%49.33%45.59%65.30%48.88%47.35%
UDA+CL [54]86.67%75.97%74.55%50.22%48.36%43.58%66.57%48.85%47.74%
SPSL-3D86.98%76.14%74.65%50.67%48.76%43.72%73.43%48.77%48.20%
PSPSL-3D87.18%76.49%74.66%52.41%49.64%44.29%74.71%49.68%49.28%
Gain from baseline +0.66% +1.43% +5.68% +2.24% +3.68% +1.20% +9.99% +1.69% +2.23%
Table 4. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 32 % .
Table 4. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 32 % .
32% Labeled3D AP of Car3D AP of Pedestrian3D AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
Baseline [21]87.20%77.14%75.44%59.13%52.55%50.49%73.85%54.21%52.26%
SESS [10]87.67%76.91%75.25%62.61%55.74%49.84%73.91%54.14%52.03%
UDA [53]87.99%77.12%75.24%60.10%54.88%49.95%74.25%57.92%56.82%
3DIoUMatch [11]87.86%77.26%75.28%63.17%56.42%50.03%74.19%55.20%52.72%
UDA+CL [54]88.13%77.46%75.63%60.49%55.59%50.02%77.84%58.42%56.91%
SPSL-3D88.30%77.81%75.58%62.84%56.97%54.59%77.90%58.13%57.02%
PSPSL-3D88.22%77.84%75.76%62.87%57.02%54.88%77.96%58.21%57.18%
Gain from baseline +1.02% +0.70% +0.32% +3.74% +4.47% +4.39% +4.11% +4.00% +4.92%
Table 5. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 64 % .
Table 5. 3D AP of current SSOD-3D methods in KITTI validation dataset at R anno = 64 % .
64% Labeled3D AP of Car3D AP of Pedestrian3D AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
Baseline [21]88.94%78.70%77.84%64.95%60.43%56.46%77.20%58.13%57.04%
SESS [10]88.97%78.83%77.55%65.21%60.37%56.15%77.38%58.44%57.65%
UDA [53]88.74%78.72%77.34%64.31%61.19%56.02%77.52%57.93%57.34%
3DIoUMatch [11]89.20%78.91%78.02%65.60%61.72%56.33%77.82%58.94%58.02%
UDA+CL [54]88.84%78.74%77.85%64.64%61.35%56.48%77.73%58.25%57.62%
SPSL-3D89.08%78.69%77.98%65.27%62.02%56.52%76.27%58.80%56.44%
PSPSL-3D89.35%78.99%78.13%65.75%62.64%56.72%78.29%60.71%57.78%
Gain from baseline +0.41% +0.29% +0.29% +0.80% +2.21% +0.26% +1.09% +2.58% +0.74%
Table 6. 3D AP of the proposed SSOD-3D at R anno = 64 % and current FSOD-3D methods in KITTI validation dataset.
Table 6. 3D AP of the proposed SSOD-3D at R anno = 64 % and current FSOD-3D methods in KITTI validation dataset.
3D AP of Car3D AP of Pedestrian3D AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
PointPillars [20]87.75%78.39%75.18%57.30%51.41%46.87%81.57%62.94%58.98%
Point-RCNN [4]88.26%77.73%76.67%65.62%58.57%51.48%82.76%62.83%59.62%
PV-RCNN [26]92.57%84.83%82.69%64.26%56.67%51.91%88.88%71.95%66.78%
3D-CVF [14]89.97%79.88%78.47%
SE-SSD [29]93.19%86.12%83.31%
EPNet [55]92.28%82.59%80.14%
TANet [56]88.21%77.85%75.62%70.80%63.45%58.22%85.98%64.95%60.40%
Voxel RCNN [21]89.17%79.25%78.33%66.43%62.59%57.14%83.02%63.87%57.62%
SPSL-3D (Our, 64%)89.08%78.69%77.98%65.27%62.02%56.52%76.27%58.80%56.44%
PSPSL-3D (Our, 64%)89.35%78.99%78.13%65.75%62.64%56.72%78.29%60.71%57.78%
PSPSL-3D (Our, 100%)89.27%79.32%78.58%68.11%65.51%59.58%83.29%64.30%58.13%
Table 7. BEV AP of the proposed SSOD-3D at R anno = 64 % and baseline FSOD-3D methods in KITTI validation dataset.
Table 7. BEV AP of the proposed SSOD-3D at R anno = 64 % and baseline FSOD-3D methods in KITTI validation dataset.
BEV AP of CarBEV AP of PedestrianBEV AP of Cyclist
MethodsEasyModerateHardEasyModerateHardEasyModerateHard
Voxel RCNN [21]90.17%88.17%87.41%67.86%63.28%59.97%82.87%64.17%57.85%
SPSL-3D (Our, 64%)90.12%87.84%87.04%69.23%63.34%58.14%76.89%64.33%56.84%
PSPSL-3D (Our, 64%)90.13%88.03%87.12%70.58%64.73%58.92%77.22%64.74%57.89%
PSPSL-3D (Our, 100%)90.40%88.30%87.86%73.97%67.83%61.45%83.38%64.51%58.30%
Table 8. 3D recall rate with an IoU threshold of 0.7 in the KITTI validation dataset.
Table 8. 3D recall rate with an IoU threshold of 0.7 in the KITTI validation dataset.
Method/ R anno 4%8%16%32%64%
SPSL-3D43.47%57.46%65.65%67.06%71.30%
PSPSL-3D44.25%58.03%66.13%67.37%71.63%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

An, P.; Liang, J.; Hong, X.; Quan, S.; Ma, T.; Chen, Y.; Wang, L.; Ma, J. Leveraging Self-Paced Semi-Supervised Learning with Prior Knowledge for 3D Object Detection on a LiDAR-Camera System. Remote Sens. 2023, 15, 627. https://doi.org/10.3390/rs15030627

AMA Style

An P, Liang J, Hong X, Quan S, Ma T, Chen Y, Wang L, Ma J. Leveraging Self-Paced Semi-Supervised Learning with Prior Knowledge for 3D Object Detection on a LiDAR-Camera System. Remote Sensing. 2023; 15(3):627. https://doi.org/10.3390/rs15030627

Chicago/Turabian Style

An, Pei, Junxiong Liang, Xing Hong, Siwen Quan, Tao Ma, Yanfei Chen, Liheng Wang, and Jie Ma. 2023. "Leveraging Self-Paced Semi-Supervised Learning with Prior Knowledge for 3D Object Detection on a LiDAR-Camera System" Remote Sensing 15, no. 3: 627. https://doi.org/10.3390/rs15030627

APA Style

An, P., Liang, J., Hong, X., Quan, S., Ma, T., Chen, Y., Wang, L., & Ma, J. (2023). Leveraging Self-Paced Semi-Supervised Learning with Prior Knowledge for 3D Object Detection on a LiDAR-Camera System. Remote Sensing, 15(3), 627. https://doi.org/10.3390/rs15030627

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop