Next Article in Journal
Stability Analysis of the Fluorescent Tracer 1-Methylnaphthalene for IC Engine Applications by Supercontinuum Laser Absorption Spectroscopy
Previous Article in Journal
Towards Urban Archaeo-Geophysics in Peru. The Case Study of Plaza de Armas in Cusco
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Large-Scale Place Recognition Based on Camera-LiDAR Fused Descriptor

1
School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
2
School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
3
Department of Mathematics, School of Science, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(10), 2870; https://doi.org/10.3390/s20102870
Submission received: 9 April 2020 / Revised: 8 May 2020 / Accepted: 10 May 2020 / Published: 19 May 2020
(This article belongs to the Section Physical Sensors)

Abstract

:
In the field of autonomous driving, carriers are equipped with a variety of sensors, including cameras and LiDARs. However, the camera suffers from problems of illumination and occlusion, and the LiDAR encounters motion distortion, degenerate environment and limited ranging distance. Therefore, fusing the information from these two sensors deserves to be explored. In this paper, we propose a fusion network which robustly captures both the image and point cloud descriptors to solve the place recognition problem. Our contribution can be summarized as: (1) applying the trimmed strategy in the point cloud global feature aggregation to improve the recognition performance, (2) building a compact fusion framework which captures both the robust representation of the image and 3D point cloud, and (3) learning a proper metric to describe the similarity of our fused global feature. The experiments on KITTI and KAIST datasets show that the proposed fused descriptor is more robust and discriminative than the single sensor descriptor.

1. Introduction

Place recognition has received a significant amount of attention in various fields including computer vision [1,2,3,4,5,6], autonomous driving systems [7,8,9,10] and augmented reality [11]. In these tasks, place recognition addresses a question of “where am I in a route”. This question is related to how to recognize places based on their appearance and surrounding structure. Therefore, this technology plays an important role in the Simultaneous Localization And Mapping (SLAM) [12] for the robotic and autonomous driving system.
Given a place recognition problem, a common and direct method is to find the most matching candidate place(s) of the current place, which is stored in a database of the environment map [13,14]. Its fundamental scientific question is to determine the appropriate representation (of a place) to distinguish similar and dissimilar places. Traditionally, features for visual information, such as Oriented FAST and Rotated Brief (ORB) [15], Scale-Invariant Feature Transform (SIFT) [16] and Speeded Up Robust Features (SURF) [17], have been proposed to describe the strong corner features in each image. Moreover, some methods like the Bag-of-Words (BoW) [18,19], Vector of Locally Aggregated Descriptors (VLAD) [20] and Fisher vector [21] have been used to aggregate these local features into a single vector representation for the entire image representing a certain place. Existing well-performing image retrieval [22,23,24,25] and place recognition [6,26] systems mainly utilize traditional hand-crafted features or processing models, which might fail to build feature representation appropriately, since hand-crafted feature representation perhaps only extract limited features.
Recently, with massive data and the development of powerful hardware, data-driven methods have emerged. Deep neural network (DNN) is the most representative method, which has a good capability to extract features, showing strong advantages in the field of computer vision and pattern recognition. The success of deep learning has been particularly noticeable on 2D images where convolution kernels can be easily applied to the regular 2D lattice grid structure of the image. Besides, these kernels can also aggregate the information of surrounding pixels to the following layer. Meanwhile, convolutional neural network (CNN) ResNet [27] and VGGNet [28] have been widely used in the image retrieval and detection field. For visual information, all kinds of networks are designed to extract visual descriptor or conduct image retrieval directly [29,30,31,32,33,34]. Reference [29] is the first work to introduce a CNN-based place recognition system, and Reference [30] provides a thorough investigation of the utility and viewpoint-invariant properties of deep learned features for the place recognition. After them, more and more CNN-based place recognition methods [31,32,33] have been proposed. NetVLAD [34], a generalized VLAD layer, combines the advantages of the classical VLAD and data-driven methods. Based on the NetVLAD, the GhostVLAD [35] proposes a variant that added ghost clusters. The network automatically learns how to distinguish the low quality images and then gives them lower weights, which helps to generate a better template. Generally speaking, these CNN-based recognition approaches have been proposed rapidly and most of them perform better than those traditional approaches.
Intuitively, the sun exposure conditions and things in the scene may continuously change. Even the images acquired in the same place often look different. It is well known that these visual-based methods will encounter problems of illumination, occlusion and the change of viewpoint. To avoid these disadvantages of visual sensors, some 3D point clouds based place recognition methods, captured by LiDAR (Light Detection And Ranging), are introduced [36]. Unfortunately, it is more challenging for convolution kernels to work on orderless 3D point clouds. Compared with the visual methods, there are indeed less methods to deal with the point cloud, no matter classical or data-driven methods, since the input 3D information is orderless and sparse in R 3 [37].
For a 3D point cloud, the 3D shape context [38] descriptor is used to obtain local 3D features. Moreover, Point Feature Histogram [39] and Signature Histogram of Orientation (SHOT) [40] generate histograms based on geometric attribution. However, these methods consume much time and computing resources. By contrast, Fast Point Feature Histogram (FPFH) [41,42], 3D SIFT [43] can extract local point cloud features with low cost in computation, but they do not perform well at sparse space locations. Some global descriptor extraction methods such as Ensemble of Shape Functions (ESF) [44], Normal Aligned Radial Feature (NARF) [45] and Viewpoint Feature Histogram (VFH) [46] are able to ignore sparse distribution of point cloud at some local locations.
Several deep networks attempted to cope with this challenge by transforming the input point clouds into regular 3D volumetric representations. For instance, the 3D ShapeNet [47], volumetric CNN [48] and Vote3Deep [49] resolve 3D object detection, classification and recognition, respectively. The recent PointNet [50], PointNet++ [51] GVCNN [52] and PVNet [53] made it possible to directly input the original point cloud into network. However, they were designed to handle small object classifications and indoor scene segmentations. In [36], the PointNetVLAD, which is the combination of PointNet and NetVLAD, utilizes the unique VLAD layer to aggregate LiDAR features extracted by the PointNet [50], which makes the feature extraction more effective. However, point clouds may encounter limited ranging distance, and also suffer from motion distortion and some disturbances.
Because of the complexity and variability of outdoor large-scale places, it may not be effective for place recognition by only using the image or point feature. In fact, only using a single sensor will have limitations in a specific scenario. For example, in Figure 1, A 1 and A 2 represent images for the same place from two different views, A 3 represents the point cloud corresponding to A 1 , while A 4 represents that of A 2 . These two images captured in the different viewpoints only have a few matching pairs. As a result, it is hard to find out the relationship between these two images only using the visual information. Meanwhile, their point clouds are almost the same, meaning that the LiDAR information is helpful for obtaining a proper conclusion. As depicted in Figure 2, the LiDAR features B 3 and B 4 are different (marked by red triangular regions), but their visual data B 1 and B 2 are sufficient to prove these two frames are captured in the same place. Actually, these two cases show that there exists complementarity between the image and point cloud. Therefore, we hope to fuse the information of image and point cloud, since both of them possess their own peculiarities. Some pioneers have tried to fuse these two kinds of input by establishing the relationship between the rectilinear feature in image and the planar intersection in point cloud [54] or combining them in the final decision [55]. Unlike the above methods, our method uses networks to learn the features fusing two kinds of input, which has a more tight coupling.
In this paper we propose a camera-LiDAR sensors fusion method to extract fused global descriptors for place recognition via a deep neural network. The contributions of our work are summarized as—(1) applying the trimmed strategy for non-informative clusters to reduce the impact of some unrepresentative information in a 3D point cloud; (2) proposing a fusion neural network concatenating both the visual descriptor and 3D spatial global descriptor; and (3) using metric learning ideology to reduce the distance of fused descriptors among similar places and extend it among dissimilar places. With a metric learning loss, the whole network is optimized to get an appropriate mapping from the 3D Euclidean space to the descriptor space, in which the descriptor can distinguish the different places easier.
This paper is organized as follows. Section 2 introduces some methods directly related to the approach in this paper. In Section 3 we introduce the framework of our neural network based method that can generate global fusion descriptor based on visual and 3D point information. Section 4 presents the datasets and experimental results. The paper is concluded in Section 5.

2. Related Works

2.1. Visual Feature Extractor

ResNet [27] is the basis of our proposed approach for extracting image features. It performs well in image recognition, and solves the degeneration problem when training deep neural networks. This method presents a residual learning framework to ease the training of networks that are substantially deeper than those used previously. A principal module named residual block is shown in Figure 3.
This method shows that deep residual learning can solve the degeneration problem. For a stacked layers structure, X is the input of the first layer of the residual block. Formally, denoting the desired output as H ( X ) , the residual block learns another nonlinear mapping F ( X ) = H ( X ) X to get the residual value. It not only solves the deep learning degeneration problem, but also apparently reduces the computational complexity. With the accumulation of residual block, the ResNet is formed naturally to obtain the deeper features of images.

2.2. Global Descriptor of 3D Point Cloud via Netvlad

VLAD [20], as mentioned before, is a classical hand-crafted feature extractor method which can aggregate these local features into a single vector representation for the entire image representing a certain place. Fortunately NetVLAD [34] can combine the advantages of VLAD and data-driven methods. Based on PointNet [50] and NetVLAD, PointNetVLAD [36] focuses on the place recognition in scene retrieval, and outperforms the maxpool layer features extracted from the original PointNet. Its framework is shown in Figure 4.
The main idea of PointNetVLAD is to view the point features of every point cloud learned from the PointNet block as the input of NetVLAD block, and then utilize the input N D-dimensional feature points to aggregate a single ( K · D ) × 1 global descriptor vector as
V k = i = 1 N e w k · p i + b k k e w k · p i + b k ( p i c k ) ,
where the cluster centers { c k } and parameters { w k } , { b k } are trainable from the network. One branch of the NetVLAD block computes the difference between local features and cluster centers, while the softmax part determines the weights of the points to each cluster centers.
The advantage of PointNetVLAD is the acquirement of the global descriptor for the whole point cloud, which has less redundant feature comparing with those local feature extraction operators. However, the global descriptor may be disturbed by some outliers and isolated points. We will avoid these problems via our modified local-to-global fusion descriptor retrieval.

3. Proposed Method

We list here some notations in our algorithm. For each scene, the source data ( I , P ^ ) include the image information I and the corresponding 3D point cloud P ^ . Note that P ^ , acquired from LiDAR, may have different sizes. For feature extraction facility in deep neural network, we apply a down-sampling preprocess to 3D source point cloud P ^ , to get the fixed size point cloud P = { p l | p l R 3 } l = 1 N before feature extraction, that is, P ^ s a m p l i n g P = { p l | p l R 3 } l = 1 N .
We then use the neural network to generate a robust and compact representation for a specific place ( I , P ) . For acquiring this representation, we need to learn a proper mapping F ( · ) from the input data space S = { ( I , P ) } to a new space for facilitating the place retrieval. The whole framework is shown in Figure 5. A good mapping should consist of two modules—the efficient feature extraction operator F (the green, yellow and blue blocks in Figure 5) and the best metric M to evaluate of similarity of feature descriptors (the red block in Figure 5). This process can be formulated with
S F ( · ) X M ( · ) X ^ S x = F ( S ) x ^ = M F ( S )
for a certain place S = ( I , P ) .
Firstly, we should build a feature extraction operator F to represent the place accurately and robustly. Note that each place has the image and the corresponding 3D point cloud. So, the feature extraction operator F has two map branches F I and F P to extract visual feature x I and 3D point cloud feature x P respectively, that is, F = F I F P . In this part, the input data of camera and LiDAR information, according to their data patterns, are fed into image feature extraction branch (the blue block) and 3D Spatial feature aggregation branch (the green and yellow blocks), and get the respective descriptors x I and x P . Specifically, we introduce a trimmed clustering method to extract the global descriptor x P for 3D point cloud via ignoring the non-informative point feature clusters.
To avoid the shortcoming of both visual and LiDAR information process, a concatenating and fully connection operation are applied to build the global fusion feature, that is, x = x I x P . Furthermore, the triplet constraints of the sampling are used to learn the best metric M for finding a good viewpoint to describe the similarity of sampling. Finally, the model parameters are optimized to get an appropriate mapping relationship between the input data and the fused global descriptor x ^ via an end-to-end training. We will introduce these modules in the following section.

3.1. Spatial Feature Aggregation with Trimmed Strategy

In this part, we apply a local-to-global methodology to aggregate the point cloud features. This procedure begins with a local feature extraction (the green block), then follows a global feature extraction (the yellow block) for point cloud.

3.1.1. Local Feature Extraction for Point Cloud

For the input point cloud P, the Multi-Layer Perception (MLP) and feature transform are applied to extract the local spatial feature information for point cloud by mapping each 3D dimensional point p l R 3 into a higher dimensional space p l R D , for D 3 . That is,
P = { p l | p l R 3 } l = 1 N P = { p l | p l R D } l = 1 N .
This part of the network (the green block) extracts some local rotation invariant spatial features for each single point p l R 3 from the input point cloud P. Thus, the original 3D point is mapped into a higher dimensional local feature vector. This redundant representation makes the points maintain much more their own information, which is widely used as the feature extraction module for point cloud processing, such as PointNet [50]. Some redundant information will be omitted to form a compact global descriptor in the following process.

3.1.2. Global Feature Extraction with Trimmed Clustering

Comparing with PointNetVLAD [36], we introduce a novel trimmed VLAD block for point clouds in this module (the yellow block in Figure 5). Originally, the K cluster centers K = { c k | c k R D } k = 1 K and their corresponding weights W k ( · ) = W k ( w k , b k ) ( · ) , k = 1 , , K are learned to represent K special and prominent parts of the current 3D space, so that the outputs related to these cluster centers are distinctive. However, this procedure is sensitive to the complex environments captured by a 3D LiDAR. For avoiding this redundant information and environment disturbance robustly, we will introduce a trimmed strategy via ignoring the non-informative clustering of 3D point cloud, that is, the trimmed VLAD block. The details of this new block are shown in Figure 6.
Notice that not all the points and clusters are useful for forming a compact and robust representation of point cloud. For instance, some outliers or isolated points do not correctly reflect the features of point cloud P , so they are not helpful to the representation of P . Moreover, points belonging to this kind of dynamic clusters in a scene may exert a negative influence on the recognition precision. Under this consideration, we assume that there exist more cluster centers G = { c k | c k R D } k = K + 1 G and its corresponding weights W k ( · ) = W k ( w k , b k ) ( · ) , k = K + 1 , , K + G to non-informative clusters for separating their influences.
For the given proper cluster size K and the non-informative cluster size G, the trimmed weights of an input descriptors P = { p l | p l R D } l = 1 N can be computed via
W k ( p l ) = e w k · p l + b k m K + G e w m · p l + b m , for k = 1 , , K ,
where { w k } , { b k } , k = 1 , , K are the weight parameters for proper clusters, and { w k } , { b k } , k = K + 1 , , G are the weight parameters for non-informative clusters. All of them are trainable via the trimmed weights sub-block (the red block in Figure 5). Note that the weight parameters for non-informative clusters are included into the summation in the denominator of Equation (3), which makes these non-informative clusters contribute to the soft assignments in the same manner as the original clusters K . Moreover, only the trimmed weights for the meaningful clusters K are computed, since the weights of the non-informative clusters will be ignored in the aggregation process.
For an input local descriptor P = { p l | p l R D } l = 1 N , the output is a K × D -dimension matrix V = [ V 1 , , V K ] after the partial aggregation, in which V k can be written as
V k = l = 1 N W k ( p l ) ( p l c k ) , for k = 1 , , K ,
where c k is the proper cluster center, W k ( p l ) is the trimmed weight of the local descriptor p l for the meaningful cluster K , and p l c k represents the difference between the local descriptor of sample and its corresponding meaningful cluster center c k . The cluster centers and the differences (partial residual ) are computed in a single branch network. Then we assign the trimmed weight to partial residuals of the meaningful clusters in partial aggregation process for getting a global descriptor V. This trimmed-weight based aggregation process is robust for complex LiDAR data, since the non-informative clusters are considered both in weights computation and the aggregation process. Moreover, we choose to do intra-normalization of each vector first, and then concatenate them, followed by the L 2 normalization. Thus, the K · D -dimensional feature vector is generated. After the trimmed VLAD block, we use the fully connected layer to select useful spacial features for getting a Q-dimension compact global descriptor.

3.2. Image Feature Extraction

As depicted in Section 1, the visual feature extraction is very important for place recognition, since vision is the main source of obtaining information for mankind. Here we choose the ResNet50 [27] for the following two reasons. Firstly, the images collected by a camera in outdoor large-scale scenes belong to natural images. The ResNet performs well at recognizing natural images. Secondly, as a deep network, the ResNet50 has a good capability for extracting the deep features of images, and it uses a unique building block to firstly reduce the dimension of the features and then to increase the dimension. By continuously utilizing such blocks, the computation consumption is acceptable. Compared with ResNet101, the ResNet50 consumes less computation and already meets our demand, since our motivation is to verify the efficiency of our fused network. That means the additional LiDAR sensor information can improve the place recognition. The original output of the ResNet is an 1000-dimension vector, here we resize it in Q-dimension. The details about the image feature extractor via ResNet can be found in [27].
Here we use the ResNet as our image feature extractor rather than that followed by a NetVLAD layer. The NetVLAD, inspired by the VLAD, is a clustering algorithm utilizing the soft assignment in the neural network instead of the hard assignment. Images contain a large amount of information, and the ResNet has the capability to obtain these sufficient and deep features by its residual blocks. These appearance based features have a mutual effect with the structural point cloud features. However, arranging these features through an additional NetVLAD layer may lose a part of information. As a result, we only use the ResNet to extract image features, followed by the L 2 normalization to make the image and the point cloud components in equal weights. The comparison of fusing these two kinds of image features with the point cloud features can be found in Section 4.3.4.

3.3. Metric Learning for Fused Global Descriptors

As mentioned, a good mapping (or descriptor) should consist of the efficient feature extraction and the best metric to evaluate sample similarity in the fused global feature space. Different from PointNetVLAD [36], our network needs to learn how to pick the appropriate viewpoint for the fused global descriptor of image and point cloud in the fusion process (the red block in Figure 5).
To deal with the fused information, we firstly concatenate the two kinds of features, f P and f I , into a long vector roughly. Then we use the fully connected layer to select useful parts in a feature vector that combines a mixture of image and point cloud features. It is worth mentioning that the global descriptor often needs to be normalized, i.e., x = 1 . So, L 2 layer is conducted to balance each feature. This operation is able to eliminate the negative influence made by dimensionality. Thus, a coarse fused global descriptor x R Q is acquired.
After we get a set of descriptors x, the supervised label can be introduced to evaluating the similarity and dissimilarity relationship between them in the fusion global feature space. Metric is often used to evaluate the similarity of samples [56]. Meanwhile, finding the best metric is equal to finding the best mapping to a new space, where we can get the proper viewpoint to describe the data. So, metric learning is utilized to optimize the parameters in the network, i.e., it learns an appropriate mapping from the original fused descriptor x i to the further updated descriptor x i ^ in this module. The place recognition here maps the original input point cloud and image into a new data space where the data have the better description and discrimination.
For well fitting the data and quantifying the similarity of the fused global descriptors, we introduce a triplet constraint in an intuitive way, that is,
T = { ( x i , x j , x k ) : δ i j < δ i k } ,
where x i is similar to x j , and dissimilar to x k ; δ i j means the distance between x i and x j , and δ i k is the distance between x i and x k . It means that the distance of samples in different places should be as large as possible, while as small as possible in the same place. Note that the supervised label information is needed in this triplet constraint. The traditional hinge loss function [57,58] is often used for the triplet constraint δ i j + α < δ i k as
L H = T [ α + δ i j δ i k ] + ,
where [ m ] + = m if m 0 and 0 otherwise, and α is the margin value to balance the inner class and intra class. So the minimization of the loss function L H makes x i closer to x j and maintains a margin between x i and x k .
Furthermore, deep neural network based metric learning is utilized to learn an appropriate mapping M from the original fused descriptor x i to the further updated descriptor x i ^ in this module. The parameters of the mapping (model) are optimized via the deep network training. Similar to traditional triplet constraint based metric learning method, the descriptors in the same place should get closer and the ones in the different places get farther. Here, we want further augment this constraint, since the outdoor large-scale place recognition needs a good discrimination for the descriptors. Thus, we apply the lazy triplet constraint loss [36] to learn a discriminative mapping M .
Taking each current frame as the anchor frame S a n c in the training dataset, its corresponding fused global descriptor is x a n c , while x p o s represents the descriptor of place that is similar to the anchor place, x n e g represents a dissimilar descriptor. Then, we use the descriptors of anchor frame, that of the corresponding positive and negative frames to construct a set of tuples T = { x a n c , { x p o s } , { x n e g } } .
Mathematically, the lazy triplet loss is calculated as:
L l a z y t r i p ( T ) = min ( [ α + sup δ p o s inf δ n e g ] + ) ,
where δ p o s is the Euclidean distance between the fused global descriptor of anchor frame and that of the one in { x p o s } , that is, δ p o s = { d ( x a n c , x j ) , x j { x p o s } } ; the definition of δ n e g is in the same way, that is, δ n e g = { d ( x a n c , x k ) , x k { x n e g } } ; sup δ p o s means the supremum in { δ p o s } , and inf δ n e g means the infimum in { δ n e g } . The sup and inf here reflect the concept of lazy in L l a z y t r i p . With such a loss, only the supremum in { δ p o s } and the infimum in { δ n e g } are involved in updating the parameters in model rather than the mean value of all the members.
The objective of L l a z y t r i p is to minimize the supremum in { δ p o s } and maximize the infimum in { δ n e g } , which is equivalent to reducing the distance between the global descriptor of x a n c and { x p o s } , and extending the distance between the global descriptor of x a n c and { x n e g } . With the closest negative distance and the farthest positive distance, the parameters can be updated efficiently. After using the metric learning to update the parameters in the network, x ^ gets more discriminative.
Finally, having a trained model, we can build the desired descriptors of all the places in the dataset. Once searching an unknown place, we can map the source data S to the space X and then query its descriptor x in the database to find the candidates in a relatively small range.

4. Experiments and Results

To make a fair comparison with different methods, we compare our approach with the existing open source algorithm NetVLAD [34] and PointNetVLAD [36] on the same device. The device is equipped with a Tesla P40 GPU with 24GB memory, and is implemented with Ubuntu16.04 operating system carried with TensorFlow. For the super parameters of our network, we set the number of original cluster size and non-informative cluster size as 64 and 4 respectively, and margin α = 0.8 . The dimension of output descriptor in every part is uniformly set as 256. The training process of our method takes 40 h.

4.1. Datasets and Pre-Processing

We choose two large-scale outdoor datasets KITTI (Karlsruhe Institute of Technology and Toyota Technological Institute) [59] and KAIST (Korea Advanced Institute of Science and Technology) [60] for experiments. The scenarios of KITTI dataset are diverse, capturing real-world traffic situations and ranging from freeways over rural areas to urban scenes with many static and dynamic objects. KAIST dataset is composed of complex urban scenes. These two datasets satisfy our demand that the frequencies of camera and LiDAR information are the same or multipled, which is able to make the pair of image and point cloud in the same place. KITTI dataset supplies 11 scenes containing accurate odometry ground truth information. These 11 scenes are written as KITTI 00, ⋯, KITTI 10, and then we utilize these scenes for our experiments. Each scene in the KAIST dataset has accurate GPS information, so we focus on 5 scenes, where most of them have loops for place recognition evaluation.

4.1.1. Kitti Dataset

KITTI Odometry split consists of 22 scenes that the images and point clouds strictly match. Half of the sequences have the ground truth poses of each frame. We only use these 11 scenes with the closed loops in our experiment.
As mentioned in Section 3, the quantity of points in each point cloud is not the same, so we need to resample the points in a fixed number. Figure 7 shows the pre-processing of the point cloud in the KITTI dataset. After checking the sample point cloud of HDL-64 LiDAR, we consider that the ground points are redundant to form a discriminative descriptor utilizing the structure information of each frame. Besides, removing these ground points can reduce the consumption of the memory. As a result, the ground points are removed using the method in [61]. Then the point cloud is downsampled into a certain quantity. Here we use the downsample API in the PCL Library, and the leafsize which represents the size of grid cells in the point cloud is set to ( 0.3 , 0.3 , 0.1 ) . Finally, we randomly pick N = 6000 points as our input of LiDAR to give a slight disturbance to the point cloud. For the image part, considering the size of raw input, we resize them into 180 × 600 on all the sequences. We choose such a large dimension of input to keep more information of raw data, while the network can also be trainable on a single GPU.

4.1.2. Kaist Dataset

The KAIST dataset mainly focuses on the urban scene. Images, point clouds and GPS information are acquired in different frequencies in the dataset, but the timestamps are available for all the sensors. We firstly generate data tuples depending on the timestamp of each frame among the different sensors. The difference on time of data tuple is under 0.1 s. After getting sequential frames like the data format in KITTI dataset, KAIST dataset can be operated in similar way. Notice that the images in KAIST dataset are in 8-bit Bayer format, so we demosaic them to recover the raw data into RGB images. Besides, the point cloud in KAIST consists of two VLP-16 LiDARs. We use the extrinsics between sensors and carrier to transform these two point clouds in carrier coordinate system. The merging process is shown in Figure 8.

4.1.3. Our Campus Data

The hardware platform is constructed from the HESAI Pandora and the Trimble GPS in Figure 9. The Pandora contains a set of sensors, which are 4 greyscale cameras, a color camera and a 3D LiDAR. While the GPS consists of a receiver (SPS461) and 2 antennas (GA530). Here we use the GPS data as ground truth. The LiDAR in Pandora has 40 scanners so that we can get a dense point cloud of surroundings. As a result, our campus data can use the same pre-processing in KITTI dataset. The related experiments can be found in Section 4.3.6.

4.1.4. Triplet Tuple

We use the ground truth trajectory to generate training tuples. Every tuple is composed of two parts, that is, image and point cloud, respectively. Both of them contain the keys S a n c , { S p o s } and { S n e g } . For each anchor frame S a n c , { S p o s } is generated in terms of position distance and the time. The distance between the candidate positive frame S p o s { S p o s } and S a n c is under 5 m and the time difference is under 10 s. Similarly, the distance between candidate negative frame S n e g and S a n c is over 50 m. Through random selection from these candidate frames, the input training tuple can be finally generated. This operation also increases the robustness of the trained models. Additionally, we take the frames passing through the same place as the evaluation frames.

4.2. Place Recognition Results

Here we evaluate our approach to the KITTI and KAIST datasets. To the best of our knowledge, few methods combine both the image and the point cloud on the place recognition task, so we compare the proposed approach with some methods on a single sensor mentioned before. The NetVLAD is an image-based method which utilizes the concept of VLAD to develop a unique layer which aggregates the image features extracted by a CNN architecture. Here we use the ResNet50 to extract visual information. Besides, the output of ResNet is a vector which can also be seen as a compact representation of an image. Therefore, the ResNet is also be included. The PointNetVLAD is a point-based approach combining the concept of PointNet and NetVLAD, which applies an image-based method to the point cloud. The PointNetVLAD with the trimmed strategy will also be included. For a fair comparison, we use the same training datasets to train the models and then do the retrieval.
The results are given in Figure 10 and Table 1. As mentioned in Section 1, the LiDAR and camera data have mutual effect on each other. Intuitively, when both the image and the point cloud parts have a good ability to distinguish the similar and dissimilar places, fusing these two features will have a better performance. Because of the lower recall rate on the LiDAR-only descriptor, in the curves of KITTI 05 and KAIST 30, the performance of our fusion feature is worse than the image-only method in the certain segment. But combining the visual information with point cloud information reduces the bad discrimination of the LiDAR-only descriptor. It can also be observed in Table 1. Although the frames in each scene are different, the results of the top 1% of candidates, as our benchmark, show the average performance of all the approaches.
In the KAIST dataset, because the point cloud only contains a part of the structure of the whole scene, the results of LiDAR are much lower than that of the image. Generally speaking, our fusion approach is more accurate and robust.

4.3. Analysis and Discussion

In this subsection, we discuss the results of our network in detail. We determine the parameters, show the impact of the non-informative clusters, compare the different image features extractor mentioned before, further demonstrate the advantages of our fusion descriptor and analyze the utility of our method.

4.3.1. Number of Points

In Table 2, we firstly discuss the quantity of the points in each point cloud P. The more points retained after downsampling from the original point cloud, a better the performance of model is. However, the more points used in the network, the more computing resources are occupied. As a compromise, we set a fixed number of points N = 6000 in our method.

4.3.2. Number of Cluster Centers

Table 3 shows the recall of top 1 candidate in different number of cluster centers K . Intuitively, this parameter is due to the points number N and the complexity of the scene. The more points or the more complex the scenes are, the more cluster centers is required. In our experiments, K = 64 achieves a better performance.

4.3.3. Effect of the Non-Informative Clusters

Here we use the recall rate of top N-number candidates in the database to examine whether the non-informative clusters can improve the PointNetVLAD. Figure 10 and Table 1 show the comparison of PointNetVLAD and our trimmed version. In conclusion, our trimmed strategy performs better than the original PointNetVLAD. Meanwhile, we rank the weights of the points to all the clusters. The cluster with the biggest weight is defined as the owner of the point. Then we marked out the points belonging to the non-informative clusters with the different colors in Figure 11. It is shown that the non-informative clusters are not so meaningful. Specifically, these points of abandoned clusters in our trimmed approach contain the outliers or isolated points. Therefore, the concept of trimmed strategy makes network learn how to ignore these non-informative clusters, which helps to improve the quality of features computed by the trimmed VLAD block.

4.3.4. Different Image Features

We fuse two different image features with the point cloud part in our fusion network. One is the features directly extracted from the ResNet50, and the other is the features extracted from the ResNet50 and followed by a NetVLAD layer. The ResNet is a deep network with stacked residual blocks. Its final layer is a fully connected layer, which has already downsampled the features. To some extent, it is also a concept of local aggregation. Therefore, in our fusion network, fusing the features from ResNet without the NetVLAD layer with the point cloud features is reasonable (Table 4).
While the point features are orderless and there are only a few methods designed on the point cloud, using the VLAD here can efficiently aggregate the point features to a global point cloud feature. In Table 4, we use the KITTI dataset to further verify our observation. Here we show the result of the top 1% of candidates and the top 1 candidate to compare the effect of fusing these two image features with the spatial feature. In our fusion architecture, ResNet without NetVLAD layer has a better performance in overall cases.

4.3.5. Effect of Learned Descriptors

Figure 12 shows the top two candidates of two test frames using our fusion method. Our method can still find the corresponding frames in database, even there are some moving objects in datasets. Besides the top candidates of test frame in database are usually adjacent frames, our fused descriptor has a good performance on retrieval.
Here we show the visualization of the features in several scenes. The numbers of frames in each scene are in different scales. In Figure 13, we use the t-SNE to represent three kinds of global descriptors—image, point cloud and fusion, which are generated from the ResNet, PointNetVLAD and our fusion way respectively. In addition, we also draw the legend with gradient colors in each scene. In the t-SNE plot, the more points in different colors are mixed, the worse the discrimination of the descriptors is. In other words, the adjacent frames are similar intuitively, so the corresponding points in the plots should be closer. From this perspective, our method has fewer mixed areas in overall cases. In Figure 14, we select short segments to explain the t-SNE plot in detail. Note that the marked area on the trajectory corresponds to the frames represented by the points in the plots. In our fusion method, there are less isolated points than those in other methods. It means our descriptor has a better performance in the feature space.

4.3.6. Usability

Finally, we further study the application of our method. In Table 5, we list the recall and the average processing time between the VLAD and our method. Here we use the ORB feature to detect the corners and aggregate the hand-crafted features in VLAD vector. Our method performs better in overall scenes. Note that the VLAD method needs to calculate the local descriptors in the image before aggregation, so the processing time here also includes the time to extract features and corresponding local descriptors. While our method can directly get a global descriptor from the model. Less time is needed in our approach to form a descriptor. Compared with the VLAD algorithm, which has been proven to be practical, our method has the potential to be applied in the field of autonomous driving or robotics systems.
Besides, we use the heat map to visualize the minimal required candidates of finding the corresponding frame in database. In Figure 15, those tested frames are marked on the trajectory by specific colors. The color represents the least required candidates to find the corresponding frame. In our method, most of the marks on the map are dark, and only a few points near the intersection are bright. Compared with the hand-crafted method, our method is also capable of the place recognition task.
Here we also test the proposed approach in our campus. In Figure 16, we show the comparison against the state-of-the-art methods in our campus, the results on the sensors rather than the public datasets show our fusion method is competitive.

5. Conclusions

In this paper, we have proposed a novel network for place recognition by fusing the information from the image and the point cloud. We apply the trimmed strategy in the point cloud global feature aggregation to avoid the perturbation of complex environments. Moreover, we fuse this compact global descriptor of the point cloud with that of the corresponding image to get a robust fused global descriptor for each place. Finally, we learn a proper metric to describe the similarity of our fused global feature to get an end-to-end place representation network. We implement our approach and some off-the-shelf methods on the open source KITTI and KAIST datasets. Experiments and visualization show that the fusion of two kinds of sensors can improve the performance, that is, the descriptor generated from two sensors is more robust than that from a single sensor.

Author Contributions

Methodology, Y.P.; Investigation, S.Y.; Supervision, S.X.; Visualization, C.P. and K.L.; Writing—original draft, C.P. and K.L.; Writing—review & editing, S.X. and Y.P. All authors have read and agreed to the published version of the manuscript.

Funding

The research is supported by the National Natural Science Foundation of China under Grant Nos. 61625304, 11771276, 11971296, Shanghai Science and technology innovation action plan(18441909000), and the capacity construction project of local universities in Shanghai (18010500600).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arandjelovi, R.; Zisserman, A. DisLocation: Scalable Descriptor Distinctiveness for Location Recognition. Lect. Notes Comput. Sci. 2015, 9006, 188–204. [Google Scholar]
  2. Cao, S.; Snavely, N. Graph-Based Discriminative Learning for Location Recognition. Int. J. Comput. Vis. 2015, 112, 239–254. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, D.M.; Baatz, G.; Koser, K.; Tsai, S.S.; Vedantham, R.; Pylvanainen, T.; Roimela, K.; Chen, X.; Bach, J.; Pollefeys, M.; et al. City-scale Landmark Identification on Mobile Devices. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 20–25 June 2011; pp. 737–744. [Google Scholar]
  4. Gronat, P.; Sivic, J.; Obozinski, G.; Pajdla, T. Learning and Calibrating Per-Location Classifiers for Visual Place Recognition. Int. J. Comput. Vis. 2016, 118, 319–336. [Google Scholar] [CrossRef] [Green Version]
  5. Sattler, T.; Havlena, M.; Radenovic, F.; Schindler, K.; Pollefeys, M. Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2102–2106. [Google Scholar]
  6. Bonardi, F.; Ainouz, S.; Boutteau, R.; Dupuis, Y.; Savatier, X.; Vasseur, P. PHROG: A Multimodal Feature for Place Recognition. Sensors 2017, 17, 1167. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Ouerghi, S.; Boutteau, R.; Savatier, X.; Tlili, F. Visual Odometry and Place Recognition Fusion for Vehicle Position Tracking in Urban Environments. Sensors 2018, 18, 939. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Cummins, M.; Newman, P. FAB-MAP: Probabilistic Localization and Mapping in the Space of Appearance. Int. J. Robot. Res. 2008, 27, 647–665. [Google Scholar] [CrossRef]
  9. Dube, R.; Dugas, D.; Stumm, E.; Nieto, J.; Siegwart, R.; Cadena, C. SegMatch: Segment Based Place Recognition in 3D Point Clouds. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 5266–5272. [Google Scholar]
  10. Fraundorfer, F.; Heng, L.; Honegger, D.; Lee, G.H.; Meier, L.; Tanskanen, P.; Pollefeys, M. Vision-Based Autonomous Mapping and Exploration Using a Quadrotor MAV. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012; pp. 4557–4564. [Google Scholar]
  11. Middelberg, S.; Sattler, T.; Untzelmann, O.; Kobbelt, L. Scalable 6-DOF Localization on Mobile Devices. Lect. Notes Comput. Sci. 2014, 8690, 268–283. [Google Scholar]
  12. Qiao, Y.; Cappelle, C.; Ruichek, Y.; Yang, T. ConvNet and LSH-based Visual Localization Using Localized Sequence Matching. Sensors 2019, 19, 2439. [Google Scholar] [CrossRef] [Green Version]
  13. He, L.; Wang, X.; Zhang, H. M2DP: A Novel 3D Point Cloud Descriptor and Its Application in Loop Closure Detection. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Daejeon, Korea, 9–14 October 2016; pp. 231–237. [Google Scholar]
  14. Kim, G.; Kim, A. Scan Context: Egocentric Spatial Descriptor for Place Recognition within 3D Point Cloud Map. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018; pp. 4802–4809. [Google Scholar]
  15. Mur-Artal, R.; Montiel, J.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  16. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  17. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  18. Sivic, J.; Zisserman, A. Video Google: A Text Retrieval Approach to Object Matching in Videos. In Proceedings of the IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003. [Google Scholar]
  19. Li, T.; Mei, T.; Kweon, I.S.; Hua, X.S. Contextual Bag-of-Words for Visual Categorization. IEEE Trans. Circ. Syst. Video Technol. 2011, 21, 381–392. [Google Scholar] [CrossRef]
  20. Jegou, H.; Douze, M.; Schmid, C.; Perez, P. Aggregating Local Descriptors into a Compact Image Representation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 3304–3311. [Google Scholar]
  21. Perronnin, F.; Liu, Y.; Sanchez, J.; Poirier, H. Large-scale Image Retrieval with Compressed Fisher Vectors. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 3384–3391. [Google Scholar]
  22. Chum, O.; Mikulik, A.; Perdoch, M.; Matas, J. Total Recall II: Query Expansion Revisited. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 20–25 June 2011; pp. 889–896. [Google Scholar]
  23. Delhumeau, J.; Gosselin, P.H.; Jegou, H.; Perez, P. Revisiting the VLAD Image Representation. In Proceedings of the 2013 ACM Multimedia Conference, Barcelona, Spain, 21–25 October 2013; pp. 653–656. [Google Scholar]
  24. Jegou, H.; Chum, O. Negative Evidences and Co-occurences in Image Retrieval: The Benefit of PCA and Whitening. Lect. Notes Comput. Sci. 2012, 7573, 774–787. [Google Scholar]
  25. Jegou, H.; Douze, M.; Schmid, C. Hamming Embedding and Weak Geometric Consistency for Large Scale Image Search. Lect. Notes Comput. Sci. 2008, 5302, 304–317. [Google Scholar]
  26. Torii, A.; Sivic, J.; Okutomi, M.; Pajdla, T. Visual Place Recognition with Repetitive Structures. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 2346–2359. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Chen, Z.; Lam, O.; Jacobson, A.; Milford, M. Convolutional Neural Network-based Place Recognition. arXiv 2014, arXiv:1411.1509. [Google Scholar]
  30. Sunderhauf, N.; Shirazi, S.; Dayoub, F.; Upcroft, B.; Milford, M. On the Performance of ConvNet Features for Place Recognition. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Hamburg, Germany, 28 September–2 October 2015; pp. 4297–4304. [Google Scholar]
  31. Gordo, A.; Almazan, J.; Revaud, J.; Larlus, D. Deep Image Retrieval: Learning Global Representations for Image Search. Lect. Notes Comput. Sci. 2016, 9910, 241–257. [Google Scholar]
  32. Chen, Z.; Jacobson, A.; Sunderhauf, N.; Upcroft, B.; Liu, L.; Shen, C.; Reid, I.; Milford, M. Deep Learning Features at Scale for Visual Place Recognition. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 3223–3230. [Google Scholar]
  33. Radenovi, F.; Tolias, G.; Chum, O. CNN Image Retrieval Learns from BoW: Unsupervised Fine-Tuning with Hard Examples. Lect. Notes Comput. Sci. 2016, 9905, 3–20. [Google Scholar]
  34. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN Architecture for Weakly Supervised Place Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1437–1451. [Google Scholar] [CrossRef] [Green Version]
  35. Zhong, Y.; Arandjelović, R.; Zisserman, A. GhostVLAD for Set-Based Face Recognition. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 35–50. [Google Scholar]
  36. Uy, M.A.; Lee, G.H. PointNetVLAD: Deep Point Cloud Based Retrieval for Large-Scale Place Recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4470–4479. [Google Scholar]
  37. Yang, Y.; Fan, D.; Du, S.; Wang, M.; Chen, B.; Gao, Y. Point Set Registration with Similarity and Affine Transformations Based on Bidirectional KMPE Loss. IEEE Trans. Cybern. 2019, 1–12. [Google Scholar] [CrossRef] [PubMed]
  38. Frome, A.; Huber, D.; Kolluri, R.; Bülow, T.; Malik, J. Recognizing Objects in Range Data Using Regional Point Descriptors. In Proceedings of the European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; pp. 224–237. [Google Scholar]
  39. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning Point Cloud Views using Persistent Feature Histograms. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
  40. Salti, S.; Tombari, F.; Di Stefano, L. SHOT: Unique Signatures of Histograms for Surface and Texture Description. Comput. Vis. Image Underst. 2014, 125, 251–264. [Google Scholar] [CrossRef]
  41. Rusu, R.B.; Blodow, N.; Beetz, M. Fast Point Feature Histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  42. Zhao, J.; Li, C.; Tian, L.; Zhu, J. FPFH-based Graph Matching for 3D Point Cloud Registration. Proc. Int. Soc. Opt. Eng. 2018, 10696, 106960M. [Google Scholar]
  43. Scovanner, P.; Ali, S.; Shah, M. A 3-Dimensional SIFT Descriptor and its Application to Action Recognition. In Proceedings of the ACM International Multimedia Conference and Exhibition, Augsburg, Germany, 25–29 September 2007; pp. 357–360. [Google Scholar]
  44. Wohlkinger, W.; Vincze, M. Ensemble of Shape Functions for 3D Object Classification. In Proceedings of the IEEE International Conference on Robotics and Biomimetics, Karon Beach, Phuket, Thailand, 7–11 December 2011; pp. 2987–2992. [Google Scholar]
  45. Steder, B.; Rusu, B.; Konolige, K.; Burgard, W. NARF: 3D Range Image Features for Object Recognition. In Proceedings of the Workshop on Defining and Solving Realistic Perception Problems in Personal Robotics at the IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 Ocotber 2010. [Google Scholar]
  46. Rusu, R.B.; Bradski, G.; Thibaux, R.; Hsu, J. Fast 3D Recognition and Pose using the Viewpoint Feature Histogram. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 Ocotber 2010; pp. 2155–2162. [Google Scholar]
  47. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  48. Qi, C.R.; Su, H.; Niebner, M.; Dai, A.; Yan, M.; Guibas, L.J. Volumetric and Multi-View CNNs for Object Classification on 3D Data. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5648–5656. [Google Scholar]
  49. Engelcke, M.; Rao, D.; Wang, D.Z.; Tong, C.H.; Posner, I. Vote3Deep: Fast Object Detection in 3D Point Clouds Using Efficient Convolutional Neural Networks. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 1355–1361. [Google Scholar]
  50. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  51. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5100–5109. [Google Scholar]
  52. Feng, Y.; Zhang, Z.; Zhao, X.; Ji, R.; Gao, Y. GVCNN: Group-View Convolutional Neural Networks for 3D Shape Recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 264–272. [Google Scholar]
  53. You, H.; Ji, R.; Feng, Y.; Gao, Y. PVNet: A Joint Convolutional Network of Point Cloud and Multi-view for 3D Shape Recognition. In Proceedings of the 2018 ACM Multimedia Conference, Seoul, Korea, 22–26 October 2018; pp. 1310–1318. [Google Scholar]
  54. Fernandez, A.; Wis, M.; Vecchione, G.; Silva, P.; Colomina, I.; Angelats, E.; Pares, M. Real-time Navigation and Mapping with Mobile Mapping Systems using LiDAR/Camera/INS/GNSS Advanced Hybridization Algorithms: Description and Test Results. In Proceedings of the International Technical Meeting of the Satellite Division of the Institute of Navigation, Tampa, FL, USA, 8–12 September 2014. [Google Scholar]
  55. Vyroubalova, J. LiDAR and Stereo Camera Data Fusion in Mobile Robot Mapping. Available online: http://excel.fit.vutbr.cz/submissions/2017/042/42.pdf (accessed on 3 May 2017).
  56. Ying, S.; Wen, Z.; Shi, J.; Peng, Y.; Peng, J.; Qiao, H. Manifold Preserving: An Intrinsic Approach for Semisupervised Distance Metric Learning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2731–2742. [Google Scholar] [CrossRef] [PubMed]
  57. Weinberger, K.Q.; Blitzer, J.; Saul, L.K. Distance Metric Learning for Large Margin Nearest Neighbor Classification. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 1473–1480. [Google Scholar]
  58. Peng, Y.; Zhang, N.; Li, Y.; Ying, S. A Local-to-Global Metric Learning Framework From the Geometric Insight. IEEE Access 2020, 8, 16953–16964. [Google Scholar] [CrossRef]
  59. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision Meets Robotics: The KITTI Dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  60. Jeong, J.; Cho, Y.; Shin, Y.S.; Roh, H.; Kim, A. Complex Urban Dataset with Multi-level Sensors from Highly Diverse Urban Environments. Int. J. Robot. Res. 2019, 38, 642–657. [Google Scholar] [CrossRef] [Green Version]
  61. Himmelsbach, M.; Hundelshausen, F.V.; Wuensche, H.J. Fast Segmentation of 3D Point Clouds for Ground Vehicles. In Proceedings of the IEEE Intelligent Vehicles Symposium, San Diego, CA, USA, 21–24 June 2010; pp. 560–565. [Google Scholar]
Figure 1. Complementarity between 3D point clouds and images. A 1 and A 2 are the images of the same place, which look very different since they are from two different views. A 3 and A 4 are the corresponding point clouds of images A 1 and A 2 .
Figure 1. Complementarity between 3D point clouds and images. A 1 and A 2 are the images of the same place, which look very different since they are from two different views. A 3 and A 4 are the corresponding point clouds of images A 1 and A 2 .
Sensors 20 02870 g001
Figure 2. Complementarity between images and 3D point clouds. B 1 and B 2 are two images at the same place, which is supplementary information to their corresponding point clouds B 3 and B 4 . B 3 and B 4 look very different (marked by red triangular regions).
Figure 2. Complementarity between images and 3D point clouds. B 1 and B 2 are two images at the same place, which is supplementary information to their corresponding point clouds B 3 and B 4 . B 3 and B 4 look very different (marked by red triangular regions).
Sensors 20 02870 g002
Figure 3. Residual block in ResNet [27].
Figure 3. Residual block in ResNet [27].
Sensors 20 02870 g003
Figure 4. The framework of PointNetVLAD [36].
Figure 4. The framework of PointNetVLAD [36].
Sensors 20 02870 g004
Figure 5. Network architecture.
Figure 5. Network architecture.
Sensors 20 02870 g005
Figure 6. Diagram of the trimmed Vector of Locally Aggregated Descriptors (VLAD).
Figure 6. Diagram of the trimmed Vector of Locally Aggregated Descriptors (VLAD).
Sensors 20 02870 g006
Figure 7. The point cloud pre-processing in the KITTI dataset.
Figure 7. The point cloud pre-processing in the KITTI dataset.
Sensors 20 02870 g007
Figure 8. Merging the left and right point clouds in KAIST dataset. The top point cloud is from the right LiDAR, and the bottom one is from the left LiDAR.
Figure 8. Merging the left and right point clouds in KAIST dataset. The top point cloud is from the right LiDAR, and the bottom one is from the left LiDAR.
Sensors 20 02870 g008
Figure 9. The devices used to acquire the campus data.
Figure 9. The devices used to acquire the campus data.
Sensors 20 02870 g009
Figure 10. Comparison of our approach with other approaches on KITTI and KAIST datasets.The ResNet and ResNetVLAD are image-based approaches. Both of them use the ResNet as the feature extractor, and the only difference is the final layer. The PointNetVLAD and PointNetVLAD (trimmed) are point-based approaches. PointNetVLAD is open source, and the trimmed version uses our non-informative clusters.
Figure 10. Comparison of our approach with other approaches on KITTI and KAIST datasets.The ResNet and ResNetVLAD are image-based approaches. Both of them use the ResNet as the feature extractor, and the only difference is the final layer. The PointNetVLAD and PointNetVLAD (trimmed) are point-based approaches. PointNetVLAD is open source, and the trimmed version uses our non-informative clusters.
Sensors 20 02870 g010
Figure 11. Visualization of non-informative clusters. The marked points in the same color belong to the same non-informative cluster. For each non-informative cluster, most of the points are gathered together, but some points are far away from the center.
Figure 11. Visualization of non-informative clusters. The marked points in the same color belong to the same non-informative cluster. For each non-informative cluster, most of the points are gathered together, but some points are far away from the center.
Sensors 20 02870 g011
Figure 12. Top 2 candidates of two scenes in the KITTI dataset. The position of the test frame and the top 2 candidates are marked on the trajectory. The candidates of each test frame are more likely to be adjacent frames.
Figure 12. Top 2 candidates of two scenes in the KITTI dataset. The position of the test frame and the top 2 candidates are marked on the trajectory. The candidates of each test frame are more likely to be adjacent frames.
Sensors 20 02870 g012
Figure 13. t-SNE plot of several scenarios in image, point cloud and fusion methods. The positions of the frames are represented by each point in a certain scene, where the adjacent relationship is represented by its color. We use the ground truth in the KITTI dataset to set all frames in a specific color, which means the adjacent frames have the similar colors.
Figure 13. t-SNE plot of several scenarios in image, point cloud and fusion methods. The positions of the frames are represented by each point in a certain scene, where the adjacent relationship is represented by its color. We use the ground truth in the KITTI dataset to set all frames in a specific color, which means the adjacent frames have the similar colors.
Sensors 20 02870 g013
Figure 14. Illustration of t-SNE plots. We highlight a short segment on the trajectory and mark the corresponding points in the plots.
Figure 14. Illustration of t-SNE plots. We highlight a short segment on the trajectory and mark the corresponding points in the plots.
Sensors 20 02870 g014
Figure 15. The retrieval map of KITTI 05 in our approach and the VLAD algorithm. Here we mark the minimal required candidates that contain the correct correspondence in the whole trajectory.
Figure 15. The retrieval map of KITTI 05 in our approach and the VLAD algorithm. Here we mark the minimal required candidates that contain the correct correspondence in the whole trajectory.
Sensors 20 02870 g015
Figure 16. Comparison of our approach with other approaches to our campus dataset.
Figure 16. Comparison of our approach with other approaches to our campus dataset.
Sensors 20 02870 g016
Table 1. Comparison against the state-of-the-art approaches (@1%). The symbol @1% means recall (%) on the top 1% candidates.
Table 1. Comparison against the state-of-the-art approaches (@1%). The symbol @1% means recall (%) on the top 1% candidates.
ResNetResNetVLADPointNetVLADPointNetVLAD (trimmed)Ours
KITTI 0095.696.396.497.698.1
KITTI 0280.683.684.384.388.1
KITTI 0587.687.978.278.688.1
KITTI 0694.795.593.693.697.7
KAIST 3084.283.772.672.487.3
KAIST 3288.392.686.793.195.2
Table 2. Comparison of differernt numbers of points (@1). The results are tested on the KITTI dataset. The symbols @1 means recall (%) on the top 1 candidate.
Table 2. Comparison of differernt numbers of points (@1). The results are tested on the KITTI dataset. The symbols @1 means recall (%) on the top 1 candidate.
N = 1024 N = 2048 N = 4096 N = 6000
KITTI 0057.571.776.176.6
KITTI 0221.634.335.143.3
KITTI 0541.052.756.358.5
KITTI 0650.670.280.083.0
Table 3. Comparison of differernt numbers of cluser centers (@1). The results are tested on the KITTI dataset. The symbols @1 means recall (%) on the top 1 candidate.
Table 3. Comparison of differernt numbers of cluser centers (@1). The results are tested on the KITTI dataset. The symbols @1 means recall (%) on the top 1 candidate.
K = 32 K = 64 K = 96 K = 128
KITTI 0078.078.978.876.6
KITTI 0238.845.546.343.3
KITTI 0559.761.760.958.5
KITTI 0681.583.482.383.0
Table 4. Comparison of fusing different image features (@1%/@1). The results are tested on the KITTI dataset. The symbols @1% and @1 mean recall (%) on the top 1% and 1 candidate(s), respectively.
Table 4. Comparison of fusing different image features (@1%/@1). The results are tested on the KITTI dataset. The symbols @1% and @1 mean recall (%) on the top 1% and 1 candidate(s), respectively.
@1%@1
ResNetResNetVLADResNetResNetVLAD
KITTI 0098.197.493.187.3
KITTI 0288.186.673.954.5
KITTI 0588.176.976.563.6
KITTI 0697.795.594.786.8
Table 5. Comparison of the VLAD and our approach. Here we list recall(%) @1 in the KITTI dataset and processing time of extracting the global descriptor of each frame in seconds.
Table 5. Comparison of the VLAD and our approach. Here we list recall(%) @1 in the KITTI dataset and processing time of extracting the global descriptor of each frame in seconds.
@1Time
VLADOursVLADOurs
KITTI 0088.692.70.0460.025
KITTI 0263.464.9
KITTI 0566.076.9
KITTI 0682.392.5

Share and Cite

MDPI and ACS Style

Xie, S.; Pan, C.; Peng, Y.; Liu, K.; Ying, S. Large-Scale Place Recognition Based on Camera-LiDAR Fused Descriptor. Sensors 2020, 20, 2870. https://doi.org/10.3390/s20102870

AMA Style

Xie S, Pan C, Peng Y, Liu K, Ying S. Large-Scale Place Recognition Based on Camera-LiDAR Fused Descriptor. Sensors. 2020; 20(10):2870. https://doi.org/10.3390/s20102870

Chicago/Turabian Style

Xie, Shaorong, Chao Pan, Yaxin Peng, Ke Liu, and Shihui Ying. 2020. "Large-Scale Place Recognition Based on Camera-LiDAR Fused Descriptor" Sensors 20, no. 10: 2870. https://doi.org/10.3390/s20102870

APA Style

Xie, S., Pan, C., Peng, Y., Liu, K., & Ying, S. (2020). Large-Scale Place Recognition Based on Camera-LiDAR Fused Descriptor. Sensors, 20(10), 2870. https://doi.org/10.3390/s20102870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop