Next Article in Journal
Sensors, Circuits, and Systems for Biomedical Applications
Previous Article in Journal
ADE-CycleGAN: A Detail Enhanced Image Dehazing CycleGAN Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mapping with Monocular Camera Sensor under Adversarial Illumination for Intelligent Vehicles

School of Automotive Studies, Tongji University, Shanghai 201804, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3296; https://doi.org/10.3390/s23063296
Submission received: 19 February 2023 / Revised: 7 March 2023 / Accepted: 17 March 2023 / Published: 21 March 2023
(This article belongs to the Section Vehicular Sensing)

Abstract

:
High-precision maps are widely applied in intelligent-driving vehicles for localization and planning tasks. The vision sensor, especially monocular cameras, has become favoured in mapping approaches due to its high flexibility and low cost. However, monocular visual mapping suffers from great performance degradation in adversarial illumination environments such as on low-light roads or in underground spaces. To address this issue, in this paper, we first introduce an unsupervised learning approach to improve keypoint detection and description on monocular camera images. By emphasizing the consistency between feature points in the learning loss, visual features in dim environment can be better extracted. Second, to suppress the scale drift in monocular visual mapping, a robust loop-closure detection scheme is presented, which integrates both feature-point verification and multi-grained image similarity measurements. With experiments on public benchmarks, our keypoint detection approach is proven robust against varied illumination. With scenario tests including both underground and on-road driving, we demonstrate that our approach is able to reduce the scale drift in reconstructing the scene and achieve a mapping accuracy gain of up to 0.14 m in textureless or low-illumination environments.

1. Introduction

The advent of big data and Internet of Things (IoT) has brought new prospects for intelligent vehicles [1,2,3,4]. A high-precision map is a prerequisite for the localization and planning of intelligent vehicles, and its creation is considered a key technology in this field and is well-researched with various sensor setups. For instance, Global Navigation Satellite System (GNSS)-based devices are integrated in many commercial vehicles. However, the satellite signal can be easily shielded by high buildings or specific materials. Lidar can directly provide accurate distance information and is thus widely used in mapping approaches. Yet the high resolution of 3D representation is at the cost of high data volume and non-negligible hardware cost. On the contrary, vision-based solutions, especially with monocular images, benefiting from the low sensor cost and effectiveness in texture-rich scenarios [5], have attracted much attention in mapping research.
Traditional visual mapping is based on the structure from motion (SfM) paradigm. It extracts many keypoints with feature descriptors from images captured in a scene. The 3D positions of these keypoints are either provided using additional measuring instruments or triangulated with multi-view geometry. These keypoints are stored and further utilized as landmarks to construct geometrical constraints for ego-pose estimation in vehicle localization tasks. In this way, the output of visual mapping can be presented with a point cloud consisting of sparsely distributed 3D keypoints along with their descriptors.
However, monocular visual mapping approaches are known to have a relatively low capacity in accurately estimating the depth information and the absolute scale. Moreover, handcrafed keypoint models such as SIFT [6], SURF [7], ORB [8], BRISK [9], etc., are prone to feature-extraction failure in low-texture and low-illumination environments such as underground spaces, where the keypoint detection and matching performance significantly degrades (Figure 1), further deteriorating the generated maps. Many researchers have been working on these problems, especially leveraging the powerful perception capability of the rising deep learning technology. For instance, the semantic features are considered as landmarks in the image [10] and also adopted to improve matching in low-textured environments such as in underground garages [11]. Features such as simple visual tags are used to reduce the computational costs and improve the location accuracy [12]. Furthermore, high-level landmarks such as visual fiducial markers are simultaneously perceived to aid the pose-estimation process [13]. In other attempts, keypoint detection and description are learned by convolutional neural networks (CNNs) to replace handcrafted ORB keypoints and integrated into simultaneous localization and mapping (SLAM) systems [14]. Nevertheless, the segmentation model requires a large amount of pixel-wise labeled image data for training, and the setup of fiducial markers should be elaborately designed, which limits the generalizability of the mapping approach. Moreover, the groundtruths (GTs) of keypoints in [14] are provided by the Shi–Tomasi detector [15]. In comparison to CNN-based methods, which can extract complex visual features [16], approaches that learn from such a traditional keypoint-detection paradigm can be suboptimal in scenarios with less texture.
To address these issues, this paper presents a monocular visual mapping approach in low-textured and illumination-changed environments using unsupervised keypoint learning and improved scale recovery (codes will be available at https://github.com/XinningC/Adversarial_mono_Mapping (accessed on 18 February 2023)). Our approach comprises the following two components: unsupervised keypoint extraction and 3D reconstruction with improved scale estimation. According to the experimental results of keypoint extraction and mapping in Section 4, we show that our approach outperforms existing mainstream methods with an average location error gap of 0.04 m on Euroc [17] and 1.86 m on KITTI [18] and manifests robustness in adversarial-illumination environments. Our contributions are as follows:
  • We introduce an unsupervised keypoint-detection and -description approach into the visual mapping process with an improved loss to enhance the learning of discriminative descriptors. Such an approach only requires monocular images as input, thus saving annotation labors.
  • We present a scheme integrating both feature-point verification and self-supervised multi-grained image similarity measurements. It effectively reduces the cumulative error and suppresses the overall scale drift at a low level.
  • We further integrate ground point features and camera height to recover the absolute scale. Using validation on both the self-collected data and public benchmarks, this mapping approach is demonstrated to be robust against illumination changes in scenarios such as underground parking and outdoor roads.
The remainder of this paper is organized as follows. An overview of the related work is given in Section 2. Section 3 introduces the proposed monocular visual mapping approach using unsupervised keypoint matching and improved scale recovery. Section 4 evaluates the mapping in both underground parking and outdoor driving. The conclusion is given in Section 5.

2. Related Work

In this section, existing studies are reviewed in three aspects mostly related to the proposed approach: (1) keypoint models; (2) monocular visual architecture; (3) mapping under adversarial illumination.

2.1. Keypoint Models

Traditional visual mapping approaches are based on handcrafted keypoint models, including SIFT [6], SURF [7], ORB [8], BRISK [9], etc. SIFT preserves a good invariance to rotation and scaling by leveraging gradient histograms, but its computational complexity is high. Although improvements such as the approximated gradient filtering are adopted in SURF, the runtime gain is still limited. Unlike SURF, ORB achieves an excellent real-time performance by employing the FAST keypoint detector [19] and the BRIEF feature descriptor [20]. Since feature points are extracted by simple comparison with surrounding pixels, the reliability in illumination-changed and low-textured environments is relatively poor. ASIFT [21] outperforms SIFT while features are undergoing large transition tilt due to its affinity invariance; nevertheless, the matching accuracy and efficiency need improvement [22,23]. BRISK employs feature pyramid and Gaussian filtering for point sampling, achieving an improved noise robustness, but at the cost of increased computation amount.
In recent years, deep-learning-based feature-point methods have gained much attention. Lift [24] proposes a full feature-point handling pipeline, including point detection, orientation estimation, and feature description, while a supervision form SfM system is required. Superpoint [25] proposes an unsupervised learning for feature extraction with a fully convolutional neural network (FCN) and a homography adaption, but it relies on labeled interest-point images for pre-training. The GCN series accelerates the processing with binarized descriptors and nested metric learning with groundtruths provided by the traditional Shi–Tomasi detector. Unsuperpoint [26] inherits the general idea of Superpoint, but relieves the reliance on labeled data by making the keypoint detection and matching totally self-supervised. However, its adaptability in the visual mapping process is unexplored.

2.2. Monocular Visual Architecture

Depending on the tasks defined, we review monocular visual architecture in terms of two main aspects, i.e., SLAM and SfM approaches. A milestone work of visual SLAM is the PTAM [27], which proposes an architecture of two parallel threads with one for motion tracking and the other for 3D feature-point mapping. However, suffering from the lack of loop closure and the low invariance to viewpoint change, its performance is limited in large environments and rapidly varying scenes. Built on the main idea of PTAM, ORBSLAM [28] utilizes ORB feature points through the pipeline of tracking, mapping, relocalization, and loop closing. It manifests an outstanding real-time performance and a good adaptability to RGB-D [29] and fisheye cameras [30]. However, the poor performance of ORB feature points in low-texture environments limits the application of ORBSLAM. VINS [31] adopts a more lightweight front-end with Harris corner detection and optical flow tracking. It also introduces an inertial sensor to minimize the scale variance of monocular-vision-only SLAM. Further improvements are also proposed with sensor adaptability, dynamic environment, online calibration, and pose-graph reuse [32,33,34,35,36].
SfM aims to estimate the 3D point positions as well as the camera poses through motion. Unlike SLAM, it can be applied on disordered images and the incremental methods are widely used. An early work was proposed by [37], which uses a pipeline of feature-point extraction, matching, and iterative pose optimization. Despite the relatively complete reconstruction obtained, this method suffers from mismatching in repeated scenes and a low processing speed due to the large number of images. Thus, subsequent research has focused more on the improvement of accuracy and efficiency. For example, ASFM [38] proposes adaptive thresholds in model estimation with a contrario methodology. It achieves better reconstruction results compared to those based on fixed global thresholds, yet at the cost of decreased calculation efficiency. In the further work of COLMAP [39], a geometric verification and a sampling-based triangulation are adopted to improve the robustness and completeness of reconstruction. Moreover, the global optimization is only then performed when the model grows to a certain extent, greatly reducing the load of computing resources and thus increasing the speed.

2.3. Mapping under Adversarial Illumination

The mapping task in low-illuminated environments such as underground mainly rely on lidar or camera sensors. Thus, corresponding approaches can be divided into two branches, depending on the used sensor. Lidar-based approaches prefer line or plane features. As the range information can be directly obtained with laser measurements, the cumulative error is relatively low. However, high-definition lidar is costly for commercialization and the positioning is difficult in structures such as long, narrow tunnels. For visual mapping, the main challenge remains in the visual feature extraction from low-texture, dim environments. Reference [13] proposed an underground mapping by adopting semantic features of parking slots in conjunction with a geometric prior to improve the point-matching accuracy. Additionally, they placed visual fiducial markers on specific positions to create a robust constraint in the back-end optimization. Despite an improved mapping performance, their dependence on visual markers restricts the usage of their method. Reference [40] proposed a self-supervised approach to learn features from both the entire image and its subregions. This method can effectively match images across daytime and thus improve the location-related image-retrieval accuracy. Reference [11] proposed a SLAM approach in indoor parking lots based on surround-view images to increase the perception range. Additionally, they built maps by exploiting semantic segmentation results of parking signs, slot lines, and bumps as robust features against illumination change, achieving a centimeter-level accuracy. Reference [41] proposed CNN models to accomplish the identification of slot marking points and classification of patterns represented by marking-point pairs that can be applied on both indoor and outdoor parking sites. Reference [42] proposed a more lightweight parking-slot segmentation model by employing network pruning. Their method shows a significantly reduced computational cost and can be applied on CPUs with real-time performance while maintaining a good mapping precision. Nevertheless, the segmentation performance strongly relies on the quantity and quality of pixel-wise labeled data, which restricts the generalizability of the above mapping approaches.
In this work, we mainly focus on the mapping accuracy. Thus, without strict runtime requirements, we adopt an SfM scheme, i.e., the COLMAP, as the back-end for camera pose optimization. The proposed approach is presented mainly based on the mapping task of an underground parking lot. To improve the mapping performance, we adopt an unsupervised learning method to address the challenge of feature-point detection and -description in dim, low-texture environments. Moreover, we integrate the multi-grained image similarity measurement along with feature-point verification into the loop-closure detection and adopt ground-point features with camera height for scale recovery, thus improving the mapping robustness.

3. Proposed Approach

The pipeline of the proposed visual mapping approach is shown in Figure 2. The input image is captured with a monocular front-view camera on a vehicle and fed into two subsequent processing branches. The first branch adopts a CNN to predict the scores, relative positions, and descriptors of the keypoints, which are learned in an unsupervised manner. The keypoints are matched across frames and fed into the SfM framework along with the corresponding image pairs to reconstruct a scene graph for camera pose estimation. In conjunction with the keypoint-based verification, it also adopts a multi-grained image similarity measurement to improve the pose-estimation accuracy, especially for loop closures. In the second branch, it detects the corners of ground parking slots and considers them as additional keypoints, which are tracked across frames by leveraging the keypoint network in the first branch. By integrating the tracks of the above-ground keypoints and the camera height, the absolute scale of camera motion can be recovered. With an additional association strategy, the parking slots on the ground can be reconstructed and thus the map for the underground parking lot can be created. The specific steps in this proposed mapping procedure are introduced below.

3.1. Unsupervised Keypoint Extraction

Regarding both the scene adaptability and the unreliability on training labels, in this work, we adopt an unsupervised-learning-based approach for keypoint extraction. The network consists of a backbone similar to reference [26] except with more convolutional layers to facilitate the learning of deep visual features. We additionally enhance the loss function by emphasizing the consistence between feature points in illumination-varied environments. The detailed architecture of the network is shown in Figure 3.
The input image is first processed by the backbone (top right in Figure 3), which consists of four stages. After each stage, the feature map is downsampled by a factor of 2 while its channel number is doubled. The generated feature map thus has a size of 1/8 of the input image and is further fed into three output heads to predict the tensor of keypoint scores S m a p , relative positions P m a p , and descriptors F m a p , respectively. Since the output tensors are the same size as the input feature map, the total predicted keypoints is H 8 × W 8 , and each score, relative position, and descriptor correspond to an 8 × 8 region of the input image.
For training, a source image I 1 is preprocessed using a random homography transform H 1 2 with an additional color conversion or noising to generate a warped version I 2 (see orange dashed box in Figure 3). Thus, we obtain an image pair ( I 1 , I 2 ), which is fed into the network to predict keypoints. Each keypoint P consists of a tuple of score, relative position, and descriptor, described as ( s , p , f ) . Similar to reference [26], descriptor f is a 256-dimensional tensor of floating-point numbers. A keypoint P 1 , i in I 1 with another keypoint P 2 , j in I 2 is defined as a good match if their Euclidean distance H 1 2 · p 1 , i p 2 , j 2 is less than a threshold α d i s .
The loss function of the whole network consists of six terms as
L t o t a l = w 1 L s c o r e + w 2 L p o s + w 3 L r e p + w 4 L u n i + w 5 L d e c o r + w 6 L d e s ,
where w 1 , , 6 indicates the corresponding weight of each term. The losses L s c o r e , L p o s , L r e p , L u n i , and L d e c o r are defined similar to reference [26] and briefly described in Table 1 and Appendix A.1.
The essence of a keypoint description is the expression of a corresponding image patch surrounding the point. Considering that similar image areas should yield highly correlated keypoint descriptors, the descriptor loss L d e s should reduce their distance in the feature space. Inspired by the contrastive learning of visual representation [43], the loss L d e s is defined as
L d e s = i log j exp ( t · f 1 , i · f 2 , j ) l j exp ( t · f 1 , i · f 2 , l )
where terms ( f 1 , i , f 2 , j ) in the nominator indicate keypoint descriptors of a good match, while keypoint descriptors ( f 1 , i , f 2 , l ) | l j in the denominator are from a bad match. The temperature t is a hyperparameter. For hard negatives, which can be easily classified as false positives, a smaller t can reduce their weights during learning, which further improves the feature-point matching.
In order to improve the illumination robustness of keypoints, we utilize traditional transformation methods such as random clipping, flipping, and brightness adjustment, with the latter having a certain impact on the performance of feature points in adversarial-illumination environments. Moreover, our adopted loss function L d e s contributes to improving the feature-point learning significantly. Detailed experimental verification can be seen in Section 4.2.

3.2. 3D Reconstruction with Improved Scale Estimation

The predicted image keypoints are tracked across frames (e.g., by descriptor matching) and then fed into the 3D reconstruction framework. Here, we use COLMAP [39] as the base reconstruction approach. Additionally, we add the multi-grained image matching along with the feature-point verification to reduce drift at loops and adopt ground keypoints for further scale recovery.

3.2.1. Base SfM Approach

COLMAP is an incremental SfM framework. Given a set of images with correspondences indicated by tracked keypoints, also known as the scene graph, the outputs are the estimated poses of camera images and the scene structure represented by a point cloud in the 3D space. For initialization, it uses a seeding with an elaborated two-view reconstruction, e.g., at a location in the image graph with multiple overlapping cameras, which has a high redundancy. It further proceeds with incremental registration of new images. Their poses are estimated by solving the Perspective-n-Point (PnP) problem based on the triangulation of matched feature points in exiting images. By extending the camera pose, new scene points can be observed and triangulated from the added images, thus incrementally increasing the scene coverage. To reduce the uncertainties propagated between the image pose and triangulated points, bundle adjustment (BA) is utilized as a joint non-linear refinement for both of them by minimizing the point reprojection errors on the image. Additionally, COLMAP leverages the multi-model to verify non-panoramic and calibrated image pairs in seeding the reconstruction. To reduce the risk of mis-registration, it chooses the next-best view based on a multi-resolution analysis. Furthermore, it adopts an efficient sampling-based triangulation and an elaborated local/global BA strategy. Although COLMAP has improved over conventional SfM approaches in the reconstruction completeness and scalability, it still shows limitations when applied to monocular images. Lacking the depth information, the scale estimation of a monocular SfM system significantly depends on the feature points in the local map. Due to the inherent uncertainties of feature points (e.g., by misdetection and mismatching), without sufficient constraints, it is difficult to maintain a long-term scale consistency and thus makes the monocular system susceptible to scale drifts. Therefore, more powerful constraints such as loop-closure detection (which is common in parking lots and urban streets) are required so that the scale information of feature points can be efficiently propagated.

3.2.2. Image Matching with Multi-Grained Similarity and Keypoint Verification

Compared to keypoint descriptors, images contain much richer visual information. Thus, we adopt an image-retrieval-based loop-closure detection, i.e., to identify images captured at the same place. However, spatially close-by images do not exactly depict the same scene, especially with varied camera poses or changed foreground objects, which results in noisy hard positives. By only using the image-level supervision, all the features of the target image are forced to be similar to those of the noisy positive image, impairing the discriminatory learning of local features. Regarding these points, we employ a multi-grained image-similarity learning strategy. Specifically, we decompose each image into 4 half-regions (including horizontal and vertical direction) and 4 quarter-regions. Thus, the similarity of the target and the positive image is learned in a region-level supervision. Compared to the image-level supervision, such a manner is more rational. It focuses more on the common regions between images during learning by pulling positive regions closer while pushing away negative regions in the feature space.
Here we adopt a VGG16-based network as in [40], which receives the query and noisy positive image as inputs and estimates their similarities. Additionally, we adopt an iterative learning strategy, i.e., using the converged model and its estimated image similarities in the current round to initialize the to-be-trained model and its supervised labels in the next round. The positive inputs are the top k ( k = 10 ) difficult images with their queries from the last round. By iteratively mining the hard positives, the accuracy of the predicted similarity labels and the discriminatory ability of the model are progressively improved in a self-supervised manner. The loss function is defined as in [40], which is
L = L h a r d + λ L s o f t ,
where λ is a trade-off factor. The first term is in the form of
L h a r d = j log exp ( f v q · f v p ) exp ( f v q · f v p ) + exp ( f v q · f v n j ) ,
where f v indicates the feature vector extracted from the last convolutional layer of VGG16 (i.e., the c o n v 5 ) while the subscripts q, p, and n j denote the query, the positive, and the hard negative j, respectively. The second loss term in Equation (3) is expressed as
L s o f t = CE s ( q , p 1 r , , p n r , 1 ) , s 1 ( q , p 1 r , , p n r , τ )
with CE denoting the cross entropy operator. The subscript “−1” indicates features extracted from the model trained in the previous round. The temperature τ is a hyperparameter. The similarity vector is calculated in a softmax form as
s ( q , p 1 r , , p n r , τ ) = softmax f v q · f v p 1 r τ , , f v q · f v p n r τ ,
where p 1 r , , p n r indicate the top k hard positive images and their 8 subregions. Thus, we have the number n = 9 k . In our experiment, the network is trained in 4 rounds with progressively reduced τ (0.07, 0.06, 0.05), while the trade-off factor is set to λ = 0.5 .
Given the fact that our keypoint network can extract an amount of robust keypoints and all images are acquired consecutively, we adopt a sequential image-matching procedure in conjunction with the loop-closure detection to suppress the uncertainty of scales. The image matching is based on the multi-grained similarity measurement and an additional feature-point verification, with the latter to suppress the false positives missed by the former. The whole image-matching process is as follows.
1.
Images are processed in chronological order. If there is still an image unprocessed, it is marked as q i . Otherwise, the process is terminated.
2.
The query q i is matched with its following N images.
3.
Based on the matching results, the images are searched with their multi-grained image similarities to q i within a threshold α m g and denoted as set O . For an empty O , the procedure goes back to step 1.
4.
For each image p j O , if the number of correspondence keypoints between q i and p j is greater than a threshold α n u m , the pair ( q i , p j ) is recorded into a database of query-positive candidates (Figure 4). Otherwise, the procedure goes back to step 1.
5.
For each p j in the candidate database, we also consider the correspondence between q i and N images after p j . If the number of correspondence keypoints in any image pair is not greater than the threshold α n u m , the candidate database is cleared. The assumption is that the keypoints across true positive images can be tracked for a period. Otherwise, the candidate database is recorded in the final database. The procedure goes back to step 1.
In the above process, we empirically set N = 3 , α m g = 0.5 , and α n u m = 250 .
Figure 4. The sequential image-matching process.
Figure 4. The sequential image-matching process.
Sensors 23 03296 g004

3.2.3. Scale Recovery

This part focuses on the reconstruction of underground parking slot instances and the recovery of the overall scale in the mapping task.
Since the corners of parking slots appear to be similar and can be easily recognized in the bird’s eye view (BEV), we first convert the front-view image to a BEV image using the process introduced in reference [44] and use a CenterNet-based network [45] to detect the corner points of parking slots. Afterwards, we project the positions of the detected corner points from the BEV back to the front-view image. The feature points related to the projected corner points are searched and tracked in the next few frames. As the corner points are strong keypoints, they can provide a more stable tracking performance. The related procedure is shown in Figure 5.
Due to the restricted driving speed, each parking slot as well as its corners can be observed in up to 20 consecutive frames. With triangulation, we can obtain the positions of the corner points in the 3D space. Since each parking slot has two corresponding corner points at the entrance, the position of the parking slot (represented by the entrance center) can be estimatedwithby two corner points within a distance threshold (i.e., 2.5 m).
The scale recovery is based on the relative distance between the ground and the camera trajectory. Specifically, for each pair of corner points, we consider a 3D cubic space located at the center of the corner points, as illustrated in Figure 6. The cube is above the ground plane with a side length of e greater than the camera height d. Given a camera position t , the ground plane can be approximated by
n t + d = 0 ,
where n indicates the plane normal vector. Given a camera trajectory within the cube, its historical positions ( t 1 , t 2 , , t k ) can be applied with the above equation as
t 1 t 2 t k · n + d = 0 .
Here we assume that the ground is flat and the camera height is fixed within the cube. The cube size is elaborately set so that it contains sufficient trajectory points and Equation (8) is over-determined. Thus, the normal vector n and the camera height d can be solved, e.g., using the least squares method. Given the real measurement d g t of the camera height, the scale correction factor is r = d g t d , which is used to recover the scale of the reconstructed map, especially the parking slot instances. The math symbols used in our paper are explained in Table A1 in Appendix A.2.

4. Experiment

4.1. Platform Configuration and Dataset Selection

In our work, we implement the keypoint extraction network and the mapping approach with Python 3.6.9, PyTorch 1.7.0 on Ubuntu 18.04 LTS using a machine with 32 GB RAM, a six-core Intel Core i7-8700 CPU, and an Nvidia GeForce GTX 1060 GPU.
For the keypoint extraction learning, we util-ze MS COCO [46] to train the network and test it on HPatches [47]. For the evaluation of the 3D reconstruction and mapping, we test our approach on EuRoC [17] and KITTI [18]. In comparison with other unused datasets such as Strecha [48] and DTU [49], the datasets selected here are more comprehensive in terms of the data amount and diversity, up-to-date, and broadly used to validate mainstream methods, which allows a fair comparison of our method with state of the art.

4.2. Evaluation of Keypoint Model

For a comprehensive evaluation of the utilized keypoint detector, it is tested on the HPatches benchmark [47], which consists of sequences from 59 scenes with viewpoint change and 57 scenes with illumination variation, each with one reference image and a number of target images. All images are scaled to the same size of 240 × 320 px for evaluation. Four metrics are chosen, i.e., the homography accuracy (HA), the repeatability score (RS), the location error (LE), and the matching score (MS), which are briefly described in Table 2. The loss weights are empirically set to ω 1 = 2 , ω 2 = 1 , ω 3 = 1 , ω 4 = 100 , ω 5 = 0.03 , and ω 6 = 0.001 (with t = 0.05 ). It is trained on the MS COCO dataset [46] for 10 epochs with a learning rate of 0.000025 and a batch size of 16. The ADAM optimizer is adopted. The inference time of our keypoint model is about 10 ms per image. The test results of the utilized keypoint model on scenes with illumination variation and viewpoint change are reported in Table 3, in comparison with handcrafted feature points such as SIFT and SURF and another self-supervision-based method, Superpoint.
As can be seen, the learning-based keypoint models significantly outperform the handcrafted approaches in handling scenarios with illumination variation. Especially in terms of the matching score, the gain is more significant, up to 38%, implying a more powerful feature representation with deep learning. Moreover, the adopted keypoint model in our approach achieves a slightly lower HA than Superpoint while outperforming it in terms of all three other metrics. The improvement in the location error is more critical, at 13%, demonstrating the effectiveness of unsupervised learning of the keypoint detector. A qualitative example is shown in Figure 7a. As for scenes with a viewpoint change, similar trends can be seen in Figure 7b. Deep-learning-based approaches achieve a comparable HA to SIFT while outperforming it in all other three metrics, with our adopted approach ranking at the top.
To assess the influence of brightness adjustment during preprocessing and the loss term L d e s on the illumination robustness of the learned feature points, we conducted ablation experiments on the HPatches illumination sequence, with the results presented in Table 4. The results indicate that the performance of feature points is influenced by both traditional transformations and L d e s . Only by combining these factors can the optimal performance of feature points be achieved.

4.3. Study on Image Matching and Keypoint Verification

To verify the performance of the adopted multi-grained image matching and keypoint verification, we collected a dataset by driving a test vehicle for several rounds in the underground parking lot of about 1500 m 2 at Tongji University Campus. The test vehicle was equipped with a camera capturing front-view images at 10 Hz and a lidar of 16 beam lines. As GNSS signals are unavailable in the underground space, we consider the map constructed using the laser measurements of lidar as the groundtruth. An ablation study on the influence of the adopted multi-grained image matching and keypoint verification was conducted, with the results reported in Table 5. For evaluation, it follows the protocol of the 7-DOF alignment [50] of the reconstructed trajectories. The mean value of absolute pose errors (APE) is chosen as the metric.
It is straightforward that the naive matching by keypoints yields a high position error of more than 12 m. By replacing it with the multi-grained image matching, the error is reduced to 0.6 m. However, image mismatching still occurs, which leads to an abrupt change in the estimated trajectory (yellow curves in Figure 8). By further integrating the keypoint verification procedure, the error decreases to 0.45 m and the abrupt trajectory change disappears (purple curves in Figure 8). Thus, the benefit brought by multi-grained image matching and keypoint verification has been validated.

4.4. Exploration of Mapping Architecture

Although the keypoint model alone was tested in the previous experiments, its influence on the SfM-based mapping architecture is still unexplored. Here we replace our adopted keypoint model with other approaches such as the handcrafted SIFT and ORB and the learning-based Superpoint, thus yielding three new mapping architectures, denoted COLMAP (SIFT), COLMAP (SURF), and COLMAP (Superpoint). We adopt the same data and metrics as in Section 4.3, with the test results shown in Table 6. Obviously, the learning approach of Superpoint improved the mapping accuracy compared to the handcrafted SIFT and SURF in terms of the mean APE. Since our keypoint model emphasizes the distinctive feature representation during training, by integrating it, the least pose error has been achieved. The estimated trajectories and keypoint matches are illustrated in Figure 9. Additionally, we show the map generated by our approach in representation of point clouds and with identified parking slots in Figure 10a,b, respectively.

4.5. Transferring on SLAM Approaches

Although our matching scheme is designed for SfM architectures, it can be integrated with SLAM systems. Here we replace COLMAP with the VINS approach. Due to the high complexity of the optimization scheme in VINS, we exclude the integration of multi-grained image matching and keypoint verification. We test the new architecture on the EuRoC and KITTI datasets. The EuRoC dataset is aimed for the evaluation of indoor SLAM and consists of 11 sequences captured in different rooms and fields along with IMU measurements. The KITTI visual odometry dataset is captured in a surrounding region of the city Karlsruhe, consisting of 22 image sequences, in which 11 sequences (00-11) are provided with temporally aligned groundtruths and measurements from lidar sensors. For comparison, we provide the performance of the original VINS approach, which is denoted as VINS (flow) due to its deployment of optical flow. Additionally, we replace the optical flow in VINS with other feature descriptors and our proposed feature-point approach and report their test performance. The evaluation metric is the root mean square error (RMSE) of the estimated trajectories.

4.5.1. Evaluation on EuRoC

The evaluation results are shown in Table 7 while examples of the estimated trajectories in the sequence MH04 and V101 are shown in Figure 11. As can be seen, the handcrafted feature points generally perform inferiorly to the learning-based approaches. Specifically, the VINS (flow) employs the conventional Harris keypoint detector with its point tracking assisted by optical flow. Its accuracy is still lower than our approach and the version with Superpoint. Superpoint dominates mainly on the “M”-sequences, which have rich textures, benefiting the accuracy and stability of the keypoints extracted by Superpoint. In contrast, our approach is more advantageous on the “V”-sequences with large camera motion and illumination variation. The adverse environmental conditions can result in reduced keypoints extracted by other approaches. However, as our method focuses more on the learning of distinctive visual features, it shows a superior performance. An example of feature-point matching with large camera motion is visualized in Figure 7b.

4.5.2. Evaluation on KITTI

Due to the fast motion in large-scale outdoor environments, the VINS versions integrated with handcrafted keypoint approaches such as SIFT and SURF all fail in the test, thus without the results reported in Table 8. The high-speed motion scenario (e.g., the highway) also explains the failure of learning-based approaches in sequence 01. Since VINS (flow) employs optical flow to assist the point tracking, it can still maintain a small position error in this sequence. Furthermore, it can be seen that our approach dominates the sequences 00, 02, 05, and 08–10 with the least error while in other sequences it performs better than Superpoint, yet with a small gap to VINS (flow). Especially in sequences 00, 02, and 08, our approach outperforms VINS (flow) more significantly, with an accuracy gain of more than 4 m, further demonstrating the effectiveness of our approach. Examples of estimated trajectories are shown in Figure 12.

5. Conclusions

In this paper, we proposed a novel monocular visual-mapping approach to address adversarial illumination conditions such as in underground parking lots. There are two main points contributing to this approach: the unsupervised learning of keypoints that enhances the discriminative feature representation and the scheme of multi-grained image matching and keypoint verification in scale drift suppression. This paper presents extensive experiments validating the individual modules and the entire SfM-based mapping architecture, demonstrating the robustness of the approach against illumination variation. The testing results on a public benchmark and our collected dataset show that our approach outperforms the existing mainstream methods, with an average location error gap of 0.05 m on collected data, 0.04 m on Euroc, and 1.86 m on KITTI. Additionally, the approach is effective not only in SfM systems but also in indoor SLAM and outdoor visual odometry tasks.

Author Contributions

Conceptualization, W.T.; methodology, W.T., Y.W. and X.C.; software, Y.W. and X.C.; validation, Y.W. and X.C.; formal analysis, Y.W. and X.C.; data curation, Y.W. and X.C.; writing—original draft preparation, Y.W. and X.C.; writing—review and editing, W.T.; visualization, W.T., Y.W. and X.C.; supervision, W.T.; project administration, W.T.; funding acquisition, W.T. All authors have read and agreed to the published version of the manuscript.

Funding

Project supported by the National Key R&D Program of China [No.2021YFB2501104].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and analysed during the current study are available in the [KITTI] repository and the [EuRoC] repository. (http://www.cvlibs.net/datasets/kitti (accessed on 9 February 2023)) and (https://projects.asl.ethz.ch/datasets/doku.php?id=kmavvisualinertialdatasets (accessed on 9 February 2023)).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Mathematical Explanations

Appendix A.1. Equation of L d e c o r

Descriptors should be decorrelated to prevent overfitting and carry as much useful information as possible within a limited dimension. The descriptors forming good matches in image pair ( I 1 , I 2 ) are denoted as G 1 = [ f 1 1 , , f 1 m , , f 1 p ] 256 × p and G 2 = [ f 2 1 , , f 2 m , , f 2 p ] 256 × p , where p stands for the number of good matches. Apparently, we can denote G 1 as [ h 1 1 , , h 1 m , , h 1 256 ] p × 256 and h 1 m is the m-th row vector of G 1 . Similar to reference [26], the correlation matrix R 1 = [ r i j 1 ] 256 × 256 can be defined as
r i j 1 = ( h 1 i h ¯ 1 i ) ( h 1 j h ¯ 1 j ) ( h 1 i h ¯ 1 i ) ( h 1 i h ¯ 1 i ) ( h 1 j h ¯ 1 j ) ( h 1 j h ¯ 1 j )
where h ¯ 1 i stands for the mean of the i-th row of G 1 . The same is true for R 2 . Thus, we can calculate L d e c o r as
L d e c o r = i j r i j 1 2 + i j r i j 2 2 .

Appendix A.2. Explanation for Math Symbols

Table A1. A notation table that explains math symbols in our used paper.
Table A1. A notation table that explains math symbols in our used paper.
SymbolDescription
S m a p Map of keypoint score
P m a p Map of keypoint relative positions
F m a p Map of keypoint descriptors
H , W Height/Width of input images
I 1 , I 2 Source image/Warped image
H 1 2 Homography transform matrix
P Keypoint
sKeypoint score
p Keypoint relative position
f Keypoint descriptor
L Loss function
w 1 , , 6 Weights of losses
tHyperparameter that controls L d e s
λ Trade-off factor between L h a r d and L s o f t
f v Feature vector
s Similarity vector
τ Hyperparameter that controls f v and s
q i Unprocessed image
NNumber of images following q i
O Set of images that are not searched
p j Image that is not searched
α m g Threshold of multi-grained image similarities
α n u m Threshold of the number of correspondence keypoints
dCamera height
eSide length of the cubic space in Figure 6
t Camera position
n Plane normal vector
rScale correction factor

References

  1. Abassi, R.; Ben Chehida Douss, A.; Sauveron, D. TSME: A trust-based security scheme for message exchange in vehicular Ad hoc networks. Hum.-Centric Comput. Inf. Sci. 2020, 10, 43. [Google Scholar] [CrossRef]
  2. Aliedani, A.; Loke, S.W.; Glaser, S. Robust cooperative car-parking: Implications and solutions for selfish inter-vehicular social behaviour. Hum.-Centric Comput. Inf. Sci. 2020, 10, 37. [Google Scholar] [CrossRef]
  3. Xu, Z.; Liang, W.; Li, K.C.; Xu, J.; Jin, H. A blockchain-based Roadside Unit-assisted authentication and key agreement protocol for Internet of Vehicles. J. Parallel Distrib. Comput. 2021, 149, 29–39. [Google Scholar] [CrossRef]
  4. Chen, C.; Li, K.; Teo, S.G.; Zou, X.; Li, K.; Zeng, Z. Citywide Traffic Flow Prediction Based on Multiple Gated Spatio-Temporal Convolutional Neural Networks. ACM Trans. Knowl. Discov. Data 2020, 14, 1–23. [Google Scholar] [CrossRef]
  5. Chen, Q.; Xie, Y.; Guo, S.; Bai, J.; Shu, Q. Sensing system of environmental perception technologies for driverless vehicle: A review of state of the art and challenges. Sens. Actuators A: Phys. 2021, 319, 112566. [Google Scholar] [CrossRef]
  6. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  7. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded Up Robust Features. In Proceedings of the European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar] [CrossRef]
  8. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An Efficient Alternative to SIFT or SURF. In Proceedings of the International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar] [CrossRef]
  9. Leutenegger, S.; Chli, M.; Siegwart, R.Y. BRISK: Binary Robust Invariant Scalable Keypoints. In Proceedings of the International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar] [CrossRef] [Green Version]
  10. Guan, H.; Lei, X.; Yu, Y.; Zhao, H.; Peng, D.; Junior, J.M.; Li, J. Road marking extraction in UAV imagery using attentive capsule feature pyramid network. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102677. [Google Scholar] [CrossRef]
  11. Qin, T.; Chen, T.; Chen, Y.; Su, Q. Avp-slam: Semantic Visual Mapping and Localization for Autonomous Vehicles in the Parking Lot. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020; pp. 5939–5945. [Google Scholar] [CrossRef]
  12. Gao, F.; Ma, J. Indoor Location Technology with High Accuracy Using Simple Visual Tags. Sensors 2023, 23, 1597. [Google Scholar] [CrossRef]
  13. Huang, Y.; Zhao, J.; He, X.; Zhang, S.; Feng, T. Vision-Based Semantic Mapping and Localization for Autonomous Indoor Parking. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Suzhou, China, 26–30 June 2018; pp. 636–641. [Google Scholar] [CrossRef] [Green Version]
  14. Tang, J.; Ericson, L.; Folkesson, J.; Jensfelt, P. GCNv2: Efficient Correspondence Prediction for Real-Time SLAM. IEEE Robot. Autom. Lett. 2019, 4, 3505–3512. [Google Scholar] [CrossRef] [Green Version]
  15. Shi, J.T. Good Features to Track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar] [CrossRef]
  16. Bibi, S.; Abbasi, A.; Haq, I.; Baik, S.; Ullah, A. Digital Image Forgery Detection Using Deep Autoencoder and CNN Features. Hum.-Centric Comput. Inf. Sci. 2021, 11, 1–17. [Google Scholar] [CrossRef]
  17. Burri, M.; Nikolic, J.; Gohl, P.; Schneider, T.; Rehder, J.; Omari, S.; Achtelik, M.W.; Siegwart, R. The EuRoC micro aerial vehicle datasets. Int. J. Robot. Res. 2016, 35, 1157–1163. [Google Scholar] [CrossRef]
  18. Geiger, A.; Lenz, P.; Urtasun, R. Are We Ready for Autonomous Driving? The Kitti Vision Benchmark Suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar] [CrossRef]
  19. Rosten, E.; Drummond, T. Machine Learning for High Speed Corner Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2006; Volume 1, pp. 430–443. [Google Scholar] [CrossRef]
  20. Calonder, M.; Lepetit, V.; Ozuysal, M.; Trzcinski, T.; Strecha, C.; Fua, P. BRIEF: Computing a Local Binary Descriptor Very Fast. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1281–1298. [Google Scholar] [CrossRef] [Green Version]
  21. Yu, G.; Morel, J.M. A fully affine invariant image comparison method. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 1597–1600. [Google Scholar] [CrossRef]
  22. Gao, J.; Sun, Z. An Improved ASIFT Image Feature Matching Algorithm Based on POS Information. Sensors 2022, 22, 7749. [Google Scholar] [CrossRef]
  23. Yum, J.; Kim, J.S.; Lee, H.J. Fast Execution of an ASIFT Hardware Accelerator by Prior Data Processing. Electronics 2019, 8, 1176. [Google Scholar] [CrossRef] [Green Version]
  24. Yi, K.M.; Trulls, E.; Lepetit, V.; Fua, P. Lift: Learned Invariant Feature Transform. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14October 2016; Springer: Berlin, Germany, 2016; pp. 467–483. [Google Scholar] [CrossRef] [Green Version]
  25. DeTone, D.; Malisiewicz, T.; Rabinovich, A. SuperPoint: Self-Supervised Interest Point Detection and Description. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 337–33712. [Google Scholar] [CrossRef] [Green Version]
  26. Hviid Christiansen, P.; Fly Kragh, M.; Brodskiy, Y.; Karstoft, H. UnsuperPoint: End-to-end Unsupervised Interest Point Detector and Descriptor. arXiv 2019, arXiv:1907.04011. [Google Scholar] [CrossRef]
  27. Klein, G.; Murray, D. Parallel Tracking and Mapping for Small AR Workspaces. In Proceedings of the IEEE and ACM International Symposium on Mixed and Augmented Reality, Washington, DC, USA, 13–16 November 2007; pp. 225–234. [Google Scholar] [CrossRef]
  28. Mur-Artal, R.; Montiel, J.M.M.; Tardós, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  29. Mur-Artal, R.; Tardós, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef] [Green Version]
  30. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.M.; Tardós, J.D. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  31. Qin, T.; Shen, S. Robust Initialization of Monocular Visual-Inertial Estimation on Aerial Robots. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 4225–4232. [Google Scholar] [CrossRef]
  32. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef] [Green Version]
  33. Chen, W.; Shang, G.; Hu, K.; Zhou, C.; Wang, X.; Fang, G.; Ji, A. A Monocular-Visual SLAM System with Semantic and Optical-Flow Fusion for Indoor Dynamic Environments. Micromachines 2022, 13, 2006. [Google Scholar] [CrossRef]
  34. Zang, Q.; Zhang, K.; Wang, L.; Wu, L. An adaptive ORB-SLAM3 System for Outdoor Dynamic Environments. Sensors 2023, 23, 1359. [Google Scholar] [CrossRef] [PubMed]
  35. Qin, T.; Shen, S. Online Temporal Calibration for Monocular Visual-Inertial Systems. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 3662–3669. [Google Scholar] [CrossRef] [Green Version]
  36. Qin, T.; Pan, J.; Cao, S.; Shen, S. A General Optimization-based Framework for Local Odometry Estimation with Multiple Sensors. arXiv 2019, arXiv:1901.03638. [Google Scholar] [CrossRef]
  37. Snavely, N.; Seitz, S.M.; Szeliski, R. Photo Tourism: Exploring Photo Collections in 3D. ACM Trans. Graph. 2006, 25, 835–846. [Google Scholar] [CrossRef] [Green Version]
  38. Moulon, P.; Monasse, P.; Marlet, R. Adaptive Structure from Motion with a Contrario Model Estimation. In Proceedings of the Asian Conference on Computer Vision (ACCV), Daejeon, Republic of Korea, 5–9 November 2012; pp. 257–270. [Google Scholar] [CrossRef] [Green Version]
  39. Schönberger, J.L.; Frahm, J.M. Structure-from-Motion Revisited. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar] [CrossRef]
  40. Ge, Y.; Wang, H.; Zhu, F.; Zhao, R.; Li, H. Self-Supervising Fine-Grained Region Similarities for Large-Scale Image Localization. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 369–386. [Google Scholar] [CrossRef]
  41. Zhang, L.; Huang, J.; Li, X.; Xiong, L. Vision-Based Parking-Slot Detection: A DCNN-Based Approach and a Large-Scale Benchmark Dataset. IEEE Trans. Image Process. 2018, 27, 5350–5364. [Google Scholar] [CrossRef] [PubMed]
  42. Yu, Z.; Gao, Z.; Chen, H.; Huang, Y. SPFCN: Select and Prune the Fully Convolutional Networks for Real-time Parking Slot Detection. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 13–19 October 2020; pp. 445–450. [Google Scholar] [CrossRef]
  43. Nguyen, K.; Nguyen, Y.; Le, B. Semi-Supervising Learning, Transfer Learning, and Knowledge Distillation with SimCLR. arXiv 2021, arXiv:2108.00587. [Google Scholar] [CrossRef]
  44. Tian, W.; Ren, X.; Yu, X.; Wu, M.; Zhao, W.; Li, Q. Vision-based mapping of lane semantics and topology for intelligent vehicles. Int. J. Appl. Earth Obs. Geoinf. 2022, 111, 102851. [Google Scholar] [CrossRef]
  45. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar] [CrossRef]
  46. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollar, P.; Zitnick, L. Microsoft COCO: Common Objects in Context. In Proceedings of the European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2014. [Google Scholar] [CrossRef] [Green Version]
  47. Balntas, V.; Lenc, K.; Vedaldi, A.; Mikolajczyk, K. HPatches: A Benchmark and Evaluation of Handcrafted and Learned Local Descriptors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3852–3861. [Google Scholar] [CrossRef] [Green Version]
  48. Strecha, C.; von Hansen, W.; Van Gool, L.; Fua, P.; Thoennessen, U. On Benchmarking Camera Calibration and Multi-View Stereo for High Resolution Imagery. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar] [CrossRef]
  49. Aanas, H.; Dahl, A.L.; Steenstrup Pedersen, K. Interesting Interest Points. Int. J. Comput. Vis. 2012, 97, 18–35. [Google Scholar] [CrossRef]
  50. Grupp, M. Evo: Python Package for the Evaluation of Odometry and SLAM. 2017. Available online: https://github.com/MichaelGrupp/evo (accessed on 18 February 2023).
Figure 1. Detected SIFT keypoints (green circles) in ordinary outdoor scenario (above) and dim textureless garage (below).
Figure 1. Detected SIFT keypoints (green circles) in ordinary outdoor scenario (above) and dim textureless garage (below).
Sensors 23 03296 g001
Figure 2. The overall architecture of our monocular visual mapping system.
Figure 2. The overall architecture of our monocular visual mapping system.
Sensors 23 03296 g002
Figure 3. The architecture of our feature-extraction network. The channel number is indicated under each block. The network takes an input image and outputs a feature-point vector ( s , p , f ) .
Figure 3. The architecture of our feature-extraction network. The channel number is indicated under each block. The network takes an input image and outputs a feature-point vector ( s , p , f ) .
Sensors 23 03296 g003
Figure 5. The process for scale recovery and parking slot instance reconstruction.
Figure 5. The process for scale recovery and parking slot instance reconstruction.
Sensors 23 03296 g005
Figure 6. Local scale recovery based on selected camera trajectories.
Figure 6. Local scale recovery based on selected camera trajectories.
Sensors 23 03296 g006
Figure 7. Qualitative comparison of matched keypoints using different models on scenes with (a) illumination change and (b) viewpoint change. Detected keypoint number is displayed on the right side of each figure.
Figure 7. Qualitative comparison of matched keypoints using different models on scenes with (a) illumination change and (b) viewpoint change. Detected keypoint number is displayed on the right side of each figure.
Sensors 23 03296 g007
Figure 8. The estimated trajectories using different matching schemes.
Figure 8. The estimated trajectories using different matching schemes.
Sensors 23 03296 g008
Figure 9. (a) The mapping results using different approaches. (b) Matched keypoints using different approaches.
Figure 9. (a) The mapping results using different approaches. (b) Matched keypoints using different approaches.
Sensors 23 03296 g009
Figure 10. (a) Generated map in representation of point clouds. (b) Generated map with parking slot identification.
Figure 10. (a) Generated map in representation of point clouds. (b) Generated map with parking slot identification.
Sensors 23 03296 g010
Figure 11. Estimated trajectories in sequence MH04 and V101 on EuRoC.
Figure 11. Estimated trajectories in sequence MH04 and V101 on EuRoC.
Sensors 23 03296 g011
Figure 12. Qualitative test results on sequence KITTI08 and KITTI09.
Figure 12. Qualitative test results on sequence KITTI08 and KITTI09.
Sensors 23 03296 g012
Table 1. Loss terms to train keypoint network.
Table 1. Loss terms to train keypoint network.
TermDescription
L s c o r e Squared score difference of paired points
L p o s Euclidean distance of paired points
L u n i Differences between the distribution of predicted
point coordinates and a uniform distribution
L d e c o r Correlation coefficients between keypoint
descriptors on the same image, further explained in Equation (A.1)
L r e p Ensuring closely located point pair with a high score,
interpreted as L r e p = k ( s 1 + s 2 ) ( d k d ¯ ) , with
s 1 , s 2 : point scores of the k-th pair
d k : distance of the k-th pair
d ¯ : mean distance of all pairs
Table 2. Evaluation metrics of the keypoint network.
Table 2. Evaluation metrics of the keypoint network.
MetricDescription
HARatio of estimated homographies under a threshold ϵ
(here set to 3 px) to all estimated homographies
RSRatio of corresponding points to all predicted points
LEAverage distance of corresponding points
MSRatio of good matches to predicted points in one image,
where a good match denotes two corresponding points
with the nearest descriptors in the feature space.
Table 3. Comparison of different keypoint detectors on the sequence with viewpoint and illumination change of HPatches.
Table 3. Comparison of different keypoint detectors on the sequence with viewpoint and illumination change of HPatches.
IlluminationViewpoint
MethodsHA ↑RS ↑LE ↓MS ↑HA ↑RS ↑LE ↓MS ↑
SURF0.770.571.160.270.580.531.410.23
SIFT0.860.501.110.250.660.521.220.29
Superpoint0.930.640.940.630.630.511.170.47
Ours0.910.650.810.640.620.551.090.47
Table 4. Ablation study results on HPatches illumination sequence. “Trans.” indicates brightness transformations and “ L d e s ” indicates the utilization of the loss function L d e s .
Table 4. Ablation study results on HPatches illumination sequence. “Trans.” indicates brightness transformations and “ L d e s ” indicates the utilization of the loss function L d e s .
Trans. L des HA ↑RS ↑LE ↓MS ↑
0.860.650.820.38
0.910.650.920.62
0.910.650.810.64
Table 5. Ablation study results. “Match by pts.” indicates the naive matching by keypoints; “Match by img.” indicates the multi-grained image matching; “Pts. verify” indicates the keypoint verification procedure; “-as” represents the 7-DOF alignments used in the evaluation.
Table 5. Ablation study results. “Match by pts.” indicates the naive matching by keypoints; “Match by img.” indicates the multi-grained image matching; “Pts. verify” indicates the keypoint verification procedure; “-as” represents the 7-DOF alignments used in the evaluation.
MatchMatchPts.APE-as
by Pts.by Img.Verify(Mean) ↓
12.81 m
0.60 m
0.45 m
Table 6. Exploration of SfM-based mapping integrated with different keypoint models.
Table 6. Exploration of SfM-based mapping integrated with different keypoint models.
MethodsAPE-as (Mean) ↓
Ours0.453 m
COLMAP (Superpoint)0.502 m
COLMAP (SIFT)0.562 m
COLMAP (SURF)0.596 m
Table 7. The RMSE (in meters) of estimated trajectories using VINS integrated with different keypoint models on EuRoC. “x” indicates the failure of approach.
Table 7. The RMSE (in meters) of estimated trajectories using VINS integrated with different keypoint models on EuRoC. “x” indicates the failure of approach.
VINS VINSVINSVINS
Seq.(Flow)Ours(Superpoint)(SIFT)(SURF)
MH010.240.220.200.760.50
MH020.220.220.180.510.48
MH030.280.240.17x0.27
MH040.430.430.470.560.62
MH050.310.320.220.530.75
V1010.1090.1080.120.230.20
V1020.100.110.090.130.13
V1030.1110.0880.090.150.19
V2010.1210.1160.140.180.21
V2020.110.090.130.270.25
V2030.300.200.790.390.49
Table 8. The RMSE (in meters) of estimated trajectories using VINS integrated with different keypoint models on KITTI. “x” indicates the failure of approach.
Table 8. The RMSE (in meters) of estimated trajectories using VINS integrated with different keypoint models on KITTI. “x” indicates the failure of approach.
VINS VINS
Seq.(Flow)Ours(Superpoint)
0013.747.0577.70
017.54xx
0220.6911.3417.71
031.752.673.17
041.332.272.68
056.645.545.78
063.874.8915.64
072.204.2711.28
089.374.9757.24
097.737.428.01
103.661.964.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tian, W.; Wen, Y.; Chu, X. Mapping with Monocular Camera Sensor under Adversarial Illumination for Intelligent Vehicles. Sensors 2023, 23, 3296. https://doi.org/10.3390/s23063296

AMA Style

Tian W, Wen Y, Chu X. Mapping with Monocular Camera Sensor under Adversarial Illumination for Intelligent Vehicles. Sensors. 2023; 23(6):3296. https://doi.org/10.3390/s23063296

Chicago/Turabian Style

Tian, Wei, Yongkun Wen, and Xinning Chu. 2023. "Mapping with Monocular Camera Sensor under Adversarial Illumination for Intelligent Vehicles" Sensors 23, no. 6: 3296. https://doi.org/10.3390/s23063296

APA Style

Tian, W., Wen, Y., & Chu, X. (2023). Mapping with Monocular Camera Sensor under Adversarial Illumination for Intelligent Vehicles. Sensors, 23(6), 3296. https://doi.org/10.3390/s23063296

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop