Next Article in Journal
Real-Time Underwater Image Recognition with FPGA Embedded System for Convolutional Neural Network
Next Article in Special Issue
Pixelwise Phase Unwrapping Based on Ordered Periods Phase Shift
Previous Article in Journal
A Method for Analyzing the Impact of Intra-System and Inter-System Interference on DME Based on Queueing Theory
Previous Article in Special Issue
3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Method for Extrinsic Calibration of Multiple RGB-D Cameras Using Descriptor-Based Patterns

1
School of Mechatronic Engineering and Automation, Shanghai University, No. 99 Shangda Road BaoShan District, Shanghai 200444, China
2
School of Mechanical Engineering, Beijing Institute of Technology, Beijing 100081, China
3
Department of Mechanical and Industrial Engineering, University of Toronto, Toronto, ON M5S 3G8, Canada
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(2), 349; https://doi.org/10.3390/s19020349
Submission received: 2 December 2018 / Revised: 21 December 2018 / Accepted: 14 January 2019 / Published: 16 January 2019
(This article belongs to the Special Issue Depth Sensors and 3D Vision)

Abstract

:
This paper presents a novel method to estimate the relative poses between RGB-D cameras with minimal overlapping fields of view. This calibration problem is relevant to applications such as indoor 3D mapping and robot navigation that can benefit from a wider field of view using multiple RGB-D cameras. The proposed approach relies on descriptor-based patterns to provide well-matched 2D keypoints in the case of a minimal overlapping field of view between cameras. Integrating the matched 2D keypoints with corresponding depth values, a set of 3D matched keypoints are constructed to calibrate multiple RGB-D cameras. Experiments validated the accuracy and efficiency of the proposed calibration approach.

1. Introduction

In recent years, low-cost and efficient depth and color (RGB-D) devices such as the Microsoft Kinect, Intel RealSense, and Structure Sensor have attracted much attention because of their applications in indoor scene reconstruction and robot navigation. The depth cameras of these devices can provide a depth map with a VGA resolution (640 × 480) at video rate (e.g., 30 Hz) using efficient light-coding technologies that avoid the challenging task of dense 3D reconstruction from color images. The 3D models reconstructed using these depth cameras have been used to generate more realistic 3D content for virtual reality (VR) [1] and help align the rendered virtual objects with real scenes for augmented reality (AR) [2]. Furthermore, the direct depth sensing capability of these depth cameras is particularly suitable for robots to navigate in an unknown environment.
With an RGB-D camera, the simultaneous localization and mapping (SLAM)-based approach is mainly used for fusing point cloud frames to reconstruct indoor scenes [3,4,5,6]. However, hundreds or even thousands of frames must be captured in state-of-the-art SLAM systems to reconstruct a common indoor environment, such as a room or an office [7] because of two problems. (1) The field of view (FoV) of depth cameras is limited; thus, only a small part of the scene is represented in a single frame. The Kinect, for example, has a horizontal FoV of 57 , which is much smaller than the horizontal 240 FoV of the Hokuyo URG-04LX-UG01, a laser scanner with a similar range and measurement accuracy to the Kinect [8]. (2) To track the poses of depth cameras to effectively fuse multiple point cloud frames, consecutive frames must be captured to have sufficient scene overlap. Typically, more than ninety percent of overlap is required, which further increases the number of frames for reconstruction.
One solution to these problems is to use a multi-camera setup in which RGB-D cameras face different directions to sample different sections of the environment [9,10,11,12]. However, complications can occur with the use of multiple cameras, such as more difficult calibration caused by the minimal overlapping FoV between cameras.
Classical extrinsic calibration strategies such as the chessboard-based method and the keypoints-based method cannot be applied to calibrate RGB-D cameras in a multi-camera setup, because their overlap requirement constitutes a very strong constraint. Martyna Poreba et al. proposed a linear feature-based method to perform accurate registration of point clouds [13]. Fernandez-Moral et al. proposed to calibrate multiple RGB-D cameras through planes or lines [9,14]. Planes and lines have large spatial spans; thus, they can be observed by cameras with little or no overlapping FoV.
Another kind of approaches that is proposed to calibrate multiple RGB-D cameras is based on per-camera odometry [15,16]. Cameras are calibrated by finding all camera odometry transforms-based on matched features from frames that are captured in the motion paths of all cameras. In these methods, SLAM or visual odometry techniques are applied to estimate camera trajectories. However, the robustness of SLAM and visual odometry techniques highly depends on the environment.
In this work, a new extrinsic calibration method that relies on descriptor-based patterns is proposed to estimate the relative poses between the RGB-D cameras with minimal overlapping FoV. The descriptor-based pattern is based on the reverse engineering of SIFT/SURF, can provide many more well-matched keypoints than nature scenes [17]. In our method, a set of matched 3D points are constructed based on extracted 2D keypoints and corresponding depth values from depth maps. The relative poses between RGB-D cameras are estimated efficiently by minimizing the alignment error of matched 3D points. If there are loop closures in a multi-RGB-D camera system, the estimated poses can be further optimized globally to decrease incremental estimation error through pose graph optimization. An efficient distributed capturing system is also proposed to provide multiple RGB-D data frames in real time to facilitate the usage of multi-RGB-D cameras system in practical applications such as 3D reconstruction and robot navigation.

2. Method Overview

Figure 1 shows the flow chart of our calibration method. The input to the method is color images and depth maps (RGB-D frames) captured in a scene with descriptor-based patterns. The descriptor-based patterns [17] are random textures generated using the reverse engineering of SIFT/SURF, can provide many more well-matched keypoints than nature scenes even in the case of minimal overlapping FoV between cameras. The SIFT keypoint detector is used in this work, as it is more stable to rotation and illumination changes than SURF [18].
Next the well-matched keypoints are extracted from color images of descriptor-based patterns using the RANSAC (random sample consensus)-based method. Then 3D point sets are constructed based on the extracted 2D keypoints and their depth values in the depth maps.
Afterwards the relative poses are estimated by directly minimizing the alignment error of 3D point sets. The last step is to optimize the estimated poses with possible loop closure, for example the first and the last RGB-D camera in a panoramic setup (see Figure 2a) have an overlapping FoV that can be seen as a loop closure. Since this is a conditional step, the last step is presented within a dashed box.

3. Detailed Methods

3.1. Keypoint Extraction

The descriptor-based calibration pattern is composed of several random noise images at different scales in accordance with the mechanism of SIFT/SURF. Compared with natural scenes, this pattern contains a high number of detectable features on multiple scales that can be easily detected by a camera at varying distances. Thus, the descriptor-based pattern can provide many and more accurately matched keypoints between two cameras in the case of minimal overlapping FoV.
In Figure 3, the detected keypoints are represented by green dots. Figure 3a,b shows the detected keypoints of the image captured in a natural scene, and the image captured in a scene with descriptor-based patterns, respectively. There are only 406 keypoints in Figure 3a, but 1231 keypoints in Figure 3b. Figure 4a,b shows the matched keypoints of a pair of images captured by two cameras with an approximately thirty percent overlapping FoV. Well-matched keypoints are connected by green lines; poorly matched keypoints are connected by red lines. The matching results are generated by the Flann-based descriptor matcher implemented in OpenCV, where matched keypoints are accepted only if its descriptor distance is less than three times the minimum descriptor distances. There are 56 well-matched keypoints and 32 poorly matched keypoints in Figure 4a, 116 well-matched keypoints and 30 poorly matched keypoints in Figure 4b. Thus the descriptor-based patterns provide more robust matched keypoints.
To further increase the efficiency and accuracy of extracting the matched keypoints, the keypoint detector can be only applied to the overlapping regions of neighboring images, and a RANSAC-based filtering method is applied to refine matches. The overlapping regions can be roughly estimated based on the FoV of cameras and the angle between two neighboring cameras. In the RANSAC-based filtering method, RANSAC is used to select the best subset of corresponding keypoint pairs through estimating the homography matrix. A re-projection error threshold is predetermined to help the RANSAC-based filtering method to distinguish inliers from outliers.
Figure 4c shows the matched keypoints when the keypoints are only detected in the regions within the blue rectangles (overlapping regions). There are 362 well-matched keypoints in Figure 4c, no obviously poorly matched keypoints. Figure 4d shows the selected 25 keypoint matches after applying the RANSAC-based method, the re-projection error threshold is set to 0.3 pixel.

3.2. Estimate the Pose from 3D Point Sets

Based on these matched keypoints, we find their corresponding depth values from the depth maps generated by the depth camera in the RGB-D camera to construct 3D point sets to estimate poses. This is a class of ICP (iterative closest point) problem that minimizes the difference between two point clouds. The critical difference is that correspondences between two point clouds are already known in our case, thus the time-consuming finding of matched 3D poits in typical ICP-based method can be avoided.
We construct a non-linear least squares formulation based on the alignment error of matched 3D points that can be solved by the Levenberg–Marquardt algorithm (LM) considering the error of selected matched keypoints and depth measurement. When the correspondences between two point clouds are known, there exists an analytic solution for this formulation [19] which guarantees the fast convergence of LM without the good initial guess. Implementation details are as follows.

3.2.1. Align the Depth Map with the Color Image

Due to the different spatial positions and intrinsic parameters of the depth camera and of the color camera in a RGB-D camera, the depth map is not aligned with the color image. The extrinsic parameters between the depth and color camera, intrinsic parameters of depth and color camera are used to align the depth map with the color image. SDKs (software development kits) of RGB-D cameras provide interfaces to obtain the intrinsic parameters and to align the depth map with color image. Since the users may want to obtain a more precise alignment result through calibrating the intrinsic and extrinsic parameters by themselves, the detailed methods are provided.
Let ( u 0 , v 0 ) denote the coordinates of the principal point of the depth camera, f x and f y denote the scale factors in image u and v axes of the depth camera, and u 0 , v 0 , f x and f y be the intrinsic parameters of the depth camera. Let [ u , v , Z ] represent a pixel in the depth map, Z represent the depth value in [ u , v ] , and [ X , Y , Z ] T represent the mapped 3D point in [ u , v ] in the depth camera coordinate system. According to the pinhole camera model, the values of X and Y can be calculated according to
X = ( u u 0 ) Z / f x , Y = ( v v 0 ) Z / f y .
Let T represent the extrinsic parameter (i.e., transformation matrix) between the depth and color camera, ( u 0 , v 0 ) denote the coordinates of the principal point of the color camera, f x and f y denote the scale factors in image u and v axes of the color camera. The aligned depth pixel [ u , v , Z ] is obtained by transforming X , Y , Z T based on T and get X , Y , Z T , then mapping X , Y , Z T to the color image coordinate system as follows
X , Y , Z T = T X , Y , Z T , u = X Z f x + u 0 , v = Y Z f y + v 0 .

3.2.2. Estimate the Pose

Let [ u j , v j ] and [ u j , v j ] denote the coordinates of selected matched keypoints, Z j and Z j denote the corresponding aligned depth value in [ u j , v j ] and [ u j , v j ] obtained from the aligned depth map respectively, j = 1 , 2 , , N , we construct two 3D point sets { p j } , { p j } based on Equation (1). The relative pose between these two 3D point sets can be found by minimizing
min ξ = 1 2 i = 1 n p j exp ( ξ ^ ) p j 2 2 ,
where ξ s e ( 3 ) is a vector with six dimensions that represents the camera pose, ξ is mapped to a matrix with four dimensions R 4 × 4 by ^ . ξ ^ is mapped to T S E ( 3 ) by the exponential map exp ( ) . We use the standard Levenberg–Marquardt algorithm to solve Equation (3). The derivative with respect to the camera pose of an error element is given by
( exp ( ξ ^ ) p ) ξ = lim δ ξ 0 exp ( δ ξ ^ ) exp ( ξ ^ ) p exp ( ξ ^ ) p δ ξ lim δ ξ 0 ( I + δ ξ ^ ) exp ( ξ ^ ) p exp ( ξ ^ ) p δ ξ = lim δ ξ 0 δ ξ ^ exp ( ξ ^ ) p δ ξ = lim δ ξ 0 ( exp ( ξ ^ ) p ) ξ ^ δ ξ δ ξ = ( exp ( ξ ^ ) p ) ^ ,
where ^ denotes the mapping from exp ( ξ ^ ) p represented by [ x , y , z ] T to the corresponding anti-symmetric matrix 0 z y z 0 x y x 0 .

3.3. Pose Graph Optimization

Assume that there are M RGB-D cameras in a panoramic 3D vision system. There is an overlapping FoV between each pair of adjacent RGB-D camera. Let T i , j denote the estimated relative poses between two adjacent RGB-D cameras, i = 1 , 2 , , M ; j = i + 1 , when i equals M, j is 1.
When the position of the first RGB-D camera x 1 is set to [ 0 , 0 , 0 , 1 ] T , we can calculate the position x i for each RGB-D camera based on T i , j . Due to the pose estimation error, the estimated position of the first RGB-D camera that is calculated by T N , 1 x N is not equal to [ 0 , 0 , 0 , 1 ] T . This accumulated error can be mitigated by the loop-closure constraint provides by the first and last RGB-D camera. In this case, the pose optimization problem can be seen as a small scale simultaneous localization and mapping (SLAM) problem. Estimated poses can be optimized using the pose graph optimization method in SLAM [20,21]. According to the pose graph optimization theory, the problem can be solved by finding the minimum of a function of this form:
x = arg min x ( i , j ) S r i , j T ( x ) Λ i , j r i , j ( x ) ,
where x = x 1 T , , x n T T is a vector of poses, r i , j is the residual of the predicted and observed relative poses between the i-th and j-th node, Λ i , j denotes the measurement information matrix, and S represents the set of edges that connect the nodes.

4. Experimental Results

4.1. System Setup

Figure 2a shows the panoramic RGB-D camera setup that we built to evaluate the proposed calibration method. It consists of 12 Kinect v1 cameras, all Kinects are vertically positioned for a more compact design. The Kinect v1 camera has an angular field of view (FoV) of 43 from the vertical. The overlap FoV of two neighboring cameras is 13 ( ( 43 × 12 360 ) / 12 ) which is only approximately 30 percent of the vertical FoV of each camera.
We also designed a distributed system built on top of a local area network (LAN) to capture the RGB-D frames from all twelve cameras in real time. The distributed capturing system (see Figure 2c) consists of twelve Raspberry Pi single board computers, a gigabit switch and a PC. Raspberry Pi is used to obtain RGB-D frames from each Kinect and send the data to the PC through the LAN using User Datagram Protocol. The PC is used to receive and process the RGB-D frames. Both the depth and color frames were set to a size of 640 × 480 ; the values of a single pixel in the depth frame and color frame have a size of 2 bytes and 3 bytes, respectively. Thus, the size of the RGB-D frames from all twelve Kinects is 17.58 million bytes, which can be sent to the PC through the gigabit switch at 7 fps. This frame rate can be increased by decreasing the size of RGB-D frames from Kinect, such as 28 fps when the frame size is set to 320 × 240 .
To evaluate the accuracy of the proposed extrinsic calibration method, we used a motion capture system to obtain the ground truth relative poses between the cameras. The motion capture system requires at least three reflective markers to track the pose of a rigid body, such as the Kinect and the chessboard in our experiments. We attached four reflective markers to both the Kinect and the chessboard (see Figure 5). We placed four markers on the outer corners of the chessboard (see Figure 5b) such that the relative poses between the chessboard and the motion capture system and the Kinect color camera were known. The motion capture system tracked the poses of the markers that were attached to the Kinect to determine the poses of the Kinect color camera. For convenience, we only attached markers to one Kinect and placed this Kinect at twelve different positions in the camera rig to capture RGB-D frames. The accuracy of extrinsic calibration was evaluated using twelve pairs of RGB-D frames from the captured twelve RGB-D frames. The intrinsic parameter of the Kinect color camera is obtained through the Kinect SDK, the depth frame is also aligned with the color image efficiently using an interface provides by the Kinect SDK. The depth maps and color images from Kinects are not undistorted as the distortion is little for a camera with a relatively small FoV of 43 . We also compared the pose estimation error with and without undistorting the depth map and color image in Section 4.3.

4.2. Evaluation Metrics

Given M pairs of 3D point sets p i j , { p i , j } , i = 1 , M , j = 1 , N , a sequence of relative poses P 1 , , P M S E ( 3 ) are estimated. Three common metrics are used to evaluate the accuracy of pose estimation through evaluating the alignment error of 3D point clouds and 2D images. The estimated poses are also compared with the ground truth that are obtained through a motion capture system.

4.2.1. Relative 2D Error (R2E)

The relative 2D error is the reprojection error that measures the geometric error corresponding to the image distance. Let K denotes the projection matrix, R2E is given by
e i j = K p i j K P i p i j , R 2 E = 1 M N i = 1 M j = 1 N ( e i j . x ) 2 + ( e i j . y ) 2 ,
where e i j . x and e i j . y represent the pixel error of x-coordinate and y-coordinate respectively.

4.2.2. Relative 3D Error (R3E)

The relative 3D error measures the alignment accuracy of each pairs of 3D point sets after calibration. R3E is calculated as
e i j = p i j P i p i j , R 3 E = 1 M N i = 1 M j = 1 N ( e i j . x ) 2 + ( e i j . y ) 2 + ( e i j . z ) 2 ,
where e i j . x , e i j . y and e i j . z represent the 3D error of x-coordinate, y-coordinate and z-coordinate respectively.

4.2.3. Accumulated 3D Error (A3E)

In a camera system with multiple RGB-D cameras, take the coordinate system of one RGB-D camera as the reference, the pose estimation error will be accumulated while constructing the whole 3D point cloud. For each pair of 3D point sets from camera i, i + 1 , (when i equals M, i + 1 is set to 1), the accumulated 3D error is defined as follows:
e i j = P 1 P i p i j P 1 P i + 1 p i j , A 3 E = 1 M N i = 1 M j = 1 N ( e i j . x ) 2 + ( e i j . y ) 2 + ( e i j . z ) 2 .

4.3. Results

While extracting the matched keypoints for each pair of RGB-D frames, we increased the distance threshold of the matched keypoints (denoted by d i s t _ t h r e s h ) from two times the minimum distance ( m i n _ d i s t ) to the maximum distance (ten times the minimum distance on average), increased the re-projection error threshold to increase the number of matched keypoints to analyze the pose estimation error with respect to the number of keypoints correspondences.
The relative 2D error (R2E) and relative 3D error (R3E) with respect to d i s t _ t h r e s h / m i n _ d i s t and re-projection error are shown in Table 1 and Table 2 respectively. From Table 1 we can see that, the number of matched keypoints before applying the RANSAC-based filtering method are dramatically increased by increasing d i s t _ t h r e s h / m i n _ d i s t . However, after applying the RANSAC-based filtering method the number of matched keypoints are only slightly increased. The R2E and R3E are also slightly increased while d i s t _ t h r e s h / m i n _ d i s t is increased. It can be seen from Table 2 that, increasing the re-projection error threshold will select more matched keypoints, R2E and R3E will also be increased. When d i s t _ t h r e s h / m i n _ d i s t is set to 2.0, re-projection error threshold is set to 0.1, the optimum value of R2E and R3E are 1.0 pixel and 4.0 mm respectively (see Table 3). When the color image and depth map are undistorted, optimum value of R2E and R3E are 1.4 pixel and 3.8 mm respectively which show slight difference. Because the depth and color camera have a small FoV (43 ), the distortion of depth map and color image is little. When the SURF detector is used, R2E and R3E are 1.14 pixel, and 4.41 mm respectively.
When the estimated poses are compared with the ground truth obtained through the motion capture system, our proposed calibration method resulted in a rotation error of 0.56 degree and a translation error of 1.80 cm. As to the accumulated 3D error, it is decreased from 100 mm to 33 mm after applying the pose graph optimization. The aligned point clouds before and after pose graph optimization with loop closure can also be observed in Figure 6. It can be seen in the red circle on the bottom of Figure 6a that the point clouds were misaligned; however, these point clouds aligned well in Figure 6b with optimized poses with the loop closure constraint.
The whole calibration process costs average 666 ms to calibrate a pair of RGB-D cameras, most of which was spent on extracting and matching feature points. When the SURF detector is used, it costs 109 ms to calibrate a pair of RGB-D cameras. The method is tested on a computer with a 3.60 GHz CPU.
In comparison, Fernández-Moral et al. proposed to calibrate a similar panoramic RGB-D camera setup with plane patterns, they got a ICP residual (relative 3D error (R3E)) of 2.9 mm [9]. However, the setup was required to be moved around to take about 200 images in order to extract enough planes for all camera pairs to estimated poses, which took too much time. Other approaches that are based on per-camera odometry [15,16] requires to generate useful trajectory for each camera which costs even much more time.
Note that, when the proposed method is applied in the practical usage, the interference between neighboring Kinects that will introduce more noises to the overlapping regions in the depth maps should be avoided. Take the calibration of the setup in Figure 2a as an example, the depth maps and color images from Kinects with sequence numbers 1,3,5,7,9,11 (the twelve Kinects are assigned to sequence numbers from 1 to 12 in order) can be firstly captured and then these Kinects are closed and source images from other Kinects are captured.
The operating range of the depth camera in Kinect v1 is between 0.5 m to 5.0 m [22], a single panoramic RGB-D frame from the panoramic RGB-D camera setup can reconstruct scenes within the circle with a radius of 5 m. Figure 7 presents the reconstruction result of a bedroom (3.3 m × 3 m) and a living room (9 m × 3.5 m) using only one output panoramic frame from the system. The proposed panoramic RGB-D camera setup together with the distributed capturing system provide a 360 FoV to perceive the color and depth information of surrounding things in real time. Compared with typical LIDAR systems, the proposed panoramic RGB-D camera system can provide a much denser point cloud with color information (see Figure 7) which makes it more suitable for some applications in indoor such as 3D reconstruction; mapping; navigation, path planning and avoid obstacles for robots etc.
The panoramic RGB-D camera system provides a 360 FoV, leading to better constraints for localization and can be potentially used to reduce localization and mapping errors.
Our next step is to investigate direct registration methods such as Refs. [23,24], which do not depend on time-consuming keypoint detectors or descriptors for large scale SLAM, using this panoramic 3D vision system.

5. Conclusions

In this paper, a new method that relies on well-matched keypoints provided by a feature descriptor-based calibration pattern was proposed to calibrate the extrinsic parameters of the RGB-D cameras in the system. A LAN-based distributed system was developed, which enabled the system to provide panoramic RGB-D frames in real time. The reconstruction of indoor scenes was efficiently and conveniently performed using the panoramic RGB-D 3D vision system. The experiments validated the accuracy and efficiency of the proposed calibration method and the efficiency of the panoramic RGB-D camera setup in 3D reconstruction, and quantitatively demonstrated a higher speed and competitive accuracy compared with existing methods.

Author Contributions

Conceptualization and methodology, H.L. (Hang Liu); Software, H.L. (Hang Liu) and X.L.; Supervision, H.L. (Hengyu Li) and Y.S.; Project administration and funding acquisition, H.L. (Hengyu Li), J.L., S.X. and Y.S.

Funding

This research was funded by the National Natural Science Foundation of China (grant numbers 61525305 and 61625304), the Shanghai Natural Science Foundation (grant numbers 17ZR1409700 and 18ZR1415300), the basic research project of Shanghai Municipal Science and Technology Commission (grant number 16JC1400900), and the Shandong Provincial Natural Science Foundation (grant number ZR2017BF021).

Acknowledgments

We thank Min Li and Zhihong Huang for assistance in constructing and testing the motion capture system, Xiang Gao et al. for the easily understandable explanation about SLAM in Ref. [19].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Salas-Moreno, R.F.; Newcombe, R.A.; Strasdat, H.; Kelly, P.H.J.; Davison, A.J. SLAM++: Simultaneous Localisation and Mapping at the Level of Objects. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  2. Newcombe, R.A.; Izadi, S.; Hilliges, O.; Molyneaux, D.; Kim, D.; Davison, A.J.; Kohi, P.; Shotton, J.; Hodges, S.; Fitzgibbon, A. KinectFusion: Real-time dense surface mapping and tracking. In Proceedings of the 2011 10th IEEE International Symposium on Mixed and Augmented Reality, Basel, Switzerland, 26–29 October 2011; pp. 127–136. [Google Scholar]
  3. Mur-Artal, R.; Tardos, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef] [Green Version]
  4. Whelan, T.; Salas-Moreno, R.F.; Glocker, B.; Davison, A.J.; Leutenegger, S. ElasticFusion: Real-time dense SLAM and light source estimation. Int. J. Robot. Res. 2016, 35, 1697–1716. [Google Scholar] [CrossRef] [Green Version]
  5. Endres, F.; Hess, J.; Sturm, J.; Cremers, D.; Burgard, W. 3-D Mapping With an RGB-D Camera. IEEE Trans. Robot. 2017, 30, 177–187. [Google Scholar] [CrossRef]
  6. Yan, Z.; Mao, Y.; Liu, R. Dense Visual SLAM with Probabilistic Surfel Map. IEEE Trans. Vis. Comput. Gr. 2017, 23, 2389–2398. [Google Scholar] [CrossRef] [PubMed]
  7. Sturm, J.; Engelhard, N.; Endres, F.; Burgard, W.; Cremers, D. A Benchmark for the Evaluation of RGB-D SLAM Systems. In Proceedings of the International Conference on Intelligent Robot Systems (IROS), Algarve, Portugal, 7–12 October 2012. [Google Scholar]
  8. Zug, S.; Penzlin, F.; Dietrich, A.; Nguyen, T.T.; Albert, S. Are laser scanners replaceable by Kinect sensors in robotic applications? In Proceedings of the 2012 IEEE International Symposium on Robotic and Sensors Environments (ROSE), Magdeburg, Germany, 16–18 November 2012; pp. 144–149. [Google Scholar]
  9. Fernandez-Moral, E.; Gonzalez-Jimenez, J.; Rives, P.; Arevalo, V. Extrinsic calibration of a set of range cameras in 5 seconds without pattern. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 429–435. [Google Scholar]
  10. Rives, P. Scene structure registration for localization and mapping. Robot. Autonomous Syst. 2016, 75, 649–660. [Google Scholar] [Green Version]
  11. Gokhool, T.; Meilland, M.; Rives, P.; Fernández-Moral, E. A dense map building approach from spherical RGBD images. In Proceedings of the International Conference on Computer Vision Theory and Applications, Lisbon, Portugal, 5–8 January 2014; pp. 656–663. [Google Scholar]
  12. Gokhool, T.; Martins, R.; Rives, P.; Despre, N. A compact spherical RGBD keyframe-based representation. In Proceedings of the IEEE International Conference on Robotics and Automation, Seattle, WA, USA, 25–30 May 2015; pp. 4273–4278. [Google Scholar]
  13. Poreba, M.; Goulette, F. A robust linear feature-based procedure for automated registration of point clouds. Sensors 2015, 15, 1435–1457. [Google Scholar] [CrossRef] [PubMed]
  14. Perez-Yus, A.; Fernandez-Moral, E.; Lopez-Nicolas, G.; Guerrero, J.J.; Rives, P. Extrinsic calibration of multiple RGB-D cameras from line observations. IEEE Robot. Autom. Lett. 2018, 3, 273–280. [Google Scholar] [CrossRef]
  15. Schneider, S.; Luettel, T.; Wuensche, H.J. Odometry-based online extrinsic sensor calibration. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 1287–1292. [Google Scholar]
  16. Brookshire, J.; Teller, S.J. Extrinsic Calibration from Per-Sensor Egomotion. Robot. Sci. Syst. 2012, 8, 504–512. [Google Scholar]
  17. Li, B.; Heng, L.; Koser, K.; Pollefeys, M. A multiple-camera system calibration toolbox using a feature descriptor-based calibration pattern. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 1301–1307. [Google Scholar]
  18. Juan Luo, G.O. A comparison of sift, pca-sift and surf. Int. J. Image Process. 2009, 3, 143–152. [Google Scholar]
  19. Gao, X.; Zhang, T.; Liu, Y.; Yan, Q. 14 Lectures on Visual SLAM: From Theory to Practice; Publishing House of Electronics Industry: Beijing, China, 2017. [Google Scholar]
  20. Donghwa, L.; Hyun, M. Solution to the SLAM problem in low dynamic environments using a pose graph and an RGB-D sensor. Sensors 2014, 14, 12467–12496. [Google Scholar]
  21. Kummerle, R.; Grisetti, G.; Strasdat, H.; Konolige, K.; Burgard, W. g2o: A General Framework for Graph Optimization. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011. [Google Scholar]
  22. Khoshelham, K.; Elberink, S.O. Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications. Sensors 2012, 12, 1437–1454. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Engel, J.; Koltun, V.; Cremers, D. Direct sparse odometry. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 611–625. [Google Scholar] [CrossRef] [PubMed]
  24. Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 15–22. [Google Scholar]
Figure 1. Flow chart of the proposed calibration method.
Figure 1. Flow chart of the proposed calibration method.
Sensors 19 00349 g001
Figure 2. (a) The setup used in this study for evaluating the proposed calibration method composed of 12 Kinect v1 RGB-D cameras. (b) The panoramic 3D color point cloud obtained by the panoramic 3D vision system. Descriptor-based calibration patterns are pasted on the walls to calibrate multiple RGB-D cameras. (c) Distributed capturing system consisting of a gigabit switch and 12 low-cost Raspberry Pi single board computers.
Figure 2. (a) The setup used in this study for evaluating the proposed calibration method composed of 12 Kinect v1 RGB-D cameras. (b) The panoramic 3D color point cloud obtained by the panoramic 3D vision system. Descriptor-based calibration patterns are pasted on the walls to calibrate multiple RGB-D cameras. (c) Distributed capturing system consisting of a gigabit switch and 12 low-cost Raspberry Pi single board computers.
Sensors 19 00349 g002
Figure 3. The images depicted in (a,b) show the detected keypoints of the image captured in a natural scene, and the image captured in a scene with descriptor-based patterns respectively. Keypoints are represented by green dots.
Figure 3. The images depicted in (a,b) show the detected keypoints of the image captured in a natural scene, and the image captured in a scene with descriptor-based patterns respectively. Keypoints are represented by green dots.
Sensors 19 00349 g003
Figure 4. (a) Shows the matched keypoints of the image pairs captured in a natural scene. (bd) shows the matched keypoints of the image pairs captured in a scene with descriptor-based calibration patterns. The keypoints are detected in the whole image in (b), and in the image regions within blue rectangles in (c,d). The matched keypoints in (ac) are accepted if its descriptor distance is below a threshold, (d) shows the matched keypoints when the RANSAC-based method is applied to refine the matched keypoints in (c). Well-matched keypoints are connected with green lines, poorly matched keypoints are connected with red lines.
Figure 4. (a) Shows the matched keypoints of the image pairs captured in a natural scene. (bd) shows the matched keypoints of the image pairs captured in a scene with descriptor-based calibration patterns. The keypoints are detected in the whole image in (b), and in the image regions within blue rectangles in (c,d). The matched keypoints in (ac) are accepted if its descriptor distance is below a threshold, (d) shows the matched keypoints when the RANSAC-based method is applied to refine the matched keypoints in (c). Well-matched keypoints are connected with green lines, poorly matched keypoints are connected with red lines.
Sensors 19 00349 g004
Figure 5. (a) Kinect with reflective markers. (b) Chessboard with reflective markers.
Figure 5. (a) Kinect with reflective markers. (b) Chessboard with reflective markers.
Sensors 19 00349 g005
Figure 6. The aligned point clouds before (a) and after (b) closing the loop. A misalignment can be observed in the red circle on the bottom of (a). After pose graph optimization with loop closure, the misalignment is resolved.
Figure 6. The aligned point clouds before (a) and after (b) closing the loop. A misalignment can be observed in the red circle on the bottom of (a). After pose graph optimization with loop closure, the misalignment is resolved.
Sensors 19 00349 g006
Figure 7. Reconstruction of a bedroom (3.3 m × 3 m) (a), and a living room (9 m × 3.5 m) (b) using only one output frame from the proposed panoramic 3D vision system. (c) Screenshot of the indoor point cloud generated from Velodyne VLP 16 (https://www.youtube.com/watch?v=Se5U2ne5eLk).
Figure 7. Reconstruction of a bedroom (3.3 m × 3 m) (a), and a living room (9 m × 3.5 m) (b) using only one output frame from the proposed panoramic 3D vision system. (c) Screenshot of the indoor point cloud generated from Velodyne VLP 16 (https://www.youtube.com/watch?v=Se5U2ne5eLk).
Sensors 19 00349 g007
Table 1. Relative 2D error (R2E) and Relative 3D error (R3E) of estimated poses with respect to the increasing distance threshold of matched keypoints. The re-projection error threshold in RANSAC is set to 0.1 pixel. The minimum errors of R2E and R3E are in bold font.
Table 1. Relative 2D error (R2E) and Relative 3D error (R3E) of estimated poses with respect to the increasing distance threshold of matched keypoints. The re-projection error threshold in RANSAC is set to 0.1 pixel. The minimum errors of R2E and R3E are in bold font.
dist _ thresh / min _ dist 2345678910
No. of matched keypoints before RANSAC6223541665210641316137914641535
No. of matched keypoints after RANSAC111921222122212121
R2E (pixel)1.101.081.071.071.191.171.171.171.15
R3E (mm)4.015.054.675.584.965.115.325.215.26
Table 2. Relative 2D error (R2E) and Relative 3D error (R3E) of estimated poses with respect to the increasing re-projection threshold in RANSAC-based keypoints refinement method. d i s t _ t h r e s h / m i n _ d i s t is set to 2.0. The minimum errors of R2E and R3E are in bold font.
Table 2. Relative 2D error (R2E) and Relative 3D error (R3E) of estimated poses with respect to the increasing re-projection threshold in RANSAC-based keypoints refinement method. d i s t _ t h r e s h / m i n _ d i s t is set to 2.0. The minimum errors of R2E and R3E are in bold font.
Re-Projection Error Threshold0.10.20.51.01.52.0
No. of matched keypoints before RANSAC626262626262
No. of matched keypoints after RANSAC111938475252
R2E (pixel)1.100.961.021.031.061.06
R3E (mm)4.014.685.035.265.335.33
Table 3. Relative 2D error (R2E) and Relative 3D error (R3E) of estimated poses of 12 pairs of RGB-D cameras in Figure 2a when d i s t _ t h r e s h / m i n _ d i s t is set to 2.0, re-projection error is set to 0.1 pixel. “Before” and “After” indicate the number of matched keypoints before and after applying the RANSAC-based filtering method respectively.
Table 3. Relative 2D error (R2E) and Relative 3D error (R3E) of estimated poses of 12 pairs of RGB-D cameras in Figure 2a when d i s t _ t h r e s h / m i n _ d i s t is set to 2.0, re-projection error is set to 0.1 pixel. “Before” and “After” indicate the number of matched keypoints before and after applying the RANSAC-based filtering method respectively.
IndexBeforeAfterR2E (pixel)R3E (mm)
183131.238.6
2119180.847.7
33270.782.1
444120.945.1
588120.925.0
62790.792.6
765101.051.4
87032.390.8
980121.013.5
1057120.952.9
111251.274.1
1264160.983.8
Average62111.13.97

Share and Cite

MDPI and ACS Style

Liu, H.; Li, H.; Liu, X.; Luo, J.; Xie, S.; Sun, Y. A Novel Method for Extrinsic Calibration of Multiple RGB-D Cameras Using Descriptor-Based Patterns. Sensors 2019, 19, 349. https://doi.org/10.3390/s19020349

AMA Style

Liu H, Li H, Liu X, Luo J, Xie S, Sun Y. A Novel Method for Extrinsic Calibration of Multiple RGB-D Cameras Using Descriptor-Based Patterns. Sensors. 2019; 19(2):349. https://doi.org/10.3390/s19020349

Chicago/Turabian Style

Liu, Hang, Hengyu Li, Xiahua Liu, Jun Luo, Shaorong Xie, and Yu Sun. 2019. "A Novel Method for Extrinsic Calibration of Multiple RGB-D Cameras Using Descriptor-Based Patterns" Sensors 19, no. 2: 349. https://doi.org/10.3390/s19020349

APA Style

Liu, H., Li, H., Liu, X., Luo, J., Xie, S., & Sun, Y. (2019). A Novel Method for Extrinsic Calibration of Multiple RGB-D Cameras Using Descriptor-Based Patterns. Sensors, 19(2), 349. https://doi.org/10.3390/s19020349

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop