Next Article in Journal
Quantifying the Variability of Ground Light Sources and Their Relationships with Spaceborne Observations of Night Lights Using Multidirectional and Multispectral Measurements
Previous Article in Journal
Robotic-Assisted Rehabilitation for Post-Stroke Shoulder Pain: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Systematic Stereo Camera Calibration Strategy: Leveraging Latin Hypercube Sampling and 2k Full-Factorial Design of Experiment Methods

1
Department of Electronic Engineering, Taiyuan Institute of Technology, Taiyuan 030008, China
2
Faculty of Engineering, Built Environment and Information Technology, SEGI University, Petaling Jaya 47810, Malaysia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(19), 8240; https://doi.org/10.3390/s23198240
Submission received: 11 September 2023 / Revised: 26 September 2023 / Accepted: 27 September 2023 / Published: 3 October 2023
(This article belongs to the Section Physical Sensors)

Abstract

:
This research aimed to optimize the camera calibration process by identifying the optimal distance and angle for capturing checkered board images, with a specific focus on understanding the factors that influence the reprojection error ( ϵ R P ). The objective was to improve calibration efficiency by exploring the impacts of distance and orientation factors and the feasibility of independently manipulating these factors. The study employed Zhang’s camera calibration method, along with the 2k full-factorial analysis method and the Latin Hypercube Sampling (LHS) method, to identify the optimal calibration parameters. Three calibration methods were devised: calibration with distance factors (D, H, V), orientation factors (R, P, Y), and the combined two influential factors from both sets of factors. The calibration study was carried out with three different stereo cameras. The results indicate that D is the most influential factor, while H and V are nearly equally influential for method A; P and R are the two most influential orientation factors for method B. Compared to Zhang’s method alone, on average, methods A, B, and C reduce ϵ R P by 25%, 24%, and 34%, respectively. However, method C requires about 10% more calibration images than methods A and B combined. For applications where lower value of ϵ R P is required, method C is recommended. This study provides valuable insights into the factors affecting ϵ R P in calibration processes. The proposed methods can be used to improve the calibration accuracy for stereo cameras for the applications in object detection and ranging. The findings expand our understanding of camera calibration, particularly the influence of distance and orientation factors, making significant contributions to camera calibration procedures.

1. Introduction

Binocular or stereo vision, used in diverse areas like autonomous driving [1,2], robot navigation [3,4], 3D reconstruction [5,6,7], and industrial inspection [8,9], employs two identical cameras to capture two images of the same target from different vantage points. The disparity between these images enables 3D reconstruction and measurement, the accuracy of which is heavily dependent on the precision of the camera calibration and stereo matching methods, thus making these processes key research subjects in stereo vision. During the camera calibration process, it is crucial to estimate both the intrinsic parameters, such as the focal length and optical centers, and the extrinsic parameters, like the camera’s position and orientation in space. These directly influence the calculation of 3D spatial information [10]. Camera calibration significantly impacts the accuracy of the reconstruction and measurement, thereby playing a crucial role in binocular vision.
Camera calibration techniques mainly fall into two categories: self-calibration and photogrammetric calibration [11]. Self-calibration methods, though not needing a calibration object, rely on sequences of images from an uncalibrated scene and often entail complex computations [12,13,14] and low in accuracy [5,15]. Conversely, photogrammetric methods, including the popular methods such as Zhang’s [16], Tsai’s [17], Bouguet’s [18], and Heikkila and Silven’s [19] methods utilize geometric information from a calibration object, such as a checkered board with known geometry, to determine camera parameters. The latter methods minimize ϵ R P through feature point detection and optimization, providing more accurate information about camera characteristics.
Zhang’s camera calibration method [16] has gained significant attention due to its simplicity, robustness, and high degree of accuracy. One of the key advantages of Zhang’s method is that it only requires a simple calibration object, such as a checkered board, which is easy to fabricate and measure. This stands in contrast to other photogrammetric methods that may require specialized calibration objects or more complicated patterns. The method is based on the fundamental matrix theory, which determines the intrinsic (i.e., focal length, lens distortion coefficients, and the position of the image center) and extrinsic (i.e., camera position and orientation) parameters of a camera by measuring the geometric relationship between the camera and the calibration board. The method requires at least three pairs of images of a checkered board with known dimensions taken from different positions and orientations within the camera’s field of view (FOV). Then, a corner detection algorithm is employed to detect the corners in each image, i.e., the intersections between the black and white squares, and the image coordinates of the detected corners are matched with their known 3D coordinates in the checkered board. Subsequently, least-squares minimization, i.e., the Levenberg–Marquardt algorithm, is used to search for determining the camera’s intrinsic and extrinsic parameters that yield minimum geometric error between the projected 3D calibration points and their corresponding 2D image points. The lens distortion coefficients, such as radial and tangential distortions, are also estimated and corrected.
Owing to the simplicity, flexibility and robustness of Zhang’s method, many researchers have developed new camera calibration algorithms that incorporate Zhang’s method to improve calibration accuracy. Rathnayaka et al., 2017, proposed two new calibration methods that use a specially designed, colored checkered board pattern and mathematical pose estimation approaches to calibrate a stereo camera setup with heterogeneous lenses, i.e., a wide-angle fish-eye lens and a narrow-angle lens [11]. One method calculates left and right transformation matrices for stereo calibration, while the other utilizes a planar homography relationship and Zhang’s calibration method to identify common object points in stereo images. Evaluations based on ϵ R P and image rectification results showed both methods successfully overcome the challenges, with the planar homography approach yielding slightly more accurate results. Wang et al., 2018, proposed a non-iterative self-calibration algorithm to compute the extrinsic parameters of a rotating (yaw and pitch only) and a non-zooming camera [13]. The algorithm utilizes a single feature point to improve the calibration speed and Zhang’s method was employed in advance to calculate the intrinsic parameters of the camera. Although their results are slightly less accurate than those calibrated by Zhang’s method alone, their work demonstrates its potential for real-time applications where extremely high calibration accuracy is not required. Hu et al., 2020, introduced a calibration approach for binocular cameras that begins with Zhang’s technique to obtain an initial guess of the rotation and translation matrix [5]. This guess is then refined using singular value decomposition and solved via the Frobenius norm, and is further refined through maximum likelihood estimation, generating a new calculation for the relative position matrix of the binocular cameras. The Levenberg–Marquardt algorithm is employed for additional refinement. Their results show significant improvement over the traditional Bouget’s and Hartley’s algorithms. Liu et al., 2022, proposed a multi-camera calibration stereo calibration method focusing on the high-precision extraction of circular calibration pattern features during the calibration process [20]. The Franklin matrix, used for sub-pixel edge detection and the image moment method, was used to obtain the characteristic circle center, while the calibration was realized using Zhang’s method. Their work shows that calibration effects can be improved by focusing on the high-precision extraction of calibration features.
Calibration targets also play a significant role in camera calibration. Although using a checkered board pattern as the calibration target does have the advantages, such as simplicity, stability, and a generally higher accuracy than self-calibration, it has the downside of being susceptible to illumination and noise in corner feature extraction and matching. Circular patterns have been proposed to address the problem [21]. Recently, Wei et al., 2023, presented a camera calibration method using a circular calibration target [22]. Their method operates in two stages. The first stage involves extracting the centers of the ellipses and formulating the real concentric circle center projection equation by leveraging the cross-ratio invariance of collinear points. The second stage involves solving a system of linear equations to find the other center projection coordinates, taking advantage of the fact that the infinite lines passing through the centers are parallel. Unlike other circular calibration algorithms that directly fit ellipses to estimate the centers of circular feature points, their method not only corrects errors that arise from using the ellipse’s center but also significantly improves the accuracy and efficiency of the calibration process. The proposed method reduces the average ϵ R P to 0.08 pixels and achieves real-world measurement accuracy of 0.03 mm, making it valuable for high-precision 3D measurement tasks. Recently, Zhang et al., 2023, proposed a flexible calibration method for large-range binocular vision systems for complex construction environments [23]. The calibration method consists of two stages. In stage 1, lenses are focused closer to the cameras, and non-overlapping fields of view (FOVs) are calibrated using virtual projection point pairs through Zhang’s method and nonlinear optimization. Fourteen sets of clear checkered board images are acquired for this purpose. Then, the lenses are focused on the measurement position in stage 2, where three sets of blurred three-phase shifted circular grating (PCG) images are taken. The final extrinsic parameters for the binocular camera are computed using state transformation matrices for calibration. The method has demonstrated robustness under varying target angles and calibration distances, with a mean ϵ R P of only 0.0675 pixels and a relative measurement error of less than 1.75% in their experiment.
While Zhang’s method ensures accurate calibration outcomes when viewing a planar target from various angles, the selection of these differing orientations relies largely on empirical knowledge, which could potentially lead to variations in the calibration results [24]. Furthermore, it is often necessary to manually adjust the calibration object’s position, a process that introduces uncertainties due to instability and shaking, resulting in inconsistent positioning of the calibration plate across images captured by different cameras, and thus introducing inherent errors that directly affect the calibration process’s accuracy [15]. Moreover, the calibration procedure requires a significant amount of time and effort [13].
The present study seeks to address numerous key issues related to camera calibration in the process of capturing checkered board images. In particular, it aims to optimize this process by identifying the optimal distance and angle for calibration, consequently enhancing the calibration results. Different from the existing literatures, this study places emphasis on gaining a comprehensive understanding of the factors that affect ϵ R P in the calibration process, specifically the roles of various displacement and orientation factors, and how they might be manipulated independently for better calibration efficiency. This leads to four principal research questions which revolve around understanding the impacts of distance factors (D, H, V) and orientation factors (R, P, Y) on ϵ R P , determining the effects of combined distance and orientation factors on ϵ R P , and exploring the feasibility and efficiency of conducting the calibration process separately for different factor groups. To address these, the research objectives include identifying the most influential distance and orientation factors, investigating the combinations of factors that minimize ϵ R P , and evaluating the efficiency of separately conducting the calibration process for different factor groups. The overarching goal is to develop more efficient and effective calibration procedures for various applications.

2. Methodology

2.1. Experiment Setup

Three different cameras are used in this study. The details of the cameras are presented in Table 1. However, only the calibration process of the ZED2i binocular camera [25] is presented in this study.
For camera calibration, a checkered board consisting of 2 cm alternating black and white squares arranged in a grid pattern of nine columns and seven rows is employed, as shown in Figure 1b. The checkered board is attached to the camera calibration rig as shown in Figure 1c. The Stereo Camera Calibration Toolbox for MATLAB [18] is used for calibration, where the inner corner points are detected using the Harris corner algorithm [26]. The checkered board size is fed to the algorithm and the actual number of detected corner points is 48. The high contrast between black and white aids in accurately detecting and localizing the inner corners of the squares, which are crucial for camera calibration algorithms.
Since binocular camera calibration necessitates capturing images from various directions and angles, a camera calibration rig was designed to securely hold the checkered board, ensuring precise distances and orientations for calibration purposes. The components of the calibration rig and its final assembly are depicted in Figure 2. The rig comprises four main components: an LD2R dual panoramic gimbal for yaw (Y) and pitch (P) adjustments; a two-way macro frame for left and right (H) and depth, i.e., front and back (D) displacements; a lift table for vertical (V) movements; and an RSP60R axis rotation stage for rotation (R). These adjustments ensure that the checkered board is distributed across different quadrants and acquired from various positions and angles.

2.2. Zhang’s Camera Calibration

Zhang’s calibration method [16] was used in this study to perform the calibration task due its simplicity and robustness. Mathematically, given the point P = X w ,   Y w ,   Z w T represents the coordinates of each corner point on the checkered board in the world coordinate system, its homogeneous coordinate is P ~ = X w ,   Y w ,   Z w ,   1 . In the pixel coordinate system, this point is denoted by m = u ,   v T , while its homogeneous coordinate is p ~ = u ,   v ,   1 T . The relationship between the world coordinate and the image coordinate is described by the camera model, as follows:
λ p ~ = A R t P ~
where λ is the proportionality coefficient and [ R | t ] and A are the extrinsic and intrinsic parameter matrices of the camera, respectively. By assuming that Z = 0 for the planar board, the equation can be further simplified:
λ p ~ = A r 1   r 2   t X w Y w 1 = H P ~
where H is a 3 × 3 homography matrix, defined as follows:
H = h 1   h 2   h 3 = s A [ r 1   r 2   t ]
and s is a proportional factor. The homography matrix describes the transformation between two images of the same planar scene. Before calculating the homography matrix for camera calibration using a checkered board, it is necessary to configure the parameters for the world coordinates of the checkered board. This includes details such as the board’s overall size, the size of each square, and the number of detected corner points on the checkered board.
The fundamental constraints for the camera’s intrinsic parameters are as follows:
h 1 T A T A 1 h 2 = 0 h 1 T A T A 1 h 1 = h 2 T A T A 1 h 2 = 1
By using the properties of the rotation matrix R , the unit orthogonal vectors r 1 and r 2 can be obtained. The camera intrinsic parameters can be determined by solving A T A 1 , using Equation (4).
ϵ R P is the error between the reprojected point, obtained by projecting the calibrated image point onto the camera coordinate system and its actual coordinate. This metric is used to assess the accuracy of camera calibration results in this study. As shown in Figure 3, after calibrating the camera using its intrinsic and extrinsic parameters, the point P is reprojected onto the image plane, resulting in a new point P , with pixel coordinates ( u ,   v ) . The calibration ϵ R P is calculated as the Euclidean distance between the reprojected pixel coordinates ( u ,   v ) and the actual calibrated pixel coordinates ( u ,   v ) . The formula for the ϵ R P is given by Equation (5).
ϵ R P = u u 2 + ( v v ) 2

2.3. Design of Experiment (DoE)

A 2k full factorial with replication design of experiment method [27] is employed to analyze the influence of each factor and their interactions on ϵ R P in the camera calibration. A 2k full-factorial DoE is a type of experimental design used to study the effect of k independent factors at 2 levels each (i.e., the low and high bounds) on a response variable. This allows for all possible combinations of the factors at all levels to be investigated, providing a comprehensive view of the entire experimental space. Figure 4 depicts the k = 3 factor design space, where each corner point of the cube represents the combination of the factors X 1 ,   X 2 ,   and X 3 . With replication, i.e., by repeating the entire experiment with more than one operating condition, the errors or uncertainties introduced by the other variables can be estimated and the significance and interactions of the independent factors can be revealed.
The factors involved in this study can be divided into two categories: orientation factors (R, P, and Y) and displacement factors (D, V, H). Given that Zhang’s method requires a minimum of three sets of images for calibration, specific variations must be incorporated into each experimental set. Consequently, the calibration experiment must be designed to exclude some factors to accommodate these variations. In this study, the experiments are configured as follows: Method A employs displacement factors for calibration while introducing variations in orientation. Method B utilizes orientation factors for calibration and introduces variations in displacement. Method C incorporates the two most significant factors from both orientation and displacement categories while introducing variations to the remaining factors.
The bounds for displacement and orientation factors are set at ±3 cm and ±10°, respectively. These displacement boundaries are designed to ensure that the checkered board stays within the field of view (FOV) of the camera, considering the minimum distance between the board and the camera. Upward, leftward, and forward displacements are categorized as positive, while downward, rightward, and backward displacements are considered negative. As for the orientation limits, the specific values are chosen based on experimental data to guarantee that all corners of the images captured by the camera can be detected using Zhang’s method. Positive orientations are defined as up pitch, right yaw, and left roll, whereas negative rotations are classified as down pitch, left yaw, and right roll. The convention for sign determination is presented in Figure 5.
The experiment process starts by generating the 2k full-factorial design points using the factors of interest. For methods A and B, each involves a factor group of three factors with two levels, resulting in a total of 2 3 = 8 combinations for each experiment. On the other hand, for method C, the two most important factors from each factor group are selected to construct the design points, resulting in 2 4 = 16 combinations.
Then, Latin Hypercube Sampling (LHS) is used to select 20 points from the remaining factors to introduce variations for Zhang’s method. LHS is used to ensure that the variations introduced by the remaining factors are well distributed within the multidimensional sampling space, allowing them to be treated as random variables in the analysis. The calibration process is presented in Figure 6.
The interaction effects response model, Y , is built to describe the relationship between the factors and the response variable, i.e., ϵ R P :
Y = β 0 + β 1 X 1 + β 2 X 2 + + β k X k + i < j β i j X i X j + ε
In Equation (6), β 0 is the intercept term, which is the baseline level of the response Y when all factors X 1 , X 2 , , X k are zero. The terms β 1 ,   β 2 , , β k are the model coefficients for the main effects of the factors X 1 , X 2 , , X k , respectively. Each β i represents the change in Y for a unit change in X i when all other factors are held constant. A high absolute value of β i indicates that its corresponding factor X i has a strong influence on Y , and vice versa. The term i < j β i j X i X j denotes the interaction effects between the factors X i and X j . Finally, ε is the random error term, describing the uncertainty or unexplained variability in the model.
Bayesian Information Criterion (BIC) is used to quantify the trade-off between the goodness of fit of the response model and its complexity [28]:
B I C = k ln ( n ) 2 ln ( L ^ )
where L ^ is the maximized value of the likelihood function of the response model, n is the sample size, and k is the number of parameters estimated by the model. The term k l n n is the penalty function for model complexity, to discourage overfitting by keeping the number of k low. The term 2 ln ( L ^ ) is the measure of the goodness of fit of the model to the data, where smaller values indicate a better model fit. In model selection, model with the lowest B I C is considered the best.
Analysis of variance (ANOVA) is used to determine which factors are statistically significant in the response. The R2, R2-adjusted, R2-predicted, and the number of significant factors ( N f ), are used to judge whether additional replication is needed to yield a stable response. In this study, MINITAB [29] is used to generate the factorial points and perform the DoE analysis. The experimental procedure is summarized in Figure 7.

3. Results and Discussion

3.1. Method A: Calibration with Displacement Factors

Method A is designed to calibrate with displacement factors, i.e., D, H, and V, on ϵ R P . LHS was employed to randomly sample 20 points within the range of ± 10° for 3 orientation factors, R, P, and Y. The experiment is conducted by adding one DoE replicate at a time until the R 2 , adjusted R 2 , predicted R 2 values, and the number of significant factors ( N f ) of the response model in the successive replicate are consistent. As shown in Figure 8, only three replicates are needed to reach convergence. An additional replicate was used to verify the results. The total number of photos taken to yield convergence is 3 replicates × 20 LHS samples × 8 factorial points = 480. The resultant hierarchical ϵ R P response model in uncoded units is presented in Equation (8).
ϵ R P = 0.085578 + 0.003441 D + 0.000211 H + 0.000341 V
The model is reduced to a linear regression. Within the limits of the factors, the minimum ϵ R P can be achieved by minimizing the values of the factors. Figure 9a displays the normal plot of the standardized effects of the response model, which confirms that only the three main factors are significant for ϵ R P . Among these factors, D exhibits the most influence on ϵ R P for method A, followed by V and H. Additionally, the normal plot reveals that all three factors have positive effects on the ϵ R P , indicating that an increase in the factor’s level will result in an increase in ϵ R P .
The residual plot in Figure 9b confirms that the residuals of the model follow a normal distribution with zero mean, indicating that the model effectively accounts for the systematic variation in the data, leaving only random variation unexplained. The factorial point where D = H = V = −3 cm exhibits the lowest ϵ R P , with a value of 0.0716 pixels. This result aligns with the predictions made by the linear regression model, where ϵ R P = 0.0733 is minimum for this combination.

3.2. Method B: Calibration with Orientation Factors

Method B calibrates with the three orientation factors, i.e., R, P, and Y on ϵ R P . LHS was employed to randomly sample 20 points within the range of ± 3 cm for 3 displacement factors, D, H, and V. The calibration is conducted by adding one DoE replicate at a time until the R 2 , adjusted R 2 , predicted R 2 values, and the number of significant factors ( N f ) of the response model in the successive replicate are consistent. As shown in Figure 10, six replicates are needed to reach convergence. The total calibration images acquired is 960. The resultant hierarchical ϵ R P response model in uncoded units is presented in Equation (9).
ϵ R P = 0.088156 + 0.000100 R 0.000166 P 0.000056 Y + 0.000014 R P
Within the limits of the factors, the minimum ϵ R P can be achieved by minimizing R and maximizing P and Y. Figure 11a displays the normal plot of the standardized effects of the response model, which confirms that the three main orientation factors and an additional R P interaction factor are significant for ϵ R P . Among these factors, P exhibits the most influence on ϵ R P for method B, while the R P interaction factor ranked the second, followed by R and Y. Additionally, factors P and Y have are inversely correlated with ϵ R P , suggesting an increase in the factor’s level will result in a decrease in ϵ R P . In contrast, an increase in the level of factor R and R P interaction will lead to an increase in ϵ R P .
The residual plot in Figure 11b confirms that the residuals of the model follow a normal distribution with zero mean, indicating that the model effectively accounts for the systematic variation in the data, leaving only random variation unexplained. The factorial point is where R = −10 ° and P = Y = 10 ° exhibits the lowest ϵ R P , with a value of 0.0825 pixels. This result aligns with the predictions made by the linear regression model, where ϵ R P = 0.0835 is minimum for this combination.

3.3. Method C: Calibration with Displacement and Orientation Factors

Differently from methods A and B, which calibrate each factor group on ϵ R P separately, method C calibrates the combined displacement and orientation factors concurrently. The two most prominent factors from each factor group are selected. LHS was employed to randomly sample 20 points from the remaining factors withing their respective range. The experiment is conducted by adding one DoE replicate at a time until the R 2 , adjusted R 2 , predicted R 2 values, and N f of the response model are consistent in the successive replication. As shown in Figure 12, 5 replicates are needed to reach convergence, resulting in a total of 1600 calibration images acquired. The non-hierarchical ϵ R P response model in coded units is presented in Equation (10).
ϵ R P = 0.085931 + 0.003511 R 0.006806 P + 0.011506 D 0.001726 R P + 0.001441   R D 0.001751 R V 0.001671 P D + 0.003841 P V 0.001091 D V 0.000886 R P D + 0.001411 R P V 0.000526 R D V + 0.001251 P D V + 0.000826 R P D V
Figure 13a presents the normal plot of the standardized effects of ϵ R P . Two main orientation factors R and P and a displacement factor D are significant for ϵ R P . Factors P and D are positively correlated to ϵ R P , while R is negatively correlated to ϵ R P . Additionally, eleven interaction factors appear to be significantly influential on ϵ R P for method C.
The residual plot in Figure 13b confirms that the residuals of the model follow a normal distribution with zero mean, indicating that the model effectively accounts for the systematic variation in the data, leaving only random variation unexplained. The minimum ϵ R P obtained from Equation (10) is 0.0622 pixels, and the corresponding R, P, D, and V values are −10°, +10°, −3 cm, and −3 cm, respectively. This result aligns with the measurement, where ϵ R P = 0.0638 pixles is minimum for this combination. Interestingly, the R, P, D, and V values that yield the minimum ϵ R P are also consistent with the values for methods A and B.

3.4. Comparisons of Methods A, B, and C

Two additional cameras, D435i and HBVCAM-W202011HD V33, are used to assess the calibration performance of methods A, B, and C. For method A, which used only displacement factors for camera calibration, it was found that factors D, H, and V are linearly correlated to ϵ R P for all three cameras involved in this study. Factor D is the most influential factor on ϵ R P for all cameras, followed by V and H for the ZED2i camera, and H and V for both D435i and HBVCAM. However, a closer examination on Figure 9a reveals that both H and V factors are equally important. For method B, which calibrates using the orientation factors, it was observed that factors R and P are more influential than Y on ϵ R P for all cameras. Therefore, factors D and H or V from the distance group and factors P and Y from the orientation group are recommended for calibration with method C.
Table 2 compares the calibration results for ZED2i camera using methods A, B, and C. It is evident that combining the two most significant factors, which yield the minimum ϵ R P for methods A and B, respectively, will also result in the minimum ϵ R P for method C. The results also show that a total of 1600 images are needed for calibration with method C, while methods A and B require only 1440 images in total, about 10% less compared to method C.
The focal lengths in the x and y directions ( f x , f y ), the principal point ( c x , c y ), and the radial distortion matrix ( k 1 ,   k 2 ,   k 3 ) of the ZED2i camera are presented in Table 3. The tangential distortion matrix ( p 1 ,   p 2 ) is set to zero in this study.
Zhang’s method is used to benchmark the performance of the calibration results. To improve the reliability of Zhang’s calibration, the calibration is repeated five times, each with 20 sets of images of the randomly positioned calibration target. The average ϵ R P values obtained are 0.0901, 0.0832, and 0.0858 pixels for cameras ZED2i, D435i, and HBVCAM-W202011HD V33, respectively. The results are presented in Figure 14.
Figure 14 shows that methods A, B, and C are able to reduce ϵ R P . For ZED2i, D435i, and HBVCAM cameras, method A is able to reduce their respective ϵ R P by 20%, 38%, and 16% compared to Zhang’s method. For method B, the ϵ R P reductions for the three cameras are 8%, 31%, and 33%. The average ϵ R P reduction by methods A and B are 25% and 24%, respectively, suggesting that both methods are equally effective. On the other hand, method C can produce a more consistent calibration result. The ϵ R P reductions for the three cameras are 31%, 36%, and 34%, respectively.
Figure 15 shows the modeling performance for the three cameras using methods A, B, and C. The results indicate that the proposed 2k full-factorial DoE is effective in modeling ϵ R P , as most of the models are able to achieve the R 2 , adjusted R 2 , and predicted R 2 values of more than 95%, except for method B for the ZED2i camera, where its R 2 , adjusted R 2 , and predicted R 2 values are less than 75%. A closer examination of the ϵ R P residuals for methods A, B, and C (see Figure 9b, Figure 11b and Figure 13b) reveals that that data for method B (Figure 11b) contain high variability. More replications can be carried out to improve its calibration accuracy in this case. However, method B is still able to capture the underlying significant factors and their interactions for ZED2i camera.
Interestingly, the factorial point with D = H = V = −3 cm exhibited the lowest ϵ R P for all cameras calibrated with method A. Similarly, the factorial point with R = −10° and P = Y = 10° resulted in the minimum ϵ R P for all cameras calibrated with method B. For method C, the minimum ϵ R P , obtained by considering the factorial point with R = −10°, P = 10°, and D = V = −3 cm, is consistent with the results from methods A and B. Since the baseline of each camera was used to determine the default depth distance for calibration to ensure the checkered board is within its FOV, it has no impact on ϵ R P .

4. Conclusions

This paper adopts Zhang’s camera calibration method, combined with the 2k full-factorial DoE and LHS methods, to minimize ϵ R P in camera calibration. In general, the results of the experiments conducted for methods A, B, and C provide valuable insights into the factors affecting ϵ R P in the calibration process.
For method A, which used only displacement factors for camera calibration, it was found that factors D, H, and V are linearly correlated to ϵ R P for all three cameras involved in this study. Factor D was found to be the most influential factor on ϵ R P for all cameras, followed by V and H for the ZED2i camera, and H and V for the D435i and HBVCAM cameras. However, a closer examination of Figure 9a reveals that both the H and V factors are equally important. The factorial point with D = H = V = −3 cm exhibited the lowest ϵ R P , confirming the predictions made by the linear regression model. The experiment demonstrated that by minimizing the values of the displacement factors, ϵ R P can be reduced by 25% on average compared to Zhang’s method.
On the other hand, for method B, which calibrates using the orientation factors, it was observed that factors R and P are more influential than Y on ϵ R P for all cameras. The factorial point with R = −10° and P = Y = 10° resulted in the minimum ϵ R P . The findings were consistent with the predictions made by the response models, highlighting the importance of controlling the orientation factors to minimize ϵ R P . On average, method B can reduce the ϵ R P by 24% compared to Zhang’s method.
For method C, where two of the most influential factors from distance and orientation factors were selected for calibration, the minimum ϵ R P was obtained at the factorial point with R = −10°, P = 10°, and D = V = −3 cm; this is consistent with the results from methods A and B. Although the method requires 10% more images compared to methods A and B combined, method C and produce a consistent minimum ϵ R P with a reduction of more than 30% compared to Zhang’s method.
Comparing the results of methods A, B, and C, it is evident that combining the two most significant factors from methods A and B would also result in the minimum ϵ R P for method C. Specifically, the combination of orientation factors R and P and displacement factors D and V or H can achieve the minimum ϵ R P value for method C. Furthermore, it was found that the calibration process could be conducted separately and more efficiently by using the two factor groups independently to achieve more than 20% improvement in ϵ R P compared to Zhang’s method alone. This approach required fewer calibration images, about 70% and 40% fewer for methods A and B, respectively, compared to method C. For applications where a lower value of ϵ R P is required, method C is recommended.
Systematic camera calibration plays an important role in 3D reconstruction as it greatly influences the accuracy and reliability of the resulting 3D models. Accurate calibration ensures the precise determination of both the intrinsic and extrinsic parameters of a camera. This precision enables the correct mapping of 2D image data to real-world 3D coordinates. Optimizing camera calibration enhances the geometric fidelity of 3D reconstruction, paving the way for a wide range of applications in fields like computer vision, robotics, and augmented reality. Thus, it serves as the cornerstone underpinning the entire process of translating 2D images into accurate 3D reconstructions.
Overall, these findings contribute to a better understanding of the factors affecting ϵ R P in calibration processes, enabling more efficient and effective calibration procedures in various applications. The proposed methods can be used to improve the calibration accuracy for stereo cameras for the applications in object detection and ranging. However, the proposed systematic camera calibration experiment involves a substantial amount of repetitive testing and the capture of numerous calibration checkered board photos, which consumes a significant amount of manpower and time resources. Additionally, camera distortions have a pronounced impact on camera calibration as they alter the geometric properties of camera imaging, subsequently affecting the precise mapping relationship between pixel coordinates and real-world coordinates. These two limiting factors will be subjected to in-depth exploration and discussion in future research.

Author Contributions

Conceptualization, Y.H. and V.C.T.; methodology, Y.H.; software, Y.H.; validation, Y.H., V.C.T. and Y.C.T.; formal analysis, Y.H.; investigation, Y.H.; resources, Y.H.; data curation, Y.H.; writing—original draft preparation, Y.H.; writing—review and editing, Y.H.; visualization, Y.H.; supervision, Y.C.T.; project administration, Y.C.T.; funding acquisition, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Scientific and Technological Innovation Programs of Higher Education Institutions in Shanxi, STIP, grant number: 2022L548.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Muhovič, J.; Perš, J. Correcting Decalibration of Stereo Cameras in Self-Driving Vehicles. Sensors 2020, 20, 3241. [Google Scholar] [CrossRef]
  2. Feng, M.; Liu, Y.; Jiang, P.; Wang, J. Object Detection and Localization Based on Binocular Vision for Autonomous Vehicles. J. Phys. Conf. Ser. 2020, 1544, 012134. [Google Scholar] [CrossRef]
  3. Zhang, M.; Cai, W.; Xie, Q.; Xu, S. Binocular-Vision-Based Obstacle Avoidance Design and Experiments Verification for Underwater Quadrocopter Vehicle. J. Mar. Sci. Eng. 2022, 10, 1050. [Google Scholar] [CrossRef]
  4. Yu, X.; Fan, Z.; Wan, H.; He, Y.; Du, J.; Li, N.; Yuan, Z.; Xiao, G. Positioning, Navigation, and Book Accessing/Returning in an Autonomous Library Robot Using Integrated Binocular Vision and QR Code Identification Systems. Sensors 2019, 19, 783. [Google Scholar] [CrossRef]
  5. Hu, G.; Zhou, Z.; Cao, J.; Huang, H. Highly Accurate 3D Reconstruction Based on a Precise and Robust Binocular Camera Calibration Method. IET Image Process. 2020, 14, 3588–3595. [Google Scholar] [CrossRef]
  6. Hao, Y.N.; Tan, Y.C.; Tai, V.C.; Zhang, X.D.; Wei, E.P.; Ng, S.C. Review of Key Technologies for Warehouse 3D Reconstruction. J. Mech. Eng. Sci. 2022, 16, 9142–9156. [Google Scholar] [CrossRef]
  7. Zhong, L.; Qin, J.; Yang, X.; Zhang, X.; Shang, Y.; Zhang, H.; Yu, Q. An Accurate Linear Method for 3D Line Reconstruction for Binocular or Multiple View Stereo Vision. Sensors 2021, 21, 658. [Google Scholar] [CrossRef] [PubMed]
  8. Zhou, G.; Sun, X.; Dong, Q.; Cao, S.; Li, M. Research on Camera Calibration Method for Visual Inspection of Excavator Working Object. J. Phys. Conf. Ser. 2020, 1678, 012022. [Google Scholar] [CrossRef]
  9. Chen, C.; Shen, P. Research on Crack Width Measurement Based on Binocular Vision and Improved DeeplabV3+. Appl. Sci. 2023, 13, 2752. [Google Scholar] [CrossRef]
  10. Zhang, B.; Zhu, D. Improved Camera Calibration Method and Accuracy Analysis for Binocular Vision. Int. J. Pattern Recognit. Artif. Intell. 2021, 35, 2155010. [Google Scholar] [CrossRef]
  11. Rathnayaka, P.; Baek, S.-H.; Park, S.-Y. An Efficient Calibration Method for a Stereo Camera System with Heterogeneous Lenses Using an Embedded Checkerboard Pattern. J. Sens. 2017, 2017, 6742615. [Google Scholar] [CrossRef]
  12. Yin, H.; Ma, Z.; Zhong, M.; Wu, K.; Wei, Y.; Guo, J.; Huang, B. SLAM-Based Self-Calibration of a Binocular Stereo Vision Rig in Real-Time. Sensors 2020, 20, 621. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, Y.; Wang, X.; Wan, Z.; Zhang, J. A Method for Extrinsic Parameter Calibration of Rotating Binocular Stereo Vision Using a Single Feature Point. Sensors 2018, 18, 3666. [Google Scholar] [CrossRef]
  14. Su, Z.; Lu, L.; Dong, S.; Yang, F.; He, X. Auto-Calibration and Real-Time External Parameter Correction for Stereo Digital Image Correlation. Opt. Lasers Eng. 2019, 121, 46–53. [Google Scholar] [CrossRef]
  15. Yin, Z.; Ren, X.; Du, Y.; Yuan, F.; He, X.; Yang, F. Binocular Camera Calibration Based on Timing Correction. Appl. Opt. 2022, 61, 1475. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, Z. A Flexible New Technique for Camera Calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  17. Tsai, R.Y. A Versatile Camera Calibration Techniaue for High-Accuracy 3D Machine Vision Metrology Using Off-the-Shelf TV Cameras and Lenses. IEEE J. Robot. Autom. 1987, 3, 323–344. [Google Scholar] [CrossRef]
  18. Bouguet, J.-Y. Camera Calibration Toolbox for Matlab (1.0) 2022. Available online: https://data.caltech.edu/records/jx9cx-fdh55 (accessed on 20 April 2023).
  19. Heikkila, J.; Silven, O. A Four-Step Camera Calibration Procedure with Implicit Image Correction. In Proceedings of the Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Juan, PR, USA, 17–19 June 1997; pp. 1106–1112. [Google Scholar]
  20. Liu, X.; Tian, J.; Kuang, H.; Ma, X. A Stereo Calibration Method of Multi-Camera Based on Circular Calibration Board. Electronics 2022, 11, 627. [Google Scholar] [CrossRef]
  21. Liang, S.; Zhao, Y. Camera Calibration Based on the Common Pole-Polar Properties between Two Coplanar Circles with Various Positions. Appl. Opt. 2020, 59, 5167. [Google Scholar] [CrossRef]
  22. Wei, L.; Zhang, G.; Huo, J.; Xue, M. Novel Camera Calibration Method Based on Invariance of Collinear Points and Pole–Polar Constraint. J. Syst. Eng. Electron. 2023, 34, 744–753. [Google Scholar] [CrossRef]
  23. Zhang, H.; Huo, J.; Yang, F.; Han, J.; Guo, D. A Flexible Calibration Method for Large-Range Binocular Vision System Based on State Transformation. Opt. Laser Technol. 2023, 164, 109546. [Google Scholar] [CrossRef]
  24. Yin, Y.; Zhu, H.; Yang, P.; Yang, Z.; Liu, K.; Fu, H. Robust and Accuracy Calibration Method for a Binocular Camera Using a Coding Planar Target. Opt. Express 2022, 30, 6107. [Google Scholar] [CrossRef] [PubMed]
  25. ZED 2i—Industrial AI Stereo Camera. Available online: https://www.stereolabs.com/zed-2i/ (accessed on 9 June 2023).
  26. Harris, C.; Stephens, M. A Combined Corner and Edge Detector. In Proceedings of the Alvey Vision Conference 1988, Manchester, UK, 31 August–2 September 1988; Alvey Vision Club: Manchester, UK, 1988; pp. 23.1–23.6. [Google Scholar]
  27. Das, A.K.; Dewanjee, S. Optimization of Extraction Using Mathematical Models and Computation. In Computational Phytochemistry; Elsevier: Amsterdam, The Netherlands, 2018; pp. 75–106. ISBN 978-0-12-812364-5. [Google Scholar]
  28. Konishi, S.; Kitagawa, G. (Eds.) Bayesian Information Criteria. In Information Criteria and Statistical Modeling; Springer: New York, NY, USA, 2008; pp. 211–237. ISBN 978-0-387-71887-3. [Google Scholar]
  29. Minitab 2010. Available online: www.minitab.com (accessed on 20 April 2023).
Figure 1. Calibration experiment. (a) ZED2i binocular camera [25]; (b) calibration checkered board; (c) experimental setup.
Figure 1. Calibration experiment. (a) ZED2i binocular camera [25]; (b) calibration checkered board; (c) experimental setup.
Sensors 23 08240 g001
Figure 2. Components of the camera calibration rig. (a) RSP60R axis rotation stage; (b) LD2R dual panoramic gimbal; (c) two-way macro frame; (d) lift table; (e) complete assembly.
Figure 2. Components of the camera calibration rig. (a) RSP60R axis rotation stage; (b) LD2R dual panoramic gimbal; (c) two-way macro frame; (d) lift table; (e) complete assembly.
Sensors 23 08240 g002
Figure 3. Schematic representation of reprojection error.
Figure 3. Schematic representation of reprojection error.
Sensors 23 08240 g003
Figure 4. Representation of the 2 3 design space [27].
Figure 4. Representation of the 2 3 design space [27].
Sensors 23 08240 g004
Figure 5. Sign convention.
Figure 5. Sign convention.
Sensors 23 08240 g005
Figure 6. Calibration processes: (a) calibration board corner detection; (b) 20 sets of calibration images and the positions of the binocular camera.
Figure 6. Calibration processes: (a) calibration board corner detection; (b) 20 sets of calibration images and the positions of the binocular camera.
Sensors 23 08240 g006
Figure 7. Experiment flow chart.
Figure 7. Experiment flow chart.
Sensors 23 08240 g007
Figure 8. Convergence plot for method A for ZED2i camera.
Figure 8. Convergence plot for method A for ZED2i camera.
Sensors 23 08240 g008
Figure 9. Calibration of ZED2i camera with method A. (a) Normal plot of the standardized effects; (b) ϵ R P residual plot.
Figure 9. Calibration of ZED2i camera with method A. (a) Normal plot of the standardized effects; (b) ϵ R P residual plot.
Sensors 23 08240 g009
Figure 10. Convergence plot for method B for ZED2i camera.
Figure 10. Convergence plot for method B for ZED2i camera.
Sensors 23 08240 g010
Figure 11. Calibration of ZED2i camera with method B. (a) Normal plot of the standardized effects; (b) ϵ R P residual plot.
Figure 11. Calibration of ZED2i camera with method B. (a) Normal plot of the standardized effects; (b) ϵ R P residual plot.
Sensors 23 08240 g011
Figure 12. Convergence plot for experiment method C.
Figure 12. Convergence plot for experiment method C.
Sensors 23 08240 g012
Figure 13. Calibration of ZED2i camera with method C. (a) Normal plot of the standardized effects; (b) ϵ R P residual plot.
Figure 13. Calibration of ZED2i camera with method C. (a) Normal plot of the standardized effects; (b) ϵ R P residual plot.
Sensors 23 08240 g013
Figure 14. Calibration performance of methods A, B, and C on 3 different cameras.
Figure 14. Calibration performance of methods A, B, and C on 3 different cameras.
Sensors 23 08240 g014
Figure 15. Modeling performance of methods A, B, and C on 3 different cameras.
Figure 15. Modeling performance of methods A, B, and C on 3 different cameras.
Sensors 23 08240 g015
Table 1. Specifications of the three cameras used in the present study.
Table 1. Specifications of the three cameras used in the present study.
SpecificationsZED2iD435iHBVCAM-W202011HD V33
Baseline120 mm50 mm61 mm
FOV110° (H) × 70° (V) × 120° (D)87° (H) × 58° (V) × 95° (D)105°
Depth range0.2–20 m0.2–10 m0.3 m-N/A
Table 2. Comparison of ZED2i camera calibration using methods A, B, and C.
Table 2. Comparison of ZED2i camera calibration using methods A, B, and C.
MethodABC
D (cm)−3LHS−3
H (cm)−3LHSLHS
V (cm)−3LHS−3
R   ( ° ) LHS−10−10
P   ( ° ) LHS1010
Y   ( ° ) LHS10LHS
No. of replicates365
No. of interactions0110
No. of images taken4809601600
ϵ R P (pixel)0.07160.08250.0638
Baseline (mm)119.07118.76117.56
Table 3. The intrinsic parameters of ZED2i camera using methods A, B, and C.
Table 3. The intrinsic parameters of ZED2i camera using methods A, B, and C.
Intrinsic ParametersMethod
ABC
Left cameraFocal length (pixel)(534.00, 534.61)(449.54, 494.71)(515.16, 512.36)
Principal point (pixel)(617.36, 351.94)(591.02, 351.96)(615.94, 363.29)
Radial distortion(−0.11, 0.10, −0.04)(−0.11, 0.08, −0.03)(−0.12, 0.10, 0.00)
Right cameraFocal length (pixel)(533.54, 533.84)(500.57, 496.92)(514.77, 515.49)
Principal point (pixel)(640.20, 360.92)(500.57, 496.92)(640.90, 369.25)
Radial distortion(−0.13, 0.14, −0.07)(−0.12, 0.14, −0.08)(−0.12, 0.10, −0.02)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hao, Y.; Tai, V.C.; Tan, Y.C. A Systematic Stereo Camera Calibration Strategy: Leveraging Latin Hypercube Sampling and 2k Full-Factorial Design of Experiment Methods. Sensors 2023, 23, 8240. https://doi.org/10.3390/s23198240

AMA Style

Hao Y, Tai VC, Tan YC. A Systematic Stereo Camera Calibration Strategy: Leveraging Latin Hypercube Sampling and 2k Full-Factorial Design of Experiment Methods. Sensors. 2023; 23(19):8240. https://doi.org/10.3390/s23198240

Chicago/Turabian Style

Hao, Yanan, Vin Cent Tai, and Yong Chai Tan. 2023. "A Systematic Stereo Camera Calibration Strategy: Leveraging Latin Hypercube Sampling and 2k Full-Factorial Design of Experiment Methods" Sensors 23, no. 19: 8240. https://doi.org/10.3390/s23198240

APA Style

Hao, Y., Tai, V. C., & Tan, Y. C. (2023). A Systematic Stereo Camera Calibration Strategy: Leveraging Latin Hypercube Sampling and 2k Full-Factorial Design of Experiment Methods. Sensors, 23(19), 8240. https://doi.org/10.3390/s23198240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop