Next Article in Journal
Feature-Based Sentimental Analysis on Public Attention towards COVID-19 Using CUDA-SADBM Classification Model
Next Article in Special Issue
Multi-Path U-Net Architecture for Cell and Colony-Forming Unit Image Segmentation
Previous Article in Journal
Measurement of Sea Waves
Previous Article in Special Issue
Improving Automatic Renal Segmentation in Clinically Normal and Abnormal Paediatric DCE-MRI via Contrast Maximisation and Convolutional Networks for Computing Markers of Kidney Function
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Influence of Pitch Angle Errors in 3D Scene Reconstruction Based on U-V Disparity: A Sensitivity Study

1
Instituto Tecnológico y de Energías Renovables, 38600 Granadilla de Abona, Spain
2
Departamento de Ingeniería Informática y de Sistemas, Facultad de Ciencias, Universidad de La Laguna, 38200 San Cristobal de La Laguna, Spain
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(1), 79; https://doi.org/10.3390/s22010079
Submission received: 20 November 2021 / Revised: 20 December 2021 / Accepted: 21 December 2021 / Published: 23 December 2021
(This article belongs to the Special Issue Advances in Image Segmentation: Theory and Applications)

Abstract

:
U-V disparity is a technique that is commonly used to detect obstacles in 3D scenes, modeling them as a set of vertical planes. In this paper, the authors describe the general lines of a method based on this technique for fully reconstructing 3D scenes, and conduct an analytical study of its performance and sensitivity to errors in the pitch angle of the stereoscopic vision system. The equations of the planes calculated for a given error in this angle yield the deviation with respect to the ideal planes (with a zero error in the angle) for a large test set consisting of planes with different orientations, which is represented graphically to analyze the method’s qualitative and quantitative performance. The relationship between the deviation of the planes and the error in the pitch angle is observed to be linear. Two major conclusions are drawn from this study: first, that the deviation between the calculated and ideal planes is always less than or equal to the error considered in the pitch angle; and second, that even though in some cases the deviation of the plane is zero or very small, the probability that a plane of the scene deviates from the ideal by the greatest amount possible, which matches the error in the pitch angle, is very high.

1. Introduction

U-V disparity is a way of representing the information contained in a pair of stereo images to extract depth information. To do this, a three-dimensional space is built in the dimensions u (vertical coordinate of a point projected in one of the images of the stereo pair), v (horizontal coordinate of the same point), and Δ (difference between the horizontal coordinates of the point projected in both images of the stereo pair). This technique provides a satisfactory solution to the problem of obstacle detection, as has been shown in numerous works. For example, in [1,2] the V-disparity and U-V disparity are used, focusing on a specific subset of components of a scene for the detection and location of obstacles in vehicle guidance and driving aid systems. In [3], the authors propose a system that detects, recognizes, and tracks vehicles and pedestrians. They do this by combining different sources of local patterns and depth information using deep learning techniques. In particular, they employ an adaptive U-V disparity algorithm, the results of which are applied to a novel vehicle and pedestrian recognition system based on deep learning. In [4], we find another interesting contribution, in which an application is used to detect and predict the movement of pedestrians in various stages. First, the position of obstacles is precisely determined by merging the information provided by a stereo camera system and a laser scanner. The pedestrians are then identified using smart algorithms based on polylines and pattern recognition, as well as dense disparity maps and U-V disparity. In [5], the authors propose a procedure based on U-V disparity to detect and recognize obstacles on the road as part of a driver assistance system. The proposed realistic U-disparity map greatly improves accuracy when detecting distant obstacles compared to that obtained with a conventional U-disparity map. Other examples may be found in [6,7,8].
The 3D reconstruction problem, which consists of modeling a scene as a set of planes, can actually be regarded as a generalization of the obstacle detection problem, which is why it can also be successfully approached by making use of U-V disparity. For example, in [9], the authors propose a system for segmenting real-world scenes relying on a novel segmentation technique based on U-V disparity analysis to differentiate objects from the surfaces that support them in noisy environments. Both in [9] and in other similar papers, it is assumed that all the parameters that come into play are known exactly.
However, it is well known that in practice, discrepancies exist or may exist between the ’theoretical’ and ’actual’ values of certain parameters. These discrepancies can have different sources. They can be due to errors in determining the value of these parameters with complete accuracy, but may also be due to the changes that said parameters undergo over time due to the normal aging that is experienced by any physical system. Therefore, it seems reasonable to conclude that being able to determine how errors or uncertainties in the knowledge of certain system parameters affect the performance of an algorithm in real conditions is as important as having a procedure that satisfactorily solves the 3D reconstruction problem under ideal conditions. In other words, we wish to measure how sensitive the method designed is to the errors considered. The literature contains numerous works that support the importance of conducting a sensitivity analysis, as in [10,11,12,13]. Specifically, in [13], the authors state that sensitivity analysis is used to measure the robustness of model-based inference, that is, the extent to which the results output by the model depend on the assumptions used to construct it.
This type of analysis is also applied to systems in which depth information is obtained from a scene, which can be achieved using different techniques. For example, in [14], the authors propose what they call “The Relative Stereo 3D Measurement Method”. This method exhibits good tolerance to the calibration errors of the stereo vision system and is very precise in relative distance. In other applications, such as those found in [15,16], Kinect sensors are used to obtain depth information. In [15], the authors present a discussion on calibrating the Kinect sensor and discuss the accuracy and resolution of the depth information it provides. The experimental results obtained show that the random error in the depth measurements increases with the distance to the sensor. Elsewhere, [16] presents a mathematical model of uncertainty for the spatial measurement of visual characteristics using Kinect sensors. Thanks to this model, a qualitative and quantitative analysis is available for using Kinect sensors as 3D perception sensors. In [17], a different approach is presented in which the authors propose a method for analyzing and evaluating the uncertainties found in vision-based reconstructed scenes. This study considers primarily the errors in the image segmentation stage, which propagate throughout the reconstruction procedure.
In [18], we find a contribution that, despite its age, is of great interest. In this work, the authors conduct a thorough study of the influence of certain calibration parameters on the performance of a stereoscopic 3D reconstruction system. Specifically, they consider a binocular vision system with two cameras mounted on a rigid support and they study the depth errors due to the pitch, roll, and yaw between the two cameras. They also study the magnitude of the errors caused by the fact that in one of the two cameras the CCD array is not parallel to the lens, as well as the errors caused by lens distortion. Their quantitative study allows the authors to determine the requirements that the aforementioned parameters must meet for the errors to stay within certain thresholds. Finally, Ref. [19] provides a procedure for correcting misalignment in order to reduce depth errors due to camera displacements. The real images obtained with a stereo camera system show that the method proposed by the authors significantly reduces misalignments caused by the pitch, roll, and yaw of the camera.
The present article outlines an analytical study of how errors in the pitch angle of a stereoscopic vision system with respect to the world reference system influence the determination of the planes used to reconstruct a 3D scene. The pitch angle was chosen to carry out the study in line with the approach made by the authors of other works in which techniques similar to ours are used for modeling scenes, such as in [1,2]. As will be explained in Section 2, the influence of this parameter is quantified by measuring how these planes deviate in the presence of a certain error in the pitch angle with respect deviations under ideal conditions (assuming a zero error in the angle). At this point, it is worth asking whether a small error in the pitch angle can yield a large distortion in the reconstructed scene and if it is possible to limit the deviation suffered by the reconstructed planes as a consequence of the error in said angle. In this paper, we intend to answer these questions, something that, to the best of our knowledge, provides a novel contribution, since we have found no previous work that studies the sensitivity of the 3D reconstruction system to errors in the pitch angle of the vision system.
This work is structured as follows. Section 2 describes the vision system considered and the application of U-V disparity for modeling a scene. We also introduce an error in the pitch angle of the vision system to analyze how it influences the orientation of the planes that define the reconstructed scene. Section 3 describes the experiments carried out to measure the sensitivity of the system to errors in the pitch angle and the results obtained. Finally, Section 4 presents the main conclusions that we draw from the study.

2. Materials and Methods

Disparity is defined as the difference between the position of a pixel in the right image and its correspondence in the left image. Thus, a disparity map is understood as an image containing one disparity value for any point of the original image. By using a stereo system whose characteristics are known, it is possible to compute a full disparity map and calculate a complete 3D map from it.

2.1. Notation

A standard binocular vision system with a single degree of freedom (pitch angle) is shown in Figure 1.
The following parameters are defined:
  • d: distance between the optical centers ( O i , i = r , l ) of both cameras;
  • h: height of the optical center of the camera above the ground;
  • θ : the angle between the optical center and the horizontal plane (pitch angle).
In addition, three different reference systems are observed: R w , which is the world coordinate system in which the real-world points P = ( X , Y , Z , 1 ) T are represented, and R l and R r , which define the coordinate systems of the left and right cameras. To completely define the framework, it is necessary to assume that the epipolar correction has been performed and that both image planes are coplanar and at the same level ( h ) above the ground.
In the R l and R r systems, each image point is designated by its coordinates ( u , v ) , resulting from the projection of real-world coordinates in the image plane. A specific case is the projections of the optical centers, denoted as ( u 0 , v 0 ) .
The intrinsic camera parameters required for the projection are denoted as:
  • f: focal length of the cameras, assumed equal in both cameras;
  • t u , t v : size of the pixels in the u and v dimensions.
For simplicity, when expanding the equations α u = f / t u and α v = f / t v are used, assuming that α = α u = α v .
The relationship between the world coordinate systems and the camera systems can be established by means of a simple transformation matrix.

2.2. Disparity Space

From the above equations, and using the epipolar constraint, we see that the coordinate v resulting from projecting point P is the same in both images, and has the form:
v = y z = Y α cos θ + v 0 sin θ + v 0 cos θ α sin θ Z Y sin θ + Z cos θ + h sin θ + α h cos θ + v 0 h sin θ Y sin θ + Z cos θ + h sin θ
Similarly, the u i coordinates for the left and right images are:
u i = x i z = α X + Y u 0 sin θ + u 0 Z cos θ α q i d 2 + u 0 h sin θ Y sin θ + Z cos θ + h sin θ
Particularizing for the left and right images, q i = 1 in the left image and q i = 1 in the right one, yields:
{ (3a) u l = u 0 + α X + α d 2 Y + h sin θ + Z cos θ (3b) u r = u 0 + α X α d 2 Y + h sin θ + Z cos θ
Hence, the disparity Δ = u l u r is:
Δ = α d Y + h sin θ + Z cos θ Δ = α d Y + h sin θ + Z + f cos θ
As a result, the Disparity Space will be as defined in ( u r , v , Δ ) .

2.3. Scene Model: U-V Disparity Generalization

Any scene in the real world can be interpreted as a combination of several planes with different orientations. With this assumption, every single object present in an arbitrary scene will be represented by a single plane whose orientation will be proportional to the position of the points in real-world space, defined by the general plane equation:
r X + s Y + t Z + u = 0
In the state of the art, the implementations and developments based on V-disparity or U-V disparity are concentrated in a set of concrete planes that respond to reduced versions of Equation (5). For example, Ref. [1] presents the application of the V-disparity technique previously described in [20]. In this article, the authors reduce the recognized planes to those that respond to the equation t Z + u = 0 . In this way, they are able to estimate the inclination of the road and the existence of obstacles making use of only the information contained in the V-disparity plane.
On the other hand, in [2], not only the oblique planes are considered as in the previous case, but a distinction is also made in the vertical planes, which respond to Z = u / t , and in those perpendicular to the plane image. In this case they make use of both the V-disparity and U-disparity projections to not only detect the elements of the image that respond to the cases considered, but also to determine their dimensions.
In this article, the authors do not limit the study carried out to a subset of planes that respond to some variation of the Equation (5), but rather propose a totally generalized treatment of the information. In this way, it is possible to carry out the composition of a complete scene and not only of certain elements that compose it.

Scene Elements Identification

It is necessary to identify which points belong to each element of the image in the U-V disparity space so that these points can be grouped and interpreted as a plane. Assuming this real-world simplification and taking into account the following theorem:
Theorem 1.
Given two planes that are parallel to a line, the intersection of those planes will be another line parallel to the first.
We can deduce that independently of the orientation of a given plane in the real world, is it always possible to find a line parallel to that plane that is also parallel to the image plane.
Introducing the restriction of Equation (5) and expanding from Equation (1), the disparity can also be calculated as:
Δ = 2 d 2 u + r d 2 s h · r u r u 0 + v v 0 s cos θ t sin θ + α s sin θ + t cos θ
The Equation (6) can be particularized with the restrictions discussed above used by other authors, verifying that it is a generalization of those in [1,2].
Notice that the disparity is a function that depends on u r and v, which are discrete domain variables. Because of this, the gradient of the disparity is defined as the derivative of Δ with respect to both variables:
Δ = d Δ d u d , d Δ d v
where:
d Δ d u d = 2 d 2 u + r d 2 s h r
d Δ d v = 2 d 2 u + r d 2 s h s cos θ t sin θ
These equations make it possible to postulate a theorem that will be fundamental for this abstract representation of the content of a scenework:
Theorem 2.
The gradient of every element ( u r , v ) belonging to the projection and transformation of coplanar points in the real-world space of coordinates is constant and only depends on the plane itself and its orientation with respect to the vision system.
An example of the application of Theorems 1 and 2 is shown in Figure 2. The colored parallel plane piped within the image plane is constructed by joining the projections of the line parallel to plane P in each image of the stereo pair. Since the line is parallel to the image plane, it will be contained in a plane described by the equation 0 X + 0 Y + t Z = u and the angle between the planes will be 0. With this configuration and according to the previous expressions, we see that Δ = ( 0 , 0 ) , which means that the disparity is constant for every point on the line, with the two projections being parallel in the image plane.
Taking this into account, it is possible to divide the disparity space based on the different disparity gradients detected. We can thus obtain the equation of the plane that generates that gradient, making it possible to completely reconstruct the scene from the information in the ( u r , v , Δ ) space.

2.4. Scene Reconstruction

The above theorems allow us to lay the foundations for reconstructing any scene by representing it by means of a set of planes. Since our goal is to reconstruct a scene in world coordinates, we need to be able to calculate them based on the points projected in the vision system. Expanding the equations for Δ , u r y v (see Equations (1), (3b) and (4)), we find the expressions to obtain the coordinates of these points with respect to the world reference system:
X = d Δ u r u 0 + d 2
Y = d Δ v v 0 cos θ + α sin θ h
Z = d Δ α cos θ v v 0 sin θ
By using the property described in Theorem 2, we can recognize those projected points that belong to the same plane in the real scene. This makes it possible to have three pairs of points ( u i j , u d j | j { 1 : 3 } ) projected in both images that, once reconstructed, form part of the same plane. Moreover, we can calculate the disparity value Δ 1 Δ 3 for each pair of points.
By specifying the Equations (10)–(12) for each pair of points, it is possible to calculate the coefficients of the equation of the plane. This can be done by simply determining those regions of space ( u r , v , Δ ) whose gradient is constant and selecting any three points within those regions.
From the expressions that represent the projected points on the image planes ((1), (3a), and (3b)) and the general equation of Plane (5), it is possible to analytically calculate the coefficients of the plane that contains the three points as described in Equations (13)–(16).
r = d 2 Δ 1 Δ 2 Δ 3 α v 2 v 1 Δ 3 + v 1 v 3 Δ 2 + v 3 v 2 Δ 1
s = d 2 Δ 1 Δ 2 Δ 3 A sin θ + B α cos θ
t = d 2 Δ 1 Δ 2 Δ 3 B α sin θ + A c o s θ
u = d 2 Δ 1 Δ 2 Δ 3 u 1
u 1 = v 2 v 1 Δ 3 2 + v 1 v 3 Δ 2 2 + v 3 v 2 Δ 1 2 + u d 1 u d 2 v 3 + u d 3 u d 1 v 2 + u d 2 u d 3 v 1 α d + B α cos θ + A sin θ h A = u d 1 u 0 v 2 + u 0 u d 2 v 1 + u d 2 u d 1 v 0 Δ 3 + u 0 u d 1 v 3 + u d 3 u 0 v 1 + u d 1 u d 3 v 0 Δ 2 + u d 2 u 0 v 3 + ( u 0 u d 3 v 2 + u d 3 u d 2 v 0 Δ 1 B = u d 2 u d 1 Δ 3 + u d 1 u d 3 Δ 2 + u d 3 u d 2 Δ 1
To illustrate the behavior of the method described in a simple way, Figure 3 shows its application on a synthetic image.

2.5. Pitch Angle Error Analysis

To introduce the effect of this error in the above expressions (13)–(16), the value of θ is replaced by θ + ε , where ε represents the difference between the calculated and ideal pitch angles. The ideal plane is obtained by assuming a zero error in the pitch angle.

2.5.1. Measuring the Deviation between the Ideal and Calculated Planes

Given any three points P 1 P 3 belonging to the same element in the real scene identified by its coordinates with respect to the world reference system [ X j , Y j , Z j ] , it is possible to determine the coefficients of the equation of the plane that contains them P i , ideal = [ r , s , t , u ] ideal . If these points are projected using Equations (1), (3a), and (3b), it is possible to determine the projected coordinates in both images for all the points. Introducing the ε error factor described in the previous section and resorting to modified Equations (13)–(16) yields the coefficients of the reconstructed plane P i , calculated = [ r , s , t , u ] calculated .
Since the planes are characterized by four parameters ( r , s , t , u ) , one possible way to compare the ideal and calculated planes would be to obtain the difference between the values of these parameters in the two cases. However, it is easier and more intuitive to carry out this comparison using a single variable: specifically, the angle between the normal vectors of the two planes in question. This angle is calculated as follows:
ε normal = arccos n ideal · n calculated n ideal n calculated
The above equation makes it possible to quantitatively evaluate how the proposed reconstruction system responds to errors in the pitch angle.

2.5.2. Construction of the Test Set

In order to conduct an exhaustive study of the behavior of the error in pitch angle and its influence on the reconstructed scene, we need to evaluate how the system responds in a variety of situations representing the cases that can occur in any a real scene. This problem is addressed by building a test set consisting of planes in a wide range of orientations with respect to the vision system. Note that from the point of view of the test set, it has no relevance that the planes that compose it come from a simple synthetic scene, such as the one shown in Figure 3, or from a more complex real scene.
The starting element for defining the set is a plane parallel to the image plane. This seed plane is rotated about the three Cartesian axes, varying the ρ X , ρ Y , and ρ Z angles between 0° and 90° at 5° intervals. The result is a test set consisting of 6859 planes with different orientations, but all at the same distance from the image plane. Therefore, in order for the test set to be able to faithfully represent any real scene, we must introduce an additional degree of freedom, namely, the distance from the planes to the image plane. This involves multiplying the size of the test set by a factor N, where N is the number of distances considered.

3. Results

To analyze how an error in the pitch angle affects the planes that are used to reconstruct a scene, we measured the difference between the normals of the ideal and calculated planes for this error for all the planes in the test set. Doing so revealed that the distance between the plane considered and the image plane has no influence on the deviation between the ideal and calculated planes. In other words, said deviation only depends on the orientation of the plane, and not on its distance from the image plane. This allows the test set to be reduced to 6859 planes with different orientations, as explained in Section 2.5.2, and at a distance of 5m from the image plane. The choice of this value for the distance is arbitrary.
Due to the large size of the test set, it is impossible to include in this paper the graphs showing the difference between the normals of the ideal and calculated planes as a function of the error in the pitch angle for all the planes in the test set. Since the qualitative behavior observed is the same for all of them, Figure 4 only shows the response for four planes with different orientations, which suffices to illustrate the diversity that exists at the quantitative level. The following planes were selected:
1.
ρ X = 0 ; ρ Y = 45 ; ρ Z = 45
2.
ρ X = 15 ; ρ Y = 75 ; ρ Z = 0
3.
ρ X = 30 ; ρ Y = 15 ; ρ Z = 90
4.
ρ X = 45 ; ρ Y = 75 ; ρ Z = 0
As Figure 4 shows, and, by extension, all the planes in the test set, there is a perfect symmetry in the deviation between the normals of the ideal and calculated planes with respect to ε . In other words, the value of the deviation only depends on the absolute value of the error in the angle θ and not on the sign. This allows us to characterize the effect of this error on the reconstruction of each plane by using a scalar parameter, which we will call deviationrate ( dr ) , defined as:
deviationrate = ε normal | ε |
This variable represents the slope of the line of the graph analogous to those shown in Figure 4 for that plane, leaving only the positive values of the error in the angle θ . This reduces the dimensionality of the problem and enables a compact representation on a single graph of the effect of ε on all the possible planes that are used to reconstruct a generic scene.
Despite this, we still have three independent variables that characterize every possible plane in the scene, ρ X , ρ Y , and ρ Z , and a dependent variable, deviationrate . Thus, for the three-dimensional representation, we need to set the value of one of the independent variables, in this case ρ X . The choice of fixed variable for the representation is arbitrary and does not adhere to any particular criterion.
Figure 5 shows the graph corresponding to an angle ρ X = 0 .
The observed behavior varies depending on the value of ρ X considered, as evidenced by Figure 6, which shows this same representation for values of ρ X between 0° and 90° with 15° increments between one graph and the next:
As Figure 6 shows, the value of deviationrate is bound between 0 and 1. Measuring this variable for all 6859 planes in the test set yields the same result in every case.

4. Discussion and Conclusions

The graphs shown in Figure 4 clearly demonstrate that the effect of an error in angle θ on the planes resulting from the reconstructed scene depends, in general, on the plane in question. Actually, as explained in the results section, it depends on its orientation with respect to the image plane, but not on its distance from it.
The results shown in Table 1 should be interpreted as follows: if we look, for example, at the row corresponding to the range 0.4 < dr <= 0.5, we see that 4.23% of the planes in the test set have a deviationrate in this range. Furthermore, 10.12% of the planes in the same set have a deviationrate less than or equal to 0.5.
In view of Figure 6 and Table 1, it holds that for all the planes in the test set, the value of deviationrate is always in the interval [0,1]. This implies that the angle formed by the normals of the ideal and calculated planes is, in the worst case, equal to the error in the angle θ .
This result is important, since it guarantees that the technique presented does not introduce errors in the reconstruction process, and limits the maximum error made during the process based on the deviation of the real vision system from the ideal.
Analyzing the data further, the data contained in Table 1 are represented graphically in Figure 7. If we focus on the analysis of the cumulative percentage of possible values for the deviationrate variable, we see that it behaves like the density function of a random variable with a beta distribution; see [21,22], whose parameters are α = 2 and β = 0.4 .
These results are supported by the Pareto chart [23] of the distribution in Figure 8. We can clearly see that, statistically, more than half of the planes will exhibit a deviation rate between 0.9 and 1, which means that the reconstruction error will be practically identical to that of the vision system used.
If we analyze the other end of the Pareto chart, we find a small percentage of planes whose deviationrate is 0 or practically zero. Focusing our analysis on these planes, we see that they correspond to those with ρ X = 90 and ρ Z = 90 ; that is to say, they are vertical planes perpendicular to the image plane and thus insensitive to errors in the pitch angle ( θ ).
The quantitative results presented earlier can be used to draw a series of conclusions to consider when using the scene reconstruction technique described. First, the sensitivity of the proposed system to errors in the pitch angle ( θ ) has been measured to reveal its upper bound. This allows us to guarantee that the plane obtained by the reconstruction will never deviate from the ideal by an angle greater than the error in the angle θ ( ε ).
In addition to the above result, it is possible to define a parameter that lets us anticipate the behavior of the reconstruction based on the error ( ε ) of the pitch angle. This parameter, the deviation rate, has been modeled as a random variable that follows a beta distribution. This result complements the previous conclusion since, in addition to having an upper limit on the error, the existence of a model allows us to determine what the expected deviation will be.
Analyzing the data in Table 2 obtained from the information shown in Table 1 and the Pareto chart in Figure 8, we may conclude that in 78.31% (8.69% + 15.56% + 54.06%) of the cases, the value of deviationrate will be greater than 0.7. This means that for almost 80% of the planes in the test set, the calculated plane will deviate from the ideal by an angle of at least 70% of the value of ε . In addition, in more than half of the cases (54.06%), the deviation will exceed this value by over 90%. It is also important to note that the results presented in this paper were obtained for a test set that contains 6859 planes with different orientations and at a fixed distance from the image plane, specifically 5m. The authors believe that the set of planes obtained in this way is sufficient to represent the diversity that can be found in a real scene, since, as has been proven, the deviationrate variable depends only on the orientation of the plane and not on its distance.

Author Contributions

Conceptualization, J.F., M.S. and L.A.; methodology, M.S. and L.A.; software, J.F.; validation, J.F., M.S. and L.A.; formal analysis, J.F., M.S. and L.A.; investigation, J.F.; resources, J.F.; data curation, J.F.; writing—original draft preparation, J.F.; writing—review and editing, J.F. and M.S.; visualization, J.F.; supervision, L.A.; project administration, M.S. and L.A.; funding acquisition, J.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors wish to acknowledge support from the Instituto Tecnologico y de Energías Renovables.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Labayrade, R.; Aubert, D. Robust and Fast Stereovision Based Obstacles Detection for Driving Safety Assistance. IEICE Trans. Inf. Syst. 2004, 87, 80–88. [Google Scholar]
  2. Hu, Z.; Lamosa, F.; Uchimura, K. A complete U-V-disparity study for stereovision based 3D driving environment analysis. In Proceedings of the Fifth International Conference on 3-D Digital Imaging and Modeling (3DIM’05), Ottawa, ON, Canada, 13–16 June 2005; pp. 204–211. [Google Scholar] [CrossRef]
  3. Nguyen, V.D.; Van Nguyen, H.; Tran, D.T.; Lee, S.J.; Jeon, J.W. Learning Framework for Robust Obstacle Detection, Recognition, and Tracking. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1633–1646. [Google Scholar] [CrossRef]
  4. Musleh, B.; García, F.; Otamendi, J.; Armingol, J.M.; De la Escalera, A. Identifying and Tracking Pedestrians Based on Sensor Fusion and Motion Stability Predictions. Sensors 2010, 10, 8028–8053. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Leng, J.; Liu, Y.; Du, D.; Zhang, T.; Quan, P. Robust Obstacle Detection and Recognition for Driver Assistance Systems. IEEE Trans. Intell. Transp. Syst. 2020, 21, 1560–1571. [Google Scholar] [CrossRef]
  6. Gao, Y.; Ai, X.; Wang, Y.; Rarity, J.; Dahnoun, N. U-V-Disparity based Obstacle Detection with 3D Camera and steerable filter. In Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), Baden-Baden, Germany, 5–9 June 2011; pp. 957–962. [Google Scholar] [CrossRef]
  7. Iloie, A.; Giosan, I.; Nedevschi, S. UV disparity based obstacle detection and pedestrian classification in urban traffic scenarios. In Proceedings of the 2014 IEEE 10th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 4–6 September 2014; pp. 119–125. [Google Scholar] [CrossRef]
  8. Benacer, I.; Hamissi, A.; Khouas, A. A novel stereovision algorithm for obstacles detection based on U-V-disparity approach. In Proceedings of the 2015 IEEE International Symposium on Circuits and Systems (ISCAS), Lisbon, Portugal, 24–27 May 2015; pp. 369–372. [Google Scholar] [CrossRef]
  9. Li, X.; Chen, L.; Li, S.; Zhou, X. Depth segmentation in real-world scenes based on U–V disparity analysis. J. Vis. Commun. Image Represent. 2020, 73, 102920. [Google Scholar] [CrossRef]
  10. Iooss, B.; Saltelli, A. Introduction to Sensitivity Analysis. In Handbook of Uncertainty Quantification; Ghanem, R., Higdon, D., Owhadi, H., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 1–20. [Google Scholar] [CrossRef]
  11. Borgonovo, E.; Plischke, E. Sensitivity analysis: A review of recent advances. Eur. J. Oper. Res. 2016, 248, 869–887. [Google Scholar] [CrossRef]
  12. Ferretti, F.; Saltelli, A.; Tarantola, S. Trends in sensitivity analysis practice in the last decade. Sci. Total Environ. 2016, 568, 666–670. [Google Scholar] [CrossRef] [PubMed]
  13. William Becker, A.S. Design for Sensitivity Analysis. In Handbook of Design and Analysis of Experiments; CRC Press: Boca Raton, FL, USA, 2015; Chapter 18. [Google Scholar] [CrossRef]
  14. Nagatomo, S.; Hayashi, J.; Hata, S. Proposal of calibration error-tolerant 3D measurement method using stereo vision for robot vision. In Proceedings of the 2009 International Conference on Mechatronics and Automation, Changchun, China, 9–12 August 2009; pp. 3263–3268. [Google Scholar] [CrossRef]
  15. Khoshelham, K.; Elberink, S.O. Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications. Sensors 2012, 12, 1437–1454. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Park, J.H.; Shin, Y.D.; Bae, J.H.; Baeg, M.H. Spatial Uncertainty Model for Visual Features Using a Kinect™ Sensor. Sensors 2012, 12, 8640–8662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Belhaoua, A.; Kohler, S.; Hirsch, E. Error Evaluation in a Stereovision-Based 3D Reconstruction System. EURASIP J. Image Video Process. 2010, 2010. [Google Scholar] [CrossRef] [Green Version]
  18. Zhao, W.; Nandhakumar, N. Effects of camera alignment errors on stereoscopic depth estimates. Pattern Recognit. 1996, 29, 2115–2126. [Google Scholar] [CrossRef]
  19. Santoro, M.; AlRegib, G.; Altunbasak, Y. Misalignment correction for depth estimation using stereoscopic 3-D cameras. In Proceedings of the 2012 IEEE 14th International Workshop on Multimedia Signal Processing (MMSP), Banff, AB, Canada, 17–19 September 2012; pp. 19–24. [Google Scholar] [CrossRef]
  20. Labayrade, R.; Aubert, D.; Tarel, J.P. Real time obstacle detection in stereovision on non flat road geometry through “v-disparity” representation. In Proceedings of the Intelligent Vehicle Symposium, Versailles, France, 17–21 June 2002; Volume 2, pp. 646–651. [Google Scholar] [CrossRef]
  21. Devore, J.L. Probabilidad y estadística para ingeniería y ciencias. In Probabilidad y Estadística Para Ingeniería y Ciencias; International Thomson Editores: Mexico, D.F., Mexico, 2001. [Google Scholar]
  22. Wackerly, D.D. Estadística matemática con aplicaciones. In Estadística Matemática Con Aplicaciones, 6th ed.; Grupo Editorial Iberoamérica: Mexico, D.F., Mexico, 2002. [Google Scholar]
  23. Wilkinson, L. Revising the Pareto Chart. Am. Stat. 2006, 60, 332–334. [Google Scholar] [CrossRef]
Figure 1. Pinhole vision system model.
Figure 1. Pinhole vision system model.
Sensors 22 00079 g001
Figure 2. The arbitrarily oriented plane P (in gray) is represented in the image plane (in yellow). The spheres behind the image plane represent the optical centers of the left and right cameras, respectively.
Figure 2. The arbitrarily oriented plane P (in gray) is represented in the image plane (in yellow). The spheres behind the image plane represent the optical centers of the left and right cameras, respectively.
Sensors 22 00079 g002
Figure 3. (a) Synthetic test scene; (b) reconstruction obtained from (a).
Figure 3. (a) Synthetic test scene; (b) reconstruction obtained from (a).
Sensors 22 00079 g003
Figure 4. Deviation between the normals of the ideal and calculated planes as a function of the error in angle θ for: (a) Plane 1, (b) Plane 2, (c) Plane 3, (d) Plane 4.
Figure 4. Deviation between the normals of the ideal and calculated planes as a function of the error in angle θ for: (a) Plane 1, (b) Plane 2, (c) Plane 3, (d) Plane 4.
Sensors 22 00079 g004
Figure 5. Deviationrate as a function of ρ Y and ρ Z and for ρ X = 0 .
Figure 5. Deviationrate as a function of ρ Y and ρ Z and for ρ X = 0 .
Sensors 22 00079 g005
Figure 6. Deviationrate varying ρ Y and ρ Z , for ρ X = 0°, 15°, 30°, 45°, 60°, 75°, 90°.
Figure 6. Deviationrate varying ρ Y and ρ Z , for ρ X = 0°, 15°, 30°, 45°, 60°, 75°, 90°.
Sensors 22 00079 g006
Figure 7. Graphical representation of the data in Table 1.
Figure 7. Graphical representation of the data in Table 1.
Sensors 22 00079 g007
Figure 8. Pareto chart of the data in Table 1.
Figure 8. Pareto chart of the data in Table 1.
Sensors 22 00079 g008
Table 1. Left column: intervals of the deviation rate (between 0 and 1); center column: percentage of occurrence or percentage of planes in the test set that fall within each interval considered; right column: cumulative percentage or percentage of planes in the test set whose deviation rate is less than or equal to the upper value of the range considered.
Table 1. Left column: intervals of the deviation rate (between 0 and 1); center column: percentage of occurrence or percentage of planes in the test set that fall within each interval considered; right column: cumulative percentage or percentage of planes in the test set whose deviation rate is less than or equal to the upper value of the range considered.
DeviationrateOccurrence %Cumulative %
0.0 < dr <= 0.10.90%0.90%
0.1 < dr <= 0.21.04%1.94%
0.2 < dr <= 0.31.57%3.51%
0.3 < dr <= 0.42.38%5.89%
0.4 < dr <= 0.54.23%10.12%
0.5 < dr <= 0.64.20%14.32%
0.6 < dr <= 0.77.38%21.69%
0.7 < dr <= 0.88.69%30.38%
0.8 < dr <= 0.915.56%45.94%
0.9 < dr <= 1.054.06%100.00%
Table 2. Left column: intervals of the deviation rate (between 0 and 1); right column: percentage of occurrence or percentage of planes in the test set that fall within each interval considered.
Table 2. Left column: intervals of the deviation rate (between 0 and 1); right column: percentage of occurrence or percentage of planes in the test set that fall within each interval considered.
DeviationrateOccurrence %
0.7 < dr <= 0.88.69%
0.8 < dr <= 0.915.56%
0.9 < dr <= 1.054.06%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Felipe, J.; Sigut, M.; Acosta, L. Influence of Pitch Angle Errors in 3D Scene Reconstruction Based on U-V Disparity: A Sensitivity Study. Sensors 2022, 22, 79. https://doi.org/10.3390/s22010079

AMA Style

Felipe J, Sigut M, Acosta L. Influence of Pitch Angle Errors in 3D Scene Reconstruction Based on U-V Disparity: A Sensitivity Study. Sensors. 2022; 22(1):79. https://doi.org/10.3390/s22010079

Chicago/Turabian Style

Felipe, Jonatán, Marta Sigut, and Leopoldo Acosta. 2022. "Influence of Pitch Angle Errors in 3D Scene Reconstruction Based on U-V Disparity: A Sensitivity Study" Sensors 22, no. 1: 79. https://doi.org/10.3390/s22010079

APA Style

Felipe, J., Sigut, M., & Acosta, L. (2022). Influence of Pitch Angle Errors in 3D Scene Reconstruction Based on U-V Disparity: A Sensitivity Study. Sensors, 22(1), 79. https://doi.org/10.3390/s22010079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop