Next Article in Journal
Building Low-Cost Sensing Infrastructure for Air Quality Monitoring in Urban Areas Based on Fog Computing
Previous Article in Journal
Quantification of DAS VSP Quality: SNR vs. Log-Based Metrics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Calibration Method and Pose Domain Determination of a Light-Pen in a 3D Vision Coordinate Measurement System

1
School of Mechanical Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250300, China
2
School of Electrical Engineering and Automation, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250300, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1029; https://doi.org/10.3390/s22031029
Submission received: 4 January 2022 / Revised: 22 January 2022 / Accepted: 26 January 2022 / Published: 28 January 2022
(This article belongs to the Section Navigation and Positioning)

Abstract

:
Light pens for 3D vision coordinate measurement systems are increasingly widely used due to their advantages, such as their small size, convenience of being carried, and widespread applicability. The posture of the light pen is an important factor that affects accuracy. The pose domain of the pen needs to be given so that the measurement system has a suitable measurement range to obtain more qualified parameters. The advantage of the self-calibration method is that the entire self-calibration process can be completed at the measurement site with no auxiliary equipment. After the system camera calibration was completed, we took several pictures of the same measurement point with different poses to obtain the conversion matrix of the picture and subsequently used spherical fitting, the generalized inverse method of least squares, and the principle of position invariance in the pose domain range. The combined stylus tip center self-calibration method calculates the actual position of the light pen probe. The experimental results verify the effectiveness of the method; the measurement accuracy of the system can satisfy basic industrial measurement requirements.

1. Introduction

With the ongoing advancement of human society, manufacturing continues to flourish, and industrial measurement has become an extremely important link in the industrial development process. With the development of the automotive industry and aerospace industry, which require high-precision real-time measurement, there are increasingly more applications and requirements for workpiece measurement, particularly in terms of accuracy. At present, mainstream measurement technology cannot satisfy the ever-increasing measurement requirements. The traditional coordinate measuring machine (CMM) [1] cannot satisfy the online measurement requirements of some industrial sites due to its large size and limited measurement range. Indoor GPS, white light scanners [2,3], optical laser trackers [4,5,6], and other common measurement equipment [7,8] also have many shortcomings such as low measurement efficiency, high cost, complex configuration, and poor portability. The shortcomings of modern measurement technology have limited the development of industry. Therefore, the development of new measurement technology has become a common expectation. Vision measurement systems [9] are mainly based on computer vision, including measurement technology, electronic technology, image processing technology, etc. to create a new measurement structure frame, reduce measurement limitations, increase flexibility and convenience, and enable workpiece measurement from multiple angles. Its small size has met with a favorable response in modern manufacturing fields such as aerospace and automobiles. The light-pen-type 3D vision measurement system [10] based on machine vision is increasingly used because of its small size, convenient portability, flexible assembly according to on-site measurement requirements, and good applicability; and related products have been put into production [11]. Of course, many scholars have unique insights into visual measurement systems and have achieved notable results [12,13,14,15,16,17].
The actual coordinates of the center of the light pen stylus, in the coordinate system, directly affect the measurement accuracy of the three-coordinate measuring system. In practical applications, the center of the light pen stylus is usually taken as a theoretical value when the light pen is used for processing. Due to the processing error of the light pen, the center value directly affects the measurement accuracy. When measuring various measurement objects such as deep holes and grooves, it is necessary to replace the light pen probe with other angles or add an extension rod for measurement. The difference in force and loss of multiple uses of the light pen also cause small changes in the coordinates, so the self-calibration of the stylus tip center is particularly important. Liu optimized an objective function based on the principle of position invariance and generalized an inverse method of the least squares solution of the nonlinear equations and obtained the position of the stylus tip center in the coordinate system [18,19]. Zheng proposed a two-step calibration method for the center of a plane target probe [20]. According to the principle of position invariance, the objective optimization function is established, and the optimized correction value of the stylus tip center in the X and Y directions in the target coordinate system is solved by the Levenberg–Marquardt algorithm [21]. Then, the error of the stylus tip center in the Z direction is solved. Zhang also adopted the principle of position invariance, upon which he provided the error constraint parameters [22].
The posture of the light pen is an important factor that affects the accuracy. The pose domain of the light pen needs to be given so that the measurement system has a suitable measurement range to obtain more qualified parameters. Therefore, a threshold measurement device was designed according to the measurement system in the experiment. The main body was composed of a light pen, connecting rod, base, motor seat, coupling, and other parts. The transmission device was composed of a Siemens servo motor and the corresponding drive. The experiment was completed on the built measurement system [23]. After the motor parameters and program settings were completed, the experiment was driven by the motor to drive the device of the mechanical structure to make precise angle changes. For each rotation of 1°, the corresponding experimental image was collected to obtain the measurement error at this angle, and the data were sorted to obtain the pose that satisfied the requirement area.
A new self-calibration method for the center of the optical pen probe was proposed. The actual position of the light pen probe was obtained by combining methods such as spherical fitting, the generalized inverse method of the least squares method, and the principle of position invariance [19]. Regardless of the changes in pose of the stylus, the position of the center of the stylus remained the same, as shown in Figure 1. After the camera calibration was completed [24,25,26,27], the light pen probe collected multiple sets of images in different poses that satisfied the threshold measurement requirements under the CCD camera and used the conversion relationship of each image to fit the same feature point on a sphere with the center of the probe as the center of the sphere, as shown in Figure 2. Thus, the actual position of the center of the probe was obtained.

2. Self-Calibration Method of the Stylus Tip Center

2.1. Light-Pen-Type 3D Vision Measurement System

The light-pen-type vision measurement system is mainly composed of three parts: a CCD industrial camera capable of imaging infrared rays, a computer, and a light pen, as shown in Figure 3. The light pen consists of eight infrared LEDs and a detachable ruby ball probe. Figure 4 shows several types of probes. A and B have different core diameters, and the B–E measuring rods have different lengths. During the measurement, the light pen image is taken by the CCD industrial camera and transmitted to the computer for related calculations; finally, the measurement result is obtained.
The measurement process of the measurement system is as follows: After the camera’s internal parameters have been calibrated [28], the user places the measured object and light pen in the range that the CCD camera can shoot, the user presses the shooting button, the CCD camera takes a set of pictures, and the world coordinates of the measured object are obtained. For the coordinates in the system, the distance between two points can be similarly measured by placing the light pen on another point.

2.2. Establishment of the Coordinate System

The visual coordinate measurement system contains three coordinate systems:
1.
Image coordinate system (O0-XY):
The coordinate system is established on the imaging surface of the CCD camera with the center of the imaging surface as the coordinate origin. The X-axis and Y-axis are parallel to the horizontal and vertical imaging directions of the CCD camera, respectively, and parallel to the O1UV plane (unit: pixel).
2.
Camera coordinate system (O1-uvw):
The camera coordinate system is established on the industrial camera, and the origin of the camera coordinate system is the optical center of the industrial camera. The plane formed by the camera coordinate system is also parallel to the O1UV plane (unit: mm).
3.
Light pen coordinate system (O2-xyz):
The theoretical value of the center of the light pen stylus is the coordinate origin, and the direction of the light pen shaft is the Y-axis; the lower plane perpendicular to the six characteristic points of the light pen is the Z-axis, and the X-axis is determined by the right-hand rule (unit: mm).

2.3. Solve Equations

Before measurement, the stylus tip center in the light pen coordinate system should be calibrated. During the measurement, the images of the LED targets are captured by a CCD camera, and the computer processes the images to obtain the center coordinates of the center position of the target as a known quantity to calculate the rotation matrix R and translation vector T [23]. The coordinates of the stylus tip center that correspond to the measured position in the camera coordinate system can be obtained by (1). The camera coordinate system is transformed into the light pen coordinate system as follows:
[ u v w ] = [ r 1 r 2 r 3 t x r 4 r 5 r 6 t y r 7 r 8 r 9 t z ] [ x y z 1 ]
[ r 1 r 2 r 3 r 4 r 5 r 6 r 7 r 8 r 9 ] = R [ t x t y t Z ] = T
We establish the relationship between the camera coordinate system and pixel coordinate system:
s [ X Y 1 ] = [ f 0 0 0 f 0 0 0 f ] [ u v w ]
The conversion between the camera coordinate system and pixel coordinate system is carried out using Equation (2); f is the focal length, X = f w u , Y = f w v ;
We establish the linear equation set of the j-th feature point in the image:
s j [ X j Y j 1 ] = t z [ a 1 a 2 a 3 a 4 a 5 a 6 a 7 a 8 a 9 a 10 a 11 1 ] [ x j y j z j 1 ]
let t z [ a 1 a 2 a 3 a 4 a 5 a 6 a 7 a 8 a 9 a 10 a 11 1 ] = [ f 0 0 0 f 0 0 0 f ] [ u v w ] = [ f 0 0 0 f 0 0 0 f ] [ r 1 r 2 r 3 t x r 4 r 5 r 6 t y r 7 r 8 r 9 t z ]
s j = r 7 x j + r 8 y j + r 9 z j + t z = t z ( a 9 x j + a 10 y j + a 11 z j + 1 ) = t z ζ j
Equation (3) is equivalent to:
ζ j [ X c j Y c j 1 ] = [ a 1 a 2 a 3 a 4 a 5 a 6 a 7 a 8 a 9 a 10 a 11 1 ] [ x j y j z j 1 ]
After solving the linear equations, the parameters are separated to obtain the initial values of R and T.
In the ideal perspective imaging process, the solution obtained by the above formula is feasible because there is no error in the coordinate values of the object point and image point in each coordinate system, and the calculated R satisfies the orthogonal system constraint. However, in the actual environment, due to various factors such as camera parameter calibration error [29], image plane position extraction error of the control point center, and coordinate value calibration error of the control point center in the coordinate system, the R matrix does not satisfy the orthogonal constraint relationship. Consequently, matrix T also has a large error, so the initial values of R and T can be obtained by the linear equation solving method. Then, the optimal solution of R and T can be obtained by the Newton–Gaussian iteration method [23].

2.4. Self-Calibration Method of the Stylus Tip Center

After R and T have been determined, because the distance between the characteristic point of the light pen and the actual position of the stylus tip center coordinate is a fixed value, according to the generalized inverse method of the nonlinear least square method, we obtain:
( x i j x 0 ) 2 + ( y i j y 0 ) 2 + ( z i j z 0 ) 2 = d i j 2
where i represents the i-th image, and j represents the j-th feature point.
The coordinate positions of the pictures in the coordinate system are different, but the distance from the actual position of the center coordinate of the probe is the same. Hence, the same feature point of n groups of pictures in the same spatial coordinate system (based on the probe and actual position of the center coordinate) is the center of the sphere, and the constant distance between the characteristic point and the stylus tip center is the radius of the sphere. The same feature point of all pictures is fitted to a sphere, and eight feature points are fitted to eight spheres. The spherical center coordinates of the fitted spheres are the actual position of the center coordinates of the stylus.
We obtain the residual formula:
F = i = 1 n [ ( x i x 0 ) 2 + ( y i y 0 ) 2 + ( z i z 0 ) 2 d 2 ] 2
The corresponding partial derivatives are
F x = 4 i = 1 n ( x i x 0 ) [ ( x i x 0 ) 2 + ( y i y 0 ) 2 + ( z i z 0 ) 2 d 2 ] F y = 4 i = 1 n ( y i y 0 ) [ ( x i x 0 ) 2 + ( y i y 0 ) 2 + ( z i z 0 ) 2 d 2 ] F z = 4 i = 1 n ( z i z 0 ) [ ( x i x 0 ) 2 + ( y i y 0 ) 2 + ( z i z 0 ) 2 d 2 ] F d = 4 d i = 1 n [ ( x i x 0 ) 2 + ( y i y 0 ) 2 + ( z i z 0 ) 2 d 2 ]
We insert the coordinate values, set the partial derivative equal to zero, construct a system of equations, and solve it to obtain the actual position of the center coordinate of the stylus.

2.5. Self-Calibration Steps

The self-calibration method is realized on the built visual coordinate measurement system. Briefly, we outline the steps to obtain the actual position of the stylus center coordinates using the self-calibration method.
  • After the measurement system has been built, use the spatial coordinates of the characteristic points of the light pen measured by CMM as the initial value.
  • Use the principle of position invariance to shoot images of different poses. Considering the particularity of the self-calibration method, the image pose should be changed as much as possible in the threshold range to make the result more accurate.
  • After obtaining at least eight sets of pictures within the threshold range, calculate R and T for each image using Equations (1)–(5).
  • After R and T have been determined, construct and solve Equations (6)–(8).
  • Obtain the actual position of the stylus tip center coordinate in the light pen coordinate system.

3. Experiment

3.1. Pose Domain Experiment

The rotation angle of the light pen in each direction affects the measurement accuracy. Considering the rotation angles of the three rotation axes of the light pen, the threshold measurement experiment was carried out according to the above method.
We performed the threshold measurement experiment according to the above method.
When the pen was rotated around the X-axis, the angle between the light pen and the positive half axis of the Y-axis (pitch) was 41°~159°, and there was a clear and recognizable image.
When the pen was rotated around the Y-axis, the angle between the light pen and the positive half axis of the X-axis (yaw) was 0°~180°, and there was a clear and recognizable image.
When the pen was rotated around the Z-axis, the angle between the light pen and the positive half axis of the X-axis (roll) was 0°~45° and 136°~180°, and there was a clear and recognizable image.
Due to the processing error of the mechanical device [30,31,32] and measurement error of the measuring equipment, we used 1° as the unit of the data to ensure that the error did not affect the experimental results. The motor drove the light pen to the measurement position, took a stable image, calculated the measurement accuracy, and rotated 1 degree to take the image. Every time it rotated by 1 degree, the camera collected 15 groups of images within the angle, analyzed and processed the data, and compared the obtained measurement results with the real values to determine differences. The graphs are shown in Figure 5, Figure 6 and Figure 7.
When the pitch angle range is 80°~97°, the yaw angle is 80°~101°, and the tilt angle is 0°~2° and 177°~180°, a certain measurement accuracy can be guaranteed to satisfy the measurement requirements. Satisfactory experimental results were obtained.

3.2. Self-Calibration Experimental Results

3.2.1. Self-Calibration Measurement Experiment

We carried out the self-calibration experiment according to the method steps in this article. The self-calibration method was incorporated into the light pen measurement system for measurement comparison experiments, and each light pen performed ten measurement experiments. The measured value was compared with the real value (30.0317 mm) to obtain the absolute error and relative error. The measurement results are shown in Table 1.
The measured values of the five light pen styluses are displayed in lines 2–11. The last three lines show the average, absolute error, and relative error. The measurement results of each stylus in the ten images are shown in lines 1–10. As shown in Table 1, the absolute error of the probe of each light pen is stabilized below 0.0737 mm, and the relative error is stabilized below 0.0025 mm. The experimental results satisfy the basic industrial measurement requirements, which verifies the effectiveness of this method.

3.2.2. Single-Point Repeatability Experiment

We put the light pen stylus in a standard cone. To maintain the position of the light pen stylus unchanged, we rotated the position of the light pen to perform a single-point repeatability experiment.
The experimental results of five light pen styluses are shown in Table 2. The coordinates of the center point of each light pen stylus in the ten images are shown in columns 1–10. The single-point repeatability is expressed by the standard deviation of ten center point coordinates. As shown in the previous column, the probe repeatability of each light pen does not exceed 0.041, which verifies the effectiveness of the self-calibration method.

4. Conclusions

A self-calibration method for the center of a light pen stylus based on spherical fitting is proposed. Multiple sets of light pen images with different poses were taken, and the actual position of the light pen probe was obtained by combining methods such as spherical fitting, the generalized inverse method of least squares, and the position invariance principle. The model and steps of the self-calibration method are given. To verify the effectiveness of the method, five different light-pen probes were used to conduct experiments, and a single-point repeatability experiment was carried out. The test results show that the measurement accuracy after using the method is obviously better than that before the method is used, and it satisfies the basic industrial measurement requirements.
The pose of the light pen is an important factor that affects the measurement accuracy of the light pen. Combined with the measurement system, the measurement experiment of the light pen pose field was carried out. In the experiment, three angles were measured, and the pose domain of the light pen was obtained, which standardized the pose of the light pen and ensured its measurement accuracy. The next step of research will involve actual industrial applications, and it will be better applied to actual measurements.

Author Contributions

Conceptualization, D.S. and C.Z.; methodology, C.Z. and Y.X.; formal analysis, C.Z.; resources, D.S.; data curation, C.Z.; writing—original draft preparation, C.Z.; writing—review and editing, P.Z., X.W., D.H.; Validation, G.Y. and M.Z.; Supervision, D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Project of Shandong Provincial Major Scientific and Technological Innovation, grant No.2019JZZY010444, No.2019TSLH0315, in part by the Project of 20 Policies of Facilitate Scientific Research in Jinan Colleges, grant No.2019GXRC063, in part by the Project of Shandong Province Higher Educational Science and Technology Program, grant No.J18KA345, and in part by the Natural Science Foundation of Shandong Province of China, grant No.ZR2020MF138.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Brezina, I. Coordinate measuring machines and systems: John A. Bosch (Ed.), Marcel Dekker, Inc., New York/Basel/Hong Kong, 1995, 1st Edition, xi + 444 pp., ISBN 0-8247-9581-4, price: 125,00 US$. Measurement 1997, 20, 287. [Google Scholar] [CrossRef]
  2. Milroy, M.J.; Weir, D.J.; Bradley, C.; Vickers, G.W. Reverse engineering employing a 3D laser scanner: A case study. Int. J. Adv. Manuf. Technol. 1996, 12, 111–121. [Google Scholar] [CrossRef]
  3. Chen, J.; Wu, X.; Yu Wang, M.; Li, X. 3D shape modeling using a self-developed hand-held 3D laser scanner and an efficient HT-ICP point cloud registration algorithm. Opt. Laser Technol. 2013, 45, 414–423. [Google Scholar] [CrossRef]
  4. Vincze, M.; Prenninger, J.P.; Gander, H. A Laser Tracking System to Measure Position and Orientation of Robot End Effectors Under Motion. Int. J. Robot. Res. 1994, 13, 305–314. [Google Scholar] [CrossRef]
  5. Mayer, J.R.R.; Parker, G.A. A portable instrument for 3-D dynamic robot measurements using triangulation and laser tracking. IEEE Trans. Robot. Autom. 1994, 10, 504–516. [Google Scholar] [CrossRef]
  6. Xie, Z.; Zhang, Z.; Jin, M. Development of a multi-view laser scanning sensor for reverse engineering. Meas. Sci. Technol. 2006, 17, 2319–2327. [Google Scholar] [CrossRef]
  7. Santolaria, J.; Guillomia, D.; Cajal, C.; Albajez, J.A.; Aguilar, J.J. Modelling and calibration technique of laser triangulation sensors for integration in robot arms and articulated arm coordinate measuring machines. Sensors 2009, 9, 7374–7396. [Google Scholar] [CrossRef] [Green Version]
  8. Zhou, N.; An, Z.; Li, L.; Zhu, Y. iGPS Measurement Network Multi-Station Arrangement Design. Appl. Mech. Mater. 2013, 443, 223–227. [Google Scholar] [CrossRef]
  9. Estler, T.; Edmundson, K.; Peggs, G.; Parker, D. Large-Scale Metrology—An Update. Cirp Ann. Manuf. Technol. 2002, 51, 587–609. [Google Scholar] [CrossRef]
  10. Zhang, Z.; Huang, Q.; Lin, W.; Che, R. Probe imaging vision coordinate measuring system using single camera. In Proceedings of the SPIE 3558, Automated Optical Inspection for Industry: Theory, Technology, and Applications II, Beijing, China, 10 August 1998. [Google Scholar]
  11. Fraser, C. Innovations in Automation for Vision Metrology Systems. Photogramm. Rec. 2003, 15, 901–911. [Google Scholar] [CrossRef]
  12. Wang, S.; Liu, S.; Mao, Q. A CMM-Based Method of Control Point Position Calibration for Light Pen Coordinate Measuring System. Sensors 2020, 20, 5592. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, R.; Liu, S.G.; Wang, H.Y. Wireless Control for Pointolite in Light-Pen CMMs. Appl. Mech. Mater. 2014, 602–605, 2217–2220. [Google Scholar] [CrossRef]
  14. Fu, S.; Zhang, L.; Ye, N.; Zhang, W.; Liu, S. A flexible approach to light pen calibration for a monocular-vision-based coordinate measuring system. Meas. Sci. Technol. 2014, 25, 125006. [Google Scholar] [CrossRef]
  15. Alblalaihid, K.; Kinnell, P.; Lawes, S.; Desgaches, D.; Leach, R. Performance Assessment of a New Variable Stiffness Probing System for Micro-CMMs. Sensors 2016, 16, 492. [Google Scholar] [CrossRef] [Green Version]
  16. Magdziak, M. Selection of the Best Model of Distribution of Measurement Points in Contact Coordinate Measurements of Free-Form Surfaces of Products. Sensors 2019, 19, 5346. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Magdziak, M. A New Method of Distribution of Measurement Points on Curvilinear Surfaces of Products. Sensors 2019, 19, 2667. [Google Scholar] [CrossRef] [Green Version]
  18. Liu, S.G.; Peng, K.; Huang, F.S.; Zhang, G.X.; Li, P. A Portable 3D Vision Coordinate Measurement System Using a Light Pen. Key Eng. Mater. 2005, 295–296, 331–336. [Google Scholar] [CrossRef]
  19. Liu, S.; Zhang, H.; Dong, Y.; Tang, S.; Jiang, Z. Portable Light Pen 3D Vision Coordinate Measuring System- Probe Tip Center Calibration. Meas. Sci. Rev. 2013, 13, 194–199. [Google Scholar] [CrossRef]
  20. Zheng, X.; Zhao, M.; Feng, S. Two-Step Calibration of Probe Tip Center of Planar Target. Laser Optoelectron. Prog. 2018, 55, 001201. [Google Scholar] [CrossRef]
  21. Du, S.-Q.; Gao, Y. The Levenberg-Marquardt-Type Methods for a Kind of Vertical Complementarity Problem. J. Appl. Math. 2011, 2011, 161853. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, R.; Liu, S.; Wang, S.; Song, X. Stylus Tip Center Position Self-Calibration Based on Invariable Distances in Light-Pen Systems. Sensors 2017, 17, 131. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Shan, D.; Bi, X.; Wang, X.; Zhang, P.; Xu, Z.; Han, M.; Liu, Y.; Zhang, C.; Xu, Y. The Impact of Control Point Dispersion on Measurement Accuracy in a New Type of Light-Pen Coordinate Measuring System. IEEE Access 2021, 9, 76015–76023. [Google Scholar] [CrossRef]
  24. Tsai, R. A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom. 1987, 3, 323–344. [Google Scholar] [CrossRef] [Green Version]
  25. Heikkila, J. Geometric camera calibration using circular control points. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1066–1077. [Google Scholar] [CrossRef] [Green Version]
  26. Svoboda, T.; Martinec, D.; Pajdla, T. A Convenient Multicamera Self-Calibration for Virtual Environments. Presence Teleoperators Virtual Environ. 2005, 14, 407–422. [Google Scholar] [CrossRef]
  27. Zhengyou, Z. Camera calibration with one-dimensional objects. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 892–899. [Google Scholar] [CrossRef]
  28. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  29. Yan, K.; Tian, H.; Liu, E.; Zhao, R.; Hong, Y.; Zuo, D. A Decoupled Calibration Method for Camera Intrinsic Parameters and Distortion Coefficients. Math. Probl. Eng. 2016, 2016, 1392832. [Google Scholar] [CrossRef] [Green Version]
  30. Koszela, W.; Pawlus, P.; Galda, L. The effect of oil pockets size and distribution on wear in lubricated sliding. Wear 2007, 263, 1585–1592. [Google Scholar] [CrossRef]
  31. Podulka, P. The Effect of Surface Topography Feature Size Density and Distribution on the Results of a Data Processing and Parameters Calculation with a Comparison of Regular Methods. Materials 2021, 14, 4077. [Google Scholar] [CrossRef]
  32. Tan, Ö.; Schmid, H.; Seyfried, V. A new evaluation method for interferometric surface topography measurements which is robust against environmental disturbances. CIRP J. Manuf. Sci. Technol. 2021, 33, 234–239. [Google Scholar] [CrossRef]
Figure 1. The principle of position invariance: (1) Object to be measured (2) Light pen stylus (3) stylus tip center.
Figure 1. The principle of position invariance: (1) Object to be measured (2) Light pen stylus (3) stylus tip center.
Sensors 22 01029 g001
Figure 2. Schematic diagram of the self-calibration method: (1) Stylus tip center (2) The object to be measured (3) The light pen (4) The spherical surface fitted.
Figure 2. Schematic diagram of the self-calibration method: (1) Stylus tip center (2) The object to be measured (3) The light pen (4) The spherical surface fitted.
Sensors 22 01029 g002
Figure 3. System composition diagram: (a) camera; (b) computer; (c) light pen.
Figure 3. System composition diagram: (a) camera; (b) computer; (c) light pen.
Sensors 22 01029 g003
Figure 4. Light pen stylus.
Figure 4. Light pen stylus.
Sensors 22 01029 g004
Figure 5. Pitch angle measurement results.
Figure 5. Pitch angle measurement results.
Sensors 22 01029 g005
Figure 6. Yaw angle measurement results.
Figure 6. Yaw angle measurement results.
Sensors 22 01029 g006
Figure 7. Tilt angle measurement results: (a) 0°~40° and (b) 136°~180°.
Figure 7. Tilt angle measurement results: (a) 0°~40° and (b) 136°~180°.
Sensors 22 01029 g007
Table 1. Self-calibration experiment.
Table 1. Self-calibration experiment.
Stylus TypeABCDE
129.98930.33130.27230.24229.923
229.87030.30530.24830.21429.853
330.08330.29430.25430.24430.058
429.73329.86529.82829.82029.750
529.90529.79229.77029.79429.939
630.13530.05830.05130.08030.127
729.99429.77329.76429.80130.018
829.87729.56029.56129.60929.869
930.18729.89229.90029.95630.209
1030.03629.94829.93129.95830.046
Average value29.98129.98229.95829.97229.979
Absolute error0.0510.0500.0740.0600.053
Relative error0.17%0.17%0.25%0.20%0.18%
Table 2. Single-point repeatability experiment.
Table 2. Single-point repeatability experiment.
Test12345678910AVESTD
Au78.99578.99579.01479.00478.99778.99778.98379.00278.99078.99578.9970.008
v79.96779.97579.96579.96879.96179.96479.98279.95679.97979.98079.9700.009
w790.029789.994790.040790.026790.047790.030789.965790.083789.986789.948790.0150.041
Bu79.06179.07479.05479.06679.06879.07179.05679.06479.06879.04779.0630.008
v80.26280.25080.27580.24780.26180.24380.26880.26080.27180.26980.2610.011
w790.209790.262790.205790.261790.220790.271790.236790.277790.216790.219790.2380.028
Cu79.09479.09679.09079.09479.10379.09479.09379.10279.09679.09779.0960.004
v80.62680.60480.61980.63480.61680.62580.62780.61180.62880.60980.6200.01
w790.286790.298790.263790.222790.283790.235790.231790.302790.250790.289790.2660.03
Du78.90778.90078.89778.90078.90678.90978.89978.88778.89978.90678.9010.006
v79.34779.35279.35279.34879.34079.34079.34579.35779.34979.34279.3470.005
w789.472789.444789.425789.474789.508789.518789.481789.404789.467789.509789.4700.037
Eu79.15579.16679.16679.15979.17879.16779.17579.16779.16479.16979.1670.007
v80.84380.86980.86080.84580.86180.85980.85680.86080.86880.88380.8600.012
w790.859790.855790.882790.908790.881790.845790.909790.907790.864790.843790.8750.026
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shan, D.; Zhang, C.; Zhang, P.; Wang, X.; He, D.; Xu, Y.; Zhou, M.; Yu, G. Self-Calibration Method and Pose Domain Determination of a Light-Pen in a 3D Vision Coordinate Measurement System. Sensors 2022, 22, 1029. https://doi.org/10.3390/s22031029

AMA Style

Shan D, Zhang C, Zhang P, Wang X, He D, Xu Y, Zhou M, Yu G. Self-Calibration Method and Pose Domain Determination of a Light-Pen in a 3D Vision Coordinate Measurement System. Sensors. 2022; 22(3):1029. https://doi.org/10.3390/s22031029

Chicago/Turabian Style

Shan, Dongri, Chenglong Zhang, Peng Zhang, Xiaofang Wang, Dongmei He, Yalu Xu, Maohui Zhou, and Guoqi Yu. 2022. "Self-Calibration Method and Pose Domain Determination of a Light-Pen in a 3D Vision Coordinate Measurement System" Sensors 22, no. 3: 1029. https://doi.org/10.3390/s22031029

APA Style

Shan, D., Zhang, C., Zhang, P., Wang, X., He, D., Xu, Y., Zhou, M., & Yu, G. (2022). Self-Calibration Method and Pose Domain Determination of a Light-Pen in a 3D Vision Coordinate Measurement System. Sensors, 22(3), 1029. https://doi.org/10.3390/s22031029

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop