1. Introduction
Autonomous vehicles are the next step in transportation. To achieve this goal, cars must be equipped with sensors that perceive the environment in detail and accurately. Modern setups ensure safe driving by combining perception sensors such as cameras, radars, and especially LiDARs. LiDAR sensors in particular are one of the key enablers of automotive driving since their output provides direct distance information on every object in the scanned field of view (FoV). LiDAR (short for light detection and ranging) sensors are most commonly used as time-of-flight sensors. They can determine the distance to any object they face by measuring the time that an emitted laser pulse takes to arrive at an object, be scattered by it, and finally return to the sensor and be detected. Modern LiDARs differentiate themselves mainly in their approach to scanning a scene. Depending on the application and light source, capturing the scene can be achieved without a scanner (FLASH-LiDAR), or a combination of one-dimensional scanning and a one-dimensional sensor, or two-dimensional scanning. While, mechanically, one-dimensional scanning systems were the first introduced to this market [
1], many big companies focus on solid-state technology such as micro-electro-mechanical systems (MEMS) to enable steering in two dimensions. The main reason for using MEMS is the scalability of the production and hence the reduction in price [
2].
The requirements for LiDAR MEMS scanners are derived from automotive-use cases, one of which is highway driving. Highway scenarios are crucial since high vehicle speeds lead to long breaking times, which means that even small obstacles can create deadly accidents. Four main parameters must to be addressed for this use case in general: range, FoV, resolution, and frame rate. For the sake of simplification, the following description focuses on 2D-scanning LiDAR sensors. With respect to range, the most substantial influence comes from the detecting aperture, which itself is highly affected by the optical size of the deflecting surface of the scanner. Thus, MEMS scanners with large apertures are needed [
3]. The FoV of a scanning LiDAR is solely created by the used scanners as they steer the illumination and detection cone into the desired direction. The FoV is defined as the maximum solid angle extent that can be covered by the sensor. For the depicted scenario of highway driving, large horizontal FoVs of around 40° to 60° are geometrically required to detect the full road ahead when including mounting tolerances and different road scenarios. Many companies have already committed themselves to buildings sensors with a horizontal FoV of 60° [
4]. The vertical FoV is typically specified as ca. 20–30° [
4,
5]. The horizontal value can be directly projected and used as a specification for the scanning range of the MEMS. Inside the covered FoV, the density of capture is defined by the resolution of the pattern with which the laser is fired while scanning. It enables the differentiation of small objects over vast distances and is typically specified as ca. 0.1° [
4,
5]. Achieving this requires a dynamic characterization of MEMS scanners with low uncertainty. Since pulsing on a sinusoidal oscillation to achieve the resolution is directly proportional to speed and thus to the maximum steering angle (FoV), an accurate measurement of the FoV with a relative uncertainty of 1% is sufficient for this requirement. In operation, this ensures that individual acquisitions are spatially grouped with the same relative uncertainty. This generates a consistent pulse pattern without gaps. Lastly, a high frame rate is needed to increase the confidence in detected objects. This figure of merit describes the rate at which the full FoV is captured exactly once. In general, it can be said that the more optical channels, the higher the oscillation frequencies of the scanner, the lower the vertical FoV, and the smaller the vertical resolution, the higher the frame rate. Thus, scanners oscillate with frequencies of around 250 Hz, enabling high frame rates above 10 Hz–20 Hz [
4,
5]. Dynamically measuring or characterizing the steering angles of such scanners with these merits in mind is a highly difficult task, especially because of the required accuracy and the overall angular range.
Calibration of 3D LiDARs has been performed in the past, mainly focusing on the calibration of the entire device and the effects on mounting [
6,
7]. This ensures that the position of detected objects is correct compared to the coordinate system of the car. The procedure is directly influenced by the specific mounting pose and is especially important whenever multiple sensors are fused. Diverging from this kind of calibration, a different approach is the characterization of the individual scanners within a sensor. As they are deeply integrated into generating the desired capturing pattern in a point cloud, it is enormously important to characterize scanners with a high degree of confidence. The following paragraphs cover possible setups and techniques to characterize such MEMS scanners.
Measuring steering angles can—in its easiest form—be achieved simply by using a laser that is steered by the mirror of the scanner onto a projection plane, which was used as a validation technique by Fujita et al. [
8]. Using trigonometric equations, a displacement of the projected spot gives the steered angle. When dynamically determining the position of a laser spot, the most accurate and most used sensor is a Position-Sensitive Device (PSD). Available as two-dimensional sensors with sizes up to 20 mm × 20 mm, they are capable of measuring the laser spot’s weighted center of intensity with micrometer uncertainties. These sensors have been used in the past for similar applications such as fast autofocus inspections [
9] as well as angle measurements of the relative story displacement [
10]. The dynamic potential of the PSD has also been observed in the past [
11], which showed that dynamic measurements with speeds up to 104.72 m/s (=ca. 10.5 mm/100 μs) are possible with a PSD.
Measurements of mirror-inclination angles or MEMS steering angles with PSDs have been performed by [
8,
12,
13,
14]. Nishibori and Nishibori [
9] focused on a static measurement in a range of ±8° with a measurement uncertainty of 0.5°. Fujita et al. [
8] designed a controlling system with an rms error of 0.21° in an angular range between ±6° for a two-dimensional MEMS mirror and a holed PSD, which showed that the dynamic characterization of MEMS with PSDs is possible. A third study focused on the evaluation of small, dynamic angular rotations with different mirror speeds which led to dynamic measurements with a measurement uncertainty of 1.9 arcseconds (≈0.032°) in a range of 400 arcseconds (≈6.7°) [
13]. The most recent results are from Han Woong Yoo et al. [
14],who described a MEMS-scanner test bench that is able to dynamically measure steering angles of up to 15° with a measurement uncertainty of 0.026°, by using adjustment mounts and a stage shifting technique.
All described methods for measuring steering angles of mirrors are insufficient for characterizing MEMS scanners for LiDAR applications because none fulfills the required combination of measuring a large FoV of at least 40° dynamically with low measurement uncertainties. Furthermore, current setups require many accurate calibration steps. To achieve the derived merits, we propose a novel measurement technique of characterizing large-FoV MEMS scanners by using a projection screen and a so-called ray trace shifting technique of a PSD camera setup. We further extend the shifting technique described in [
14] to reduce the necessary alignment steps of the scanner and enable a fast characterization procedure. The system was calibrated using a translation stage and the camera model of Zhang [
15]. The capability of the built measurement system was then verified statically with a rotation stage. A final validation step compared the measured FoV for three different driving amplitudes of a MEMS scanner with a reference to quantify the results.
2. Setup and Methods
This chapter explains the components and the general setup of the system in detail. It further describes how the setup is calibrated and validated. Finally, it provides the equations that are necessary to derive the steering angle of the MEMS scanner from the measured sensor data.
2.1. General Setup Description
The proposed setup captures a projected laser spot on an optical diffuser screen with a calibrated imaging system consisting of a PSD lens combination. The steering angle of the MEMS scanner can be derived from the knowledge of the imaging system and the distance between the scanner and the diffuser screen. This enables the measurement of large steering angles, high dynamic bandwidth, and small measurement uncertainties thanks to the quality of the PSD as a detector. A schematic of the setup for characterizing MEMS-based LiDAR scanners is depicted in
Figure 1.
As can be seen in
Figure 1, the setup consists of two main parts: the emission path (a) and the detection path (b). The emission path comprises a laser, two mirrors, and the MEMS scanner. The laser, a 200 mW CW laser (RLDH980-200-3, Roithner Lasertechnik GmbH, Vienna, Austria) with a center wavelength of 980 nm and a collimating lens, generates a laser spot with a diameter of approx. 8 mm at the projection screen. The mirrors adjust the beam to be perpendicularly aligned in regards to the base plate. The laser beam is steered by the MEMS scanner under test, which is a MEMS scanning module similar to [
16] with a variable scanning amplitude. Here, the driving amplitude is set to reach ca. 50°. The steered laser beam hits the screen, where a projection figure is created. For a perfectly sinusoidal oscillating scanner, a straight, horizontal line is created as a projection figure.
The screen itself, which is 23 mm to 48 mm away from the MEMS scanner, is a transmissive glass diffuser (Broadband Hybrid Diffusers, #36619, Edmund Optics Inc, Barrington, NJ, United States) with a Lambertian scattering distribution. The scattered light from the diffuser is captured by an aspheric lens (A240TM-B, Thorlabs GmbH, Munich, Germany) with a focal length of 8 mm, which creates an image of the projection figure on a two-dimensional PSD sensor (PDP90A, Thorlabs GmbH, Munich, Bavaria, Germany) with a total sensing area of 9.1 mm × 9.1 mm and a bandwidth of 16 kHz. The PSD is connected to an oscilloscope (RTB2004, Rohde & Schwarz GmbH & Co. KG, Munich, Germany), which is triggered by the control unit of the MEMS scanner. The oscilloscope captures the PSD data and saves them for postprocessing. The measured voltages are converted into sensor positions according to the following Equation (1) from [
17]. The symbols are slightly altered to be consistent throughout.
Here, the positions and on the PSD can be calculated with the length of the PSD sensor = 10 mm and the three voltages (, , and ) that are output by the PSD. Here, and are proportional to the laser spot’s position. The voltage is proportional to the laser spot’s intensity.
The calculated PSD position is transformed onto the coordinate system of the screen based on camera calibration. This calibration used the pinhole model of Zhang [
15] in combination with the full Brown–Conrady distortion model [
18]. The process and calibration steps are explained in more detail in
Section 2.2.
To compensate for misalignments of the scanner, a special method, herein denoted as ray trace shifting technique (RTST), is used. Therefore, the detection part of the system (see
Figure 1b) is mounted on a translation stage (MTS25/M-Z8, Thorlabs GmbH, Munich, Germany) with a total traveling range of 25 mm, which is oriented parallel to the static beam path. This enables measurements of the projection figure at different distances between the screen and the scanner and generates additional information about the correct distance between the scanner’s pivotal point in regards to the measurement setup. It further allows measuring static vector-based as well as dynamic steering angles. The method is explained in more detail in
Section 2.3.
In a final step, the sensor reading, the calibration data, and the results of the RTST are used to calculate the horizontal steering angle of the MEMS scanner. The detailed calculation is described in
Section 2.4.
2.2. Camera Calibration
The PSD in combination with the lens must be calibrated since both elements inherently have a distortion, which would decrease the performance of the system. To compensate for this distortion, the PSD camera is calibrated with the use of Zhang’s camera model [
15] and the Brown–Conrady distortion model [
18] against a series of laser spot positions similar to [
19]. Widely used in the field of object detection, this method applies a series of matrix multiplications to transforming the camera’s coordinate system into a real-world coordinate system.
2.2.1. Calibration Process
The measured position on the diffuser screen is calibrated against a grid of laser spots by translating a single laser spot similar to the grid of laser diodes in [
19]. The setup of calibration is depicted in
Figure 2.
Here, the main CW laser beam is translated with two additional mirrors, each fixed to one translation stage. The horizontal stage (Thorlabs LTS300/M) moves across a range of 30 mm, and the vertical stage moves across a range of 5 mm (Thorlabs PT1/M). The laser beam is translated in the and directions along with the measurement range. When setting up the system, it was carefully ensured that the and stages are aligned along the axes of the setup, which are set by the baseplate. For a detailed calibration, a rectangular grid of 29 × 7 measurement positions with a spacing of 1 mm was chosen. The PSD’s signal was captured and processed according to Equation (1) for every position of the and stages. The -stage of the PSD was not moved during the process.
2.2.2. Calibration Arithmetic
The evaluation of the taken measurements is based on the pinhole model for cameras. A simplified and adapted version for this setup from [
15] is expressed in Equation (2).
The parameters and are the horizontal and vertical undistorted coordinates on the PSD sensor surface, which become transformed to the coordinates on the screen and . The first matrix on the right side of the equation is the so-called intrinsic camera matrix, which adjusts for translations , , scaling and , and skew . The second matrix is a simplified extrinsic matrix, which describes the positioning of the scene in regards to the camera in the form of translation and and the rotation around the Z-axis. This expression of the extrinsic matrix is a simplification that can be made due to the fixed distance and orientation between the PSD, the lens, and the projection plane. The parameter s is a scaling factor.
An extension to the model, which is explained in [
18], suggests a series of terms to model the distortion of the lens (Equation (3)):
Here, , and , are constants, and is the length of the radius from the sensor’s origin and to the image of the detected laser spot , . This model calculates the distorted image from an undistorted image.
The measured positions of captured points described in
Section 2.2.1 are optimized against the set positions of the calibration stages with a least-square approach for the
X and
Y directions simultaneously. The resulting parameters from Equations (2) and (3)
. create a full description of the transformation from the PSD to the screen coordinate system. This is used to convert the PSD position to the desired screen position for the measurement.
2.2.3. Calibration Results
Figure 3 shows the results of the calibration.
Figure 3a shows the set positions (black) and the measured positions by the system using the calibrated camera model from
Section 2.2.2 (red). The deviations are depicted in
Figure 3b.
Figure 3a qualitatively shows good agreement of the set and measured data over the entire range of the screen in both directions. The largest deviations occur at positive horizontal screen positions for vertical values far away from the center. A quantitative statement about the quality of the method can be derived from
Figure 3b. Here, it can be seen that the maximum deviations are ca. ± 80 μm in horizontal and ca. ± 50 μm in the vertical direction, respectively. The standard deviations are calculated to be ±29 μm and ±16 μm, respectively.
2.3. Ray Trace Shifting Technique (RTST)
The setup is designed in such a way that the screen calibration must be performed only once. The accurate position and mounting angle of the scanner are not relevant since a unique technique is used to compensate for these effects. This marks the main advantage and benefit of the system because it reduces alignment effort and speeds up the process. The ray trace shifting technique (RTST) enables the measurement of the static steering angles of each scanner. Woong Yoo et al. [
14] used a similar kind of shifting to measure the distance accurately. They used a stage with six degrees of freedom to align each scanner in front of the PSD sensor. The herein-proposed technique, however, is a wide-ranging extension that removes the need to mount each mirror precisely with optical stages. The following paragraphs cover the detailed process of the RTST.
After the first measurement of the projection figure on the screen, the translation stage is shifted to measure the same oscillation at different distances. This changes the position on the screen and generates a scaled projection figure, whose extent is proportionally larger with increasing distance. The concept of the RTST is illustrated in
Figure 4.
As all captures for each stage position are triggered at the same point in time after starting up the scanner, measured positions with the same time stamps can be used to calculate a vector of the steered laser ray. For this, linear regression is used (e.g., red dotted lines). By calculating the inverse tangent (arctangent) of the slope of any vector, a steering angle can be derived. This method is called “vector-based” from here on. The most significant merit is the static orientation angle , which describes the rotational mounting errors of the scanner.
Furthermore, the RTST provides the means to calculate the positioning error of the scanner compared to the system’s coordinates (yellow). Based on the vector calculation of each ray, the intersection of each measurement vector over time with the static steering ray can be evaluated for an oscillating scanner. The intersection distance is equal to the distance between the scanner’s pivotal point and the zero position of the stage. The intersection offset is equal to the difference between the detection axis and the scanning axis in the horizontal plane. Both parameters are retrieved by calculating the mean value across the fully captured time over several oscillations. They provide vital information on the orientation of the scanner in the setup.
When using the RTST as described here with a calibrated screen, we can compensate for misalignments of the scanners directly during the measurement process. This greatly reduces the calibration effort and thus the required time. Furthermore, it allows for dynamic steering-angle measurements.
2.4. Trigonometric Calculation of the Steering Angle
The measurement of the dynamic position of the laser spot on the screen and the results of the RTST are used to dynamically calculate the steering angle of the scanner dynamically. It is based on the trigonometric relationship between the steered ray, the position, and the orientation of the screen. The horizontal steering angle of the scanner with this trigonometric method is calculated as follows (Equation (4))
Here,
X(
t) is the time-dependent horizontal laser-spot position on the screen, while
is the horizontal offset of the detector axis compared to the scanner axis. The parameter
describes the distance between the scanner’s pivotal point and the screen’s surface for a stage position of
Z = 0. It is therefore also equal to the maximum distance between the scanner and the screen. The angle
is the offset steering angle for the non-oscillating scanner in the horizontal direction. The parameters
,
, and
result from the ray trace shifting technique described in
Section 2.3.
4. Discussion
This chapter analyzes the results of the three areas of this paper: the calibration, the static validation with a rotational stage, and the dynamic validation in the form of a proof of concept with an oscillating MEMS scanner.
Calibrating the measured position on the PSD sensor surface against the position on the screen with a laser on several positions resulted in a standard deviation of ±29 μm in the horizontal direction and ±16 μm in the vertical direction, respectively. The camera model of Zhang et al. [
15] that was used therefore seems reasonable and well suited for the setup. Further improvements could be made with an imaging lens with a longer focal length or by using a better controlled single-mode laser that would result in a wider usable area on the screen.
To validate the proposed system, we performed two different validations. The first—a static validation—compared the measured steering angles against the set steering angles by rotating a mirror with a manual rotation stage. The linear regression of the FoVs resulted in a slope of 0.998 for the vector-based and 0.999 for the trigonometric approach, respectively. This verified the correctness of the setup. The remaining absolute deviation of the regression of +0.2° and −0.3° are systematic errors of the absolute steering angle. Considering the maximum relative angle of ca. 50°, the uncertainty of the system is less than 1%.
To show that the system is capable of measuring time-dependent steering angles—especially of MEMS scanners—with the same precision as static ones, a dynamic validation was performed. Therefore, a scanner was operated in a constant sinusoidal mode, and the total absolute FoV was compared to a reference. The vector-based calculation showed the highest deviations of up to 1° compared to the reference. For a dynamic “one-shot” measurement, the trigonometric approach is required. It resulted in a maximum absolute deviation of 0.3°. The traced oscillation was sinusoidal, as was expected. Since all three actuation amplitudes showed the same magnitude of deviation, it can be assumed that dynamic measurements in this angular range are also valid. This would also fit previous results [
11], which demonstrated that dynamic measurements with movements slower than 10 mm per 100 μs are possible when using a similar PSD as with the proposed setup, since the maximum speed of this setup is less than 0.5 mm per 100 μs.
It is evident that the uncertainties for both validations are in the same magnitude of 0.3°. This is supported by the fact that two different references and validation methods were used to validate the performance of this setup.
The remaining uncertainty of the system for dynamic “one-shot” measurements (trigonometric method) of 0.3° does match the required metric of 1% for all tested FoVs statically and dynamically, respectively. For future work, to improve this further, we suggest calibrating a wider area on the screen as mentioned above. This would allow for measurements at greater distances between the scanner and the screen, which reduce the overall uncertainty.