Next Article in Journal
Cloud-Based User Behavior Emulation Approach for Space-Ground Integrated Networks
Next Article in Special Issue
Development and Calibration of Pressure-Temperature-Humidity (PTH) Probes for Distributed Atmospheric Monitoring Using Unmanned Aircraft Systems
Previous Article in Journal
The Identification of Non-Driving Activities with Associated Implication on the Take-Over Process
Previous Article in Special Issue
Sensors and Measurements for UAV Safety: An Overview
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extraction of a Weak Flow Field for a Multi-Rotor Unmanned Aerial Vehicle (UAV) Using High-Speed Background-Oriented Schlieren (BOS) Technology

1
Key Laboratory for Urban Geomatics of National Administration of Surveying, Mapping and Geoinformation, Beijing University of Civil Engineering and Architecture, Beijing 100044, China
2
Shanghai Key Laboratory for Planetary Mapping and Remote Sensing for Deep Space Exploration, Tongji University, Shanghai 200092, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(1), 43; https://doi.org/10.3390/s22010043
Submission received: 22 November 2021 / Revised: 15 December 2021 / Accepted: 20 December 2021 / Published: 22 December 2021
(This article belongs to the Special Issue Advanced UAV-Based Sensor Technologies)

Abstract

:
To optically capture and analyze the structure and changes of the flow field of a weak airflow object with high accuracy, this study proposes novel weak flow field extraction methods based on background-oriented schlieren. First, a fine background pattern texture and a sensor network layout were designed to satisfy the requirement of weak flow field extraction. Second, the image displacement was extracted by calculating the correlation matrix in the frequency domain for a particle image velocimetry algorithm, and further calculations were performed for the density field using Poisson’s equation. Finally, the time series baseline stacking method was proposed to obtain the flow field changes of weak airflow structures. A combustion experiment was conducted to validate the feasibility and accuracy of the proposed method. The results of a quad-rotor unmanned aerial vehicle experiment showed that the clear, uneven, and continuous quantitative laminar flow field could be obtained directly, which overcame the interference of the weak airflow, large field of view, and asymmetrical steady flow.

1. Introduction

At present, the unmanned aerial vehicle (UAV) industry has a wide range of application prospects in both military and civilian operations [1]. With the effectiveness advantages such as taking off and landing without a runway, autonomous operation, and a wide scope of work, multi-rotor UAVs, as the most commonly used type of UAV, are widely employed in military reconnaissance [2,3,4], military strikes [5], disaster rescue, tilt photography imaging, agricultural applications [6,7], logistics transportation, exploration, and other fields [8,9]. However, due to the many rotating lift surfaces [10], strong coupling interference [11], obvious unsteady flow, and poor stability [12] of multi-rotor UAVs, it is essential to understand the mechanisms of their aerodynamic characteristics. Although many engineering prototypes and products exist for current multi-rotor UAVs, the mechanism research is still at the stage of simulation experiments, and it is necessary to obtain real flow field data to further expand the research [13,14].
In general, the structural change characteristics of a flow field are important parameters for UAV stability analysis and structural optimization design. Therefore, special types of equipment and technology have been used to study the characteristics of UAV flow fields, including flow sensors and flow visualization methods. Because it can directly obtain a continuous flow field and a detailed vortex structure, flow visualization technology has become the main research method for studying a flow field. Moreover, in addition to the visual expression of fluid changes, qualitative and quantitative analyses have further gradually emerged to achieve the acquisition and analysis of flow phenomena. In general, flow visualization technologies are mainly divided into three categories [15]: (1) Adding tracer particles: Some particles that are easy to track and locate in the area of influence of the experimental fluid are added, and the movement state of the fluid is displayed by these tracer particles [16]. (2) Surface adhesion coating: Some coating materials (such as pressure-sensitive coating) are attached to the surface of the object to be tested in a wind tunnel experiment. The surrounding flow will change the color of the coating, thereby enabling the analysis of the flow characteristics of the fluid [17]. (3) Optical-based flow visualization method: Compared with the above two types of flow visualization methods, the optical-based flow visualization method uses the deflection characteristics of light in a flow field to display the flow of a fluid and the physical parameters of a flow field through optical changes [18,19,20]. The optical-based flow visualization method has many advantages: (1) Non-contact measurement does not require the addition of tracer particles or additional sensors in the flow field; so, the flow field changes are not affected; (2) Quantitative analysis, which can not only provide an intuitive and vivid flow map but also allow some quantitative measurement data to be obtained for the flow field; (3) with the upgrading of optical equipment such as the improvement of the resolution of cameras, it has become easier to obtain high-precision measurement results [21].
In general, optical-based flow field visualization methods include traditional schlieren technology and background-oriented schlieren (BOS) technology [22]. The traditional schlieren method uses the light deflected by the knife edge cutting diaphragm so that the change of the projected light intensity can be observed in the field of view to achieve the visualization of the flow field, but there are significant limitations on the actual field of view size. Moreover, special customized light paths and equipment are also required, which are expensive [23]. BOS technology is an emerging flow visualization method that can be used to obtain the first derivative of density based on the light deflection and flow field density changes caused by flow field disturbances, in combination with high-speed video measurement methods [24]. It is important to obtain the first derivative of density to perform a quantitative analysis of the flow field changes and correction of aero-optical effects [25]. Moreover, for the density gradient through a given deflection field, the boundary conditions can be solved with Poisson’s equation obtained with the finite element method. Therefore, the quantitative refractive index distribution can be obtained, and the quantitative density field distribution can be obtained using the Gladstone–Dale equation. Over the past 10 years, BOS technology has been frequently used for flow field qualitative analysis and quantitative measurement [26] and has thus become an important basic method for flow field measurement.
The core of BOS technology is to accurately obtain the relationship between the background deflection and the first derivative of the density. Therefore, it is important to obtain the background deflection, which should affect the accuracy of the flow field. At present, particle image velocimetry (PIV) and optical flow are the two main methods used to determine BOS deflections [27,28,29]. Two mainstream optical flow algorithms are Lucas–Kanade (L–K) optical flow and Farneback optical flow methods [30,31]. The L–K optical flow method calculates optical flow based on feature points and two basic assumptions, and its calculation results are accurate, simple and efficient. However, due to the sparse and uneven distribution of feature points, only sparse optical flow field changes can be obtained. The Farneback optical flow method flow uses quadratic polynomial to approximate each neighborhood of frames to obtain the field changes of dense optical flow. However, the calculation is complex and time-consuming, and the global calculation also introduces errors, resulting in the decline of accuracy. PIV is an efficient and convenient particle image processing method that can quickly extract the displacement and velocity fields of particles. In principle, high particle density is required for the PIV algorithm, which is difficult to accomplish in complex or high-strength flow fields [32]. However, there are high-density artificial speckles for the BOS method that satisfy the above conditions. The effect of the cross-correlation algorithm calculated with the PIV is relatively good, and the method is used to calculate the offset in the experiment.
However, there are two limitations for the current BOS methods. One limitation is that the resolution of the BOS system is easy to be affected by the experimental parameters, and the detection objects need to cause large flow field structure disturbance (super wind tunnels and flame or high-temperature objects in the laboratory). These objects are relatively easy to detect and less affected by the experimental parameters, as the characteristics of flow field structure are obvious or basically symmetrical. However, for the weak flow field, there is no appropriate standard to determine the experimental parameters, including the parameters of the background plate and the parameters of the sensor arrangement, so it is difficult to arrange an appropriate experimental device for the experiment. The other limitation is that high frame rate BOS data have a high spatio-temporal resolution, but there are few methods to conduct spatio-temporal analysis and verification of the results, including visual interpretation directly according to the displacement change for the comparison of multi-frame analysis results [33,34] and single baseline numerical analysis in the area with severe change or obvious characteristics [35]. As the structural changes of UAV weak flow field are not obvious between the adjacent frames, it is difficult to make spatio-temporal analysis and verification using visual interpretation and simple baseline analysis methods. Therefore, to extract the flow field structure of the weak wind object accurately, a fine background pattern texture and sensor network layout are presented to build a BOS system construction standard suitable for weak flow field in this study. Moreover, a time series baseline accumulation verification analysis method is proposed to make spatio-temporal analysis and verification for weak flow field.

2. BOS Technology Measurement Principle

The air medium should satisfy the Gladstone–Dale equation between the refractive index and the density of the flow field. Therefore, the density of the flow field can be obtained by calculating the refractive index of the flow field, as follows:
n 1 = K ρ
where n is the refractive index, K is the Gladstone–Dale constant, and ρ is the density of the flow field.
Figure 1 shows a schematic view of the optical path for BOS measurement, which contains two steps. First, a reference image is generated by recording the background pattern without the undisturbed flow field. As shown in Figure 1, the black solid line is an optical path without refraction, which is the projection of the background pattern without the disturbance of the flow field. Then, the background pattern affected by the density of the flow field is photographed as an analysis image. As shown in Figure 1, the blue solid line is an optical path with the refraction field, which is the deflection of the light caused by the disturbance of the flow field. Finally, the image difference of background pattern can be determined between the reference image and the analysis image using the PIV or OF methods, and the displacement Δ y of the background pattern can be obtained with the deflection angle α y in the direction y , which is caused by light deflection, as follows:
Δ y = L D L i L B α y
where L D is the distance between the background plate and the refractive field, L B is the distance between the background plate and the camera lens, and L i is the distance between the camera lens and the image plane.
The light passes through the flow field along the Z axis. d z is the differential along the Z axis. The deflection angle α is projected on the X-Z coordinate plane to obtain the deflection angle of the light in the y direction. For the air medium, the change of the refractive index caused by the flow field disturbance can be ignored. When calculating the deflection angle α y , the refractive index can be regarded as an environmental constant n 0 .
α y = 1 n 0 n y d z
The imaging system has to be focused on the background pattern, which satisfies the following formula:
1 f = 1 L i + 1 L B L i = L B f L B f
where f is the focal length, L i can be represented with L B and f , and the displacement Δ y of the background pattern can be rewritten as follows:
Δ y = L D f L B f α y
By combining Equations (3) and (5), the relationship between the displacement Δ y of the background pattern and the refractive index n of the flow field can be obtained:
Δ y = ( L D L B f ) f n 0 n y d z
According to Equation (6), the larger the L D , the greater the displacement of the background pattern and the higher the sensitivity of the system, which makes it easier to detect the low-density gradients of the flow field. However, with the increase of the system sensitivity, the resolution of the background pattern decreases, which reduces the calculation accuracy of the displacement in the search window area with the cross-correlation algorithm. Therefore, it is necessary to ensure a suitable experimental distance according to the flow field and the actual layout of the measurement object.
The relationship between the displacement Δ y of the background pattern and the density gradient can be obtained by Equation (1), as follows:
Δ y = ( L D K L B f ) f n 0 ρ y d z
When the density field is two-dimensional or approximately two-dimensional, it can be considered that the density gradient remains unchanged along the optical axis. When the width of the disturbed flow field is L , the relationship between the density gradient and the displacement Δ y of the background pattern can be obtained as follows:
ρ y = L ( L B f L D K ) n 0 f Δ y
Similarly, the relationship between the displacement and the density of the background pattern in the X direction can be obtained as follows:
ρ x = L ( L B f L D K ) n 0 f Δ x
Consequently, the density gradient of the flow field satisfies the quantitative relationship shown in Equations (7) and (8); so, the density gradient of the flow field can be calculated by obtaining the displacement of the background pattern in the experimental images.

3. Methodology

The overall concept of the proposed method for obtaining the mechanism of the weak airflow field is shown in Figure 2. The concept can be summarized with the following four aspects: (1) Fine background pattern texture and sensor network layout, (2) Image displacement extraction using the PIV algorithm, (3) Calculation of the density field, and (4) flow field analysis using the time series baseline stacking method.

3.1. Fine Background Texture and Sensor Network Layout

To extract the flow field structure of the weak wind object accurately, an appropriate background plate design (including background contrast, number of points, and point size) and the layout of the background plate and camera were two important factors. At present, different experiments have their own appropriate experimental background pattern and layout setups [36]. However, these setups should satisfy the following basic requirements: (1) The background pattern should have a clear contrast, such as black dots on a white background or white dots on a black background. (2) The number of generated background points should be sufficient to ensure accuracy. (3) A sufficient distance should be kept between the experimental background and the flow field. Therefore, it may be more effective to set a high-density speckle background for the flow field monitoring of weak airflow and further validate the suitable instrument layout.
According to Equation (4), the resolution of a BOS system is affected by the distance L D between the background plate and the disturbed flow field. Increasing L D will increase the light deflection, which should make it easier to detect the low-density gradient of the flow field. Therefore, in this study, to extract the flow field structure of the weak wind object, the speckle background was designed, as shown in Figure 3. The parameter requirements of the background pattern and the experimental layout distance are presented in Table 1.

3.2. Image Deflection Extraction Using PIV Algorithm

3.2.1. Cross-Correlation Algorithm

The cross-correlation algorithm is the key technique for background deflection calculation using PIV. Each image is divided into partially overlapping small sub-images (search area). The search area selected from the reference image is represented as W 1 , and the search area selected from the analysis image is represented as W 2 . Essentially, the cross-correlation algorithm uses a statistical image matching method to find the background pattern deflection from the reference image search area W 1 to the analysis image search area W 2 . By calculating the correlation function C ( m , n ) between each image sequence and the search window of the reference image, the position of the peak of the correlation coefficient represents the displacement of the background pattern. The correlation function C ( m , n ) is obtained as follows:
C ( m , n ) = i j W 1 ( i , j ) W 2 ( i m , j n )
where i and j denote the image plane coordinates in the Χ and Y directions. Equation (9) is too sensitive to the change of the search window in practical applications, and it can be easily affected by the background noise. Therefore, in this study, the mean normalized cross-correlation was adopted to highlight the peak and reduce noise interference.
C ( m , n ) = i j [ W 1 ( i , j ) W 1 ¯ ] [ W 2 ( i m , j n ) W 2 ¯ ] i j [ W 1 ( i , j ) W 1 ¯ ] 2 i j [ W 2 ( i m , j n ) W 2 ¯ ] 2
where W 1 ¯ and W 2 ¯ are the mean values of W 1 and W 2 , respectively. Figure 4 shows the correlation plane obtained with the mean normalized cross-correlation algorithm. The peak position is the position with the highest correlation, and it represents the most likely background pattern displacement from the reference image W 1 to the analysis image W 2 .
In contrast to correlation matrix calculation directly in the spatial domain, PIV is used to calculate the correlation matrix in the frequency domain using a Fast Fourier Transform, which can greatly improve the search efficiency. The correlation theorem states that the cross-correlation of two functions is equivalent to a complex conjugate multiplication of their Fourier transforms:
C ( m , n ) F F T 1 ( W ^ 1 W ^ 2 )
where W ^ 1 and W ^ 2 are the Fourier transforms of the functions W 1 and W 2 , respectively, W ^ 2 is the conjugate complex root of W ^ 2 , and F F T 1 is the inverse Fourier transform.

3.2.2. Peak Finding Algorithm

It requires not only the correlation algorithm but also the sub-pixel peak finding algorithm to obtain the fine displacement of the background pattern. Therefore, the peak finding technology is used to ensure the sub-pixel precision of image deflections. The background images’ intensity profile formed by the particles on the imaging array can be best described by a Gaussian shape [32]. BOS background speckle is similar to the particle imaging model, so the shape of its correlation peak is also a Gaussian function. The integer displacement can be determined directly between the two search windows based on the location of the peak position, and further calculate the non-integer intensity distribution by a twice three-point fit (Equation (13)) to refine the location of peak with sub-pixel precision. The twice three-point method performs one-dimensional “fitting” through the highest value and its two neighborhoods (three-point estimator), and measures the X-axis and Y-axis, respectively, for twice fitting, which can better conform to the peak function, as shown in Figure 5.
It is assumed that an integer peak was found at x , y in a correlation function C ( m , n ) and that the correlation peak satisfies the Gaussian distribution. Then, the accurate location of the correlation peak ( x s u b , y s u b ) can be obtained as follows:
x s u b = x + { log C ( x 1 , y ) log C ( x + 1 , y ) } × { 2 [ log C ( x 1 , y ) + log C ( x + 1 , y ) 2 log C ( x , y ) ] } 1 y s u b = y + { log C ( x , y 1 ) log C ( x , y + 1 ) } × { 2 [ log C ( x , y 1 ) + log C ( x , y + 1 ) ] 2 log C ( x , y ) } 1
After obtaining the accurate location of the correlation peak ( x s u b , y s u b ) , the displacement Δ x Δ y of the background pattern can be obtained by subtracting the center position of the window (half of the window).

3.3. Calculation of Density Field

The displacements Δ y and Δ x of the background pattern are obtained with the PIV algorithm described in the previous section, and the density gradient field of the flow field is obtained by Equations (8) and (9). However, density is an indispensable and important piece of information in practical engineering and experimental research. Therefore, as described in this section, the density field information for the flow field to be measured is calculated using the displacement of the background pattern. If the partial derivatives of the whole displacement vector field are obtained in the X and Y directions, the following Poisson equation can be obtained:
2 ρ x 2 + 2 ρ y 2 = L ( L B f L D K ) n 0 f ( Δ x x + Δ y y )
The constant on the right side of the equation is related to the parameters of the experimental system instrument, and it can be replaced by P , and the density gradient can be replaced by ρ ; hence, the equation can be expressed as follows:
2 ρ = P ( Δ x x + Δ y y )
For the calculated displacement vector field and boundary conditions, the above formula can be solved with the finite difference method or the finite element method to obtain the quantitative refractive index field distribution of the projection integral effect in the measurement area. The quantitative density field information can be calculated with the Gladstone–Dale formula.

3.4. Flow Field Analysis Using Time Series Baseline Stacking Method

For high-speed BOS, although it has the characteristics of high spatial and temporal resolution, the information for the spatial and temporal changes of the flow field structure in the image is separated. Therefore, it is difficult to display and extract the spatial changes of the continuous time series of flow field structure intuitively. In general, by setting the analysis baselines for the regions of interest and the regions with strong flow field changes, the characteristics of the displacement field changes along the baselines can be obtained, and the temporal and spatial changes of the flow field structure can be displayed with the time series broken line graph of the baseline position changes. However, this kind of analysis requires a large inter-frame difference to create an obvious effect. Because the change of the weak airflow is not obvious, it is difficult to obtain the significant time series change characteristics between two adjacent frames. Therefore, in this study, the baseline stacking analysis method is adopted to obtain the change of the weak airflow, and multiple background schlieren images of the continuous time series are selected for baseline extraction. The extracted baseline matrix is superimposed to form a new time series baseline stacking matrix, which contains the change information for the selected baseline in n consecutive frames. The time series variation characteristics of the flow field structure can be displayed, and the high-precision space–time analysis of the flow field structure can be achieved. Figure 6 shows the model construction and analysis process for the time series baseline stacking method.
The pixel matrix of the flow field analysis image is denoted as A ij . The time series baseline is divided into the horizontal baseline a Rj and the vertical baseline a iR .
A ij = a 11 a 1 j a i 1 a i j
a Rj = a R 1 a R j
a iR = a 1 R a i R T
The horizontal baseline analysis is used to successively extract the horizontal baselines a Rj 1 a Rj n of the region of interest in the image sequence matrices A ij 1 A ij n with a frame interval of d .
a Rj 1 = a R j 1 a R j 1 a Rj 1 + d = a R j 1 + d a R j 1 + d a Rj n d = a R j n d a R j n d a Rj n = a R j n a R j n
A new matrix A R j is formed by stacking a Rj 1 a Rj n from top to bottom, which denotes the horizontal baseline stacking graph for n images, and is given by
A R j = a R j 1 a R j n
The vertical baseline analysis is used to successively extract the vertical baseline a i R 1 a i R n of the region of interest in the image sequence matrices A ij 1 A ij n with a frame interval of d .
a i R 1 = a 1 R 1 a i R 1 T a i R 1 + d = a 1 R 1 + d a 1 R 1 + d T a i R n d = a i R n d a i R n d T a i R n = a 1 R n a i R n T
A new matrix A i R is formed by stacking a i R 1 a i R n from left to right, which denotes the vertical baseline stacking graph for n images, and is given by
A i R = a i R 1 a i R n

4. Experiment Results and Analysis

In this study, to obtain a weak airflow field with higher accuracy, a high-speed Complementary Metal–Oxide–Semiconductor (CMOS) transistor camera (CP80-4-M-500, Optronis company, Kehl, Germany), along with a Nikon lens with a focal length of 35 mm (Nikon Corporation, Tokyo, Japan), a high-speed data acquisition card, a workstation, and a tripod, was used to capture the image sequence. The key parameters of the CP80-4-M-500 camera are presented in Table 2.

4.1. Displacement Extraction Algorithm Verification

To verify the accuracy and reliability of the background pattern displacement extraction algorithm, the checkerboard background and artificial speckle background patterns were used to simulate the background displacement. The image size of the checkerboard background was 479 × 624 pixels. To verify the difference between horizontal and vertical directions for the displacement extraction algorithm, three displacement parameters were set for the analyzed image, including moving only 2 pixels vertically downward, moving 2 pixels horizontally to the right, and moving 2 pixels horizontally and vertically together. The commonly used L–K and Farneback optical flow algorithms were compared with the cross-correlation algorithm used in the study. The displacement vector diagrams (expanded by 12 times) are shown in Figure 7, and the evaluation index of displacement extraction is shown in Table 3. The inspection from this figure and table highlights that: (1) The L–K optical flow method produced some larger mismatched results, the resulting root mean squared error (RMSE) was 4.0434 pixels, which greatly reduced the data quality; (2) Farneback optical flow method also produced many wrong displacement value calculations. Its extraction accuracy was only 87.20% for the correct displacement value interval with the threshold of 2 ± 0.2 pixels, and it needed a consumed time of 2.484 s (Table 3). Compared with the above two optical flow methods, the RMSE of the cross-correlation algorithm was less than 0.1 pixel, the extraction accuracy was 100%, and the shortest time was 0.624 s.
The image size of the artificial speckle background was 793 × 982 pixels. To ensure the measurement accuracy of micro flow field distribution, the analyzed background image was set to move 1 pixel horizontally and vertically. The displacement vector diagrams are shown in Figure 8 (expanded by 20 times), and the evaluation index of displacement extraction is shown in Table 4. The inspection from this figure and table highlights that: (1) For the L–K optical flow method, there was no large mismatch with an extraction accuracy of 98.37%, and the RMSE was 0.2511 pixels. However, due to the limitations of feature matching, only some uneven sparse optical flows can be obtained in the regions with obvious features. (2) For the Farneback optical flow method, its extraction accuracy was only 90.68% for the correct displacement value interval with the threshold of 1 ± 0.2 pixels, and it needed a consumed time of 4.848 s with the increase of image size and background texture. (3) Compared with the two optical flow methods, the RMSE of cross-correlation algorithm was only 0.1432 pixels, the extraction accuracy was 100%, and the shortest time was 1.024 s. The simulated results of the two kinds of background plates indicated that the cross-correlation algorithm has a good extraction accuracy and time efficiency in displacement extraction for BOS background pattern.

4.2. Combustion Experiment

In order to establish a non-uniform refractive index field of the flame, the turbulent flow field above the flame of the lighter was used as the experimental object. According to the requirements of the background schlieren technique and the experimental layout, the distribution density of the experimental background pattern exceeded 60% of the background plate, which could improve the resolution of the weak airflow. The flame density measurement experimental equipment setup is shown in Figure 9. The size of the background plate was 238.5 mm × 195.5 mm, the background plate was directly placed on the transparent glass, and sunlight was used for illumination, which increased the brightness of the background and did not require special light–source illumination. Moreover, the size of the background pattern point in the image had to be greater than 2 pixels. Therefore, the size of the background pattern point setting was about 2–4 pixels in this study. A high-speed CP80-4-M-500 camera was used, along with a Nikon lens with a focal length of 35 mm. The distance between the lens and the background plate was set to about 1 m, and the flame of the lighter was about 0.55 m away from the background, which developed a field of view of 471.0 mm × 351.6 mm. The optical axis of the camera was perpendicular to the background plate, and the camera was focused on the background pattern. The combustion experiment lasted for about 5 s, and the resulting 2500 images were acquired from static to burning for the flame.
To obtain the accurate and dense slow convective flows, the general search area size was 64 × 64 pixels, and the search area size was reduced to 16 × 16 pixels in the weak airflow experiment with a step size of 8 pixels. If the size of the search area was reduced or the step size was increased, the inaccurate results would be sharply increased, and the signal-to-noise ratio would be reduced. Moreover, to avoid analysis algorithm irritations, disturbing elements such as the measurement object and the instruments were masked, for which the values of the flows were equal to zero.
With the PIV algorithm described in Section 3.2, a 16 × 16 pixel search area was used to obtain the displacement vector diagram of the background pattern, as shown in Figure 10a. The green arrow in the figure points to the direction of displacement, and the length represents the size of the displacement. The size is displayed in the figure in an eight-fold magnification relationship. The specific numerical relationship was usually split and displayed in the displacement cloud diagram in the horizontal and vertical directions. The main transformation of the flame combustion analysis was the horizontal direction. Therefore, the horizontal displacement that was obtained is shown in Figure 10b. The yellow component in the figure represents the positive displacement to the right, and the blue component in the figure represents the negative displacement to the left. The maximum displacement in the figure is about 1150 pixels in the X direction of positive 1.2 pixels and about 1350 pixels in the X direction of negative 0.8 pixels. Then, the density field was reconstructed, and the density field distribution diagram was obtained, as shown in Figure 10c. It was obvious that the density of the two positions with large displacement transformations changed greatly, which might have been caused by the supplement of the surrounding air after the flame burned oxygen.
The flow field information for the flame combustion was clearly obtained in the schlieren image, and the flow field changed obviously from ignition to combustion stability. Since the environment was not closed flow, the flame was not a symmetrical structure. The boundary conditions of the flow field above the flame were easy to distinguish, and the structure of the flow field with deflection changed significantly with time and had a certain regularity. However, when the combustion was stable, due to the high shooting frame rate of the high-speed camera, the change of the flow field per frame was small, which was not easy to distinguish. Therefore, in order to extract the temporal change information of the flow field, the analysis method of a frame skipping the extracting baseline and generating a broken line graph was used to extract the maximum value of the baseline position and the overall change trend. As shown in Figure 11, the sequence schlieren images with stable combustion after 550 frames were compared and analyzed at intervals of 10 frames (from frame 550 to frame 600), and the red dotted line was set at 490 pixels (14.3 cm above the flame) in the vertical direction, which was the horizontal analysis baseline.
The flow field 14.3 cm above the flame was selected as the region of interest for the horizontal analysis. As shown in Figure 12, the peak of the outer flame boundary on the right side of the images from frame 550 to frame 570 of the flow field showed a downward trend, that from frame 580 to frame 590 showed an upward trend, that at frame 600 showed a downward trend again, and the change interval of its flow field was about 10–30 frames. The reason for this change might have been the change of the full combustion position since the indoor airflow caused the combustion air supplement to vary. For a more obvious analysis, the fine structural changes of the flow field above the flame in the continuous frames are shown with the baseline stacking analysis method to obtain the change of the airflow of the flame, which was selected from frames 550 to 600 of the background schlieren images of the continuous time series for baseline extraction.
Figure 13 shows the stacking diagram of the horizontal baselines from frames 550 to 600 stacked from top to bottom, and the figure shows the temporal change characteristics of the image and the intensity of the change. The maximum displacement of the flow field was 1.0–1.2 pixels/frame, and the flow field had to be stable in a static state in a symmetrical state. With the action of the indoor circulating flow field, the left and right deviations relative to the center line occurred, which was a normal phenomenon. The method successfully showed the boundary conditions of the flow field, and the intensity change trend was also consistent with the previous analysis of the line graph. The variation law showed that the maximum position of the flow field (baseline analysis position) weakened first, then became stronger, and then became weaker again. The results showed that the time series baseline stacking method could accurately reflect the obvious changes of the weak airflow.

4.3. UAV Flow Field Visualization

In order to establish a non-uniform refractive index field for a quad-rotor UAV, an experimental system was constructed to obtain the flow field structure of a quad-rotor UAV for visualization. As shown in Figure 14, a high-speed camera CP80-4-M-500 was used, along with a Nikon lens with a focal length of 35 mm. The size of the background plate was 1200 mm × 960 mm, and four pieces were used to form a large background plate. The background plate (attached to the wall) was perpendicular to the optical axis of the camera. The distance between the lens of the camera and the background plate was adjusted to 7800 mm, and the distance from the UAV flow field to the background plate was about 4000 mm, which resulted in a field of view of 2516.3 mm × 1878.2 mm. It was almost impossible to achieve that using the traditional schlieren method. The large field of view was more conducive to observing the changes in the UAV’s airflow. In the experiment, two sets of 5 s, 2500-frame UAV flight sequence images were obtained.
In this study, to obtain the displacement of the background pattern with the frequency domain cross-correlation algorithm, the search area was 16 × 16 pixels with a step size of 8 pixels, which could ensure that accurate and dense flow field data were obtained. Since the flow field caused by the four-rotor UAV was complex and difficult to analyze, for the experimental handheld UAV, only the wind flows of the two left rotors were collected and analyzed. The obtained UAV flow field displacement grayscale diagram is shown in Figure 15. Moreover, due to the moving interference of boundary, gap and UAV, it was easy to produce false matching. Therefore, the red area was the masked area to prevent the wrong data from affecting the flow field structure analysis.
To clearly show the obvious flow field changes, the frame difference of 25 frames was used to obtain the horizontal displacement field of the left double rotor, as shown in the red area in Figure 16. In the figure, the red component on the left in the red circle represents the airflow moving to the right, and the blue component on the right in the red circle represents the airflow moving to the left. Comparison and analysis of the three images revealed that the horizontal flow field of the handheld UAV was relatively stable during the selected time period. Moreover, the fine structural changes of the flow field under the UAV in the continuous frames were shown with the time series baseline stacking analysis method, and frames 570 to 620 of the background schlieren images of the continuous time series were selected for baseline extraction. Two continuous 50-frame images at the two baselines (baselines a and b shown in Figure 16) from frames 570 to 620 were superimposed from top to bottom in the horizontal direction, as shown in Figure 17a(1),b(1). Simultaneously, to facilitate the visual analysis of baseline stacking analysis chart, the analysis chart was stained with imagesc function, as shown in Figure 17a(2),b(2). The maximum displacement of the flow field was 0.1 pixel/frame. The flow field structures of the two baseline superposition diagrams were almost stable within 50 frames, which indicated that the flow field of the horizontal direction under the UAV was relatively stable. Therefore, the horizontal structure of the flow field was continuous and stable at almost the same position. However, there was only a slight deviation, which might have been caused by the vibration of the handheld or indoor disturbed airflow. Moreover, to further evaluate the preciseness of time series baseline stacking method for flow field analysis, the data set was expanded 100 frames, from 545 to 645. As shown in Figure 18, the flow field structure of the two baseline overlay images of 545 to 645 frames was also almost stable, which proved the stable state of the horizontal structure of the UAV flow field.
The vertical displacement field of the left double-rotor was obtained with the frame difference of 25 frames, as shown in the red area in Figure 19. In the figure, the red component on the left in the red circle represents the airflow moving downwards, and the blue component on the right in the red circle represents the airflow moving upwards. Through the comparative analysis of the three images, it was found that the vertical flow field of the UAV changed significantly, the vertical structure of the flow field changed upward at an interval of 25 frames, and the structure position changed by more than 50 pixels, which represented the vertical lift change of the UAV to maintain its own suspension. Moreover, two continuous 50-frame images at the two baselines (baselines a and b shown in Figure 19) from frames 570 to 620 were superimposed from left to right in the vertical direction, as shown in Figure 20a(1),b(1). Simultaneously, to facilitate the visual analysis of baseline stacking analysis chart, the analysis chart was stained with imagesc function, as shown in Figure 20a(2),b(2). It was found that the characteristics of the strong eddy current generated by the UAV rose continuously with time, which was consistent with the change trend of the 25-frame skipping analysis diagram. The indirect velocity measurement of the BOS flow field could be achieved by using the eddy current positioning analysis method (the displacement was extracted during 50 frames of the strong eddy current position, as the red line shown in Figure 20b(2), which was difficult to achieve with the traditional BOS. The average 50 frames of the strong eddy current structure changed by 205 pixels, and a wind speed of about 4.969 m/s was generated by the two propellers of the UAV. Moreover, to demonstrate the accuracy of the results, the data set was also expanded to 100 frames from 545 to 645. As shown in Figure 21, it changed continuously and maintained a steady rise for the strong eddy current characteristics of the two baseline superimposed images from 545 to 645 frames. The strongly eddy structures changed 288 pixels in 70 frames, as the red line shown in the Figure 21b(2), and the two propellers of the UAV produced wind speeds of about 4.9869 m/s, with only a 17 mm error.
The displacement field generated in the experiment showed that the maximum offset caused by the UAV flow field was about a sub-pixel, and the value of the UAV flow field was small, which made it extremely difficult to capture and display the UAV flow field. The difference between the flow field structure and the environmental disturbance was only one order of magnitude. This made it easy for the flow field to be confused with the environment, and it was also difficult to distinguish the flow field from the environmental noise. However, the structural change law of a strong eddy current was successfully extracted with the time series baseline stacking method, which also fully proved the stability and continuity of the flow field. Therefore, the experimental results showed that the proposed method was effective. The experimental results proved that the weak flow field extraction methods successfully distinguished between environmental disturbance and UAV wind flow, and clear, continuous, and stable results were obtained.

5. Conclusions

Conventional measurement methods are generally constrained by space and equipment for weak wind flow extraction, and strict optical path conditions make it very difficult to perform measurements in outdoor environments. BOS technology is sensitive to the refractive index gradient in a transparent medium, and it can be used for the visualization of objects with weak wind flows. Therefore, in this paper, a novel flow field structure analysis method is proposed for weak airflow. The results presented in the paper clearly highlight the following points.
(1) A fine background pattern texture and a sensor network layout were designed to satisfy the requirement of weak flow field extraction. The results of the combustion and a quad-rotor UAV experiment indicated that the clear, uneven, and continuous quantitative laminar flow field could be obtained directly, which overcame the interference of the weak airflow, large field of view, and asymmetrical steady flow. The results were also validated using the continuity and regularity of the flow field and the consistency of the difference.
(2) By comparing the performance of cross-correlation algorithm with L–K optical flow and Farneback optical flow methods on checkerboard background and artificial speckle background, it was concluded that the cross-correlation algorithm has a good extraction accuracy and time efficiency in displacement extraction for BOS background pattern.
(3) The proposed baseline stacking method was used to obtain the time series change characteristics of the flow field, and the flow field structural change feature was used to inversely calculate the wind speed of the flow field.
(4) With the current goal of weak wind flow experiments in small indoor and outdoor scenarios, the flow field obtained with the weak flow field extraction method could meet the visualization requirements, the sequence data could be stacked and demonstrated by using the relevant baseline, and the temporal and spatial information for the flow field could be displayed with intuitive correlation.

Author Contributions

Conceptualization, X.L. and X.T.; methodology, X.L., T.G., P.Z. and Z.J.; validation, X.L., T.G. and X.T.; formal analysis, X.L., T.G., P.Z. and Z.J.; writing—original draft preparation, T.G.; writing—review and editing, X.L.; project administration, X.T.; funding acquisition, X.L. and X.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of the People’s Republic of China, grant number 2018 YFE0206100, the National Natural Science Foundation of China, grant number 41871367, and the BUCEA Postgraduate Innovation Project.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mukhamediev, R.I.; Symagulov, A.; Kuchin, Y.; Zaitseva, E.; Bekbotayeva, A.; Yakunin, K.; Assanov, I.; Levashenko, V.; Popova, Y.; Akzhalova, A.; et al. Review of Some Applications of Unmanned Aerial Vehicles Technology in the Resource-Rich Country. Appl. Sci. 2021, 11, 10171. [Google Scholar] [CrossRef]
  2. Li, L.; Sheng, W.; Hu, C. Research on Formation Keeping of Multi-rotor UAVs Based on Improved Virtual Structure Method. J. Phys. Conf. Ser. 2020, 1631, 012106. [Google Scholar] [CrossRef]
  3. Liu, J.; Xiang, J.; Jin, Y.; Liu, R.; Yan, J.; Wang, L. Boost Precision Agriculture with Unmanned Aerial Vehicle Remote Sensing and Edge Intelligence: A Survey. Remote Sens. 2021, 13, 4387. [Google Scholar] [CrossRef]
  4. Petrova, T.; Petrov, Z. Long Term Development Perspectives for UAV Potential. Int. E-J. Adv. Soc. Sci. 2020, 6, 45–53. [Google Scholar]
  5. Oh, D.; Han, J. Smart Search System of Autonomous Flight UAVs for Disaster Rescue. Sensors 2021, 21, 6810. [Google Scholar] [CrossRef]
  6. Hassler, S.C.; Baysal-Gurel, F. Unmanned Aircraft System (UAS) Technology and Applications in Agriculture. Agronomy 2019, 9, 9618. [Google Scholar] [CrossRef] [Green Version]
  7. Mogili, U.R.; Deepak, B.B.V.L. Review on Application of Drone Systems in Precision Agriculture. Procedia Comput. Sci. 2018, 133, 502–509. [Google Scholar] [CrossRef]
  8. Lee, S.; Choi, Y. Reviews of unmanned aerial vehicle (drone) technology trends and its applications in the mining industry. Geosyst. Eng. 2016, 19, 197–204. [Google Scholar] [CrossRef]
  9. Greenwood, W.; Lynch, J.; Zekkos, D. Applications of UAVs in Civil Infrastructure. J. Infrastruct. Syst. 2019, 25, 04019002. [Google Scholar] [CrossRef]
  10. Han, D.; Gwak, D.Y.; Lee, S. Noise prediction of multi-rotor UAV by RPM fluctuation correction method. J. Mech. Sci. Technol. 2020, 34, 1429–1443. [Google Scholar] [CrossRef]
  11. Chan, C.W.; Kam, T.Y. A procedure for power consumption estimation of multi-rotor unmanned aerial vehicle. J. Phys. Conf. Ser. 2020, 1509, 012015. [Google Scholar] [CrossRef]
  12. Fu, Z.; Jiang, J.; Gao, Y.; Krienke, B.; Wang, M.; Zhong, K.; Cao, Q.; Tian, Y.; Zhu, Y.; Cao, W.; et al. Wheat Growth Monitoring and Yield Estimation based on Multi-Rotor Unmanned Aerial Vehicle. Remote Sens. 2020, 12, 508. [Google Scholar] [CrossRef] [Green Version]
  13. Kose, O.; Oktay, T. Simultaneous quadrotor autopilot system and collective morphing system design. Aircr. Eng. Aerosp. Technol. 2020, 92, 1093–1100. [Google Scholar] [CrossRef]
  14. Kose, O.; Oktay, T. Investigation of the effect of differential morphing on forward flight by using PID algorithm in quadrotors. J. Aviat. 2020, 4, 15–21. [Google Scholar]
  15. Adrian, R.J. Twenty years of particle image velocimetry. Exp. Fluids 2005, 39, 159–169. [Google Scholar] [CrossRef]
  16. Xu, P.; Babanezhad, M.; Yarmand, H.; Marjani, A. Flow visualization and analysis of thermal distribution for the nanofluid by the integration of fuzzy c-means clustering ANFIS structure and CFD methods. J. Vis. 2020, 23, 97–110. [Google Scholar] [CrossRef]
  17. Masato, D.; Sorgato, M.; Babenko, M.; Whiteside, B.; Lucchetta, G. Thin-wall injection molding of polystyrene parts with coated and uncoated cavities. Mater. Des. 2018, 141, 286–295. [Google Scholar] [CrossRef] [Green Version]
  18. Raissi, M.; Yazdani, A.; Karniadakis, G.E. Hidden fluid mechanics: A Navier-Stokes informed deep learning framework for assimilating flow visualization data. arXiv 2018, arXiv:1808.04327. [Google Scholar]
  19. Günther, T.; Gross, M.; Theisel, H. Generic objective vortices for flow visualization. ACM Trans. Graph. (TOG) 2017, 36, 1–11. [Google Scholar] [CrossRef]
  20. Kumar, R.; Pancholi, V.; Bharti, R.P. Material flow visualization and determination of strain rate during friction stir welding. J. Mater. Process. Technol. 2018, 255, 470–476. [Google Scholar] [CrossRef]
  21. Settles, G.S.; Hargather, M.J. A review of recent developments in schlieren and shadowgraph techniques. Meas. Sci. Technol. 2017, 28, 042001. [Google Scholar] [CrossRef]
  22. Dalziel, S.B.; Hughes, G.O.; Sutherland, B.R. Whole-field density measurements by ‘synthetic schlieren’. Exp. Fluids 2000, 28, 322–335. [Google Scholar] [CrossRef]
  23. Richard, H.; Raffel, M. Principle and applications of the background oriented schlieren (BOS) method. Meas. Sci. Technol. 2001, 12, 1576. [Google Scholar] [CrossRef]
  24. Venkatakrishnan, L.; Meier, G.E.A. Density measurements using the background oriented schlieren technique. Exp. Fluids 2004, 37, 237–247. [Google Scholar] [CrossRef] [Green Version]
  25. Zhao, Y.X.; Yi, S.H.; Tian, L.F.; He, L.; Cheng, Z. An experimental study of aero-optical aberration and dithering of supersonic mixing layer via BOS. Sci. China Phys. Mech. Astron. 2010, 53, 81–94. [Google Scholar] [CrossRef]
  26. Raffel, M. Background-oriented schlieren (BOS) techniques. Exp. Fluids 2015, 56, 1–17. [Google Scholar] [CrossRef] [Green Version]
  27. Brox, T.; Bruhn, A.; Papenberg, N.; Weickert, J.; Weickert, J. High accuracy optical flow estimation based on a theory for warping. In Proceedings of the European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; Volume 3024, pp. 25–36. [Google Scholar]
  28. Richard, H.; Raffel, M.; Rein, M.; Kompenhans, J.; Meier, G.E.A. Demonstration of the applicability of a background oriented schlieren (BOS) method. In Techniques for Fluid Mechanics; Springer: Berlin/Heidelberg, Germany, 2002; pp. 145–156. [Google Scholar] [CrossRef]
  29. Meier, G.E.A. Computerized background-oriented schlieren. Exp. Fluids 2002, 33, 181–187. [Google Scholar] [CrossRef]
  30. Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision (darpa). Proc. IJCAI 1981, 81, 674–679. [Google Scholar]
  31. Farnebck, G. Two-Frame Motion Estimation Based on Polynomial Expansion. In Proceedings of the Scandinavian Conference on Image Analysis, Halmstad, Sweden, 29 June–2 July 2003; Volume 2749, pp. 363–370. [Google Scholar]
  32. Willert, C.E.; Gharib, M. Digital particle image velocimetry. Exp. Fluids 1991, 10, 181–193. [Google Scholar] [CrossRef]
  33. Goldhahn, E.; Seume, J. The background oriented schlieren technique: Sensitivity, accuracy, resolution and application to a three-dimensional density field. Exp. Fluids 2007, 43, 241–249. [Google Scholar] [CrossRef]
  34. Rajendran, L.K.; Bane, S.P.M.; Vlachos, P.P. Dot tracking methodology for background-oriented schlieren (BOS). Exp. Fluids 2019, 60, 162. [Google Scholar] [CrossRef] [Green Version]
  35. Becher, L.; Voelker, C.; Rodehorst, V.; Kuhne, M. Background-oriented schlieren technique for two-dimensional visualization of convective indoor air flows. Opt. Lasers Eng. 2020, 134, 106282. [Google Scholar] [CrossRef]
  36. Gardner, A.D.; Raffel, M.; Schwarz, C.; Braukmann, J.N.; Wolf, C.C. Reference-free digital shadowgraphy using a moving bos background. Exp. Fluids 2020, 61, 1–5. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Simplified model of light deflection in the refractive index field in a background-oriented schlieren setting. Black solid line: optical path without refractive index field. Blue solid line: light path with refraction field. The blue dotted line represents the virtual path of the light.
Figure 1. Simplified model of light deflection in the refractive index field in a background-oriented schlieren setting. Black solid line: optical path without refractive index field. Blue solid line: light path with refraction field. The blue dotted line represents the virtual path of the light.
Sensors 22 00043 g001
Figure 2. Framework of the entire methodology.
Figure 2. Framework of the entire methodology.
Sensors 22 00043 g002
Figure 3. Designed speckle background.
Figure 3. Designed speckle background.
Sensors 22 00043 g003
Figure 4. Mean normalized correlation function (sample image of size 65 × 65 pixels).
Figure 4. Mean normalized correlation function (sample image of size 65 × 65 pixels).
Sensors 22 00043 g004
Figure 5. Principle of the Gaussian 2·3-point fit.
Figure 5. Principle of the Gaussian 2·3-point fit.
Sensors 22 00043 g005
Figure 6. Time series baseline stacking method.
Figure 6. Time series baseline stacking method.
Sensors 22 00043 g006
Figure 7. Displacement vector diagrams of checkerboard background. (a) Only 2 pixels downward, (b) only 2 pixels horizontally to the right, and (c) 2 pixels horizontally and vertically together; 1 L–K optical flow, 2 Farneback optical flow, and 3 Cross-correlation.
Figure 7. Displacement vector diagrams of checkerboard background. (a) Only 2 pixels downward, (b) only 2 pixels horizontally to the right, and (c) 2 pixels horizontally and vertically together; 1 L–K optical flow, 2 Farneback optical flow, and 3 Cross-correlation.
Sensors 22 00043 g007
Figure 8. Displacement vector diagrams of the artificial speckle background. (a) L–K optical flow, (b) Farneback optical flow, and (c) cross-correlation.
Figure 8. Displacement vector diagrams of the artificial speckle background. (a) L–K optical flow, (b) Farneback optical flow, and (c) cross-correlation.
Sensors 22 00043 g008
Figure 9. Flame measurement setup. (1) High-speed camera with a 35 mm lens. (2) Handheld lighter. (3) Speckle background plate. (4) Data operation and storage industrial control machine (including input and output accessories).
Figure 9. Flame measurement setup. (1) High-speed camera with a 35 mm lens. (2) Handheld lighter. (3) Speckle background plate. (4) Data operation and storage industrial control machine (including input and output accessories).
Sensors 22 00043 g009
Figure 10. Displacement of flame. (a) Background displacement vector diagram, (b) horizontal displacement diagram, and (c) density field distribution diagram.
Figure 10. Displacement of flame. (a) Background displacement vector diagram, (b) horizontal displacement diagram, and (c) density field distribution diagram.
Sensors 22 00043 g010
Figure 11. Schematic diagram of frame skipping baseline analysis.
Figure 11. Schematic diagram of frame skipping baseline analysis.
Sensors 22 00043 g011aSensors 22 00043 g011b
Figure 12. Frame skipping baseline analysis diagram.
Figure 12. Frame skipping baseline analysis diagram.
Sensors 22 00043 g012
Figure 13. Time series horizontal baseline stacking diagram. (a) A grayscale image, and (b) a jet color image.
Figure 13. Time series horizontal baseline stacking diagram. (a) A grayscale image, and (b) a jet color image.
Sensors 22 00043 g013
Figure 14. UAV measurement setup. (1) High-speed camera and 35 mm lens. (2) Handheld UAV. (3) Speckle background plate. (4) Data operation and storage industrial control machine (including input and output accessories).
Figure 14. UAV measurement setup. (1) High-speed camera and 35 mm lens. (2) Handheld UAV. (3) Speckle background plate. (4) Data operation and storage industrial control machine (including input and output accessories).
Sensors 22 00043 g014
Figure 15. Displacement grayscale diagram of UAV flow field.
Figure 15. Displacement grayscale diagram of UAV flow field.
Sensors 22 00043 g015
Figure 16. Frame skipping analysis diagram and horizontal baseline setting. (a) T = frame 570, (b) T = frame 595, (c) T = frame 620.
Figure 16. Frame skipping analysis diagram and horizontal baseline setting. (a) T = frame 570, (b) T = frame 595, (c) T = frame 620.
Sensors 22 00043 g016
Figure 17. The stacked baseline analysis graphs in the horizontal direction of 50 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Figure 17. The stacked baseline analysis graphs in the horizontal direction of 50 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Sensors 22 00043 g017
Figure 18. The stacked baseline analysis graphs in the horizontal direction of 100 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Figure 18. The stacked baseline analysis graphs in the horizontal direction of 100 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Sensors 22 00043 g018
Figure 19. Frame skipping analysis diagram and vertical baseline setting. (a) T = frame 570, (b) T = frame 595, (c) T = frame 620.
Figure 19. Frame skipping analysis diagram and vertical baseline setting. (a) T = frame 570, (b) T = frame 595, (c) T = frame 620.
Sensors 22 00043 g019
Figure 20. The stacked baseline analysis graphs in the vertical direction of 50 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Figure 20. The stacked baseline analysis graphs in the vertical direction of 50 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Sensors 22 00043 g020
Figure 21. The stacked baseline analysis graphs in the vertical direction of 100 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Figure 21. The stacked baseline analysis graphs in the vertical direction of 100 frames. a(1) Baseline a (gray scale), and a(2) baseline a (RGB color). b(1) Baseline b (gray scale), and b(2) baseline b (RGB color).
Sensors 22 00043 g021
Table 1. Parameters of the requirements of the speckle background and the experimental layout.
Table 1. Parameters of the requirements of the speckle background and the experimental layout.
ParametersNumber of Background PointsSize of Background PointsLayout Distance
>60% of the background2–3 pixels 1 / 2   L B   >   L D > 2/3
Table 2. CP80-4-M-500 specification.
Table 2. CP80-4-M-500 specification.
ParameterInformation
Resolution2304 × 1720
Max Framerate (8 bit)500
Active sensor area16.13 mm × 12.04 mm
Pixel size in μm7 μm × 7 μm
Shortest exposure time2 ms
Sensitivity9 V/lux.s
Storage formatBMP, JPG, TIFF, AVI
Table 3. Evaluation index of displacement extraction for the checkerboard background.
Table 3. Evaluation index of displacement extraction for the checkerboard background.
AlgorithmRMSEConsumed TimeAccuracy
L–K optical flow4.0434 pixels0.823 s98.60%
Farneback optical flow0.6509 pixels2.484 s87.20%
Cross-correlation0.0828 pixels0.624 s100%
Table 4. Evaluation index of displacement extraction for the artificial speckle background.
Table 4. Evaluation index of displacement extraction for the artificial speckle background.
AlgorithmRMSEConsumed TimeAccuracy
L–K optical flow0.2511 pixels1.261 s98.37%
Farneback optical flow0.1884 pixels4.848 s90.68%
Cross-correlation0.1432 pixels1.024 s100%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, X.; Guo, T.; Zhang, P.; Jia, Z.; Tong, X. Extraction of a Weak Flow Field for a Multi-Rotor Unmanned Aerial Vehicle (UAV) Using High-Speed Background-Oriented Schlieren (BOS) Technology. Sensors 2022, 22, 43. https://doi.org/10.3390/s22010043

AMA Style

Liu X, Guo T, Zhang P, Jia Z, Tong X. Extraction of a Weak Flow Field for a Multi-Rotor Unmanned Aerial Vehicle (UAV) Using High-Speed Background-Oriented Schlieren (BOS) Technology. Sensors. 2022; 22(1):43. https://doi.org/10.3390/s22010043

Chicago/Turabian Style

Liu, Xianglei, Tongxin Guo, Pengfei Zhang, Zhenkai Jia, and Xiaohua Tong. 2022. "Extraction of a Weak Flow Field for a Multi-Rotor Unmanned Aerial Vehicle (UAV) Using High-Speed Background-Oriented Schlieren (BOS) Technology" Sensors 22, no. 1: 43. https://doi.org/10.3390/s22010043

APA Style

Liu, X., Guo, T., Zhang, P., Jia, Z., & Tong, X. (2022). Extraction of a Weak Flow Field for a Multi-Rotor Unmanned Aerial Vehicle (UAV) Using High-Speed Background-Oriented Schlieren (BOS) Technology. Sensors, 22(1), 43. https://doi.org/10.3390/s22010043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop