Next Article in Journal
Inter-Comparison of S-NPP VIIRS and Aqua MODIS Thermal Emissive Bands Using Hyperspectral Infrared Sounder Measurements as a Transfer Reference
Previous Article in Journal
A Review of the Application of Optical and Radar Remote Sensing Data Fusion to Land Use Mapping and Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fast and Reliable Matching Method for Automated Georeferencing of Remotely-Sensed Imagery

1
The Institute of Remote Sensing and Digital Earth (RADI), Chinese Academy of Sciences, No.9 Dengzhuang South Road, Haidian District, Beijing 100094, China
2
Hainan Key Laboratory for Earth Observation, Sanya 572029, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2016, 8(1), 56; https://doi.org/10.3390/rs8010056
Submission received: 10 December 2015 / Revised: 2 January 2016 / Accepted: 4 January 2016 / Published: 19 January 2016

Abstract

:
Due to the limited accuracy of exterior orientation parameters, ground control points (GCPs) are commonly required to correct the geometric biases of remotely-sensed (RS) images. This paper focuses on an automatic matching technique for the specific task of georeferencing RS images and presents a technical frame to match large RS images efficiently using the prior geometric information of the images. In addition, a novel matching approach using online aerial images, e.g., Google satellite images, Bing aerial maps, etc., is introduced based on the technical frame. Experimental results show that the proposed method can collect a sufficient number of well-distributed and reliable GCPs in tens of seconds for different kinds of large-sized RS images, whose spatial resolutions vary from 30 m to 2 m. It provides a convenient and efficient way to automatically georeference RS images, as there is no need to manually prepare reference images according to the location and spatial resolution of sensed images.

Graphical Abstract

1. Introduction

Direct geo-location of remotely-sensed (RS) images is based on the initial imaging model, e.g., rigorous sensor model and Rational Polynomial Coefficients (RPC) model without ground control, and the accuracy of the model is limited by the interior and exterior orientation parameters. The accurate interior orientation parameters can be achieved by performing on-board geometric calibration, but the exterior orientation parameters, which are directly observed by on-board GPS, inertial measuring units and star-trackers, usually contain variable errors. Even the most modern satellite geo-positioning equipment results in varying degrees of geo-location errors (from several meters to hundreds of meters) on the ground [1]. In practical applications, the reference image is of great importance to collect ground control points (GCPs) and to perform precise geometric rectification. However, the reference images are commonly difficult or expensive to obtain, and an alternative approach is to use GCPs obtained by GPS survey, which is time consuming and labor intensive. In recent years, many online aerial maps (e.g., Google satellite images [2], Bing aerial images [3], MapQuest satellite maps [4], Mapbox satellite images [5], etc.) and interactive online mapping applications (e.g., Google Earth [6], NASA World Wind [7], etc.) have become available, and they show high geometric accuracy according to the authors’ recent GPS survey experiments. The surveyed GCPs are distributed in 17 different areas around China, where the latitude varies from 18 ° N to 48 ° N and the longitude varies from 75 ° E to 128 ° E. The accuracy of the online satellite maps (Google satellite images, Bing aerial images and Mapbox satellite images) in the surveyed areas is shown in Table 1. Note that the accuracy of MapQuest satellite maps is not included, as MapQuest satellite maps of high zoom levels (higher than 12) are not available in China. Although some areas lack high resolution images or the positioning errors of the images are around 10 m, most of the surveyed areas are of high geometric accuracy, and the root mean square (RMS) values of the positioning errors of these online resources are less than 5 m. Moreover, the areas lacking high resolution images are decreasing, and the geometric accuracy of the online resources is increasingly improving. These online resources provide another alternative to manually collecting GCPs, and they should be used more widely in the future as their accuracies increase. As far as we know, however, automatic solutions have not been reported yet.
Table 1. Accuracy of the online aerial maps, i.e., root mean square (RMS) values of the positioning errors according to our GPS survey results.
Table 1. Accuracy of the online aerial maps, i.e., root mean square (RMS) values of the positioning errors according to our GPS survey results.
Map SourceRMS Errors (Meters)
Google3.89
Bing4.12
Mapbox4.23
Automatic image matching is one of the most essential techniques in remote sensing and photogrammetry, and it is the basis of various advanced tasks, including image rectification, 3D reconstruction, DEM extraction, image fusion, image mosaic, change detection, map updating, and so on. Although it has been extensively studied during the past few decades, image matching remains challenging due to the characteristics of RS images. A practical image matching approach should have good performance in efficiency, robustness and accuracy, and it is difficult to perform well in all of these aspects, as the RS images are usually of a large size and scene and are acquired in different conditions of the spectrum, sensor, time and geometry (viewing angle, scale, occlusion, etc.).
The existing image matching methods can be classified into two major categories [8,9]: area-based matching (ABM) methods and feature-based matching (FBM) methods.
Among the ABM methods, intensity correlation methods based on normalized cross-correlation (NCC) and its modifications are classical and easy to implement, but the drawbacks of high computational complexity and flatness of the similarity measure maxima (due to the self-similarity of the images) prevent them from being applied to large-scale and multi-source images [9]. Compared to intensity correlation methods, phase correlation methods have many advantages, including high discriminating power, numerical efficiency, robustness against noise [10] and high matching accuracy [11]. However, it is difficult for phase correlation methods to be extended to match images with more complicated deformation, although Fourier–Mellin transformation can be applied to deal with translated, rotated and scaled images [12]. Moreover, as phase correlation methods depend on the statistical information of the intensity value of the image, the template image should not be too small to provide reliable phase information, and phase correlation may frequently fail to achieve correct results if the template image covers changed content (e.g., a newly-built road). In least squares matching (LSM) methods, a geometric model and a radiometric model between two image fragments are modeled together, and then, least squares estimation is used to find the best geometric model and matched points [13]. LSM has a very high matching accuracy potential (up to 1/50 pixels [14]) and is computationally efficient and adaptable (can be applied to complicated geometric transformation models and multispectral or multitemporal images [15]). However, LSM requires good initial values for the unknown parameters, as the alignment/correspondence between two images to be matched generally has to be within a few pixels or the process will not converge [14,16].
In contrast to the ABM methods, the FBM methods do not work directly with image intensity values, and this property makes them suitable for situations when illumination changes are expected or multisensor analysis is demanded [9]. However, FBM methods, particularly line- and region-based methods, are commonly less accurate than ABM methods [15] (fitting these high-level features usually introduces additional uncertainty [17] to the matching result). FBM methods generally include two stages: feature extracting and feature matching. As automatic matching of line- and region-features is more difficult and less accurate, the point-based methods are much more widely used. Among the point-based methods, scale-invariant feature transform (or SIFT) [18] is one of the most important ones, which is invariant to image rotation and scale and robust across a substantial range of affine distortion, the addition of noise and changes in illumination, but imposes a heavy computational burden. More recently-proposed point detectors, e.g., Speeded Up Robust Features (SURF) [19], Features from accelerated segment test (FAST) [20], Binary Robust Invariant Scalable Keypoints (BRISK) [21], Oriented FAST and Rotated BRIEF (ORB) [22] and Fast Retina Keypoint (FREAK) [23], provide fast and efficient alternatives to SIFT, but they are proven not as robust as SIFT. However, SIFT-based methods face the following challenges when directly used in RS images: large image size, large scene, multi-source images, accuracy, distribution of matched points, outliers, etc.
During the last ten years, many improvements have been made to cope with the drawbacks of SIFT:
Efficiency: In the PCA-SIFT descriptor [24], the 3042-dimensional vector of a 39 × 39 gradient region is reduced to a 36-dimensional descriptor, which is fast for matching, but it is proven to be less distinctive than SIFT [25] and to require more computation to yield the descriptor. Speeded-up robust features (SURF) is one of the most significant speeded up versions of SIFT, but can only slightly decrease the computational cost [26] when becoming less repeatable and distinctive [22]. Some GPU (graphic process unit)-accelerated implementations of SIFT (e.g., SiftGPU [27] and CudaSift [28]) can get comparable results as Lowe’s SIFT [18], but are much more efficient. However, these implementations require particular hardware, such as the GPU, which is not available for every personal computer (PC), and they are not robust enough when applied to very large satellite images.
Multi-source image: [29] refined the SIFT descriptor to cope with the different main orientations of corresponding interesting points, which are caused by the significant difference in the pixel intensity and gradient intensity of sensed and reference images. The work in [30] proposed an improved SIFT to perform registration between optical and SAR satellite images. The work in [31] introduced a similarity metric based on local self-similarity (LSS) descriptor to determine the correspondences between multi-source images.
Distribution control: Uniform robust SIFT (UR-SIFT) [32] was proposed to extract high-quality SIFT features in the uniform distribution of both the scale and image spaces, while the distribution of matched points is not guaranteed. More recently, the tiling method was used to deal with large RS images [26,33] and to yield uniform, distributed ground control points.
Outliers’ elimination: Scale restriction SIFT (SR-SIFT) [34] was proposed to eliminate the obvious translation, rotation and scale differences between the reference and the sensed image. The work in [35] introduced a robust estimation algorithm called the HTSC (histogram of TARsample consensus) algorithm, which is more efficient than the RANSAC algorithm. The mode-seeking SIFT (MS-SIFT) algorithm [36] performs mode seeking (similarity transformation model) to eliminate outlying matched points, and it outperformed SIFT-based RANSAC according to the authors’ experiments. The similarity transformation, nevertheless, is not suitable for all kinds of RS images when the effects of image perspective and relief are serious.
In summary, despite the high matching accuracy, ABM methods do not have good performance for RS images due to the complex imaging conditions and geometric distortions. On the other hand, FBM methods are more suitable for multisensor analysis. SIFT is one of the most successful FBM methods, but it still faces many difficulties when directly applied to RS images. Although a number of improved versions of SIFT have been proposed to cope with the drawbacks, all of these methods do not make full use of the prior information (initial imaging model and possible geometric distortions) of the RS image and the requirement of a specific task. In this work, we focus on the task of image rectification (e.g., geometric correction, orthorectification and co-registration), while the tasks of 3D reconstruction and DEM extraction, which require densely-matched points, are not considered. Commonly, tens of uniform, distributed and accurate control points are sufficient to perform rectification of RS images, and more control points do not necessarily improve the accuracy of the result [37]. The purpose of this paper is to overcome the difficulties of SIFT and to develop a practical online matching method, which is efficient, robust and accurate, for the georeferencing task of RS images. The original contribution of this work mainly includes the following aspects: (i) a convenient approach to perform point matching for RS images using online aerial images; (ii) a technical frame to find uniformly-distributed control points for large RS images efficiently using the prior geometric information of the images; and (iii) an improved strategy to match SIFT features and eliminate false matches.
The rest of this paper is organized as follows. Section 2 introduces the technical frame of the proposed matching method, and Section 3 states the approach to utilize online aerial images in detail. Experimental evaluation is presented in Section 4, and the conclusion is drawn in Section 5.

2. Technical Frame

The proposed point matching method is mainly based on the following scheme:
(1) Image tiling:
The geometric distortion of the RS image is complicated, resulting from the distortion of the camera, projective deformation, affect of interior and exterior orientation parameters, Earth curvature, reliefs, and so on, and the rational function model (RFM) of 78 coefficients (RPCs) is usually used to model the deformation of the RS image [38]. However, the local distortion, e.g., that of a small image patch of 256 × 256 , can be approximated by much simpler transformations (affine or similar transformation).
In a remotely-sensed image of a large scene, SIFT may be computationally difficult and error-prone, and dividing the large image into small tiles can avoid this drawback.
The tilling strategy also helps to control the distribution and quantity of the matched points, and the computational cost can be notably saved if the number of target matches is limited.
(2) Make use of prior geometric information:
The prior geometric information of RS images, e.g., ground sample distance (or spatial resolution) and coarse geographic location, can be utilized to make the image matching process more efficient and robust.
(3) Make use of the attributes of SIFT feature:
The attributes of a SIFT feature, including location, scale, orientation and contrast, can be used to eliminate false matches and evaluate the quality of the feature.
(4) Refine the results of SIFT:
The matched points of SIFT are extracted from the sensed and reference image independently and are less accurate than those of area-based methods. However, the results of SIFT provide good initial values for least squares matching (LSM) and can be refined to achieve very high accuracy by LSM.
The process of the proposed matching method can be summarized as the flowchart in Figure 1, and the detailed techniques of the method will be introduced in the following sections (Section 2.1, Section 2.2, Section 2.3, Section 2.4, Section 2.5 and Section 2.6).

2.1. Image Tiling

In the proposed method, image tiling consists of three steps:
  • The region of interest (the whole region of the sensed image or the intersection region of the sensed and reference image) is divided into blocks according to the number of target matches.
  • Each block of the image is divided into small tiles (processing unit) to perform SIFT matching, and in this work, the size of the image tile is 256 × 256 .
  • The corresponding tile is extracted from the reference image (online aerial maps) according to the tile in the sensed image and the initial geometric model.
Figure 1. Flowchart of the proposed matching method.
Figure 1. Flowchart of the proposed matching method.
Remotesensing 08 00056 g001
Figure 2 illustrates the blocks of an image and the tiles of a block. The aim of image matching is to achieve a reliable control point in each block, and the process will move on to the next block once any tile of the current block succeeds to yield a reliable control point.
Figure 2. Blocks of an image and tiles of a block.
Figure 2. Blocks of an image and tiles of a block.
Remotesensing 08 00056 g002
When extracting the corresponding tile from the reference image, the initial geometric model should be utilized, which can be various types: the affine transformation model contained in a georeferenced image or all kinds of imaging models, such as a rigorous sensor model, a polynomial model, a direct linear transformation model, a rational function model (RFM), etc.
Commonly, these imaging models can be defined as a forward model (from the image space to the object space) or an inverse model (from the object space to the image space).
X = F X ( x , y , Z ) Y = F Y ( x , y , Z )
x = F x ( X , Y , Z ) y = F y ( X , Y , Z )
where:
( x , y ) are the coordinates in image space,
( X , Y , Z ) are the coordinates in object space,
Z is the elevation,
F X and F Y are the forward transforming functions of the X and Y coordinates, respectively,
F x and F x are the inverse transforming functions of x and y coordinates, respectively.
In the forward model, image coordinates ( x , y ) and elevation Z are needed to determine the ground coordinates ( X , Y , Z ) . With the help of DEM data, however, the ground coordinates ( X , Y , Z ) can be determined by the image coordinates ( x , y ) after several iterations. Therefore, the forward model can also be denoted by Equation (3) if DEM data are available.
X = F X ( x , y ) Y = F Y ( x , y )
With the help of the initial geometric model of the sensed image, the reference image tile can be extracted by calculating its approximate extent. Moreover, to make SIFT matching more efficient and robust, the reference image tile is resampled to a similar resolution as the sensed image tile. The detailed techniques of fetching the reference image tile from online aerial maps will be introduced in Section 3.

2.2. Extracting SIFT Features

As the reference image tile is resampled to a similar resolution as the sensed image tile, the SIFT detector can be performed in only one octave to get the expected results, and the process becomes much more efficient. In the only octave, the scale space of the image tile is defined as a function, L ( x , y , σ ) , that is produced from the convolution of a variable-scale Gaussian, G ( x , y , σ ) , with the input image tile, I ( x , y ) [18]:
L ( x , y , σ ) = G ( x , y , σ ) * I ( x , y )
where G ( x , y , σ ) = 1 2 π σ 2 e ( x 2 + y 2 ) / 2 σ 2 and * is the convolution operation.
Then, D ( x , y , σ ) , the convolution of the difference-of-Gaussian (DoG) function and the image tile, which can also be computed from the difference of two nearby scales separated by a constant multiplicative factor k, is used to detect stable keypoint locations in the scale space by searching the scale space extrema.
D ( x , y , σ ) = ( G ( x , y , k σ ) G ( x , y , σ ) ) * I ( x , y ) = L ( x , y , k σ ) L ( x , y , σ )
Once a keypoint candidate has been found, its location ( x , y ) , scale σ, contrast c and edge response r can be computed [18], and the unstable keypoint candidates whose contrast c is less than threshold T c (e.g., T c = 0 . 03 ) or whose edge response r is greater than threshold T r (e.g., T r = 10 ) will be eliminated. Then, image gradient magnitudes and orientations are sampled around the keypoint location to compute the dominant direction θ of local gradients and the 128-dimensional SIFT descriptor of the keypoint.

2.3. Matching SIFT Features

In standard SIFT, the minimum Euclidean distance between the SIFT descriptors is used to match the corresponding keypoints, and the ratio of closest to second-closest neighbors of a reliable keypoint should be greater than an empirical threshold T d r , e.g., T d r = 0 . 8 [18]. However, [29,32] pointed out that the T d r constraint was not suitable for RS images and would lead to numerous correctly-matched eliminations.
In this work, both the T d r constraint and a cross matching [32] strategy are applied to find the initial matches. Denoting by P and Q the keypoint sets in the sensed and reference image tiles, once either of the following two conditions is satisfied, the corresponding keypoints p i P and q j Q will be included in the match candidates.
T d r
constraint: The ratio of closest to second-closest neighbors of the keypoint p i is greater than T d r = 0 . 75 , and the keypoint q j is the closest neighbor of p i . Here, we chose a smaller T d r rather than 0.8, which is recommended by [18], to reduce the chance of including too many false matches for RS images.
Cross
matching: The keypoint p i is the closest neighbor of q j in P, and the keypoint q j is also the closest neighbor of p i in Q.
Of course, the match candidates usually include a number of false matches, which will be eliminated in the following step.

2.4. Eliminating False Matches

Commonly, some well-known robust fitting methods, such as RANSAC or least median of squares (LMS), are applied to estimate an affine transformation, as well as the inliers from the match candidates. However, these methods perform poorly when the percent of inliers falls much below 50%. In this work, the false matches are eliminated by four steps, i.e., rejecting by scale ratio, rejecting by rotation angle, rejecting by the coarse similarity transformation (Equation (6)) using RANSAC and rejecting outliers by the precise affine transformation (Equation (7)) one by one.
x r = s ( x s cos θ + y s sin θ ) + t x y r = s ( x s sin θ + y s cos θ ) + t y
x r = a 0 + a 1 x s + a 2 y s y r = b 0 + b 1 x s + b 2 y s
where:
s and θ are the scale parameter and rotation angle parameter of similarity transformation,
t x and t y are the translation parameters of similarity transformation in the x direction and the y direction,
a 0 , a 1 , a 2 , b 0 , b 1 , b 2 are the parameters of affine transformation.
There are a number of reasons for choosing similarity transformation to perform RANSAC estimation instead of affine transformation. Firstly, it is possible for a similarity transformation to model the geometric deformation coarsely in a small tile of an RS image. Secondly, the similarity transformation solution requires less point matches than the affine transformation solution and is also more robust. In addition, the similarity transformation can make full use of the geometric information, such as the scale and dominant direction, of the SIFT keypoints.
(1) Rejecting by scale ratio:
The scale has been computed for each keypoint in the phase of extracting SIFT features (Section 2.2) and the scale ratio of a pair of corresponding keypoints in the sensed image tile and reference image tile indicates the scale factor between the two image tiles. By computing a histogram of the scale ratios of all match candidates, the peak of the histogram will locate around the true scale factor between the two image tiles [36]. The match candidates whose scale ratio is far from the peak of the histogram are not likely to be correct matches and, therefore, are rejected from the match candidates. Denoting the peak scale ratio by σ p e a k , the acceptable matches should satisfy the following criterion:
T σ < Δ σ σ p e a k < 1 T σ
where Δ σ is the scale ratio of a match candidate, T σ is the scale ratio threshold and T σ = 0 . 8 is used in this work. The selection of T σ will be discussed later at the end of Section 2.4.
Note that the reference image tile is resampled to a similar resolution as the sensed image tile; the computation of the scale ratio histogram is not necessary. The σ p e a k is expected to be around 1 . 0 , even if we do not check the scale ratio histogram.
(2) Rejecting by rotation angle:
Similarly, as the difference of the dominant direction of corresponding keypoints indicates the rotation angle between the two image tiles, a rotation angle histogram can be computed using the dominant directions of the SIFT keypoints in all match candidates. The rotation angle histogram has 36 bins covering the 360 degree range of angles, and the match candidates whose difference of dominant direction is far from the peak of the histogram are rejected. Denoting the peak rotation angle by θ p e a k , the acceptable matches should satisfy the following criterion:
Δ θ θ p e a k < T θ
where Δ θ is the dominant directions of the SIFT features in a match candidate, T θ is the rotation angle threshold and T θ =15 ° is used in this work. The selection of T θ will be discussed later at the end of Section 2.4.
(3) Rejecting by similarity transformation:
After the first two steps, most of the false matches will be rejected, and the RANSAC algorithm is quite robust to estimate a coarse similarity transformation from the remaining match candidates. Meanwhile, outliers for similarity transformation are also excluded.
(4) Rejecting by affine transformation:
In order to achieve accurate matching results, the remaining match candidates should be further checked by an affine model. Specifically, all of the remaining match candidates are used to find the least-squares solution of an affine transformation, and inaccurate matches, which do not agree with the estimated affine transformation, should be removed. The process will iterate until none of the remaining matches deviates from the estimated affine transformation by more than one pixel.
Note that once fewer than four match candidates remain before or after any of the four steps, the match will be terminated for this tile immediately.
Next, we will provide a discussion on the recommended values of T σ and T θ , and this is based on a matching task using a collection of 1000 pairs of real image tiles that were extracted from different sources of RS images, including Landsat-8, ZY-3, GF-1, etc.
The matching task includes two groups of tests: (1) set T θ =15 ° , and let T σ vary from zero to one; (2) set T σ = 0 . 8 , and let T θ vary from 0 ° to 360 ° . A pair of image tiles is called a “matched pair” if the matching process of this image pair yields at least four matches after all four filtering steps. However, for a matched pair, it is possible that not all of the false matches were excluded after the filtering steps, and the results will be untrustworthy. Therefore, only the correctly-matched pairs, whose output matches are all correct, are reliable, and we refer to the percentage of correctly-matched pairs out of all of the matched pairs as the “correct match rate”. In each group of tests, the numbers of matched pairs and correct match rates were obtained for different values of T σ or T θ . Figure 3a shows the matching results with respect to different values of T σ , while Figure 3b shows the matching results with respect to different values of T θ . According to Figure 3a, as T σ increases from zero to one, the number of matched pairs declines, but the correct match rate ascends. According to Figure 3b, as T θ increases from 0 ° to 360 ° , the number of matched pairs ascends, while the correct match rate declines. To ensure a high correct match rate and enough matches, the value of T σ should be between 0.70 and 0.85, and the value of T θ should be between 15 ° and 30 ° .
Figure 3. The numbers of matched pairs and correct match rates for different values of T σ and T θ . (a) The results with respect to different values of T σ ; (b) the results with respect to different values of T θ .
Figure 3. The numbers of matched pairs and correct match rates for different values of T σ and T θ . (a) The results with respect to different values of T σ ; (b) the results with respect to different values of T θ .
Remotesensing 08 00056 g003

2.5. Refining Position

After the step of eliminating false matches, all of the remaining matches have good agreement (within one pixel) with the local affine transformation. However, the accuracy of the matched points is not high enough considering that they are extracted from the sensed image and reference image independently. Consequently, least squares matching (LSM) is applied to refine the matching results.
It is possible to further include any matches that agree with the final affine transformation from those rejected by error in the phase of eliminating false matches, and then, the new set of matches will be more complete. However, only a pair of matched points is needed for a sensed image tile in the proposed method, even if a large number of matches are found. Therefore, the step of adding missed matches is not included in this work for the sake of efficiency.
Considering that the features having high contrast are stable to image deformation [32], the keypoint with the highest contrast is chosen from the output of the phase of eliminating false matches. Actually, the high contrast not only guarantees the stability of the keypoints, but also benefits the accuracy of LSM.
LSM is performed in a small region around the SIFT keypoint in sensed image tile, e.g., a template of 11 × 11 , and it is quite efficient. In order to cope with both the geometric deformation and radiometric difference, a geometric model and a radiometric model between two image fragments are modeled together [16], and the condition equation of a single point is:
k 1 I s ( x s , y s ) + k 2 I r ( x r , y r ) = 0
where x r = a 0 + a 1 x s + a 2 y s and y r = b 0 + b 1 x s + b 2 y s , a 0 , a 1 , a 2 , b 0 , b 1 , b 2 are six parameters of geometric transformation, k 1 and k 2 are two radiometric parameters for contrast and brightness (or equivalently gain and offset), I s ( x s , y s ) and I r ( x r , y r ) are the gray values of a pixel in a source and reference image tile.
The geometric model and the radiometric model are estimated by least squares, and then, we can accurately locate the corresponding point in the reference image tile.
As Equation (8) is nonlinear, good initial values are required to find the optimal models. Fortunately, the previously-calculated affine transformation provides very good initial values for the geometric parameters, and those of the radiometric parameters can be set as k 1 = 1 and k 2 = 0 [16]. Finally, the Levenberg–Marquardt algorithm [39] is applied to solve the problem.
Below is an example to show the effect of position refinement. Figure 4 illustrates matched keypoints in the sensed image tile and the reference image tile, and Figure 5 shows the image fragments around the keypoints, as well as the the matched points before and after the phase of position refinement. Figure 5a,b is the original matching result of SIFT, and it is very difficult to tell whether the matched points in the sensed image and the reference image are corresponding points exactly. However, by applying least squares matching, the warped image fragment in Figure 5c is in good agreement with the search image fragment in Figure 5b, both in geometry and radiometry. Consequently, it is very clear that the marked point in Figure 5c (transformed from the keypoint in Figure 5a) and that in Figure 5d (refined by LSM) are corresponding. Meanwhile, one can see that the original SIFT keypoint in Figure 5b is not accurate enough when comparing to the point in Figure 5d. Note that the images in Figure 5 are enlarged by eight times using the cubic interpolation method, and the actual deviation between Figure 5b,d is about one pixel.
Figure 4. The matched keypoints in a sensed image tile and a reference image tile. (a) The sensed image tile; and (b) the reference image tile.
Figure 4. The matched keypoints in a sensed image tile and a reference image tile. (a) The sensed image tile; and (b) the reference image tile.
Remotesensing 08 00056 g004
Figure 5. (a) The template image around the SIFT keypoint (marked with a cross) in the sensed image tile; (b) the search image around the SIFT keypoint (marked with a cross) in the reference image tile; (c) an image fragment warped from the template image using the geometric model and the radiometric model in Equation (8), and the cross denotes the SIFT keypoint in the sensed image tile after geometric transformation; (d) the search image, and the cross denotes the refined keypoint in the reference image tile.
Figure 5. (a) The template image around the SIFT keypoint (marked with a cross) in the sensed image tile; (b) the search image around the SIFT keypoint (marked with a cross) in the reference image tile; (c) an image fragment warped from the template image using the geometric model and the radiometric model in Equation (8), and the cross denotes the SIFT keypoint in the sensed image tile after geometric transformation; (d) the search image, and the cross denotes the refined keypoint in the reference image tile.
Remotesensing 08 00056 g005

2.6. Summary

According to the number of required control points, the sensed image will be divided into a number of blocks evenly, and only one control point is needed for each block. Then, each block is divided into a number of tiles according to the previously-defined tile size, and once any of the tiles succeeds to produce a control point, the process will move on to the next block. We do not intend to exhaust all of the possible matches, but find a moderate number of control points that are very reliable. Obviously, this method is greedy and therefore efficient.
After the matching of all image blocks is finished, it is easy to further identify potential outliers by checking for agreement between each matched point and a global geometric model, e.g., rigorous sensor model, rational function model, etc. However, hardly any false matches were found in the final matches according to the results of all our tests.
We call the process of matching a pair of a sensed image tile and a reference image tile a “tile trial” (as shown in Figure 1), and the actual efficiency of the method is decided by the number of tile trial times. If the success rate of tile trial is 100 % (the best case), only one tile trial is performed for each control point, and the number of tile trial times is not related to the size of the sensed image; if the success rate of tile trial is 0 % (the worst case), all of the tiles of the sensed image will be tested, and the number of tile trial times is decided by the size of the sensed image. The success rate of tile trial is related to the similarity between the sensed image and the reference image and the distinction, which can be affected by a number of factors, e.g., the quality (cloud coverage, contrast, exposure, etc.) of the images, the scale difference, the spectral difference, changes caused by different imaging times, etc. Additionally, as the tile trial is based on SIFT matching, the success rate is limited if the test images cover a region of low texture, such as water, desert and forest.
Furthermore, the high independence among the image blocks enables a parallel implementation, which can further accelerate the proposed method. The processing of image blocks can be assigned to multiple processors and nodes (computers) in a cluster and, therefore, run concurrently. Parallelization makes full use of the computing resources and exponentially shortens the consumed time of image matching. In this work, we implemented a parallel version on multiple processors, but not on multiple computers.
In addition, the SIFT implementation designed on a graphic process unit (GPU) [27] may considerably accelerate the process of the tile trial, but it is not yet included in this work.

3. Fetch Reference Image from Online Aerial Maps

In Section 2.1, we mentioned that the reference image tile can be extracted by calculating its approximate extent, and this section will introduce the detailed techniques to fetch reference image tiles from online aerial maps, i.e., Google satellite images, Bing aerial images, MapQuest satellite maps and Mapbox satellite images.

3.1. Static Maps API Service

In this work, we use the Static Maps API Service of online aerial maps, i.e., sending a URL (Uniform Resource Locator) request, to fetch the required reference image tiles automatically. For example, the formats of URL request of Google satellite images, Bing aerial images, MapQuest satellite maps and Mapbox satellite images are listed below.
Google satellite images:
Remotesensing 08 00056 i001
Bing aerial images:
Remotesensing 08 00056 i002
MapQuest satellite maps:
Remotesensing 08 00056 i003
Mapbox satellite images:
Remotesensing 08 00056 i004
In these URLs, the parameters inside “{}” should be specified, i.e., the longitude and latitude of the center point, zoom level, width and height of the image tile and the API keys. One can apply either free or enterprise API keys from corresponding websites, freely or with a low cost, and the calculation of the other parameters will be introduced in the following sections.

3.2. Zoom Level

For online global maps, a single projection, Mercator projection, is typically used for the entire world, to make the map seamless [40]. Moreover, the aerial maps are organized in discrete zoom levels, from 1 to 23, to be rendered for different map scales. At the lowest zoom level, i.e., 1, the map is 512 × 512 pixels, and once the zoom level is increased by one, the width and height of the map expand twice.
Consequently, in order to fetch the corresponding reference image tile, we need to firstly determine the zoom level, which is related to the ground sample distances (GSD) of the sensed image tile. Similar to the relative scale between the sensed image and the reference image, the GSDs of the sensed image are not necessarily constant in a whole image, and the local GSDs of an image tile can be calculated by the formula:
G S D x = ( F X ( x c + 1 , y c ) F X ( x c , y c ) ) 2 + ( F Y ( x c + 1 , y c ) F Y ( x c , y c ) ) 2 G S D y = ( F X ( x c , y c + 1 ) F X ( x c , y c ) ) 2 + ( F Y ( x c , y c + 1 ) F Y ( x c , y c ) ) 2
where:
G S D x and G S D y are the ground sample distances in x direction and y direction,
x c and y c are the image coordinates of the center point of the sensed image tile,
F X and F Y are the forward transforming functions of the X and Y coordinates, which are described in Equation (3).
On the other hand, the GSDs (in meters) of online aerial maps vary depending on the zoom level and the latitude at which they are measured, and the conversion between the GSD and nearest zoom level is described by Equation (10),
G S D = 2 π R e a r t h 512 × 2 n 1 cos ϕ n = log 2 2 π R e a r t h cos ϕ 512 × G S D + 1
where:
R e a r t h is the earth radius, for which 6,378,137 meters is used,
ϕ is the latitude at which it is measured,
G S D is the ground sample distance (in meters), both in the x direction and y direction,
n is the zoom level,
· is an operator to find the nearest integer.
Equation (10) can be applied to find the nearest zoom level according to the GSD of the sensed image tile (the mean value of those in the x direction and the y direction) calculated by Equation (9).

3.3. Width and Height

By providing the rectangle extent of the sensed image tile, the corresponding geographic coordinates (longitude and latitude) of the four corners can be calculated by the initial forward transforming functions in Equation (3). In order to find the extent of the required reference image tile, the geographic coordinates, λ and ϕ, should be converted to the image coordinates, x and y, in the map of the nearest zoom level, n, according to Equation (11) [40],
x = λ + 180 360 × 512 × 2 n 1 y = 0 . 5 1 4 π ln 1 sin ϕ 1 + sin ϕ × 512 × 2 n 1
where λ and ϕ are the longitude and latitude.
Then, the extent of reference image tile is the minimum boundary rectangle of the four corners in the map of zoom level n, and the width and height of the tile are known, accordingly.

3.4. Center Point

Next, we need to calculate the geographic coordinates of the center point of reference image tile, and the following inverse transformation from image coordinates, x and y, in the map of the nearest zoom level, n, to geographic coordinates, λ and ϕ, will be used,
λ = 360 512 × 2 n 1 x 180 ϕ = 180 π arcsin exp [ 4 π ( 0 . 5 y 512 × 2 n 1 ) ] 1 exp [ 4 π ( 0 . 5 y 512 × 2 n 1 ) ] + 1
Equation (12) is derived from Equation (11) directly, and it will be used again when the matched points in the sensed and reference image tile are found, as the image coordinates in the reference image tile should be converted to geographic coordinates to obtain ground control points.

3.5. Resizing

Given the nearest zoom level, the width and height of the image tile, the longitude and latitude of the center point and the API keys, the Static Maps API service can be used to download the required reference image tile from the online aerial images. However, the GSD of the downloaded reference image tile may not be very close to that of the sensed image tile, since the zoom level is discrete. The downloaded image tile needs to be further resampled to a similar resolution as the sensed image tile for the sake of efficiency and robustness, according to the relative scale between the two image tiles, which can be calculated by dividing the GSD of the sensed image tile by that of the online reference image tile.

3.6. Summary

The scene of a RapidEye image captured in 2010 is used to show an example of online matching. The image is of Oahu, Hawaii, and the spatial resolution is 5 m. Figure 6 shows a tile of the RapidEye image matched with different online aerial maps, including Google satellite images, Bing aerial images, MapQuest satellite maps and Mapbox satellite images. Note that we intentionally chose a scene in the USA, as the MapQuest satellite maps of high zoom levels (higher than 12) are provided only in the United States.
From Figure 6, one can see that in the same range, the data sources of the four kinds of online aerial maps are not the same. In practically applications, different online aerial maps can be used for complementation.
Figure 6. Example of online matching, and the matched points are marked by cross. (a) to (d) are matching results using Google satellite images, Bing aerial maps, MapQuest satellite maps and Mapbox satellite images, respectively. In each figure, the left is the RapidEye image tile, while the right is the online aerial map.
Figure 6. Example of online matching, and the matched points are marked by cross. (a) to (d) are matching results using Google satellite images, Bing aerial maps, MapQuest satellite maps and Mapbox satellite images, respectively. In each figure, the left is the RapidEye image tile, while the right is the online aerial map.
Remotesensing 08 00056 g006

4. Experiments and Analysis

In this section, several groups of experiments are carried out to check the validity of the proposed method, and all experiments are performed on a 3.07-GHz CPU with four cores.

4.1. Robustness

To show the superiority of the matching strategy of the proposed method, we carry out comparative tests with three methods: the proposed method, the ordinary SIFT [18] and SR-SIFT [34], which is claimed to be more robust than ordinary SIFT. In the ordinary SIFT matching method, match candidates are found by using the distance ratio constraint of closest to second-closest neighbors, and outliers are eliminated by using the RANSAC algorithm and affine transformation. In the SR-SIFT method, scale restriction is applied to exclude unreasonable matches before the RANSAC filtering. The distance ratio threshold T d t = 0 . 75 is applied in all of the methods. Figure 7 shows the results of the three methods when applied to a pair of AVIRIS image tiles (visible and infrared).
Figure 7. Matching results of the AVIRIS visible image tile (left) and the AVIRIS infrared image tile (right), using three different methods. (a) The result of original SIFT (four matches are found, including a wrong match); (b) the result of SR-SIFT (six correct matches are found); and (c) the result of the proposed method (20 correct matches are found).
Figure 7. Matching results of the AVIRIS visible image tile (left) and the AVIRIS infrared image tile (right), using three different methods. (a) The result of original SIFT (four matches are found, including a wrong match); (b) the result of SR-SIFT (six correct matches are found); and (c) the result of the proposed method (20 correct matches are found).
Remotesensing 08 00056 g007
From Figure 7, we can see that the original SIFT yields the poorest matching results, while SR-SIFT provides more correct matches. However, the best results come from the proposed method, not only the quantity of correct matches, but also the distribution of matched points.
We also test more than 100 successfully-matched tiles, from six pairs of RS images, including Landsat-5 vs. Landsat-5 (captured at different times), Huanjing-1 (HJ-1) vs. Landsat-8, Gaofen-1 (GF-1) vs. Bing aerial maps, Ziyuan-3 (ZY-3) vs. RapidEye, GF-1 Multispectral (MSS) vs. GF-1 Panchromatic (PAN) (captured simultaneously), Kompsat-2 vs. Worldview-1, and the numbers of remaining matches after each step of the methods are noted. Table 2 shows the results of 12 pairs of tiles (two pairs of tiles are randomly selected from each dataset).
Table 2. The number of remaining matches after each step in the three methods.
Table 2. The number of remaining matches after each step in the three methods.
DatasetsTilesProposed Method Ordinary SIFT Ordinary SIFT
S1 b S2S3S4S5S6 S1S2 S1S2S3
D1 a Tile 11258833231717 1914(11) c 191714(11)
Tile 2986017855 106(4) 1097(6)
D2Tile 114291201589 88(6) 888(6)
Tile 21248818151011 1513(9) 151513(9)
D3Tile 112272311478 1410(5) 141410(5)
Tile 21249152442628 5549(26) 555250(25)
D4Tile 11308740967 2516(5) 252317(6)
Tile 2153114691689 3734(3) 373634(3)
D5Tile 12152061137575105 195104(102) 195190103(101)
Tile 22162001006968101 184102(98) 184182102(98)
D6Tile 11297619755 30(0) 330(0)
Tile 21157326867 104(2) 1095(3)
a D1 to D6 denote the 6 datasets. D1: Landsat-5 vs. Landsat-5 (captured at different times); D2: Huanjing-1 (HJ-1) vs. Landsat-8; D3: Gaofen-1 (GF-1) vs. Bing aerial map; D4: Ziyuan-2 (ZY-3) vs. RapidEye; D5: GF-1 Multispectral (MSS) vs. GF-1 Panchromatic (PAN) (captured simultaneously); D6: Kompsat-2 vs. Worldview-1; b S1 to S6 denote the 6 steps. In the proposed method: S1: matching with the distance ratio constraint and cross matching; S2: rejecting by the scale ratio; S3: rejecting by the rotation angle; S4: rejecting by similarity transformation; S5: rejecting by affine transformation; S6: adding missed matches with the help of the affine transformation estimated in S5. In ordinary SIFT: S1: matching with the distance ratio constraint; S2: rejecting by RANSAC and affine transformation. In scale restriction (SR)-SIFT, S1: matching with the distance ratio constraint; S2: rejecting by scale restriction; S3: rejecting by RANSAC and affine transformation. c The number in brackets denotes the number of correct matches.
From Table 2, the following points can be drawn:
  • The results of ordinary SIFT and SR-SIFT are similar, and a simple scale restriction filter seems not helpful to find correct matches. Specifically, in most cases (except for Dataset 5), the distance ratio constraint excludes a number of correct matches and sometimes results in failure (e.g., in Dataset 6). By applying cross matching, the proposed method includes much more initial match candidates, although many of them are false matches.
  • The percentage of outliers in the initial match candidates is usually greater than 90 % , and the RANSAC algorithm is not robust enough to identify correct subsets of matches; thus, it frequently fails or yields untrustworthy results. On the other hand, in the proposed method, four steps of outlier rejecting can eliminate all of the false matches. Actually, after the first two steps of rejecting (by the scale ratio and by the rotation angle), most of the outliers will be cast out.
  • Commonly, only a few matches will be added in Step 6 of the proposed method, except for Dataset 5, in which the correct matches are plentiful. Consequently, Step 6 can be omitted, without affecting the final result too much.
  • The ordinary SIFT matching method performs quite well for Dataset 5, in which the sensed image and reference image were captured simultaneously from the same aircraft and in the same view angle. With little variation in content, illumination and scale, the SIFT descriptor is very robust and distinctive, and the distance ratio constraint identified most of the correct matches. Then RANSAC algorithm manages to find reliable results, since the match candidates contain fewer than 50 % outliers.

4.2. Efficiency

In order to show the efficiency of the proposed method, SiftGPU [27], which is the fastest version of SIFT to our best knowledge, is applied to carry out comparative tests with the proposed method. The implementation of SiftGPU is provided by Wu C. in [27].
Two scenes of GF-2 PAN images in Beijing, China, which were acquired on 3 September 2015 and 12 September 2015, respectively, are used to perform the matching tests. GF-2 is a newly-launched Chinese resource satellite; the spatial resolution of its panchromatic image is around 0.8 m, and the size of an image scene is around 29,000 × 28,000 pixels.
Figure 8. Matching results of GF-2 PAN images, using SiftGPU and the proposed method. In each sub-figure, the left is the image captured on 3 September 2015 and the right is the image captured on 12 September 2015. The matched points are labeled by the same numbers, and red crosses stand for correct matches, while yellow crosses stand for wrong matches. (a) The result of SiftGPU (71 correct matches and 11 wrong matches); and (b) the result of proposed method (30 correct matches are found).
Figure 8. Matching results of GF-2 PAN images, using SiftGPU and the proposed method. In each sub-figure, the left is the image captured on 3 September 2015 and the right is the image captured on 12 September 2015. The matched points are labeled by the same numbers, and red crosses stand for correct matches, while yellow crosses stand for wrong matches. (a) The result of SiftGPU (71 correct matches and 11 wrong matches); and (b) the result of proposed method (30 correct matches are found).
Remotesensing 08 00056 g008
SiftGPU spends 19.4 s to find 82 matched points (including 11 incorrectly matched points) from the two scenes of GF-2 images, while the proposed method spends 20.6 s to find 30 matched points, and the matching results are shown in Figure 8. Although SiftGPU yields more matches within less time, the distribution and the correctness of the results of the proposed method (as shown in Figure 8b) are obviously superior to those of SiftGPU (as shown in Figure 8a). SiftGPU makes full use of the computing resource of the computer devices and is quite efficient when processing large images, and 15,030 and 16,291 SIFT keypoints are extracted from the two images respectively within a dozen seconds. However, finding the corresponding keypoints is difficult, as the large scene makes the descriptors of SIFT keypoints less distinctive. Moreover, the serious distortion of the satellite images also makes it difficult to identify the outliers from matched points; the residual errors of the 71 correctly matched points are more than 100 pixels when fitted by a three-degree 2D polynomial function.
Actually, SiftGPU frequently fails to provide sound results for very large satellite images according to our experimental results, despite its outstanding efficiency.
In summary, the proposed method is almost as fast as SiftGPU, but provides more reliable results.

4.3. Accuracy

In Section 2.5, an example has shown the effect of least squares match (LSM) refinement, and in this section, 42 scenes of the GF-1 MSS image are used to further evaluate the accuracy of the proposed method.
Firstly, the original GF-1 images are rectified based on the vendor-provided RPC model with constant height zero and projected in longitude and latitude. Secondly, 25 check points, ( x , y , λ , ϕ , 0 ) , are collected between each pair of the original image and the rectified image using the proposed method. Finally, the geographic coordinates, ( λ , ϕ , 0 ) , of the check point in the rectified images are transformed into the image coordinates, ( x , y ) , in the original images using the inverse transformation of the vendor-provided RPC model, and then, the biases between the matched points can be measured by the difference of the image coordinates, ( x x , y y ) . The results before and after LSM refinement are compared to show the accuracy improvement of the proposed method.
Figure 9 shows the root mean square biases of matched points before and after LSM refinement in each image, and one can see that in most of the tests, the accuracy of the matched points is notably improved after LSM refinement.
There are several reasons to use this experimental scheme. Geometric transformation, especially longitude and latitude projection, usually results in distortion of the image and then increases the uncertainty of the position of detected SIFT features, while image distortion commonly exists in practical image matching tasks. In addition, by performing image matching between the original image and the rectified image, the ground truth of the bias should be zero, since the parameters of geometric transformation are already known. Moreover, the least squares match is performed in a small patch and is independent of the global transformation (vendor-provided RPC model); thus, the agreement between the matched points and the global transformation can be applied to evaluate the accuracy of matching, objectively. Note that the position biases between matched points may not be exactly zero, due to the errors introduced by interpolation.
Figure 9. Root mean square biases of matched points before and after least squares match (LSM) refinement.
Figure 9. Root mean square biases of matched points before and after least squares match (LSM) refinement.
Remotesensing 08 00056 g009

4.4. Practical Tests

To evaluate the proposed method, four scenes of RS images, including Landsat-5, China-Brazil Earth Resources Satellite 2 (Cbers-2), Cbers-4, ZY-3, GF-2, Spot-5, Thailand Earth Observation System (Theos) and GF-1, are used to perform online matching, while Bing aerial images are utilized as the reference images. For each image, the proposed matching method is used to automatically collect control points, which are then applied to rectify the sensed image, and finally, 20 well-distributed check points are manually collected from the rectified image and reference images (Bing aerial images) to evaluate the matching accuracy. Table 3 shows the general information of the test images used in these experiments.
Table 3. General information of the test images.
Table 3. General information of the test images.
Test No.Data SourceBandImage SizeGSD (m)Acquisition TimeElevationLocationInitial Model
1Landsat-5Band 46850 × 5733302007566 to 4733China-SichuanRigorous
2Cbers-2Band 36823 × 6813202008486 to 1239Brazil-Grão MogolAffine
3Cbers-4 P10Band 36000 × 6000102015813 to 1514South Africa-BuysdorpRigorous
4ZY-3Band 18819 × 92795.82014769 to 2549China-ShaanxiRPC
5GF-2Band 17300 × 69083.22015273 to 821China-NingxiaRPC
6Spot-5Pan33,730 × 29,7562.5201017 to 155China-WuhanAffine
7TheosPan14,083 × 14,115220141285 to 1736China-XinjiangAffine
8GF-1Pan18,192 × 18,000220143080 to 4159China-QinghaiRPC
As shown in Table 3, different initial image models of the sensed image are utilized, including the rigorous sensor model, affine transformation contained in georeferenced images, the vendor-provided RPC model, etc. The rigorous sensor model of Landsat-5 is provided by the open source software OSSIM [41] and the rigorous sensor model of the Cbers-4 image is built according to the 3D ray from the image line-sample to ground coordinates in the WGS-84 system. The RPC models of the ZY-3, GF-2 and GF-1 images are provided by the vendors. The Spot-5, Cbers-2 and Theos images are processed to the L2 level of correction, and the affine transformation models contained in images are used as initial imaging models.
Cbers-2 (Test 2), Cbers-4 P10 (Test 3), Spot-5 (Test 6) and Theos (Test 7) images are rectified based on the terrain-dependent RPC (TD-RPC) model. Landsat-5 image in Test 1 is rectified based on the rigorous sensor model, and RPC refinement is applied to rectify the ZY-3 (Test 4), GF-2 (Test 5) and GF-1 (Test 8) images. Therefore, we intend to find 100 GCPs for each image in Tests 2, 3, 6 and 7, while 30 GCPs are required for each image in Tests 1, 4, 5 and 8. Note that all of the TD-RPC models in the tests are calculated using 1 -Norm-Regularized Least Squares (L1LS) [42] estimation method, to cope with the potential correlation between the parameters.
Figure 10 shows the distribution of matched points in Test 1 – Test 8, and Table 4 shows more results for each test, including the number of correct matches, consumed time, the model used for geometric rectification and the RMSE of check points. Figure 11 is the comparison between the sensed images and the online Bing aerial maps in Tests 4, 6, 7 and 8 using the “Swipe Layer” tool in ArcGIS Desktop, and ArcBruTile [43] is used to display the Bing aerial maps in ArcGIS Desktop.
In this section, the spatial resolution of the test image varies from 30 m to 2 m, but the very high resolution (less than 1 m) RS images are not included, as the geometric accuracy of the online aerial maps is limited. In this sense, we can successfully find enough GCPs for very high resolution images from the online aerial maps, but the accuracy of the GCPs is not guaranteed.
According to Figure 3, and the test results in Figure 10, Figure 11 and Table 4, one can see that:
  • Various kinds of geometric models can be used as the initial imaging model, including rigorous models for different sensors, the vendor-provided RPC model and the affine transformation model contained in georeferenced images.
  • The proposed method is successfully applied to match images captured in different imaging conditions, e.g., by different sensors, at different times, at different ground sample distances, etc.
  • Sufficient and well-distributed GCPs are efficiently collected for sensed images of different spatial resolutions, and the biases between the sensed images and online aerial maps are corrected after the process of image matching and geometric rectification.
  • It is a very convenient and efficient way to automatically collect GCPs for the task of geometric rectification of RS images, as there is no need to manually prepare reference images according to the location and spatial resolution of sensed images.
Note that the process of matching online is commonly a bit more time consuming than matching locally, since sending requests and downloading image tiles from online aerial maps may take more time than extracting image tiles from local reference images.
Figure 10. Distribution of matched points (marked by a cross) and sensed images in online image matching tests. (ah) are the results of Landsat-5, Cbers-2, Cbers-4, ZY-3, GF-2, Spot-5, Theos and GF-1 images, respectively.
Figure 10. Distribution of matched points (marked by a cross) and sensed images in online image matching tests. (ah) are the results of Landsat-5, Cbers-2, Cbers-4, ZY-3, GF-2, Spot-5, Theos and GF-1 images, respectively.
Remotesensing 08 00056 g010
Table 4. Test results of online image matching.
Table 4. Test results of online image matching.
Test No.Required GCPsBlock No.Correct MatchesRun-Time (S)Rectification ModelRMSE (Pixel)
1306 × 6327.11Rigorous1.24
210010 × 107541.00TD-RPC1.01
310010 × 109742.94TD-RPC0.84
4306 × 63523.48RPC-Refine1.82
5306 × 63616.01RPC-Refine1.35
610010 × 108245.93TD-RPC1.53
710010 × 107767.41TD-RPC1.46
8306 × 63621.96RPC-Refine1.77
Figure 11. Layer swiping between sensed images and online aerial maps in Tests 4, 6, 7 and 8. (a) and (b) are from Test 4, and the top ones are Bing aerial maps, while the lower one in (a) is the warped ZY-3 image using vendor-provided RPC and the lower one in (b) is the rectified ZY-3 image using RPC refinement; (c) and (d) are from Test 6, and the left are Bing aerial maps, while the right one in (c) is the Spot-5 image of Level 2 and the right one in (d) is the rectified Spot-5 image using terrain-dependent RPC; (e) and (f) are from Test 7, and the right are Bing aerial maps, while the left one in (e) is the Theos image of Level 2 and the left one in (f) is the rectified Theos image using terrain-dependent RPC; (g) and (h) are from Test 8, and the upper are Bing aerial maps, while the lower one in (e) is the warped GF-1 image using vendor-provided RPC and the lower one in (f) is the rectified GF-1 image using RPC refinement.
Figure 11. Layer swiping between sensed images and online aerial maps in Tests 4, 6, 7 and 8. (a) and (b) are from Test 4, and the top ones are Bing aerial maps, while the lower one in (a) is the warped ZY-3 image using vendor-provided RPC and the lower one in (b) is the rectified ZY-3 image using RPC refinement; (c) and (d) are from Test 6, and the left are Bing aerial maps, while the right one in (c) is the Spot-5 image of Level 2 and the right one in (d) is the rectified Spot-5 image using terrain-dependent RPC; (e) and (f) are from Test 7, and the right are Bing aerial maps, while the left one in (e) is the Theos image of Level 2 and the left one in (f) is the rectified Theos image using terrain-dependent RPC; (g) and (h) are from Test 8, and the upper are Bing aerial maps, while the lower one in (e) is the warped GF-1 image using vendor-provided RPC and the lower one in (f) is the rectified GF-1 image using RPC refinement.
Remotesensing 08 00056 g011

5. Conclusions

In this paper, we proposed a convenient approach to automatically collect GCPs from online aerial maps, which focuses on automated georeferencing of remotely-sensed (RS) images and makes use of the prior information of the RS image. The proposed method is based on SIFT feature, and the improvements accomplished in this work help to overcome the difficulties of SIFT when directly used in RS images, e.g., large image size, distribution of matched points, limited accuracy, outliers, etc. Both local reference images and online aerial maps can be utilized to collect control points. Different kinds of large-sized RS images, whose spatial resolutions vary from 30 m to 2 m, are included in the experiments, and the results show that the matching process can be finished within tens of seconds, yielding a sufficient number of reliable ground control points (GCPs). With the help of these reliable GCPs and DEM data, the root mean square errors (RMSEs) of the check points from the georeferenced images are less then two pixels. Moreover, by utilizing the online aerial maps, there is no need to manually prepare reference images according to the location and spatial resolution of sensed images.
Although we can successfully find enough GCPs for very high resolution (less than 1 m) RS images from the online aerial maps, the accuracy of the GCPs is not guaranteed. However, we believe the proposed approach will become even more useful as the accuracy of online aerial maps improves.

Acknowledgments

The RapidEye 5 m image in Hawaii was acquired from MAPMART [44], and the authors would like to thank the open source software, such as OSSIM, GDAL, OpenCV, VLFeat, etc., which provided great convenience for carrying out the experiments described in this paper. This work was supported by the grant from the National Natural Science Foundation of China (No. 61271013), the Director Foundation of the Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences (Y5ZZ10101B) and the National Natural Science Foundation of China (No. 61401461).

Author Contributions

Tengfei Long developed the methods, carried out the experiments and analyzed the results. Guojin He and Weili Jiao supervised the research. Zhaoming Zhang prepared the data. Tengfei Long and Weili Jiao wrote the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ozcanli, O.C.; Dong, Y.; Mundy, J.L.; Webb, H.; Hammoud, R.; Victor, T. Automatic Geo-location Correction of Satellite Imagery. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Columbus, OH, USA, 23–28 June 2014; pp. 307–314.
  2. Google Satellite Images. Available online: https://www.google.com/maps (accessed on 30 November 2015).
  3. Bing Aerial Images. Available online: http://www.bing.com/maps/ (accessed on 30 November 2015).
  4. MapQuest Satellite Maps. Available online: http://www.mapquest.com/satellite-maps/ (accessed on 30 November 2015).
  5. Mapbox Satellite Images. Available online: https://www.mapbox.com/mapbox-gl-js/example/satellite-map/ (accessed on 30 November 2015).
  6. Google Earth. Available online: https://www.google.com/earth/ (accessed on 30 November 2015).
  7. NASA World Wind. Available online: http://worldwind.arc.nasa.gov (accessed on 30 November 2015).
  8. Lemmens, M. A survey on stereo matching techniques. Int. Arch. Photogramm. Remote Sens. 1988, 27, 11–23. [Google Scholar]
  9. Zitova, B.; Flusser, J. Image registration methods: A survey. Image Vision Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef]
  10. Chen, Q.S.; Defrise, M.; Deconinck, F. Symmetric phase-only matched filtering of Fourier-Mellin transforms for image registration and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 1156–1168. [Google Scholar] [CrossRef]
  11. Zakharov, I.; Toutin, T. Subpixel image matching based on Fourier phase correlation for Radarsat-2 stereo-radargrammetry. Can. J. Remote Sens. 2012, 38, 487–495. [Google Scholar]
  12. Reddy, B.S.; Chatterji, B.N. An FFT-based technique for translation, rotation, and scale-invariant image registration. IEEE Trans. Image Process. 1996, 5, 1266–1271. [Google Scholar] [CrossRef] [PubMed]
  13. Gruen, A. Adaptive least squares correlation: A powerful image matching technique. S. Afr. J. Photogram. Remote Sens. Cartogr. 1985, 14, 175–187. [Google Scholar]
  14. Remondino, F.; El-Hakim, S.; Gruen, A.; Zhang, L. Turning images into 3-D models. IEEE Signal Process. Mag. 2008, 25, 55–65. [Google Scholar] [CrossRef]
  15. Gruen, A. Development and status of image matching in photogrammetry. Photogram. Rec. 2012, 27, 36–57. [Google Scholar] [CrossRef]
  16. Bethel, J. Least Squares Image Matching for CE604; Purdue University: West Lafayette, IN, USA, 1997. [Google Scholar]
  17. Long, T.; Jiao, W.; He, G.; Zhang, Z.; Cheng, B.; Wang, W. A Generic Framework for Image Rectification Using Multiple Types of Feature. ISPRS J. Photogramm. Remote Sens. 2015, 102, 161–171. [Google Scholar] [CrossRef]
  18. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vision 2004, 60, 91–110. [Google Scholar] [CrossRef]
  19. Bay, H.; Tuytelaars, T.; van Gool, L. Speeded-up robust features (SURF). Comput. Vision Image Understand 2008, 3, 346–359. [Google Scholar] [CrossRef]
  20. Rosten, E.; Porter, R.; Drummond, T. Faster and better: A machine learning approach to corner detection. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 105–119. [Google Scholar] [CrossRef] [PubMed]
  21. Leutenegger, S.; Chli, M.; Siegwart, R. BRISK: Binary robust invariant scalable keypoints. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Florence, Italy, 7–13 October 2011; pp. 2548–2555.
  22. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Florence, Italy, 7–13 October 2011; pp. 2564–2571.
  23. Alahi, A.; Ortiz, R.; Vandergheynst, P. FREAK: Fast Retina Keypoint. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 510–517.
  24. Ke, Y.; Sukthankar, R. PCA-SIFT: A more distinctive representation for local image descriptors. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 27 June–2 July 2004; Volume 2, pp. 2–506.
  25. Mikolajczyk, K.; Schmid, C. A performance evaluation of local descriptors. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1615–1630. [Google Scholar] [CrossRef] [PubMed]
  26. Sun, Y.; Zhao, L.; Huang, S.; Yan, L.; Dissanayake, G. L-2-SIFT: SIFT feature extraction and matching for large images in large-scale aerial photogrammetry. ISPRS J. Photogramm. Remote Sens. 2014, 91, 1–16. [Google Scholar] [CrossRef]
  27. Wu, C. SiftGPU: A GPU Implementation of Scale Invariant Feature Transform (SIFT). Available online: http://cs.unc.edu/ ccwu/siftgpu (accessed on 30 November 2015).
  28. Björkman, M.; Bergström, N.; Kragic, D. Detecting, segmenting and tracking unknown objects using multi-label MRF inference. Comput. Vision Image Understand. 2014, 118, 111–127. [Google Scholar] [CrossRef]
  29. Li, Q.; Wang, G.; Liu, J.; Chen, S. Robust scale-invariant feature matching for remote sensing image registration. IEEE Geosci. Remote Sens. Lett. 2009, 6, 287–291. [Google Scholar]
  30. Bin, F.; Chunlei, H.; Chunhong, P.; Qingqun, K. Registration of Optical and SAR Satellite Images by Exploring the Spatial Relationship of the Improved SIFT. IEEE Geosci. Remote Sens. Lett. 2013, 10, 657–661. [Google Scholar] [CrossRef]
  31. Ye, Y.; Shan, J. A local descriptor based registration method for multispectral remote sensing images with non-linear intensity differences. ISPRS J. Photogramm. Remote Sens. 2014, 90, 83–95. [Google Scholar] [CrossRef]
  32. Sedaghat, A.; Mokhtarzade, M.; Ebadi, H. Uniform robust scale-invariant feature matching for optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4516–4527. [Google Scholar] [CrossRef]
  33. Wang, L.; Niu, Z.; Wu, C.; Xie, R.; Huang, H. A robust multisource image automatic registration system based on the SIFT descriptor. Int. J. Remote Sens. 2012, 33, 3850–3869. [Google Scholar] [CrossRef]
  34. Yi, Z.; Zhiguo, C.; Yang, X. Multi-spectral remote image registration based on SIFT. Electron. Lett. 2008, 44, 107–108. [Google Scholar] [CrossRef]
  35. Zhili, S.; Shuigeng, Z.; Jihong, G. A Novel Image Registration Algorithm for Remote Sensing Under Affine Transformation. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4895–4912. [Google Scholar] [CrossRef]
  36. Kupfer, B.; Netanyahu, N.S.; Shimshoni, I. An Efficient SIFT-Based Mode-Seeking Algorithm for Sub-Pixel Registration of Remotely Sensed Images. IEEE Geosci. Remote Sens. Lett. 2015, 12, 379–383. [Google Scholar] [CrossRef]
  37. Tao, C.V.; Hu, Y. A comprehensive study of the rational function model for photogrammetric processing. Photogramm. Eng. Remote Sens. 2001, 67, 1347–1358. [Google Scholar]
  38. Fraser, C.; Dial, G.; Grodecki, J. Sensor orientation via RPCs. ISPRS J. Photogramm. Remote Sens. 2006, 60, 182–194. [Google Scholar] [CrossRef]
  39. Lourakis, M. Levmar: Levenberg-Marquardt Nonlinear Least Squares Algorithms in C/C++. 2004. Available online: http://www.ics.forth.gr/ lourakis/levmar/ (accessed on 30 November 2015).
  40. Bing Maps Tile System. Available online: https://msdn.microsoft.com/en-us/library/bb259689.aspx (accessed on 30 November 2015).
  41. OSSIM. Available online: http://www.ossim.org (accessed on 30 November 2015).
  42. Long, T.; Jiao, W.; He, G. RPC Estimation via 1-Norm-Regularized Least Squares (L1LS). IEEE Trans. Geosci. Remote Sens. 2015, 53, 4554–4567. [Google Scholar] [CrossRef]
  43. ArcBruTile. Available online: https://arcbrutile.codeplex.com/releases/view/616921 (accessed on 30 November 2015).
  44. MAPMART Sample Data. Available online: http://www.mapmart.com/Samples.aspx (accessed on 30 November 2015).

Share and Cite

MDPI and ACS Style

Long, T.; Jiao, W.; He, G.; Zhang, Z. A Fast and Reliable Matching Method for Automated Georeferencing of Remotely-Sensed Imagery. Remote Sens. 2016, 8, 56. https://doi.org/10.3390/rs8010056

AMA Style

Long T, Jiao W, He G, Zhang Z. A Fast and Reliable Matching Method for Automated Georeferencing of Remotely-Sensed Imagery. Remote Sensing. 2016; 8(1):56. https://doi.org/10.3390/rs8010056

Chicago/Turabian Style

Long, Tengfei, Weili Jiao, Guojin He, and Zhaoming Zhang. 2016. "A Fast and Reliable Matching Method for Automated Georeferencing of Remotely-Sensed Imagery" Remote Sensing 8, no. 1: 56. https://doi.org/10.3390/rs8010056

APA Style

Long, T., Jiao, W., He, G., & Zhang, Z. (2016). A Fast and Reliable Matching Method for Automated Georeferencing of Remotely-Sensed Imagery. Remote Sensing, 8(1), 56. https://doi.org/10.3390/rs8010056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop