Next Article in Journal
The Variability of Snow Cover and Its Contribution to Water Resources in the Chinese Altai Mountains from 2000 to 2022
Next Article in Special Issue
SatellStitch: Satellite Imagery-Assisted UAV Image Seamless Stitching for Emergency Response without GCP and GNSS
Previous Article in Journal
Digitization of the Built Cultural Heritage: An Integrated Methodology for Preservation and Accessibilization of an Art Nouveau Museum
Previous Article in Special Issue
Tree Species Classification from Airborne Hyperspectral Images Using Spatial–Spectral Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Modal Image Registration Based on Phase Exponent Differences of the Gaussian Pyramid

1
School of Undergraduate Education, Shenzhen Polytechnic University, Shenzhen 518055, China
2
Institute of Applied Artificial Intelligence of the Guangdong-Hong Kong-Macao Greater Bay Area, Shenzhen Polytechnic University, Shenzhen 518055, China
3
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
4
School of Artificial Intelligence, Shenzhen Polytechnic University, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(24), 5764; https://doi.org/10.3390/rs15245764
Submission received: 23 October 2023 / Revised: 4 December 2023 / Accepted: 13 December 2023 / Published: 17 December 2023

Abstract

:
In multi-modal images (MMI), the differences in their imaging mechanisms lead to large signal-to-noise ratio differences, which means that the matching of geometric invariance and the matching accuracy of the matching algorithms often cannot be balanced. Therefore, how to weaken the signal-to-noise interference of MMI, maintain good scale and rotation invariance, and obtain high-precision matching correspondences becomes a challenge for multimodal remote sensing image matching. Based on this, a lightweight MMI alignment of the phase exponent of the differences in the Gaussian pyramid (PEDoG) is proposed, which takes into account the phase exponent differences of the Gaussian pyramid with normalized filtration, i.e., it achieves the high-precision identification of matching correspondences points while maintaining the geometric invariance of multi-modal matching. The proposed PEDoG method consists of three main parts, introducing the phase consistency model into the differential Gaussian pyramid to construct a new phase index. Then, three types of MMI (multi-temporal image, infrared–optical image, and map–optical image) are selected as the experimental datasets and compared with the advanced matching methods, and the results show that the NCM (number of correct matches) of the PEDoG method displays a minimum improvement of 3.3 times compared with the other methods, and the average RMSE (root mean square error) is 1.69 pixels, which is the lowest value among all the matching methods. Finally, the alignment results of the image are shown in the tessellated mosaic mode, which shows that the feature edges of the image are connected consistently without interlacing and artifacts. It can be seen that the proposed PEDoG method can realize high-precision alignment while taking geometric invariance into account.

1. Introduction

With the rapid development of computer vision and multi-sensor technology [1] in recent years, the vast amount of multi-modal images has not only provided rich information but also poses significant challenges for image analysis [2]. Image registration is one of the fundamental tasks in the analysis of multi-modal images, aimed at geometrically aligning two or more images with overlapping regions. Multi-modal image (MMI) registration is an indispensable step in processes such as image fusion, image stitching, and change detection, and it has become a hot topic in international academic research [3].
The significant differences in illumination, contrast, geometric distortions, and nonlinear radiometric variations in MMIs make them more sensitive to image gradients and orientations, making it challenging to extract reliable common features and often resulting in subpar feature matching results. When registering MMIs, traditional parameter optimization methods can easily become trapped in local optima, leading to lower registration accuracy. In recent years, researchers have conducted extensive studies on MMI matching [4], continuously improving algorithms and evaluating accuracy to enhance the precision and reliability of registration. Despite substantial efforts to enhance the robustness of multi-modal image matching, challenges still persist in achieving both geometric invariance and high-precision registration.
To address the limitations of feature matching, as well as challenges posed by issues such as signal-to-noise ratio interference and low matching efficiency in multi-modal images. Conventional feature-based image registration methods like SIFT (the scale invariant feature transform) [5] and SURF (speeded-up robust features) [6] typically rely on gradient information to extract and describe features, which makes them sensitive to radiometric differences and unsuitable for multi-modal image registration. In contrast to gradients, image phase congruency (PC) features reflect structural information in images and are insensitive to radiometric variations [7]. Therefore, this paper combines the phase congruency model with differences in the Gaussian image pyramid and optimizes the SIFT feature detector along with normalized logarithmic polar descriptors, collectively referred to as the PEDoG algorithm. This algorithm captures more robust feature points, thereby improving the efficiency of multi-modal image matching while suppressing image noise and maintaining strong geometric invariance, facilitating robust matching in multi-modal images.
The main contributions of this study are as follows:
(1)
A lightweight multi-modal image registration method was introduced that considers phase-indexed difference pyramids and normalized filtering. This approach achieves the high-precision recognition of corresponding points while maintaining geometric invariance in multi-modal matching.
(2)
A method for constructing phase-indexed difference pyramids was proposed. By combining the phase congruency model with differences in the Gaussian pyramid and optimizing it using exponential functions, we established the phase-indexed difference pyramid image space. We also employed an SIFT-like feature extractor to extract robust feature points.
(3)
A method for normalized filtering in logarithmic polar descriptors was introduced. This involves incorporating normalized filtering functions to enhance structural information in images, constructing second-order gradient-oriented features, and ultimately using a logarithmic polar coordinate framework to generate efficient descriptors that represent features in multi-modal images effectively.
The structure of this paper is as follows: In Section 1, we introduce the significance of multi-modal image matching and the importance of this research. Section 2 briefly reviews the limitations of previous studies. Section 3 outlines the complete process of the proposed PEDoG method. In Section 4, we provide a detailed analysis of the experiments conducted. Section 5 discusses the impact of different parameter settings on the performance of PEDoG and its performance concerning rotation and scale variations. Finally, in Section 6, we summarize the contributions of this study.

2. Relate Work

Image registration is one of the critical preprocessing steps in practical applications of remote sensing imagery and holds significant practical value. Its outcomes directly impact the accuracy of tasks such as image stitching, image fusion, and object detection. Image registration methods can be categorized into three main types: region-based registration methods, feature-based registration methods, and deep learning-based methods. The following section provides a detailed overview of each of these approaches.
Region-based registration methods establish registration relationships between image pairs using the raw pixel values and specific similarity metrics. These methods can be broadly categorized into three classes, including correlation-based methods [8], Fourier-based methods [9], and mutual information-based methods [10]. In recent years, some researchers have introduced phase correlation techniques, leading to rapid advancements in multi-modal image matching. Building upon this, experts and scholars have developed methods that combine phase congruency models, self-similarity features, and enhanced gradient features. These include techniques such as directional gradient distance histograms combined with grey wolf optimization [11], log–Gabor filter-optimized matching HOPC (histogram of orientated phase congruency) [12], CFOG (channel features of orientated gradients) [13], angle-weighted orientation gradient (AWOG) [14], and multi-orientation tensor index feature (MoTIF) [15]. These approaches effectively overcome nonlinear radiometric distortions and contrast differences between multi-modal images and exhibit strong robustness in handling displacement variations in images. However, it is worth noting that such methods primarily focus on translational shifts, and if images involve complex geometric transformations, registration methods may encounter challenges.
Feature-based methods in the field of image registration began with Lowe et al.’s introduction of scale invariant feature transform (SIFT) matching [5], leading to the rapid development of various SIFT-like techniques [16]. However, gradient features are unable to adapt to the modality differences in multi-modal images, making such methods unsuitable for multi-modal image matching. Ma et al. introduced the PSO-SIFT (position scale orientation—SIFT) algorithm [17], which performs well in handling nonlinear brightness differences and rotation changes through the establishment of new image gradient features, but it remains sensitive to contrast differences and signal-to-noise ratio variations. Sedaghat et al. proposed the HOSS (histogram of oriented self-similarity) algorithm [18] based on self-similarity features, which ensures good rotational invariance but performs poorly in multi-modal images with high contrast and nonlinear radiometric distortions. The RRSS (rank-based ratio self-similarity) method [19] proposed by Xiong et al. effectively addresses differences between multi-modal images but incurs some loss in rotational invariance.
Some experts and researchers have tackled multi-modal image matching from the perspective of phase congruency models. However, it relies on a strategy involving circular feature calculations to overcome rotational differences, which results in lower technical efficiency. Xiang et al. [20] enhanced the PC model, keypoint extraction, and similarity measurement methods, constructing features for optical and SAR image matching. Fan et al. [21] designed a multi-scale PC descriptor known as multi-scale adaptive block phase congruency (MABPC). This descriptor utilizes multi-scale phase congruency features encoded with an adaptive block spatial structure. Yao et al. [22]. proposed absolute phase orientation histogram matching, designing absolute phase-oriented features to accommodate differences between multi-modal images and resist scale, displacement, and rotational differences. However, this method is limited to matching tasks with small rotational differences. Yang et al. [23] introduced local phase sharpness-oriented features to adapt to MMRSI matching and improve the applicability to rotational differences in multi-modal images but still lacks complete local rotational invariance.
Recently, various algorithms have been proposed to address multi-modal image differences by improving image scale space, such as the CoFSM algorithm [24], which reduces multi-modal image differences by improving the image scale space. Other approaches include the MS-HLMO algorithm [25], which is based on multiscale joint mean minimal gradient features, and multi-modal image matching through local normalized image filtering [26]. These methods have improved the matching performance of multi-modal remote sensing images, demonstrating strong rotational invariance. However, they still suffer from issues such as the low accuracy of corresponding points and poor matching robustness to varying degrees.
With the rapid development of artificial intelligence theory, deep learning techniques have been introduced into multi-modal image matching. Examples include convolutional neural network-based matching [27] and graph neural network-based matching (SuperGlue) [28]. These methods have demonstrated excellent performance in homogenous image matching but have shown limitations when applied to multi-modal images. In response to this challenge, researchers have explored methods specifically designed for multi-modal image matching. These include the D2-Net network for multi-source image feature extraction and description [29], the LoFTR algorithm based on transformer networks [30], and its improved version, SE2-LoFTR, with rotational invariance [31]. M2DT-Net [32] is a method that combines learned features with Delaunay triangulation constraints, but it has certain limitations in handling radiometric invariance. Deep learning-based approaches offer speed and strong feature learning capabilities, especially in the context of multi-modal image matching, where they have shown significant potential. However, due to substantial variations in land features between multi-modal images and the difficulty in obtaining training samples, the generalization and applicability of such methods are currently limited.
In summary, MMIs exhibit significant differences in illumination, contrast, nonlinear geometric and radiometric distortions, making it challenging to extract reliable similarity features and often leading to matching failures. When registering multi-modal remote sensing images, traditional parameter optimization methods are prone to becoming stuck in local optima, resulting in lower registration accuracy that cannot meet the requirements of high-precision applications. We proposed a method that starts by extracting features through the design of a phase difference pyramid and then constructs feature descriptors using normalized filtering in logarithmic polar coordinates. This approach aims to improve the success rate and accuracy of multi-modal image matching. The method presented in this paper enhances the precision and robustness of feature matching and registration in multi-modal remote sensing images. It also satisfies the demands of high-precision tasks such as image fusion, image stitching, and change detection, holding significant theoretical and practical value.

3. Methods

The proposed PEDoG method consists of four steps: (1) the construction of the phase-indexed difference of gaussian pyramid, (2) improved SIFT-like feature point extraction, (3) the construction of normalized filtering in logarithmic polar coordinate descriptors, and (4) matching and outlier removal. Of these steps, (1) and (3) are the focal points of this paper. In step (1), we build the phase-indexed Gaussian difference pyramid by first calculating the maximum moment feature of the image using the phase congruency model. Next, we construct a Gaussian difference pyramid for the moment feature and then optimize the Gaussian phase pyramid using exponential equations, ultimately generating the phase-indexed Gaussian difference pyramid feature. The feature matching and outlier removal processes employ the nearest-neighbor ratio matching strategy [33] and the fast sample consensus algorithm [34] to eliminate outliers, respectively. Following in Figure 1, shows the overall flow chart of the proposed method.

3.1. Phase Exponent of Difference of Gaussian-Pyramid

To address the issues of nonlinear radiometric distortions and noise between multi-modal remote sensing images, a phase-indexed Gaussian difference pyramid feature detector is introduced. This detector utilizes the structure tensor-based phase congruency (PC) to construct a difference image pyramid that is robust to nonlinear radiometric distortions (NRD) and noise. The following sections provide a detailed explanation.
(1) Phase coherence model solving. The PC value represents the importance of features and is robust to NRD and noise [35]. Therefore, the PC structure map is very suitable for feature point detection in multi-modal remote sensing images. The PC structure map is calculated as follows. First, the image I(x, y) is convolved with the odd-symmetric wavelet Lodd(x, y, s, o) and even-symmetric wavelet Leven(x, y, s, o) to obtain the response components Eso(x, y) and Oso(x, y).
L x , y , s , o = L e v e n x , y , s , o + i L o d d ( x , y , s , o ) E s o ( x , y ) , O s o ( x , y ) = [ I ( x , y ) L e v e n x , y , s , o , I ( x , y ) L o d d ( x , y , s , o ) ]
where s and o denote the scale and orientation of Log–Gabor wavelets, respectively; L e v e n x , y , s , o represents the even-symmetric filter of the Log–Gabor filter; L o d d ( x , y , s , o ) represents the odd-symmetric filter; i-th represents the imaginary unit of the retest; I(x, y) is the image; E s o ( x ,   y ) represents the response result of the image on the real part filter; O s o ( x , y ) represents the response result of the image on the imaginary part filter; and represents the convolution operation.
Then, the amplitude component Aso(x, y) at scale s and direction o is calculated. By combining all directions and scales and introducing a noise compensation term T, the value of P C ( x , y ) is calculated as follows:
P C ( x , y ) = S O w O ( x , y ) A S O ( x , y ) Δ Φ S O ( x , y ) T S O A S O ( x , y ) + ξ
where wo is the weighting factor and ξ is a small constant that prevents the denominator from being zero, and when its value is less than zero, ⌊.⌋ means it is equal to zero, otherwise it is unchanged. ∆Φso(x, y) is the phase deviation function and Aso(x, y) ∆Φso(x, y) can be calculated from the response components Eso(x, y) and Oso(x, y). Also, it can be calculated using Equation (2), which can calculate the importance of the phase feature at each pixel and is then defined as the maximum moment.
(2) Construction of a Gaussian difference image pyramid. In order to extract scale-invariant feature points in multimodal remote sensing images, this paper constructs a differential image pyramid using the phase coherence structure map. First, the scale space L(x,y,σ) is constructed from the PC structure map PC(x, y) and the Gaussian function G(u, v, σ) with the formula shown in (3) and the Gaussian function G(u, v, σ) shown in (4):
L ( x , y , σ ) = M ( x , y ) G ( u , v , σ )
G u , v , σ = e u 2 + v 2 2 σ 2
In the scale space, Gaussian-smoothed images are computed using a Gaussian function with continuous scale σn = knσ0, where k is a constant multiplication factor between two neighboring scales. Then, the phase difference image pyramid is computed as shown in (5):
D ( x , y , σ ) = L ( x , y , k σ ) L ( x , y , σ )
In Equation (5), denotes the phase Gaussian difference pyramid result. However, in order to further enhance the structural feature information of the image and reduce the contrast difference of the image, this paper also introduces the exponential function [36] to optimize the phase Gaussian difference pyramid features, the expression of which is shown in (6):
P E ( x , y , σ ) = exp D ( x , y , k σ ) exp D ( x , y , σ )
In Equation (6), P E x , y , σ denotes the phase-exponential Gaussian difference pyramid result. In order to better demonstrate the results of the phase-exponential Gaussian difference pyramid, a set of infrared–optical images are selected to be significant in this paper, and the results are shown in Figure 2.

3.2. Improved Feature Detection

After constructing the difference image pyramid, the detection of extremal points within its multi-scale octave range is performed. Typically, it involves comparing a pixel with 26 points in its neighborhood, as seen in algorithms like SIFT [5]. However, the texture of the PC structural map is quite sparse, and the Gaussian function further blurs it, resulting in the detection of only a few feature points. Therefore, it is possible to directly use an 8-neighborhood detection method to ensure an adequate number of feature points.
First, in each layer of the DoG image pyramid, candidate feature points with locally very large or very small values within their 8 neighborhoods are detected. In addition, the feature point extraction threshold (Ct) is set as its mathematical expression is shown in (7):
P E ( x , y , σ n ) > P E ( x , y , σ n ) P E ( x , y , σ n ) < P E ( x , y , σ n ) P E ( x , y , σ n ) > C t
where ( x , y ) denotes the coordinates of the 8 neighborhood pixel points of (x, y). Then, it is verified whether the phase index of the candidate feature point in the Gaussian difference detector in the scale direction is greater than the feature point extraction threshold value.
Based on the above three steps, the phase-exponential Gaussian differential feature detector can extract scale-invariant feature points. As can be seen in Figure 3, the feature points are basically distributed over the image structure and a large number of corresponding feature points are detected in multimodal remote sensing image pairs.

3.3. Improved GLOH-like Feature Descriptors

In this paper, we design improved log-polar descriptors for normalized filtering. Firstly, normalized filtering is introduced to weaken the modal differences of the image, and then the second-order oriented gradient features after normalized enhancement are solved. Then, the joint log-polar coordinate descriptor framework calculates the descriptor vector features to complete the descriptor construction. The flowchart of an improved log-polar descriptor is shown in Figure 4.

3.3.1. Normalized Filtering of Image-Oriented Gradient Features

Image-oriented gradient features are a crucial step in feature description. However, traditional gradient features of images are sensitive to the nonlinear radiometric distortions in MMIs, making them less robust and less favorable for robustly representing feature points. Normalized filtering is a filtering method proposed by Li et al., 2022 [26], which effectively enhances the similarity information in MMIs. Building on this, we introduce this method into the computation of oriented gradient features for MMIs, thereby reducing modality feature differences and better expressing the common information in the images. The mathematical expression for normalized filtering is shown below:
P N ( x , y ) = P ( x , y ) 1 W x , y , s W x , y , s P ( x , y )
where P N ( x , y ) denotes the normalized image; P ( x , y ) denotes the original image; (x, y) denotes the coordinates of the image; W x , y , s is a local window centered on (x, y) with a size of (2 × s + 1) × (2 × s + 1); and denotes the taking of an absolute value. Essentially, the local normalization filter is equivalent to the original image minus its average filtering result. Therefore, this filter retains only the detailed components, i.e., the normalized image contains most of the structural information of the image, which is very important for multimodal image matching. This is the important theoretical reason why we introduce this method.
Image-oriented gradient computation. The normalized filtered image is passed through the Soble operator [37] to eliminate the nonlinear brightness difference of the image. Therefore, in order to further highlight the feature information, the second- and third-order gradient computation of the image is carried out based on the joint Sobel template with the following equation.
P S σ 2 = ( P N ( x , y ) x , σ 1 Λ x ) 2 + ( P N ( x , y ) y , σ 1 Λ y ) 2 A n g l e σ 2 = arctan ( P N ( x , y ) y , σ 1 Λ y P N ( x , y ) x , σ 1 Λ x )
P S σ 3 = ( P S x , σ 2 Λ x ) 2 + ( P S y , σ 2 Λ y ) 2 A n g l e σ 3 = arctan ( P S y , σ 2 Λ y P S x , σ 2 Λ x )
Ʌ x = 1 2 0 1 0 2 1 0 1    Ʌ y = 1 0 2       1 0       0 1 2 1
In Equations (9)–(11), P S σ 2 denotes the new second-order gradient amplitude; A n g l e σ 2 denotes the new second-order gradient direction; P S σ 3 denotes the new third-order gradient amplitude; A n g l e σ 3 denotes the new third-order gradient direction; Λ x and Λ y are the gradient order amplitude function, where Λ x denotes the X-direction Sobel operator template; Λ y denotes the Y-direction Sobel operator template; and σ denotes the image scale.

3.3.2. Log-Polar Descriptive Feature Solving

After extracting key points, describing them is a critical subsequent step for successful matching. The log-polar histogram framework is a classical framework that has been successfully applied in multi-modal image matching. The log-polar descriptor approach using the gradient location-orientation histogram (GLOH) algorithm is notably advantageous and stable [38]. However, the log-polar descriptor method is not the only one available, and it heavily depends on the partitioning of the polar grid. Different partitioning methods can generate different descriptors. Therefore, it exhibits better versatility when constructing descriptors.
Therefore, considering the stability and robustness of the descriptor, this paper, starting from the zero-degree direction on the right end of the polar grid, divides the circular neighborhood into 12 equal parts, creating a new polar coordinate grid consisting of (12 × 3 + 1) sub-region grids. Each sub-region in the polar coordinate grid has approximately equal area. Within each grid, the horizontal direction represents the polar angle of the pixels within the circular neighborhood. After calculating the orientation histogram for each feature point, the direction is divided into 8 dimensions at 45-degree intervals, covering a range from 0 to 360 degrees. As a result, each sub-region grid has adjacent points with 8-dimensional gradient location-orientated histograms. Finally, multiplying the number of sub-region grids in the log-polar coordinates (38) by the dimensions of the orientation histogram (8 dimensions) generates a novel 296-dimensional log-polar descriptor.

4. Results

To evaluate the performance of PEDoG, it was compared to five state-of-the-art algorithms (SIFT, PSO-SIFT, RIFT2 [39], OSS (oriented self-similarity) [40], and HOWP (histogram of the orientation of the weighted phase descriptor) [3]). To ensure a fair comparison, the code provided by the authors and their recommended parameter settings were used. The paper employed the fast sample consensus algorithm to eliminate mismatches and obtain corresponding points. PEDoG was implemented in MatlabR2020a, and the experimental platform used an AMD Ryzen 9 5900HX with Radeon Graphics processor clocked at 3.30 GHz, 64 GB of RAM, and the Windows 11 X64 operating system.

4.1. Experimental Datasets

In this study, three commonly used multi-modal data types were selected as data sources, including multi-temporal images, infrared and optical images, and electronic navigation maps with optical images, totaling 60 image pairs, with each modality consisting of 20 image pairs. The image sizes ranged from 381 pixels to 750 pixels. One pair of images from each data type is shown in Figure 5. This dataset encompasses diverse scenarios for multi-modal image matching applications, with ground resolutions ranging from 1 m to 300 m, making it representative. Additionally, significant nonlinear radiometric differences and geometric transformation differences exist between the image pairs.

4.2. Evaluation of Indicators

To comprehensively evaluate the performance of the proposed PEDoG method, this paper provides both qualitative and quantitative comparative results. The quantitative evaluation employs three metrics to assess the matching performance of the methods: success rate (SR), NCM, and RMSE. SR measures the ratio of successfully matched image pairs to the total number of image pairs. This metric reflects the robustness of the matching method for specific types of multi-modal image pairs. Image pairs with an NCM of less than 20 and matching errors greater than seven pixels are considered matching failures. The number of matches within seven pixels are counted as the correct matches. RMSE is a commonly used metric to assess the performance of matching algorithms. RMSE reflects the precision of correct matches. A smaller RMSE value indicates higher accuracy. The mathematical expression for RMSE is given in Equation (12):
R M S E = 1 N i = 1 N x i x i 2 + y i y i 2
where N is the number of homonymous points, x i , y i is the coordinate of the i-th truth point x i , y i after the corresponding matching transformation.

4.3. Results

Table 1 displays the average results for the four metrics. In Figure 6a, “Image Type 1” represents multi-temporal optical–optical images, “Image Type 2” represents infrared–optical images, and “Image Type 3” represents map–optical images. In Figure 6b, the symbol “+∞” indicates matching failures or RMSE 7 pixels. The units for SR are percentage (%), NCM is in point numbers, and RMSE is in pixels.
Figure 6 presents a comparison of the matching results between the PEDoG algorithm and five other algorithms using two metrics, NCM and RMSE. The results of the SIFT method are represented by a black dashed line. Its SR is the lowest, standing at only 58.3%, with an average NCM of 79.53, the lowest among the six algorithms. The RMSE is 4.08 pixels, indicating the lowest matching accuracy. The SIFT algorithm achieves feature description by computing the gradient features of images, but it is sensitive to multi-modal remote sensing images with nonlinear radiometric distortions and contrast differences, leading to matching failures. Therefore, the applicability of the SIFT algorithm in multi-modal remote sensing image matching is relatively poor.
The results of the PSO-SIFT algorithm are shown as a blue dashed line. It slightly outperforms the SIFT algorithm in terms of matching performance with an SR of 88.3% and an average NCM of 139.95. It has good MRSI scaling and rotation invariance, as evidenced by an average RMSE of 2.36 pixels.
The results of the OSS algorithm are indicated by a green dashed line. The algorithm utilizes offset mean filtering to quickly compute self-similar features. SR results are better at 91.7%, which is in the middle of the range. The average NCM is 175.6, which is lower than the HOWP, RIFT2, and PEDoG algorithms. This may be due to the low discriminative power of the LSS descriptors, which cannot maintain a robust matching performance in multimodal matching situations such as e-navigation maps and optical images. The RMSE of this algorithm is lower among the six algorithms with an average RMSE of 2.08 pixels.
The results of the HOWP algorithm are indicated by the purple dashed line. The successful matching rate SR of this algorithm is 66.7%, and the reason for the low successful matching rate is that the algorithm does not support multimodal remote sensing image matching with too large a rotation angle and lacks robust rotational invariance. The average NCM is 236.92 and the average RMSE is 3.63 pixels.
The results of the RIFT2 algorithm are indicated by the brown dashed line. The algorithm uses image frequency domain features to complete the matching and greatly optimizes the matching performance by maximizing the indexed map descriptors. The SR is increased to 96.7% and the average NCM is 209.38. the average RMSE is 2.07 pixels. the algorithm is altered from the RIFT algorithm to support rotational invariance but does not support scale differences, so the RMSE results are volatile.
The results of the PEDoG algorithm are represented by a solid red line, and it produced the most robust matching outcomes. The PEDoG algorithm successfully matched all 60 pairs of multi-modal remote sensing images, achieving a 100% SR. The average NCM reached the highest value at 259.43, and the average RMSE was 1.69 pixels. The PEDoG algorithm is capable of handling geometric variations, nonlinear radiometric differences, illumination discrepancies, and contrast differences in MRSI, while also possessing scale and rotation invariance. In summary, the PEDoG algorithm generates more robust matching results in multi-modal remote sensing image matching tasks and exhibits superior overall performance compared to the other five algorithms.
(1)
Qualitative results
In order to better demonstrate the performance of the proposed PEDoG algorithm, this paper also qualitatively evaluates it against five comparison algorithms, and the results are shown in Figure 7.
Figure 7 shows the matching results of the six matching algorithms on the three types of data; the green line in the figure represents a correct match for the homonymous points, and the red line indicates a failed match. The optical–optical data show two sets of optical image pairs of different years; the four images shown in the infrared–optical data, where the image on the top left is an infrared image, the image on the bottom left is a thermal infrared image, and the matching images on their right are all optical images; in the map–optical data, the two on the left are electronic navigation maps, and the two on the right are the corresponding optical images. As shown in Figure 7, the proposed PEDoG algorithm can obtain the most abundant matching keypoints among the three modalities and exhibits good scale and rotation invariance. In contrast, the SIFT algorithm and PSO-SIFT algorithm perform well with multi-temporal images and electronic navigation maps but have only moderate performance with infrared images. The RIFT2 algorithm is an improved version of RIFT, demonstrating good rotation invariance but not ideal scale adaptation. The OSS algorithm employs local self-similarity descriptors and performs well with both multi-temporal images and infrared images but yields fewer matching keypoints. Lastly, the HOWP algorithm uses weighted phase-oriented features for description and performs well, although it is sensitive to significant rotational differences.
(2)
Results of the registration
To further demonstrate the performance of the PEDoG algorithm, a checkerboard visualization method was employed to showcase the registration results, as depicted in Figure 8. In the displayed image checkerboards, it is evident that the overlapping regions between images are fully aligned, effectively avoiding artifacts and misalignment issues. This further underscores the algorithm’s high registration accuracy. The exceptional matching performance of the PEDoG algorithm can be attributed to two main factors. Firstly, the PEDoG algorithm extracts keypoints using the phase-difference Gaussian pyramid, effectively leveraging the similarity features in multi-modal images, resulting in keypoints with higher repeatability and saliency. Secondly, the designed normalized log-polar coordinate descriptors, by introducing normalized filtering, effectively reduce the sensitivity of image gradient features to nonlinear radiometric distortions, thus achieving the superior representation of multi-modal images and enhancing registration accuracy.

5. Analysis and Discussion

5.1. Parameter Setting

To comprehensively evaluate the performance of the PEDoG algorithm, an analysis of its two core parameters was conducted: the number of layers in the differences of the Gaussian (DOG) pyramid, denoted as Dl, and the feature point extraction threshold C t . RMSE and NCM were employed as evaluation metrics to assess the impact of these parameters. The specific parameter settings are presented in Table 2. Through parameter tuning and experimental evaluations, we identified the optimal parameter configuration that achieves the best matching performance. Such parameter selection not only leads to ideal results in terms of RMSE and NCM but also ensures that the matching task can be completed within an acceptable time frame. RMSE is measured in pixels, while NCM is measured in point counts.
In this paper, 60 sets of MMI data are used to test the performance of PEDoG under different parameter settings. Figure 9a demonstrates the matching results under different DOG pyramid layers Dl. As Dl increases from 2 to 5, the NCM results show a steady upward trend to 268. However, when Dl exceeds 5, the NCM starts to decrease slightly until Dl reaches 8, at which point the NCM stays around ~250. For RMSE, the lowest value was reached at Dl = 4, which was 1.69 pixels. Considering NCM and RMSE together, the highest accuracy and rich matching of homonymous points can be obtained when Dl is set to 4, so it is recommended to set Dl to 4.
Figure 9b demonstrates the results for different coarseness rejection threshold Ct parameter settings. From the figure, it can be observed that the number of NCMs is decreasing as Ct increases, and considering the RMSE factor, the optimal parameter setting should be 0.3. The RMSE is at a low level while more matches are obtained. Therefore, in this paper, the number of pyramid layers Dl = 4 and the coarseness rejection threshold Ct = 0.3 are chosen as the parameter settings that match the richness of corresponding points with optimal accuracy.

5.2. Analysis of Rotational-Invariance

To verify the performance of the PEDoG algorithm in terms of rotation invariance, the same set of images was used for matching tests. Initially, the reference image was rotated in both clockwise and counterclockwise directions at 30-degree intervals, resulting in a total of 12 simulated images across six different orientations. Subsequently, these 12 sets of simulated images were used for matching tests, and the matching result images were displayed. As shown in Figure 10 (“-” represents counterclockwise rotation), in a range of [−180°, 180°] for the rotation difference setting, the PEDoG algorithm consistently achieved successful matches and obtained abundant NCM. This is attributed to the fact that the proposed PEDoG algorithm first utilizes the exponential phase difference pyramid to extract feature points. This approach ensures the extraction of significant feature points in multi-modal images while reducing the sensitivity of gradient features to nonlinear radiometric distortions, thus ensuring the accurate calculation of feature point primary directions. Furthermore, the design of normalized log-polar descriptors further mitigates the modality differences between multi-modal images, enhancing the feature description capability between such images and ensuring robustness in matching.
In summary, the PEDoG algorithm possesses rotational invariance. The proposed PEDoG algorithm exhibits the most robust MMI matching performance in terms of handling rotational invariance as well as matching accuracy compared to traditional methods.

5.3. Analysis of Scale-Invariance

To verify the performance of the PEDoG method in terms of scale invariance, we used a set of map–optical image pairs for matching tests. First, we processed the reference images at 0.2× image scale intervals to generate six sets of simulated images, ranging from 0.6× to 1.8×. The matching results are shown in Figure 11. The observation of the results shows that the number of NCMs obtained decreases as the scale difference increases. Among them, the 1.8-fold scale yields the least number of matching points, but it is sufficient for the alignment operation. Therefore, the PEDoG method is effective in achieving scale invariance for MMI matching.

6. Conclusions

In multi-modal images, the differences in imaging mechanisms often lead to significant variations in signal-to-noise ratios (SNR) among different modalities. This, in turn, poses a challenge for matching algorithms to maintain both matching geometric invariance and accuracy. Therefore, achieving high-precision matching correspondences while mitigating SNR interference in multi-modal remote sensing image matching is a challenge. To address this challenge, this paper proposes a lightweight multi-modal image registration algorithm that incorporates phase-indexed difference pyramids and normalized filtering. It aims to preserve geometric invariance while achieving high-precision matching correspondences. The proposed PEDoG matching method consists of three main parts.
The introduction of the phase consistency model into the difference Gaussian pyramid, resulting in the creation of new phase indices. The selection of four types of multi-modal remote sensing images (multi-temporal images, infrared–optical images, electronic navigation, and map–optical images) as experimental datasets. Comparison with state-of-the-art matching methods (SIFT, PSO-SIFT, OSS, RIFT2, and HOWP). The results indicate that the PEDoG algorithm achieves significant improvements, with an SR increase of up to 41.7% compared to the SIFT algorithm. Furthermore, its NCM is 3.3 times that of the SIFT method, 1.9 times that of the PSO-SIFT method, 1.5 times that of the OSS method, 1.1 times that of the HOWP method, and 1.2 times that of the RIFT2 method. Additionally, the RMSE achieved by PEDoG is the lowest among all matching methods, with a value of 1.69 pixels.
Finally, the registration results are demonstrated using a checkerboard mosaic pattern, showing consistent alignment of image features without artifacts or misalignment. Thus, the PEDoG algorithm proposed in this paper achieves high-precision registration while preserving geometric invariance.

Author Contributions

Conceptualization, X.Y. and Y.Y. (Yongxiang Yao); methodology, X.Y.; software, X.Y.; validation, X.Y., Y.Y. (Yongxiang Yao) and Y.C.; formal analysis, Y.C.; investigation, X.Y.; resources, X.Y.; data curation, Y.C.; writing—original draft preparation, X.Y.; writing—review and editing, X.Y.; visualization, Y.Y. (Yongxiang Yao); supervision, Y.Y. (Yijun Yangand); project administration, Y.Y. (Yijun Yangand); funding acquisition, Y.Y. (Yijun Yangand). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grant 62102268; the Stable Supporting Program for Universities of Shenzhen under grant 20220812102547001; the Research Foundation of Shenzhen Polytechnic University under grants 6022312044K, 6023310030K, 6021310008K, and 6019310010K; and the Post-doctoral Later-stage Foundation Project of Shenzhen Polytechnic University under grants 6020271005K and 6021271004K.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Cao, S.; Shen, H.; Chen, S.; Li, C. Boosting structure consistency for multispectral and multimodal image registration. IEEE Trans. Image Process. 2020, 29, 5147–5162. [Google Scholar] [CrossRef]
  2. Zhang, F.; Yang, T.; Liu, L.; Liang, B.; Bai, Y.; Li, J. Image-only real-time incremental uav image mosaic for multi-strip flight. IEEE Trans. Multimed. 2020, 23, 1410–1425. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Yao, Y.; Wan, Y.; Liu, W.; Yang, W.; Zheng, Z.; Xiao, R. Histogram of the orientation of the weighted phase descriptor for multi-modal remote sensing image matching. ISPRS J. Photogramm. Remote Sens. 2023, 196, 1–15. [Google Scholar] [CrossRef]
  4. Zhang, X.; Leng, C.; Hong, Y.; Pei, Z.; Cheng, I.; Basu, A. Multimodal remote sensing image registration methods and advancements: A survey. Remote Sens. 2021, 13, 5128. [Google Scholar] [CrossRef]
  5. Lowe, G. Sift-the scale invariant feature transform. Int. J. 2004, 2, 2. [Google Scholar]
  6. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  7. Kovesi, P. Phase congruency: A low-level image invariant. Psychol. Res. 2000, 64, 136–148. [Google Scholar] [CrossRef]
  8. Helor, Y.; Helor, H.; David, E. Matching by Tone Mapping: Photometric Invariant Template Matching. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 317–330. [Google Scholar] [CrossRef]
  9. Reddy, B.S.; Chatterji, B.N. An FFT-based technique for translation, rotation, and scale-invariant image registration. IEEE Trans. Image Process. 1996, 5, 1266–1271. [Google Scholar] [CrossRef]
  10. Sotiras, A.; Davatzikos, C.; Paragios, N. Deformable medical image registration: A survey. IEEE Trans. Med. Imaging 2013, 32, 1153–1190. [Google Scholar] [CrossRef]
  11. Xu, X.; Li, X.; Liu, X.; Shen, H.; Shi, Q. Multimodal registration of remotely sensed images based on Jeffrey’s divergence. ISPRS J. Photogramm. Remote Sens. 2016, 122, 97–115. [Google Scholar] [CrossRef]
  12. Ye, Y.; Shan, J.; Bruzzone, L.; Shen, L. Robust registration of multimodal remote sensing images based on structural similarity. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2941–2958. [Google Scholar] [CrossRef]
  13. Ye, Y.; Bruzzone, L.; Shan, J.; Shen, L. A Fast and Robust Matching Framework for Multimodal Remote Sensing Image Registration. arXiv 2018, arXiv:1808.06194. [Google Scholar]
  14. Fan, Z.; Zhang, L.; Liu, Y.; Wang, Q.; Zlatanova, S. Exploiting High Geopositioning Accuracy of SAR Data to Obtain Accurate Geometric Orientation of Optical Satellite Images. Remote Sens. 2021, 13, 3535. [Google Scholar] [CrossRef]
  15. Yao, Y.; Zhang, B.; Wan, Y.; Zhang, Y. Motif: Multi-orientation tensor index feature descriptor for sar-optical image registration. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2022, XLIII-B2-2022, 99–105. [Google Scholar] [CrossRef]
  16. Chien, H.J.; Chuang, C.C.; Chen, C.Y.; Klette, R. When to use what feature? SIFT, SURF, ORB, or A-KAZE features for monocular visual odometry. In Proceedings of the 2016 International Conference on Image and Vision Computing New Zealand (IVCNZ), Palmerston North, New Zealand, 21–22 November 2016; pp. 1–6. [Google Scholar]
  17. Ma, W.; Wen, Z.; Wu, Y.; Jiao, L.; Gong, M.; Zheng, Y.; Liu, L. Remote sensing image registration with modified SIFT and enhanced feature matching. IEEE Geosci. Remote Sens. Lett. 2016, 14, 3–7. [Google Scholar] [CrossRef]
  18. Sedaghat, A.; Mohammadi, N. Illumination-Robust Remote Sensing Image Matching Based on Oriented Self-Similarity. ISPRS J. Photogramm. Rem. Sens. 2019, 153, 21–35. [Google Scholar] [CrossRef]
  19. Xiong, X.; Jin, G.; Xu, Q.; Zhang, H. Robust SAR image registration using rank-based ratio self-similarity. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2358–2368. [Google Scholar] [CrossRef]
  20. Xiang, Y.; Tao, R.; Wang, F.; You, H.; Han, B. Automatic Registration of Optical and SAR Images Via Improved Phase Congruency Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5847–5861. [Google Scholar] [CrossRef]
  21. Fan, J.; Ye, Y.; Li, J.; Liu, G.; Li, Y. A Novel Multiscale Adaptive Binning Phase Congruency Feature for SAR and Optical Image Registration. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  22. Yao, Y.; Zhang, Y.; Wan, Y. Heterologous Images Matching Considering Anisotropic Weighted Moment and Absolute Phase Orientation. Geomat. Inf. Sci. Wuhan Univ. 2021, 46, 1727–1736. [Google Scholar]
  23. Yang, W.; Xu, C.; Mei, L.; Yao, Y.; Liu, C. LPSO: Multi-source image matching considering the description of local phase sharpness orientation. IEEE Photonics J. 2022, 14, 7811109. [Google Scholar] [CrossRef]
  24. Yao, Y.; Zhang, Y.; Wan, Y.; Liu, X.; Yan, X.; Li, J. Multi-modal Remote Sensing Image Matching Considering Co-occurrence Filter. IEEE Trans. Image Process. 2022, 31, 2584–2597. [Google Scholar] [CrossRef] [PubMed]
  25. Gao, C.; Li, W.; Tao, R.; Du, Q. MS-HLMO: Multi-scale Histogram of Local Main Orientation for Remote Sensing Image Registration. arXiv 2022, arXiv:2204.00260. [Google Scholar] [CrossRef]
  26. Li, J.; Xu, W.; Shi, P.; Zhang, Y.; Hu, Q. LNIFT: Locally Normalized Image for Rotation Invariant Multimodal Feature Matching. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5621314. [Google Scholar] [CrossRef]
  27. Yi, K.M.; Trulls, E.; Lepetit, V.; Fua, P. Lift: Learned Invariant Feature Transform. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 467–483. [Google Scholar]
  28. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
  29. Dusmanu, M.; Rocco, I.; Pajdla, T.; Pollefeys, M.; Sivic, J.; Torii, A.; Sattler, T. D2-net: A trainable cnn for joint description and detection of local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8092–8101. [Google Scholar]
  30. Sun, J.; Shen, Z.; Wang, Y.; Bao, H.; Zhou, X. LoFTR: Detector-free local feature matching with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8922–8931. [Google Scholar]
  31. Bökman, G.; Kahl, F. A case for using rotation invariant features in state of the art feature matchers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5110–5119. [Google Scholar]
  32. Zhang, Y.; Liu, Y.; Zhang, H.; Ma, G. Multimodal Remote Sensing Image Matching Combining Learning Features and Delaunay Triangulation. IEEE Trans. Geosci. Remote. Sens. 2022, 60, 5635517. [Google Scholar] [CrossRef]
  33. Muja, M.; Lowe, D.G. Fast approximate nearest neighbors with automatic algorithm configuration. VISAPP 2009, 2, 2. [Google Scholar]
  34. Wu, Y.; Ma, W.; Gong, M.; Su, L.; Jiao, L. A novel point-matching algorithm based on fast sample consensus for image registration. IEEE Geosci. Remote Sens. Lett. 2014, 12, 43–47. [Google Scholar] [CrossRef]
  35. Kovesi, P. Image Features FromPhase Congruency. Videre J. Comput. Vis. Res. 1999, 1, 1–26. [Google Scholar]
  36. Schraudolph, N.N. A fast, compact approximation of the exponential function. Neural Comput. 1999, 11, 853–862. [Google Scholar] [CrossRef]
  37. Gao, W.; Zhang, X.; Yang, L.; Liu, H. An improved Sobel edge detection. In Proceedings of the 2010 3rd International Conference on Computer Science and Information Technology, Chengdu, China, 9–11 July 2010; Volume 5, pp. 67–71. [Google Scholar]
  38. Mikolajczyk, K.; Schmid, C. A performance evaluation of local descriptors. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1615–1630. [Google Scholar] [CrossRef] [PubMed]
  39. Li, J.; Hu, Q.; Ai, M. RIFT: Multi-modal image matching based on radiation-variation insensitive feature transform. IEEE Trans. Image Process. 2019, 29, 3296–3310. [Google Scholar] [CrossRef] [PubMed]
  40. Xiong, X.; Jin, G.; Xu, Q.; Zhang, H. Self-similarity features for multimodal remote sensing image matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12440–12454. [Google Scholar] [CrossRef]
Figure 1. Overall flow chart.
Figure 1. Overall flow chart.
Remotesensing 15 05764 g001
Figure 2. Phase-exponential Gaussian differential scale space.
Figure 2. Phase-exponential Gaussian differential scale space.
Remotesensing 15 05764 g002
Figure 3. Schematic distribution of feature points.
Figure 3. Schematic distribution of feature points.
Remotesensing 15 05764 g003
Figure 4. Normalized log-polar coordinate descriptor flowchart. (a) is the neighborhood window, (b) is the result of the normalization process, (c) is the feature principal direction, (d) is the description subframe, and (e) is the feature vector value.
Figure 4. Normalized log-polar coordinate descriptor flowchart. (a) is the neighborhood window, (b) is the result of the normalization process, (c) is the feature principal direction, (d) is the description subframe, and (e) is the feature vector value.
Remotesensing 15 05764 g004
Figure 5. Partial multi-modal image data. (a) optical–optical; (b) infrared–optical; (c) map–optical.
Figure 5. Partial multi-modal image data. (a) optical–optical; (b) infrared–optical; (c) map–optical.
Remotesensing 15 05764 g005
Figure 6. Test results of the six methods. (a) NCM results for six methods, (b) RMSE results for six methods.
Figure 6. Test results of the six methods. (a) NCM results for six methods, (b) RMSE results for six methods.
Remotesensing 15 05764 g006aRemotesensing 15 05764 g006b
Figure 7. Qualitative matching results for six matching algorithms. (a) SIFT; (b) PSO-SIFT; (c) OSS; (d) HOWP; (e) RIFT2; (f) PEDoG.
Figure 7. Qualitative matching results for six matching algorithms. (a) SIFT; (b) PSO-SIFT; (c) OSS; (d) HOWP; (e) RIFT2; (f) PEDoG.
Remotesensing 15 05764 g007aRemotesensing 15 05764 g007b
Figure 8. Registration results for selected images. (a) multi-temporal–optical; (b) infrared–optical; (c) map–optical.
Figure 8. Registration results for selected images. (a) multi-temporal–optical; (b) infrared–optical; (c) map–optical.
Remotesensing 15 05764 g008aRemotesensing 15 05764 g008b
Figure 9. Parameter analysis results. (a) Results for parameter Dl; (b) Results for parameter Ct.
Figure 9. Parameter analysis results. (a) Results for parameter Dl; (b) Results for parameter Ct.
Remotesensing 15 05764 g009
Figure 10. Rotation invariant matching results.
Figure 10. Rotation invariant matching results.
Remotesensing 15 05764 g010
Figure 11. Scale invariant matching results.
Figure 11. Scale invariant matching results.
Remotesensing 15 05764 g011
Table 1. Findings of the six methodologies for the three evaluation indicators.
Table 1. Findings of the six methodologies for the three evaluation indicators.
SIFTPSO-SIFTOSSHOWPRIFT2PEDoG
SR58.3%88.3%91.7%66.7%96.7%100%
NCM79.53139.95175.6236.92209.38259.43
RMSE4.082.362.083.632.071.69
Table 2. Recommended parameter settings for the PEDoG algorithm.
Table 2. Recommended parameter settings for the PEDoG algorithm.
ParameterVariable ValuesFixed Parameters
DlDl = [2, 3, 4, 5, 6, 7, 8]Ct = 0.3
CtCt = [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]Dl =4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, X.; Cao, Y.; Yang, Y.; Yao, Y. Multi-Modal Image Registration Based on Phase Exponent Differences of the Gaussian Pyramid. Remote Sens. 2023, 15, 5764. https://doi.org/10.3390/rs15245764

AMA Style

Yan X, Cao Y, Yang Y, Yao Y. Multi-Modal Image Registration Based on Phase Exponent Differences of the Gaussian Pyramid. Remote Sensing. 2023; 15(24):5764. https://doi.org/10.3390/rs15245764

Chicago/Turabian Style

Yan, Xiaohu, Yihang Cao, Yijun Yang, and Yongxiang Yao. 2023. "Multi-Modal Image Registration Based on Phase Exponent Differences of the Gaussian Pyramid" Remote Sensing 15, no. 24: 5764. https://doi.org/10.3390/rs15245764

APA Style

Yan, X., Cao, Y., Yang, Y., & Yao, Y. (2023). Multi-Modal Image Registration Based on Phase Exponent Differences of the Gaussian Pyramid. Remote Sensing, 15(24), 5764. https://doi.org/10.3390/rs15245764

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop