1. Introduction
ISAR has been used for space target monitoring for many years, which can provide useful information about space targets such as shape structure, body attitude, and working mode, etc. [
1,
2,
3,
4]. To improve ISAR image quality and exploit more complete information, image fusion [
5] and the three-dimensional (3D) target reconstruction technique [
6,
7,
8] have been extensively investigated, which usually involve jointly processing radar data of the same target taken at different times from different viewpoints and/or by different sensors. Image registration is a crucial step in integrating or comparing different measurements in these applications. For example, different images need to be rotated to a common coordinate system prior to high resolution image formation in the image fusion; the same scatterers from different images need to be correctly registered to compare position differences or phase differences in 3D reconstruction. According to the imaging mechanism of radar systems, an ideal ISAR image can be represented by a two-dimensional (2D) point cloud, which is significantly different from an optical image [
9]. Low SNR, sparse scatterer distribution, scatterer defocusing, scatterer glinting, and secondary reflections of targets in ISAR images impact the performance of image registration algorithms. Therefore, image registration methods should be reevaluated and modified according to the characteristics of ISAR images, even if they work well in matching synthetic aperture radar (SAR) images [
10], optical images [
11], and medical images [
12].
ISAR image registration algorithms based on joint translational compensation have been proposed to resolve the translation mismatch problem in the signal domain [
13,
14,
15,
16,
17,
18]. These algorithms utilize the relative rotation of the line of sight (LOS) to compensate for image shifts. In [
13], the relationship between time-varying rotation parameters and the compensation phase is derived from the observation geometry in a multiple antenna configuration, such as an L-shaped structure. Several studies have focused on image registration algorithms for non-cooperative space targets, where the relative rotation of target is unknown. In [
13,
14,
15], the rotation trajectory is estimated by tracking dominant scatterers in one-dimensional range profiles along the cross direction. To solve the problem of cross-term phase interference, the authors in [
16] proposed measuring the phases of joint-channel phase difference profiles generated by raw echo data from different receiving antennas. Reference [
17] combined joint range alignment and joint Doppler centroid tracking to register multichannel ISAR images. In [
18], a joint wave path difference compensation algorithm was developed to enable precise image registration for InISAR imaging with SFB-SA signals.
Registration techniques processed in the image domain have also been developed, which can avoid the impact of fluctuating quality in range profiles, especially in low SNR scenarios or when there are few dominant scatterers in the range profiles. Image alignment was achieved using cross-correlation techniques by iteratively searching for the translation and rotation parameters [
19]. However, the drawback of this method is its high computational complexity. Since cross-correlation can be calculated in the frequency domain based on the Fourier theorem, computational efficiency can be improved by using the fast Fourier transform (FFT). The authors of [
20] utilized phase correlation in the frequency domain to estimate the relative translation offset between two ISAR images. For image rotation parameter estimation, [
21] decomposed the image rotation using a three-step FFT-based convolution interpolation process to improve computational efficiency, and [
20] applied the polar mapping approach, which turns the rotation of an image into a translation. Since this type of method uses the intensity information of the image, the content of the image has a significant influence on the registration result. In [
22], a registration method was proposed by tracking and matching scatterers based on the Kanade–Lucas–Tomasi (KLT) algorithm. The KLT algorithm makes use of spatial intensity information to directly search for the position that yields the best match. The rapid advancement of deep learning has led to significant progress in SAR image registration [
23,
24]. Deep learning’s feature extraction and matching capabilities have greatly improved image registration accuracy and efficiency. However, there has been limited research on the application of deep learning methods for ISAR image registration, which requires further investigation.
Feature-based methods are powerful tools in remote sensing, medical imaging, and computer vision applications. By focusing on distinctive features within images, these methods can effectively align images despite changes in viewpoint or partial overlaps. Researchers have explored feature-based registration techniques in the field of ISAR image registration [
25,
26,
27,
28,
29,
30,
31]. For instance, line features [
25] and rectangular features [
26] are applied for the image registration of continuously changing ISAR image sequences. The extraction of these advanced features primarily relies on the target’s contours. For a discrete ISAR image, image preprocessing techniques, such as morphological operations, are necessary to produce continuous contours. Reference [
27] proposed an image registration method by integrating SIFT and a local correlation matching method. In [
28], a combination of scale-invariant feature transform (SIFT) and speeded up robust features (SURF) was employed to align ISAR image sequences and achieve large-angle imaging of ISAR targets. While effective, SIFT and SURF are optimized for images with distinct, well-defined features such as corners, edges, and blobs, but they are sensitive to image noise [
32,
33]. A well-focused ISAR image can be taken as a composition of many 2Ds sinc functions, with slight blur, located at the projected positions of the 3D scatterers in the ISAR imaging plane [
1]. The lack of rich textures in many ISAR images poses challenges, necessitating alternative feature detection methods to achieve optimal results.
Sparsely distributed scatterers resemble point clouds more than color patches. Thus, using scatterers rather than other features to describe an ISAR image is more appropriate. Additionally, using structural information instead of boundary or gradient information in ISAR image registration is more reasonable.
In this work, we propose utilizing the center points of the sinc envelope as feature points (FPs) and constructing a spatial-based feature descriptor by employing the characteristics of distance, angle, and SNR of scatterers. This descriptor exploits essential information from ISAR images and is expected to provide a more stable performance in parameter estimation for image registration.
The innovations of this article are listed as follows:
- (1)
A spatial feature of dominant scatterers is proposed to describe the spatial information of ISAR images, offering a more fundamental approach compared to boundary-based or gradient-based features in ISAR image registration.
- (2)
The proposed method projects the two-dimensional spatial structure of the target into a one-dimensional feature vector, translating image rotation into a shift of the feature vector, which can be easily matched through circular cross-correlation.
- (3)
The traditional random sample consensus (RANSAC) method was enhanced with a consistency check for the rough rotation angle, allowing closer integration with the proposed features and thereby improving the accuracy and stability of mapping parameter estimation, even when the percentage of outliers approaches 50%.
The remainder of this paper is organized as follows:
Section 2 analyzes the transform model of ISAR images from different viewpoints.
Section 3 introduces the proposed ISAR image registration method, including spatial feature model construction, feature matching, and transform model estimation.
Section 4 presents simulated and experimental results to validate the efficacy of the proposed method. Finally,
Section 5 provides the discussion, and
Section 6 presents the conclusions.
2. Transform Model of ISAR Images
The conventional ISAR imaging methodology generates two-dimensional representations of targets by projecting the reflectivity function onto an imaging projection plane (IPP) [
34]. By analyzing the relationship between imaging coordinate systems from two distinct viewpoints, an ISAR image registration model can be derived. This section will analyze the transformation model for registering two ISAR images obtained from different arcs by a ground-based radar imaging system.
Figure 1 illustrates the imaging geometry of a spatial target. An ISAR image sequence is generated for the entire visible orbit. Without a loss of generality, we assume that
and
represent the imaging moments for the
n-th and
n+1-th images in the sequence, with
being the time interval between these moments. Let
and
denoting the imaging coordinate systems corresponding to these moments.
and
are the range axes aligned with the imaging LOS.
and
are the Doppler axes perpendicular to the range axis and the effective rotation vector.
and
are the normal vectors of the IPP, consistent with the direction of the effective rotation vector.
During target observation, LOS rotates with the target’s movement, while the target adjusts its attitude to maintain three-axis stability or accomplish specific missions. This results in relative rotation between the radar and the target. In the imaging model described in [
9], a relative LOS vector is defined within the target’s body coordinate system, assuming that the target’s body coordinate system aligns with the instantaneous imaging system
. The imaging LOS vector
can be expressed as
where
is the azimuth angle of the LOS,
is the elevation angle of the LOS, and the superscript T represents the transpose of the matrix.
By differentiating and normalizing LOS vector, the Doppler vector is derived, as shown in Equation (2). This allows for the determination of the instantaneous IPP at time
. The normal vector of the IPP is obtained by taking the cross-product of the range axis and the Doppler axis, as shown in Equation (3).
The transformation matrix between the imaging coordinate systems at these two different time instances is
where
and
are the derivatives of
and
, respectively.
Suppose P is an arbitrary scatterer on the target. Its coordinates in the imaging coordinate system are and in the imaging coordinate system . Based on the analysis above, the coordinate transformation relationship for the scatterer P in these two imaging systems is .
In ISAR imaging, target motion compensation techniques, such as envelope alignment and phase focusing, are employed to compensate for target motion and enhance image quality. However, the self-focusing procedure in translational motion compensation causes misalignment of ISAR images. We denote the induced offsets in ISAR processing as
, where
and
represent the shifts in the cross-range and range dimensions of the image, respectively. The registration model can be represented as follows:
From Equation (5), it can be noted that the registration model is related with the and . is always the primary rotational component of LOS, providing the foundation for achieving high resolution in the azimuth direction. In contrast, represents the spatial variation of the IPP, resulting in a height-dependent displacement.
Let
denote the target’s maximum height. The cross-range displacement
satisfies
, and the range displacement
satisfies
. According to [
35], when the coordinate error is less than one-eighth of a resolution unit, its impact on image correlation and the interferometric phase can be neglected. Thus, the constraint can be derived as
, where
and
are the range and cross-range resolution, respectively. Assuming
is 10 m, with a bandwidth of 1 GHz and a resolution of 0.15 m, the permissible angular deviation is
.
For slow-rotating space targets, changes in IPP occur more gradually and do not shift rapidly over short periods. By limiting the interval between arc segments, the variation in IPP can be ignored.
Section 4.1 provides a detailed analysis of
and
. Both
and
are influenced by various factors, such as the satellite’s orbital parameters, the radar station’s position, and the satellite’s attitude. Notably,
is much larger than
, particularly when the satellite is passing through the zenith. When the interval between two arcs is less than 10 s, the influence of
and
on image registration can be ignored.
Based on the previous analysis, it can be assumed that
,
, and
. Therefore, the mathematical registration model between two ISAR images from adjacent imaging arcs can be simplified as a rigid transformation that includes both translation and rotation. This can be expressed as
Through the registration of consecutive images, sequential image registration is ultimately employed to achieve long interval image registration.
3. Method
The image registration of ISAR images is the process of overlaying different ISAR images obtained from different viewpoints [
11]. The principle of image registration involves extracting and matching representative features of ISAR images to establish mapping parameters between these images. In this section, a feature-based image registration method is proposed by exploiting the concordance of spatial structures of multiple scatterers in various ISAR images.
3.1. Feature Detection
Unlike optical images, ISAR images are not composed of color blocks. Instead, they are a set of slightly blurred 2D sinc functions due to system errors and imperfections in ISAR imaging processing. Assuming that the radar transmits a broadband signal with bandwidth
and center frequency
,
is the coherent processing time under the
n-th viewing. After imaging processing, the obtained ISAR image can be expressed as:
where
is the fast time,
is the Doppler frequency,
represents the number of scatterers on the target,
is the scattering coefficient,
is the velocity of light, and
and
denote the 2D location of scatterer
Pi in the ISAR image, while
is its 3D coordinate in the body coordinate system. For in-orbit space targets, the cumulative rotation angle can be estimated using high-precision orbital measurements and attitude information. This ensures that two adjacent ISAR images have similar resolution. After cross-range scaling, the images are further adjusted to the same size through scaling and cropping.
Traditional salient structures and features, such as prominent regions, lines, region corners, line intersections, and points on curves with high curvature, used for optical image registration, are based on variations in color regions of images. In ISAR imaging, many edges form between dominant scatterers and the background. However, these edges are unstable due to noise, defocusing, and sidelobe interference from adjacent scatterers. Consequently, the performance of methods that rely on edge inflection points or corners as FPs degrades for ISAR image registration. This instability worsens when scatterers are sparsely distributed within well-focused ISAR images. Therefore, it is more appropriate to use the peak information of scatterers rather than their boundary or gradient information for ISAR image registration. In addition to amplitude and phase, the spatial structure of dominant scatterers is a key identifying feature in ISAR images. This paper extracts scatterers with high SNR as FPs and analyzes their relative distribution to construct a feature descriptor.
3.2. Spatial Feature Descriptor
A feature descriptor is a vector representation of an image feature, encapsulating essential information about the feature’s characteristics in a format useful for further processing or analysis. When a reference scatterer is selected for each set of FPs, each FP relative to this reference FP can be modeled as a distance vector. Considering counterclockwise as the positive direction, the angle-distance distribution of these distance vectors can describe the spatial relationships of these FPs.
Figure 2 illustrates a schematic diagram of the reference and sensed images used in image registration. The sets of FPs in the reference and sensed images are denoted as
and
, where
and
correspond to the number of dominant scatterers extracted from each image.
Let FP
be the reference. We can denote the vector
as the distance vectors of FPs relative to the reference
in the set.
where
denotes the modulus of the vector
and
denotes the angle.
FP
and FP
are the projections of the same scatterer on the IPPs of the reference and sensed images, respectively. Similarly, the distance vectors of the FPs set relative to the reference
in the sensed image are
where
denotes the modulus of the vector
, and
denotes the angle.
The indices of the same scatterer in the two sets can differ significantly, and even the total number of FPs may not be the same. Elements with the same index in
and
may be different, namely
. Therefore, computing the conjugate cross-correlation of
and
may lead to failure in feature matching. To resolve this problem, the distance vectors of the FPs should be sorted and padded before feature matching, and then we have
where
is the sorting function that rearranges the distance vectors of FPs in increasing order according to the vector angle,
or
.
is the process of scatterer superposition and zero-padding in the generation of
and
. After this process, scatterers with the same angle will have the same index in
and
, enabling the use of cross-correlations of
and
to describe their similarity. In practical applications, the correlation of such discrete features is highly sensitive to angular errors. Therefore, an improved continuous feature is considered.
Noise affects the estimation accuracy of scatterer coordinates, which subsequently impacts the angular precision of the corresponding distance vectors of FPs. According to [
36], the coordinate errors in range and cross-range are independent Gaussian random variables. Under the small error approximation, the angular error is a linear combination of the coordinate errors and can be approximated as Gaussian noise. Therefore, the angle value in the feature descriptor can be replaced with a parameterized Gaussian function to achieve an angular-error-tolerant feature. Finally, the continuous feature descriptor of the scatterer
is expressed as the accumulation of N-1 Gaussian functions, which is
where the range of variable
is from −
to
,
is the mean of the Gaussian function, and
is the standard deviation, which is calculated based on the SNR and the modulus of the distance vectors of FPs. According to [
36], the 2D coordinate errors between two scatterers in the ISAR image are expressed as
where
denote the cross-range error and range error, respectively, and
and
represent the peak SNR of scatterer
and
.
Figure 3 shows the schematic diagram of angle error
. The standard deviation of
can be derived as
We present an example that utilizes this new feature descriptor. We assume that each line segment in
Figure 2 is 5 m. The lengths and angles of each vector from
to other points are listed in
Table 1. Given that the SNR of each scatterer is 10 dB and the resolutions,
and
, are 1 m, we can calculate the variance
of the angle
using Equations (13) and (14). The feature of scatterer
is then obtained using Equation (12).
Figure 4 illustrates the feature curves of scatterer
along with its 10 components. It is worth noting that the modulus of the distance vector and the amplitude of the FPs are incorporated into the feature descriptor through
. As shown, a longer modulus corresponds to a taller and narrower Gaussian waveform, while a shorter modulus results in a shorter and wider Gaussian waveform.
In
Figure 5, the feature curve of
is compared with those of
and
. It can be observed that the
curve and the curve
are quite different, whereas the
curve and the
curve show a stronger correlation after an angle shift, despite some outlier pairings. The proposed method projects the two-dimensional structure into a one-dimensional vector and maps the image rotation into a translation of the one-dimensional feature vector.
3.3. Feature Matching
After obtaining the feature descriptors of the FPs, the next step is feature matching. Image rotation causes a shift in the feature curves. Therefore, this paper employs the circular cross-correlation method for feature matching [
27,
37]. The circular cross-correlation of scatterer
and scatterer
is
Performing circular cross-correlation on all the FPs yields a cross-correlation matrix
.
where
. In this circular cross-correlation matrix, the row number represents the index of the scatterer in the reference image, and the column number represents the index of the scatterer in the sensed image. Furthermore, the circular cross-correlation can be obtained via the FFT, which greatly improves computational efficiency.
For scatterer index
in the reference image, its matching scatterer index
in the sensed image can be obtained by searching for the maximum value in the
-th row. Let
stand for the argument of the maxima, it can be expressed as
Feature matching is typically achieved by searching for the maximum value in each row. However, false matches are inevitable in image feature matching due to factors such as symmetrical structures, repetitive patterns, and noise. Additionally, viewpoint changes and occlusion contribute to these errors. Estimating registration parameters accurately is challenging in the presence of outliers because they introduce noise and bias, leading to incorrect estimates. To ensure precise registration, it is crucial to identify and remove outliers before estimating the registration parameters.
3.4. Two-Step Transform Model Estimation Method
In this subsection, we provide a two-step transform model estimation process, which includes outlier detection, removal, and parameter estimation.
The first step uses a statistical method for outlier filtering. Since the angular shift that corresponds to the maximum circular cross-correlation is close to the image’s rotation angle, a histogram analysis is applied to the angle shifts of all matching pairs. This histogram-based method automatically identifies incorrect matches, such as mirror mismatches caused by symmetrical structures. The angle intervals can be set based on a rough estimate of the image rotation. By adjusting the histogram’s central angle, most correct matches can be grouped within one interval, while matches outside this range are considered outliers. However, this method is less effective at excluding redundant adjacent scatterer pairs with similar rotation angles.
In the second step, a model-based approach identifies adjacent scatterer pairs using the RANSAC algorithm. The RANSAC algorithm fits a model to the data and identifies redundant outliers based on their deviation from the model. The parameters of the transformation model are estimated from the residual inliers.
RANSAC is an iterative method used to estimate parameters of a mathematical model from a set of observed data containing outliers [
38]. According to the RANSAC algorithm, a subset of matched scatterer pairs is randomly selected, and the related parameters are estimated by solving the transform model equation. This allows redundant matches and mismatches to be excluded through consistency checks. It iterates through two steps: generating a hypothesis from random samples and verifying it against the data [
39]. The most probable hypothesis, supported by the most inlier candidates, is chosen during the iterations. The RANSAC algorithm can be formulated by
where
,
,
are the registration parameters as defined in
Section 2,
,
,
are their estimated values,
and
are scatterer pairs for matching the reference image and sensed image, respectively,
represents the number of matched scatterer pairs after the initial outlier exclusion,
is the arguments of the minima,
is the error function, and
is the loss function. The explicit definitions of these functions and parameter estimation procedure are provided below.
Assuming
L scatterers are randomly selected,
L must be at least three due to the need to estimate three parameters. The choice of
L should balance model accuracy and algorithm efficiency. By substituting the scatterers’ coordinates into the transformation model in Equation (6), it can be derived
where
is the column vector related to parameters to be solved,
is a Matrix related to scatterer’s coordinates, and the 2
l-1th row and 2
lth row are
where
and
denote the scatterer’s coordinates in the
n-th ISAR image, while
and
denote the coordinates of the corresponding scatterer in the
n+1-th ISAR image.
Through SVD decomposition of matrix
, we obtain
. The singular vector
corresponding to the smallest singular value is the non-zero solution that minimizes
[
40]. Thus,
. The mapping function parameters between different ISAR images can be obtained as follows:
The coordinates of scatterers in the sensed image can be recalculated based on the estimation results of this randomly selected subset of matched scatterer pairs. Let
represent the estimated transformation matrix, which is shown in Equation (6), then the new coordinates of scatterer
can be obtained as
. The Euclidean distance between scatterers is used as the error function
, and the distance error
is expressed as
where
and
denote the coordinates of scatterer
, and
and
denote the coordinates of scatterer
.
The subset of matched scatterer pairs is considered an inlier candidate, when its error
is within a predefined threshold, formulated as
This threshold, , depends on the image resolution and SNR and can be set to or . is the standard deviation of the distance between the same scatterers in the two images when it is ideally registered, and it can be calculated by Equation (13).
Once the outliers are identified, they can be eliminated from the dataset before estimating the registration parameters. The remaining scatterers can be taken as inliers and used to estimate the registration parameters using least squares estimation (LSE).
3.5. Flowchart of the ProposedMethod
The entire procedure of the proposed ISAR image registration method can be summarized as feature detection, feature matching, transform model estimation, and image transformation. The flowchart is shown in
Figure 6.
The specific processing steps are as follows.
Step 1: Analyze the noise power in the noise region to set the threshold for scatterer extraction and extract dominant scatterers as FPs in each image using the CLEAN technology.
Step 2: Calculate the distance vectors between any two scatterers within the image and construct the feature descriptor for each FP based on shape structures.
Step 3: Calculate the circular cross-correlation matrix and the angle shift of all matching scatterer pairs via FFT and then obtain the feature matching results.
Step 4: Discard false matched pairs by a consistency check of the estimated rotation angle.
Step 5: Discard redundant adjacently outliers and estimate the parameters of transform model by RANSAC.
Step 6: Image registration based on the estimated parameters.
4. Experiments and Results
In this section, several experiments based on both simulated and measured ISAR images are conducted to validate the effectiveness of the proposed method. Additionally, the well-known feature registration algorithms, SIFT, SURF and SIFT+SURF, are compared with the proposed method. The real ISAR images used in this section are sourced from a report published by the German FGAN Lab in 2012 (@Fraunhofer FHR) [
21]. This report provides an ISAR image sequence of the US space shuttle, obtained with the TIRA system after long arc imaging.
4.1. Azimuth and Elevation Angles of LOS
In this subsection, the azimuth and elevation angles for different visible passes, imaging times, and time intervals are simulated. In this simulation, the radar station is set in Beijing (39.9° N, 116.4° E, 88 m), and the satellite orbit is the TIANHE space station. The two-line orbital elements (TLE) of the TIANHE space station are listed in
Table 2. The observation time for the first visible pass is from 2:22:07 to 2:32:31 on 1 March 2023, and for the second visible pass, from 3:59:00 to 4:09:26 on 1 March 2023.
Figure 7a,b illustrate the variations in azimuth and elevation angles, while
Figure 7c,d depict the derivation of these angles. According to the definition in [
9], the center moment of coherent processing interval is defined as the imaging moment. In
Figure 7, the time values on the horizontal axis represent the imaging moment of the reference image, while the imaging moment of the sensed image is offset by a time interval
with respect to the reference moment. During the observation, the LOS form a curved surface in the target’s body coordinate system, which expands more in the horizontal plane than the vertical change. As a result, the azimuth angle variation is significantly larger than the elevation angle variation, particularly near the zenith-passing time. When the arc segment interval is less than 10 s, with alpha less than 10° and beta less than 0.05°, and a resolution of 0.15 m, the displacement for a target at a height of 10 m is 0.0086 m. This displacement, being less than one-tenth of a resolution unit, is negligible.
4.2. Experiment with Simulated Images
This section presents the experimental results based on simulated ISAR images. The scatterer model contains 80 dominant scatterers, extracted from a real ISAR image of the US space shuttle given in
Figure 8a. It is assumed that there is a 20% outlier rate between the reference image and the sensed image to simulate phenomena such as angle glint and occlusion effects in ISAR imaging. So, scatterers are randomly divided into three groups according to the number of outliers: set1 (54 scatterers), set2 (13 scatterers), and set3 (13 scatterers), as shown in
Figure 8a. The scatterers in set1 and set2 are used to generate the reference image shown in
Figure 8b, and the scatterers in set1 and set3 are used to generate the sensed image shown in
Figure 8c. The sensed image has been affine-transformed according to the parameters listed in
Table 3.
The proposed image registration algorithm is validated using the sensed and reference images from
Figure 8. Complex Gaussian noise is added to each image to simulate a noise environment with a 24 dB SNR. After extracting FPs, the feature descriptor of each point is constructed according to Equation (12), and the circular cross-correlation of the feature descriptors between any two reference FPs is calculated.
Figure 9a shows the maximum circular cross-correlation matrix, where the brightness of each cell indicates the cross-correlation of two feature curves, implying the reliability of the matching. It is observed that a few bright points are sparsely distributed in the maximum circular cross-correlation matrix, which can be taken as candidates for matched scatterer pairs. Finally, scatterer pairs are obtained by searching for the maximum value row by row, and these are taken as the coarse matches before the consistency check. The histogram of shift feature curves for matched pairs is shown in
Figure 9b, with an interval of 20° between adjacent bins. It is observed that a tall column in the histogram is contributed by the correct matches and adjacent matches, while the scattered small values correspond to the false matches.
Figure 10 shows the feature matching results: red lines mark the mismatched pairs identified from the histogram, green lines mark the redundant adjacent matches identified during the RANSAC process, and yellow lines mark the correctly matched pairs used to estimate the mapping parameters. As per outlier detection, there are 8 false matches, 5 redundant adjacent matches and 54 correct matches.
Figure 11a depicts the superposition of the reference image and the sensed image before registration, which shows an obvious displacement.
Figure 11b depicts the superposition after registration, where the two images are overlapped perfectly. Furthermore, the estimation results listed in
Table 3 are compared with those of classical image registration methods, such as the SIFT algorithm, the SURF algorithm and the SIFT+SURF algorithm. To ensure the objectivity and fairness of the various experiments, all four methods used the RANSAC algorithm during the parameter estimation phase, randomly selecting four scatterer pairs per iteration with a total of 10,000 iterations. It can be seen that the estimation error of the proposed method is much smaller.
4.3. Experiment with Real Images
In this subsection, two different real ISAR images of a US Space Shuttle are used to test the effectiveness of the proposed method. The reference and sensed images are shown in
Figure 12a,b, respectively. We sequentially extracted 200 dominate scatterers according to their amplitude from each ISAR image.
The matched pairs are shown in
Figure 13. There are 65 pairs of false matches marked with red lines, 112 pairs of redundant adjacently matches marked with green lines, and 23 pairs of best matches marked with yellow lines.
The sensed image is transformed based on the estimated parameters, and the ISAR images before and after registration are compared in
Figure 14.
Figure 14a shows the superposition of the two ISAR images before registration, and
Figure 14b shows the registration results. It can be observed that the main body is perfectly overlapped, although the tails caused by secondary reflection are displaced. Finally, the proposed method is compared with the SIFT, SURF, and SIFT + SURF methods. The estimation errors and the normalized image correlation of the two images after registration are listed in
Table 4. The filtered image, with its tail removed, ensures that the results of image correlation are not affected by the secondary reflections in ISAR images.
4.4. Robustness Analysis
To further evaluate the robustness of the proposed image registration algorithm, we conducted 1000 Monte Carlo simulations under varying SNR levels and different outlier ratios. In the robustness versus SNR experiments, noise was added to the original ISAR image, and the SNR was adjusted from 10 dB to 30 dB in 2 dB increments using the scatterer model and mapping parameters described in
Section 4.2. The mapping parameter estimation results and the image registration success rates for the three methods were compared.
Successful image registration is evaluated based on a preset estimation error threshold. The threshold for range and cross-range shifts is set to half a resolution unit, while the estimation error for the rotation angle is set at 1.6°, ensuring that the shift of the farthest scatterer remains within half a resolution unit.
Figure 15 shows the success probabilities of the four image registration methods at different SNR levels. The results indicate that the proposed method achieves the highest success rate under low SNR conditions, whereas the SIFT method has the lowest success rate. The mean error (ME) and root mean squared error (RMSE) for range shift, cross-range shift, and rotation are depicted in
Figure 16. Although the success rates of the SURF and SIFT + SURF methods are comparable to the proposed method when the SNR exceeds 18 dB,
Figure 16 shows that the proposed method provides higher accuracy in parameter estimation, as indicated by the lower mean and variance of the registration parameters. The presence of noise reduces the number of detectable scatterers, introduces more outliers, and increases the localization error, consequently lowering both registration accuracy and method stability.
Similarly, the impact of outliers is analyzed using 1000 Monte Carlo simulations at 24 dB SNR. The ratio of outliers is varied from 0% to 50% in steps of 5%.
Figure 17 illustrates the success probabilities of the four registration methods at different outlier ratios. It can be noted that the success rate of the SIFT+SURF method is lower than that of the SURF method when the proportion of outliers is high. Although the combination of SIFT and SURF increases the number of FPs, the mismatch pairs introduced by SIFT make it more difficult for RANSAC to find inliers when the same number of iterations is used, resulting in reduced registration accuracy.
Figure 18 depicts the estimation ME and RMSE of the transformation parameters. It can be observed that the proposed method maintains its performance even as the outlier ratio increases to 50%.
4.5. Computational Complexity Analysis
In this subsection, the computational complexity of the proposed method is analyzed. It is evaluated for four core stages: feature point extraction, feature descriptor construction, feature matching, and model parameter estimation. Assuming the image size is , containing scatterers, the computational complexity of each stage is denoted as , , and respectively.
- (1)
At the feature point extraction stage, the sinc-based CLEAN method is used to extract scatterers. Assuming
is the number of pixels in the segmented area, the computational complexity is approximately
[
41].
- (2)
At the feature descriptor construction stage, the computation is primarily focused on calculating the vector angles and magnitudes as well as the summation of Gaussian functions. Assuming the length of the Gaussian function is , the computational complexity is approximately .
- (3)
Feature matching is implemented using the circular cross-correlation method. Assuming the feature vector length is , the computational complexity using the conventional time-domain correlation is approximately . By applying FFT for frequency-domain processing, the complexity can be reduced to approximately .
- (4)
Model parameter estimation is performed iteratively. Each iteration involves singular value decomposition of the observation matrix, transformation of feature point coordinates, and distance loss calculation. Assuming the observation matrix size is and the number of scatterers after consistency checking is , the computational complexity for iteration is approximately .
From the above analysis, it can be concluded that the total computational complexity is approximately .
The computational complexities of different methods are compared through simulations.
Table 5 lists the relevant parameters and the specific values used in
Section 4.3.
Table 6 presents the computational complexity formulas for each method at various stages.
Figure 19 shows the simulation results of the computational complexity using different methods.
The SIFT method exhibits relatively high computational complexity, primarily because it requires generating a multi-scale Gaussian pyramid during the keypoint extraction stage. The SURF method effectively reduces computational complexity by employing integral images, Hessian matrix approximation, simplified feature descriptors, and accelerated orientation computation, making it the least complex of the four methods. The combination of the SIFT and SURF methods retains relatively high computational complexity. The computational complexity of the proposed method lies between that of SIFT, SIFT+SURF and SURF. In the proposed method, due to the use of longer feature descriptors and the circular cross-correlation method for matching, the feature matching stage contributes the second most to the overall computational complexity. However, with the introduction of preliminary outlier filtering, the complexity of the proposed method in the parameter estimation stage is relatively lower compared to the other three methods.
5. Discussion
In the above research, it was demonstrated that using the spatial distribution of dominant scatterers for registration is a more effective approach in ISAR image registration. For sparse and discrete ISAR images, the effectiveness of the SIFT and SURF methods decreases. This is because SIFT and SURF identify inflection points, corners, and textured areas as FPs, but in ISAR images, the edges formed between scatterers and the background are often unstable and unreliable, leading these methods to detect numerous invalid FPs. Additionally, SIFT and SURF rely on local gradient information around FPs for feature description, which restricts the amount of information that can be extracted from ISAR images. The proposed method uses the distance and angular information between all scatterers, adopting a global approach. Therefore, it is more effective for ISAR registration.
The proposed method encounters limitations in scenarios with significant IPP variations, such as target maneuvers, large viewing angle differences between distant stations, or images captured from widely separated arc segments at the same location. Variations in IPP result in projection discrepancies and image decorrelation. Improving radar image registration under such conditions remains a challenging but essential task and will be addressed in future research.