Homography Matrix-Based Local Motion Consistent Matching for Remote Sensing Images
Abstract
:1. Introduction
1.1. Resampling-Based Methods
1.2. Non-Parametric Model-Based Methods
1.3. Graph Matching-Based Methods
1.4. Learning-Based Methods
1.5. Local Geometric Constraint-Based Methods
- We propose a local geometric constraint based on the homography matrix for feature matching in remote sensing images. Compared to other methods based on local geometric constraints, LMC has more strict constraints that aim to utilize the properties of the homography matrix to represent local motion consistency, thereby retaining correct matches. This constraint is projectively invariant and applicable to images with various rigid or non-rigid deformations;
- We design a jump-out mechanism that can exit the loop without searching through all possible cases, thereby reducing the runtime. LMC can process more than 1000 putative matches within 100 ms;
- To avoid outliers affecting the search for neighborhoods with correct motion, we introduce a resampling method to construct neighborhoods.
2. Materials and Methods
2.1. Problem Formulation
2.2. Building Neighborhoods Based on RANSAC
2.3. Calculation of Reprojection Error Based on Homography Matrix
2.4. Jump-Out Mechanism
2.5. Time Complexity
Algorithm 1 The LMC Algorithm. |
|
3. Experimental Results
3.1. Datasets
- SUIRD: This dataset contains 60 pairs of low-altitude remote sensing images captured by small drones, covering natural landscapes and urban streets. The geometric transformations between image pairs are caused by viewpoint changes during drone shooting, resulting in a low overlap rate, non-rigid image deformation, and extreme viewpoint changes. The dataset is divided into five groups according to the types of drone viewpoint changes: horizontal rotation (termed Horizontal), vertical rotation (termed Vertical), scaling (termed Scaling), mixed viewpoint changes (termed Mixture), and extreme viewpoint changes (termed Extreme). To better demonstrate the experimental data, we integrated horizontal rotation, vertical rotation, and scaling into simple rotation and scaling (termed R&S) in the dataset;
- HPatches: This dataset consists of images selected from multiple datasets such as VGG, DTU, and AMOS and is used to evaluate the matching performance of matching methods on common images. The dataset contains 116 sets of image sequences, with 6 images in each set. In each image sequence, the first image is matched with the remaining five images. The images involve variations in rotation, scaling, extreme viewpoint changes, lighting changes, and image compression. The dataset is divided into two groups based on the types of image variations, viewpoint changes (termed HPatches-v), and lighting changes (termed HPatches-i);
- RS: This dataset consists of four different types of remote sensing image datasets, including SAR, CIAP, UAV, and PAN. The image pairs of SAR were captured by synthetic aperture radar and drones, totaling 18 pairs. The images of CIAP were corrected, but with a small overlap area, totaling 57 pairs. The images of UAV were captured by drones with projection distortion, totaling 35 pairs. The images of PAN were captured at different times, with projection distortion, totaling 18 pairs;
- DTU: The dataset consists of 131 pairs of images with large viewpoint changes. Due to the presence of large viewpoint changes, there are complex non-rigid deformations between image pairs;
- Retina: This dataset consists of 70 pairs of multimodal medical retinal images, which exhibit non-rigid deformations between the image pairs.
3.2. Parameter Settings and Ablation Study
- Model_2: Neither RANSAC nor a jump-out mechanism were used.
- Model_12: Used RANSAC but did not use a jump-out mechanism.
- Model_23: Used a jump-out mechanism but did not use RANSAC.
- Model_123: Both RANSAC and a jump-out mechanism were used.
3.3. Qualitative Analysis
3.4. Quantitative Analysis
3.5. Robustness Analysis
3.6. Impact of Neighborhood Construction Methods on Performance
4. Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image matching from handcrafted to deep features: A survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
- Li, X.; Luo, X.; Wu, Y.; Li, Z.; Xu, W. Research on stereo matching for satellite generalized image pair based on improved SURF and RFM. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2739–2742. [Google Scholar]
- Jiang, S.; Jiang, W.; Wang, L. Unmanned aerial vehicle-based photogrammetric 3d mapping: A survey of techniques, applications, and challenges. IEEE Geosci. Remote Sens. Mag. 2021, 10, 135–171. [Google Scholar] [CrossRef]
- Xiao, G.; Luo, H.; Zeng, K.; Wei, L.; Ma, J. Robust feature matching for remote sensing image registration via guided hyperplane fitting. IEEE Trans. Geosci. Remote Sens. 2020, 60, 1–14. [Google Scholar] [CrossRef]
- Quan, D.; Wang, S.; Gu, Y.; Lei, R.; Yang, B.; Wei, S.; Hou, B.; Jiao, L. Deep feature correlation learning for multi-modal remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
- Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
- Lin, C.C.; Pankanti, S.U.; Natesan Ramamurthy, K.; Aravkin, A.Y. Adaptive as-natural-as-possible image stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1155–1163. [Google Scholar]
- Wei, C.; Xia, H.; Qiao, Y. Fast unmanned aerial vehicle image matching combining geometric information and feature similarity. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1731–1735. [Google Scholar] [CrossRef]
- Wang, C.; Wang, L.; Liu, L. Progressive mode-seeking on graphs for sparse feature matching. In Computer Vision—ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part II 13; Springer: Berlin/Heidelberg, Germany, 2014; pp. 788–802. [Google Scholar]
- Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
- Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
- Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision; 2011; pp. 2564–2571. [Google Scholar]
- Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
- Torr, P.H.; Zisserman, A. MLESAC: A new robust estimator with application to estimating image geometry. Comput. Vis. Image Underst. 2000, 78, 138–156. [Google Scholar] [CrossRef] [Green Version]
- Chum, O.; Matas, J. Matching with PROSAC-progressive sample consensus. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 220–226. [Google Scholar]
- Barath, D.; Noskova, J.; Ivashechkin, M.; Matas, J. MAGSAC++, a fast, reliable and accurate robust estimator. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1304–1312. [Google Scholar]
- Barath, D.; Matas, J. Graph-cut RANSAC: Local optimization on spatially coherent structures. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4961–4974. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; Hu, Z. Rejecting mismatches by correspondence function. Int. J. Comput. Vis. 2010, 89, 1–17. [Google Scholar] [CrossRef]
- Myronenko, A.; Song, X. Point set registration: Coherent point drift. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2262–2275. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zhao, J.; Ma, J.; Tian, J.; Ma, J.; Zhang, D. A robust method for vector field learning with application to mismatch removing. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2977–2984. [Google Scholar]
- Liu, H.; Yan, S. Common visual pattern discovery via spatially coherent correspondences. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1609–1616. [Google Scholar]
- Zhou, F.; De la Torre, F. Factorized graph matching. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 127–134. [Google Scholar]
- Liu, Z.Y.; Qiao, H. Gnccp—Graduated nonconvexityand concavity procedure. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1258–1267. [Google Scholar] [CrossRef]
- Loiola, E.M.; De Abreu, N.M.M.; Boaventura-Netto, P.O.; Hahn, P.; Querido, T. A survey for the quadratic assignment problem. Eur. J. Oper. Res. 2007, 176, 657–690. [Google Scholar] [CrossRef]
- Jiang, X.; Xia, Y.; Zhang, X.P.; Ma, J. Robust image matching via local graph structure consensus. Pattern Recogn. 2022, 126, 108588. [Google Scholar] [CrossRef]
- Ma, J.; Fan, A.; Jiang, X.; Xiao, G. Feature matching via motion-consistency driven probabilistic graphical model. Int. J. Comput. Vis. 2022, 130, 2249–2264. [Google Scholar] [CrossRef]
- Ma, J.; Jiang, X.; Jiang, J.; Zhao, J.; Guo, X. LMR: Learning a two-class classifier for mismatch removal. IEEE Trans. Image Process. 2019, 28, 4045–4059. [Google Scholar] [CrossRef]
- Yi, K.M.; Trulls, E.; Ono, Y.; Lepetit, V.; Salzmann, M.; Fua, P. Learning to find good correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Munich, Germany, 8–14 September 2018; pp. 2666–2674. [Google Scholar]
- Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
- Jiang, X.; Wang, Y.; Fan, A.; Ma, J. Learning for mismatch removal via graph attention networks. ISPRS J. Photogramm. Remote Sens. 2022, 190, 181–195. [Google Scholar] [CrossRef]
- Chen, J.; Chen, S.; Chen, X.; Dai, Y.; Yang, Y. Csr-net: Learning adaptive context structure representation for robust feature correspondence. IEEE Trans. Image Process. 2022, 31, 3197–3210. [Google Scholar] [CrossRef]
- Jiang, B.; Sun, P.; Luo, B. GLMNet: Graph learning-matching convolutional networks for feature matching. Pattern Recogn. 2022, 121, 108167. [Google Scholar] [CrossRef]
- Jiang, Z.; Rahmani, H.; Angelov, P.; Black, S.; Williams, B.M. Graph-context Attention Networks for Size-varied Deep Graph Matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2343–2352. [Google Scholar]
- Ma, J.; Zhao, J.; Jiang, J.; Zhou, H.; Guo, X. Locality preserving matching. Int. J. Comput. Vis. 2019, 127, 512–531. [Google Scholar] [CrossRef]
- Bian, J.; Lin, W.Y.; Matsushita, Y.; Yeung, S.K.; Nguyen, T.D.; Cheng, M.M. Gms: Grid-based motion statistics for fast, ultra-robust feature correspondence. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4181–4190. [Google Scholar]
- Shao, F.; Liu, Z.; An, J. A discriminative point matching algorithm based on local structure consensus constraint. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1366–1370. [Google Scholar] [CrossRef]
- Jiang, X.; Jiang, J.; Fan, A.; Wang, Z.; Ma, J. Multiscale locality and rank preservation for robust feature matching of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6462–6472. [Google Scholar] [CrossRef]
- Chen, J.; Yang, M.; Gong, W.; Yu, Y. Multi-neighborhood Guided Kendall Rank Correlation Coefficient for Feature Matching. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
- Ma, J.; Li, Z.; Zhang, K.; Shao, Z.; Xiao, G. Robust feature matching via neighborhood manifold representation consensus. ISPRS J. Photogramm. Remote Sens. 2022, 183, 196–209. [Google Scholar] [CrossRef]
- Shen, L.; Xu, Z.; Zhu, J.; Huang, X.; Jin, T. A frame-based probabilistic local verification method for robust correspondence. ISPRS J. Photogramm. Remote Sens. 2022, 192, 232–243. [Google Scholar] [CrossRef]
- Ye, X.; Ma, J.; Xiong, H. Local Affine Preservation with Motion Consistency for Feature Matching of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
- Sedaghat, A.; Mokhtarzade, M.; Ebadi, H. Uniform robust scale-invariant feature matching for optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4516–4527. [Google Scholar] [CrossRef]
- Balntas, V.; Lenc, K.; Vedaldi, A.; Mikolajczyk, K. HPatches: A benchmark and evaluation of handcrafted and learned local descriptors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5173–5182. [Google Scholar]
- Aanæs, H.; Jensen, R.R.; Vogiatzis, G.; Tola, E.; Dahl, A.B. Large-scale data for multiple-view stereopsis. Int. J. Comput. Vis. 2016, 120, 153–168. [Google Scholar] [CrossRef] [Green Version]
- Jiang, X.; Ma, J.; Xiao, G.; Shao, Z.; Guo, X. A review of multimodal image matching: Methods and applications. Inform. Fusion 2021, 73, 22–71. [Google Scholar] [CrossRef]
Method | LPM | mTopKRP | RANSAC | NMRC | LSCC | LMC (Ours) | |
---|---|---|---|---|---|---|---|
Extreme | AR (%) | 94.71 | 94.80 | 95.02 | 98.53 | 80.96 | 98.74 |
AP (%) | 95.41 | 94.69 | 99.97 | 98.03 | 98.11 | 99.25 | |
AF (%) | 95.03 | 94.67 | 96.84 | 98.27 | 87.13 | 98.97 | |
ART (ms) | 14.04 | 860.95 | 5.88 | 956.04 | 442.66 | 81.45 | |
Mixture | AR (%) | 95.38 | 96.16 | 97.07 | 99.12 | 84.81 | 99.56 |
AP (%) | 96.89 | 96.48 | 99.97 | 98.92 | 98.84 | 99.27 | |
AF (%) | 96.05 | 96.31 | 98.18 | 99.02 | 88.94 | 99.41 | |
ART (ms) | 18.62 | 1088.45 | 6.87 | 1281.97 | 727.71 | 107.92 | |
R&S | AR (%) | 96.18 | 96.32 | 96.16 | 98.99 | 84.38 | 99.12 |
AP (%) | 96.43 | 96.02 | 99.98 | 98.42 | 98.38 | 99.26 | |
AF (%) | 96.28 | 96.12 | 97.64 | 98.70 | 89.67 | 99.17 | |
ART (ms) | 13.63 | 851.73 | 4.41 | 949.24 | 415.00 | 83.54 |
Method | LPM | mTopKRP | RANSAC | NMRC | LSCC | LMC_RANSAC (Ours) | LMC_LPM (Ours) | |
---|---|---|---|---|---|---|---|---|
DTU | AR (%) | 95.05 | 95.57 | 44.07 | 95.78 | 88.73 | 67.48 | 96.19 |
AP (%) | 95.33 | 91.71 | 97.12 | 93.60 | 95.16 | 96.49 | 96.03 | |
AF (%) | 95.10 | 93.21 | 58.15 | 94.59 | 91.64 | 78.21 | 96.07 | |
ART (ms) | 10.46 | 650.22 | 33.84 | 739.61 | 451.22 | 326.85 | 95.34 | |
Retina | AR (%) | 93.70 | 91.85 | 44.98 | 92.08 | 75.90 | 86.24 | 93.92 |
AP (%) | 83.10 | 88.27 | 98.05 | 91.54 | 93.03 | 89.59 | 91.06 | |
AF (%) | 87.61 | 89.31 | 60.69 | 91.55 | 82.85 | 87.72 | 92.24 | |
ART (ms) | 3.25 | 129.16 | 34.30 | 133.94 | 36.44 | 123.96 | 17.47 | |
RS | AR (%) | 98.82 | 99.25 | 76.87 | 99.22 | 77.52 | 93.48 | 99.10 |
AP (%) | 98.11 | 99.08 | 99.53 | 99.15 | 92.96 | 99.50 | 99.22 | |
AF (%) | 98.40 | 99.15 | 84.74 | 99.17 | 82.25 | 96.06 | 99.16 | |
ART (ms) | 11.98 | 129.16 | 15.64 | 133.94 | 36.44 | 123.96 | 86.78 | |
HPatches | AR (%) | 47.45 | 71.05 | 71.50 | 75.20 | 28.47 | 73.74 | 58.91 |
AP (%) | 63.51 | 68.12 | 78.86 | 70.45 | 43.32 | 74.14 | 63.77 | |
AF (%) | 47.57 | 66.39 | 72.79 | 71.21 | 29.70 | 73.37 | 58.57 | |
ART (ms) | 12.74 | 406.46 | 26.29 | 351.31 | 111.49 | 289.61 | 207.69 | |
SUIRD | AR (%) | 96.40 | 96.67 | 96.25 | 99.12 | 85.46 | 99.29 | 98.06 |
AP (%) | 96.73 | 96.48 | 99.98 | 98.60 | 98.57 | 99.28 | 98.03 | |
AF (%) | 96.53 | 96.54 | 97.79 | 98.86 | 90.36 | 99.27 | 97.92 | |
ART (ms) | 15.61 | 964.49 | 4.64 | 969.60 | 533.99 | 89.51 | 90.70 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, J.; Liang, A.; Zhao, E.; Pang, M.; Zhang, D. Homography Matrix-Based Local Motion Consistent Matching for Remote Sensing Images. Remote Sens. 2023, 15, 3379. https://doi.org/10.3390/rs15133379
Liu J, Liang A, Zhao E, Pang M, Zhang D. Homography Matrix-Based Local Motion Consistent Matching for Remote Sensing Images. Remote Sensing. 2023; 15(13):3379. https://doi.org/10.3390/rs15133379
Chicago/Turabian StyleLiu, Junyuan, Ao Liang, Enbo Zhao, Mingqi Pang, and Daijun Zhang. 2023. "Homography Matrix-Based Local Motion Consistent Matching for Remote Sensing Images" Remote Sensing 15, no. 13: 3379. https://doi.org/10.3390/rs15133379
APA StyleLiu, J., Liang, A., Zhao, E., Pang, M., & Zhang, D. (2023). Homography Matrix-Based Local Motion Consistent Matching for Remote Sensing Images. Remote Sensing, 15(13), 3379. https://doi.org/10.3390/rs15133379