Transformer-Based Subject-Sensitive Hashing for Integrity Authentication of High-Resolution Remote Sensing (HRRS) Images
Abstract
:Featured Application
Abstract
1. Introduction
- To the best of our knowledge, this was the first study to apply Transformer to integrity authentication of HRRS images, and the first research on Transformer-based subject-sensitive hashing;
- We modified the Swin-Unet structure to make the model more suitable for HRRS subject-sensitive hashing, which helped the algorithm comprehensively outperform existing algorithms, including the original Swin-Unet;
- We proposed a feature encoding method combining the mapping mechanism and principal component analysis (PCA) for the generation of hash sequences.
2. Related Work
2.1. Perceptual Hashing
2.2. Subject-Sensitive Hashing
3. Method
3.1. Architecture of Improved Swin-Unet
- (1)
- In order to improve the algorithm’s robustness, the patch expanding layer of the last module of the Swin-Unet decoder was canceled. After all, higher-level Transformers focus on encoding relatively complex high-level semantic information, and overly complex information could weaken algorithm robustness;
- (2)
- The position of the Skip connection between the first module of the encoder and the last module of the decoder was changed to cater to the cancellation of the Patch Expanding layer of the last module of the decoder, which reduced the impact of extraneous features on the tampering sensitivity of the algorithm;
- (3)
- Due to the network structure, the output image size of the improved Swin-Unet was 128 × 128 pixels, while the output image size of the original Swin-Unet was 224 × 224 pixels—the input image size for both models was 224 × 224 pixels. This input-output asymmetry helped to reduce the impact of redundant information and improve the performance of hashing algorithm.
3.2. Feature Coding Based on Mapping Mechanism and PCA
3.3. Overview of the Subject-Sensitive Hashing Algorithm
- (1)
- Preprocessing was designed to process the HRRS image so that it met the requirements of the input size of the improved Swin-Unet, which was 224 × 224 pixels. If the size of the HRRS image was large (for example, larger than 512 × 512 pixels), it could be divided into non-overlapping grid cells by grid division to ensure the accuracy of integrity authentication. Since grid division-based methods have been discussed repeatedly by existing algorithms [6,26], we chose not to repeat them in this paper.
- (2)
- For feature extraction, the preprocessed HRRS image was input into the trained improved Swin-Unet to obtain the feature map of the corresponding image. The training process of the improved Swin-Unet was discussed in Section IV.
- (3)
- Feature coding: the feature extracted by the improved Swin-Unet was essentially a two-dimensional matrix of pixel gray values. After feature encoding, based on the mapping mechanism and PCA, the obtained one-dimensional 0–1 sequence was encrypted by the AES (Advanced Encryption Standard) algorithm [40,41] to get the hash sequence, denoted as SH.
3.4. Integrity Authentication Process
4. Experiments
4.1. Implementation Details and Datasets
4.2. Evaluation Indicator
- (1)
- Robustness. For a single HRRS image, robustness means that, after the image undergoes an operation, the hash sequence does not change, or the change is lower than a preset threshold. Due to the strong chance of a single or a small number of test data, we measured the robustness using the proportion of HRRS images whose hash sequence variations were lower than the preset threshold T. The calculation method was as follows:
- (2)
- Tampering sensitivity. Like cryptographic hash and fragile watermarking, a subject-sensitive hash algorithm has to detect whether the image content has been tampered with, which means that tampering sensitivity is an important evaluation indicator for subject-sensitive hashing. For a single instance of image tampering, the hash sequences before and after the image is tampered with should change by a greater magnitude than the threshold T. Similar to the robustness test, tampering sensitivity also requires more test instances to be more convincing. We used the proportion of instances in which tampering was detected to describe tampering sensitivity, as follows:
- (3)
- Digestibility. The storage space occupied by the hash sequence should be as small as possible—that is, the hash sequence should be as short as possible.
- (4)
- Computational performance. Computational performance requires that the time to calculate and compare hash sequences should be as short as possible. In fact, as the comparison of hash sequences is very efficient due to digestibility, computational performance generally focuses on the efficiency of generating hash sequences.
- (5)
- Security. The security of subject-sensitive hashing means that the content of the image cannot be obtained from the subject-sensitive hash sequence.
4.3. Examples of Integrity Authentication
4.4. Robustness Testing of the Algorithms
4.5. Tampering Sensitivity Testing of Algorithms
4.6. Computational Performance
5. Discussion
- Robustness
- 2.
- Tampering sensitivity
- 3.
- Security, and Digestibility Analysis
- 4.
- Computational performance
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Niu, X.; Jiao, Y. An Overview of Perceptual Hashing. Acta Electron. Sin. 2008, 36, 1405–1411. [Google Scholar]
- Khelifi, F.; Jiang, J. Analysis of the Security of Perceptual Image Hashing Based on Non-Negative Matrix Factorization. IEEE Signal Process. Lett. 2010, 17, 43–46. [Google Scholar] [CrossRef]
- Du, L.; Ho, A.; Cong, R. Perceptual hashing for image authentication: A survey. Image Commun. 2020, 81, 115713. [Google Scholar] [CrossRef]
- Samanta, P.; Jain, S. Analysis of Perceptual Hashing Algorithms in Image Manipulation Detection. Procedia Comput. Sci. 2021, 185, 203–212. [Google Scholar] [CrossRef]
- Li, X.; Qin, C.; Wang, Z.; Qian, Z.; Zhang, X. Unified Performance Evaluation Method for Perceptual Image Hashing. IEEE Trans. Inf. Forensics Secur. 2022, 17, 1404–1419. [Google Scholar] [CrossRef]
- Ding, K.; Liu, Y.; Xu, Q.; Lu, F. A Subject-Sensitive Perceptual Hash Based on MUM-Net for the Integrity Authentication of High Resolution Remote Sensing Images. ISPRS Int. J. Geo-Inf. 2020, 9, 485. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
- Zhang, Q.; Li, Y.; Hu, Y.; Zhao, X. An Encrypted Speech Retrieval Method Based on Deep Perceptual Hashing and CNN-BiLSTM. IEEE Access 2020, 8, 148556–148569. [Google Scholar] [CrossRef]
- Khelifi, F.; Bouridane, A. Perceptual Video Hashing for Content Identification and Authentication. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 50–67. [Google Scholar] [CrossRef] [Green Version]
- Rajeshwaran, K.; Anil Kumar, K. Cellular Automata Based Hashing Algorithm (CABHA) for Strong Cryptographic Hash Function. In Proceedings of the IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT), Coimbatore, Tamil Nadu, India, 20–22 February 2019; pp. 1–6. [Google Scholar]
- Gupta, D.N.; Kumar, R. Sponge based Lightweight Cryptographic Hash Functions for IoT Applications. In Proceedings of the International Conference on Intelligent Technologies (CONIT), Hubbali, Karnataka, India, 25–27 June 2021; pp. 1–5. [Google Scholar]
- Qin, C.; Sun, M.; Chang, C. Perceptual hashing for color images based on hybrid extraction of structural features. Signal Process. 2018, 36, 194–205. [Google Scholar] [CrossRef]
- Tang, Z.; Li, X.; Zhang, X.; Zhang, S.; Dai, Y. Image hashing with color vector angle. Neurocomputing 2018, 308, 147–158. [Google Scholar] [CrossRef]
- Hamid, H.; Ahmed, F.; Ahmad, J. Robust Image Hashing Scheme using Laplacian Pyramids. Comput. Electr. Eng. 2020, 84, 106648. [Google Scholar] [CrossRef]
- Biswas, R.; González-Castro, V.; Fidalgo, E.; Alegre, E. A new perceptual hashing method for verification and identity classification of occluded faces. Image Vis. Comput. 2021, 113, 104245. [Google Scholar] [CrossRef]
- Wang, X.; Zhang, Q.; Jiang, C.; Xue, J. Perceptual hash-based coarse-to-fine grained image tampering forensics method. J. Vis. Commun. Image Represent. 2021, 8, 103124. [Google Scholar] [CrossRef]
- Huang, Z.; Liu, S. Perceptual Image Hashing With Texture and Invariant Vector Distance for Copy Detection. IEEE Trans. Multimed. 2021, 23, 1516–1529. [Google Scholar] [CrossRef]
- Romero, A.; Gatta, C.; Camps-Valls, G. Unsupervised Deep Feature Extraction for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1349–1362. [Google Scholar] [CrossRef] [Green Version]
- Quan, D.; Wei, H.; Wang, S.; Lei, R.; Duan, B.; Li, Y.; Hou, B.; Jiao, L. Self-Distillation Feature Learning Network for Optical and SAR Image Registration. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4706718. [Google Scholar] [CrossRef]
- Chen, H.; Qi, Z.; Shi, Z. Remote Sensing Image Change Detection with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5607514. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
- Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
- Adiga, V.; Sivaswamy, J. FPD-M-net: Fingerprint Image Denoising and Inpainting Using M-Net Based Convolutional Neural Networks. arXiv 2019, arXiv:1812.10191. [Google Scholar]
- Ibtehaz, N.; Rahman, M. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Net. 2020, 121, 74–87. [Google Scholar] [CrossRef] [PubMed]
- Ding, K.; Chen, S.; Wang, Y.; Liu, Y.; Zeng, Y.; Tian, J. AAU-Net: Attention-Based Asymmetric U-Net for Subject-Sensitive Hashing of Remote Sensing Images. Remote Sens. 2022, 13, 5109. [Google Scholar] [CrossRef]
- Xu, Y.; Xu, W.; Cheung, D.; Tu, Z. Line Segment Detection Using Transformers without Edges. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 21–24 June 2021; pp. 4255–4264. [Google Scholar]
- Lin, A.; Chen, B.; Xu, J.; Zhang, Z.; Lu, G.; Zhang, D. DS-TransUNet: Dual Swin Transformer U-Net for Medical Image Segmentation. IEEE Trans. Instrum. Meas. 2022, 71, 4005615. [Google Scholar] [CrossRef]
- Wang, W.; Tang, C.; Wang, X.; Zheng, B. A ViT-Based Multiscale Feature Fusion Approach for Remote Sensing Image Segmentation. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4510305. [Google Scholar] [CrossRef]
- Song, R.; Feng, Y.; Cheng, W.; Mu, Z.; Wang, X. BS2T: Bottleneck Spatial–Spectral Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5532117. [Google Scholar] [CrossRef]
- Xue, X.; Zhang, H.; Fang, B.; Bai, Z.; Li, Y. Grafting Transformer on Automatically Designed Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531116. [Google Scholar] [CrossRef]
- Zhang, W.; Zhou, M.; Ji, C.; Sui, X.; Bai, J. Cross-Frame Transformer-Based Spatio-Temporal Video Super-Resolution. IEEE Trans. Broadcast. 2022, 68, 359–369. [Google Scholar] [CrossRef]
- Lei, S.; Shi, Z.; Mo, W. Transformer-Based Multistage Enhancement for Remote Sensing Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5615611. [Google Scholar] [CrossRef]
- Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 21–24 June 2021; pp. 1833–1844. [Google Scholar]
- Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 2844–2853. [Google Scholar]
- Ye, T.; Zhang, J.; Li, Y.; Zhang, X.; Zhao, Z.; Li, Z. CT-Net: An Efficient Network for Low-Altitude Object Detection Based on Convolution and Transformer. IEEE Trans. Instrum. Meas. 2022, 71, 2507412. [Google Scholar] [CrossRef]
- Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021. [Google Scholar] [CrossRef]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical vision Transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Online, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
- Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation. arXiv 2021, arXiv:2105.05537. [Google Scholar]
- Zhang, X.; Parhi, K. High-speed VLSI architectures for the AES algorithm. IEEE Trans. Very Large Scale Integr. (VLSI) Syst. 2004, 12, 957–967. [Google Scholar] [CrossRef] [Green Version]
- Masoumi, M.; Rezayati, M.H. Novel Approach to Protect Advanced Encryption Standard Algorithm Implementation Against Differential Electromagnetic and Power Analysis. IEEE Trans. Inf. Forensics Secur. 2015, 10, 256–265. [Google Scholar] [CrossRef]
- Ding, K.; Zhu, C.; Lu, F. An adaptive grid partition based perceptual hash algorithm for remote sensing image authentication. Wuhan Daxue Xuebao 2015, 40, 716–720. [Google Scholar]
- Kingma, D.P.; Ba, J. ADAM: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Alom, M.Z.; Yakopcic, C.; Hasan, M.; Taha, T.M.; Asari, V.K. Recurrent residual U-Net for medical image segmentation. J. Med. Imaging. 2019, 6, 014006. [Google Scholar] [CrossRef]
- Zhao, S.; Liu, T.; Liu, B.W.; Ruan, K. Attention residual convolution neural network based on U-net (AttentionResU-Net) for retina vessel segmentation. IOP Conf. Ser. Earth Environ. Sci. IOP Publ. 2020, 440, 032138. [Google Scholar] [CrossRef]
- Li, R.; Zheng, S.; Duan, C.; Su, J.; Zhang, C. Multistage Attention ResU-Net for Semantic Segmentation of Fine-Resolution Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 8009205. [Google Scholar] [CrossRef]
- Ji, S.P.; Wei, S.Y. Building extraction via convolutional neural networks from an open remote sensing building dataset. Acta Geod. Cartogr. Sin. 2019, 48, 448–459. [Google Scholar]
- Xia, G.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 3974–3983. [Google Scholar]
- Chefer, H.; Gur, S.; Wolf, L. Transformer Interpretability Beyond Attention Visualization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 21–24 June 2021; pp. 782–791. [Google Scholar]
- Wang, S.; Yin, Y.; Wang, D.; Wang, Y.; Jin, Y. Interpretability-Based Multimodal Convolutional Neural Networks for Skin Lesion Diagnosis. IEEE Trans. Cybern. 2022, 52, 12623–12637. [Google Scholar] [CrossRef]
- Chi, S.; Tian, Y.; Wang, F.; Wang, Y.; Chen, M.; Li, J. Deep Semisupervised Multitask Learning Model and Its Interpretability for Survival Analysis. IEEE J. Biomed. Health Inform. 2021, 25, 3185–3196. [Google Scholar] [CrossRef] [PubMed]
Raw Value of Pixel | Mapped Value of Pixel |
---|---|
60 | 33.27 |
64 | 37.62 |
110 | 100.22 |
128 | 128.28 |
160 | 177.19 |
255 | 255 |
Model Each Algorithm Was Based on | Figure 5b | Figure 5c | Figure 5d | Figure 5e | Figure 5f | Figure 5g | Figure 5h |
---|---|---|---|---|---|---|---|
Format Conversion | Watermark Embedding | JPEG Compression | Subject-Unrelated Tampering | Subject-Related Tampering | 8 × 8 Random Tampering | 16 × 16 Random Tampering | |
MUM-Net | 0 | 0 | 0.0585 | 0.1679 | 0.2500 | 0.2578 | 0.2460 |
MultiResUnet | 0 | 0 | 0.0234 | 0.0546 | 0.0781 | 0.0859 | 0.2578 |
U-net | 0 | 0 | 0.0312 | 0.0234 | 0.0585 | 0.1835 | 0.2460 |
M-net | 0 | 0 | 0.0625 | 0.0976 | 0.0859 | 0.1210 | 0.2578 |
Attention U-Net | 0 | 0 | 0.0234 | 0.0078 | 0.0625 | 0.2226 | 0.2617 |
Attention ResU-Net | 0 | 0 | 0 | 0 | 0 | 0.0273 | 0.2734 |
Attention R2U-Net | 0 | 0 | 0.0273 | 0.0664 | 0.0781 | 0.2226 | 0.2382 |
AAU-Net | 0 | 0.0039 | 0.0195 | 0.0898 | 0.1054 | 0.2656 | 0.2578 |
Swin-Unet | 0 | 0 | 0 | 0.0078 | 0.0390 | 0.0546 | 0.1562 |
Improved Swin-Unet (Our algorithm) | 0 | 0 | 0.0039 | 0.0937 | 0.2656 | 0.2539 | 0.2773 |
The Model That Each Algorithm Based on | Figure 5b | Figure 5c | Figure 5d | Figure 5e | Figure 5f | Figure 5g | Figure 5h |
---|---|---|---|---|---|---|---|
Format Conversion | Watermark Embedding | JPEG Compression | Subject-Unrelated Tampering | Subject-Related Tampering | 8 × 8 Random Tampering | 16 × 16 Random Tampering | |
MUM-Net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
MultiResUnet | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
U-net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
M-net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
Attention U-Net | Not tampered | Not tampered | Tampered | Not tampered | Tampered | Tampered | Tampered |
Attention ResU-Net | Not tampered | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered |
Attention R2U-Net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
AAU-Net | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
Swin-Unet | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered |
Improved Swin-Unet (Our algorithm) | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
The Model That Each Algorithm Based on | Figure 5b | Figure 5c | Figure 5d | Figure 5e | Figure 5f | Figure 5g | Figure 5h |
---|---|---|---|---|---|---|---|
Format Conversion | Watermark Embedding | JPEG Compression | Subject Unrelated Tampering | Subject Related Tampering | 8 × 8 Random Tampering | 16 × 16 Random Tampering | |
MUM-Net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
MultiResUnet | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
U-net | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered |
M-net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
Attention U-Net | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered |
Attention ResU-Net | Not tampered | Not tampered | Not tampered | Not tampered | Not tampered | Not tampered | Tampered |
Attention R2U-Net | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
AAU-Net | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
Swin-Unet | Not tampered | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered |
Improved Swin-Unet (Our algorithm) | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
The Model That Each Algorithm Based on | Figure 5b | Figure 5c | Figure 5d | Figure 5e | Figure 5f | Figure 5g | Figure 5h |
---|---|---|---|---|---|---|---|
Format Conversion | Watermark Embedding | JPEG Compression | Subject-Unrelated Tampering | Subject-Related Tampering | 8 × 8 Random Tampering | 16 × 16 Random Tampering | |
MUM-Net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
MultiResUnet | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
U-net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
M-net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
Attention U-Net | Not tampered | Not tampered | Tampered | Not tampered | Tampered | Tampered | Tampered |
Attention ResU-Net | Not tampered | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered |
Attention R2U-Net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
AAU-Net | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered | Tampered |
Swin-Unet | Not tampered | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered |
Improved Swin-Unet (Our algorithm) | Not tampered | Not tampered | Not tampered | Tampered | Tampered | Tampered | Tampered |
Model Each Algorithm Was Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 79.57% | 90.92% | 97.61% | 99.88% | 100% |
MultiResUnet | 76.26% | 85.95% | 94.73% | 98.11% | 99.59% |
U-net | 63.28% | 75.06% | 88.85% | 96.74% | 99.81% |
M-net | 70.22% | 80.08% | 92.23% | 97.58% | 99.66% |
Attention U-Net | 85.94% | 92.59% | 97.14% | 98.88% | 100% |
Attention ResU-Net | 86.02% | 91.08% | 96.07% | 99.26% | 99.87% |
Attention R2U-Net | 58.53% | 69.22% | 82.97% | 95.52% | 99.26% |
AAU-Net | 73.57% | 81.01% | 93.33% | 98.76% | 99.82% |
Swin-Unet | 99.23% | 99.85% | 99.97% | 100% | 100% |
Improved Swin-Unet (our algorithm) | 95.88% | 99.02% | 100% | 100% | 100% |
Model Each Algorithm Was Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 62.09% | 74.27% | 88.18% | 97.32% | 99.69% |
MultiResUnet | 87.48% | 92.12% | 96.84% | 99.06% | 99.85% |
U-net | 61.66% | 81.75% | 85.59% | 95.84% | 99.28% |
M-net | 68.67% | 88.49% | 89.97% | 97.45% | 99.59% |
Attention U-Net | 90.07% | 94.16% | 97.25% | 98.99% | 99.87% |
Attention ResU-Net | 93.58% | 96.06% | 98.02% | 99.77% | 99.98% |
Attention R2U-Net | 48.69% | 60.24% | 77.45% | 91.98% | 98.51% |
AAU-Net | 95.09% | 97.07% | 98.33% | 99.58% | 99.85% |
Swin-Unet | 99.66% | 99.95% | 100% | 100% | 100% |
Improved Swin-Unet (Our algorithm) | 98.07% | 99.74% | 100% | 100% | 100% |
Model Each Algorithm Was Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 42.05% | 52.27% | 71.09% | 90.38% | 98.26% |
MultiResUnet | 79.69% | 88.45% | 95.98% | 98.68% | 99.82% |
U-net | 51.99% | 64.22% | 90.61% | 93.36% | 98.57% |
M-net | 53.52% | 63.39% | 78.54% | 92.77% | 98.64% |
Attention U-Net | 52.03% | 64.24% | 80.58% | 93.39% | 98.56% |
Attention ResU-Net | 81.49% | 86.61% | 92.07% | 96.43% | 98.88% |
Attention R2U-Net | 19.54% | 27.02% | 41.27% | 64.24% | 86.12% |
AAU-Net | 90.91% | 93.65% | 97.38% | 99.44% | 100% |
Swin-Unet | 94.55% | 97.09% | 99.36% | 100% | 100% |
Improved Swin-Unet (Our algorithm) | 92.82% | 97.25% | 99.54% | 100% | 100% |
Model Each Algorithm Was Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 95.02% | 90.98% | 82.19% | 56.24% | 14.88% |
MultiResUnet | 83.95% | 78.25% | 62.52% | 35.01% | 11.89% |
U-net | 96.98% | 94.94% | 85.63% | 61.39% | 31.08% |
M-net | 96.27% | 94.31% | 89.15% | 66.77% | 33.16% |
Attention U-Net | 93.68% | 91.29% | 84.42% | 68.61% | 36.24% |
Attention ResU-Net | 67.94% | 58.09% | 40.66% | 23.15% | 7.79% |
Attention R2U-Net | 98.75% | 98.07% | 95.36% | 80.93% | 52.84% |
AAU-Net | 92.76% | 90.92% | 83.04% | 65.57% | 29.52% |
Swin-Unet | 84.61% | 76.42% | 56.01% | 23.96% | 6.92% |
Improved Swin-Unet (Our algorithm) | 94.29% | 90.16% | 81.73% | 55.42% | 19.24% |
The Model That Each Algorithm Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 98.79% | 98.52% | 95.25% | 78.71% | 42.77% |
MultiResUnet | 99.05% | 97.73% | 92.43% | 79.26% | 54.29% |
U-net | 98.85% | 97.98% | 96.24% | 86.41% | 59.17% |
M-net | 99.16% | 99.03% | 96.22% | 86.98% | 60.71% |
Attention U-Net | 98.86% | 97.95% | 96.16% | 86.39% | 59.21% |
Attention ResU-Net | 83.92% | 76.68% | 57.52% | 34.74% | 12.11% |
Attention R2U-Net | 99.92% | 99.90% | 99.85% | 98.03% | 81.54% |
AAU-Net | 99.54% | 99.19% | 96.43% | 87.01% | 64.24% |
Swin-Unet | 93.23% | 89.29% | 75.22% | 32.61% | 7.64% |
Improved Swin-Unet (Our algorithm) | 99.72% | 99.41% | 98.35% | 88.62% | 38.68% |
Model Each Algorithm Was Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 100% | 100% | 99.0% | 98.0% | 60.5% |
MultiResUnet | 96.0% | 94.5% | 84.0% | 56.0% | 10.5% |
U-net | 100% | 100% | 99.5% | 95.0% | 65.0% |
M-net | 100% | 100% | 99.5% | 94.0% | 58.5% |
Attention U-Net | 100% | 100% | 98.0% | 92.5% | 50.0% |
Attention ResU-Net | 75.5% | 67.0% | 43.0% | 19.5% | 1.5% |
Attention R2U-Net | 100% | 100% | 100% | 98.0% | 77.0% |
AAU-Net | 100% | 100% | 98.0% | 89.5% | 36.0% |
Swin-Unet | 97.5% | 90.0% | 67.5% | 20.0% | 3.5% |
Improved Swin-Unet (Our algorithm) | 100% | 100% | 99.5% | 96.0% | 41.5% |
Model Each Algorithm Was Based on | T = 0.02 | T = 0.03 | T = 0.05 | T = 0.1 | T = 0.2 |
---|---|---|---|---|---|
MUM-Net | 100% | 100% | 97.0% | 80.5% | 37.5% |
MultiResUnet | 92.0% | 86.5% | 63.5% | 27.0% | 8.0% |
U-net | 99.5% | 98.0% | 95.0% | 76.0% | 28.5% |
M-net | 100% | 98.0% | 94.5% | 76.0% | 27.0% |
Attention U-Net | 100% | 97.0% | 93.0% | 78.5% | 36.5% |
Attention ResU-Net | 90.0% | 88.5% | 76.0% | 38.5% | 11.0% |
Attention R2U-Net | 100% | 100% | 97.5% | 65.0% | 36.0% |
AAU-Net | 96.5% | 94.0% | 85.0% | 56.5% | 46.5% |
Swin-Unet | 95.0% | 92.0% | 82.5% | 32.5% | 2.5% |
Improved Swin-Unet (Our algorithm) | 100% | 100% | 98.0% | 81.0% | 43.0% |
Model Each Algorithm Was Based on | 300 Images | 1000 Images | 10,000 Images | |||
---|---|---|---|---|---|---|
Average Time (ms) | Total Time (s) | Average Time (ms) | Total Time (s) | Average Time (ms) | Total Time (s) | |
MUM-Net | 35.13 | 10.54 | 24.20 | 24.20 | 23.43 | 234.30 |
MultiResUnet | 44.07 | 13.22 | 33.89 | 33.89 | 32.37 | 323.68 |
U-net | 21.20 | 6.36 | 13.34 | 13.34 | 12.78 | 127.79 |
M-net | 27.57 | 8.27 | 17.79 | 17.79 | 15.75 | 157.50 |
Attention U-Net | 21.47 | 6.44 | 15.60 | 15.60 | 13.38 | 133.82 |
Attention ResU-Net | 46.83 | 14.05 | 26.60 | 26.60 | 22.12 | 221.24 |
Attention R2U-Net | 30.40 | 9.12 | 20.84 | 20.84 | 19.24 | 192.36 |
AAU-Net | 19.70 | 5.91 | 14.89 | 14.89 | 11.73 | 117.34 |
Swin-Unet | 59.20 | 17.76 | 38.98 | 38.98 | 36.08 | 360.77 |
Improved Swin-Unet (Our algorithm) | 49.73 | 14.92 | 33.11 | 33.11 | 31.24 | 312.37 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ding, K.; Chen, S.; Zeng, Y.; Wang, Y.; Yan, X. Transformer-Based Subject-Sensitive Hashing for Integrity Authentication of High-Resolution Remote Sensing (HRRS) Images. Appl. Sci. 2023, 13, 1815. https://doi.org/10.3390/app13031815
Ding K, Chen S, Zeng Y, Wang Y, Yan X. Transformer-Based Subject-Sensitive Hashing for Integrity Authentication of High-Resolution Remote Sensing (HRRS) Images. Applied Sciences. 2023; 13(3):1815. https://doi.org/10.3390/app13031815
Chicago/Turabian StyleDing, Kaimeng, Shiping Chen, Yue Zeng, Yingying Wang, and Xinyun Yan. 2023. "Transformer-Based Subject-Sensitive Hashing for Integrity Authentication of High-Resolution Remote Sensing (HRRS) Images" Applied Sciences 13, no. 3: 1815. https://doi.org/10.3390/app13031815
APA StyleDing, K., Chen, S., Zeng, Y., Wang, Y., & Yan, X. (2023). Transformer-Based Subject-Sensitive Hashing for Integrity Authentication of High-Resolution Remote Sensing (HRRS) Images. Applied Sciences, 13(3), 1815. https://doi.org/10.3390/app13031815