AFMSFFNet: An Anchor-Free-Based Feature Fusion Model for Ship Detection
Abstract
:1. Introduction
- (1)
- The complex background conditions and strong clutter interference make it difficult to distinguish targets from the background, especially in inshore areas where ships are easily affected by noise interference from the land and other equipment, resulting in false alarm problems.
- (2)
- The detection performance of small-size targets is poor. Due to the unclear features, the features of small ships are difficult to accurately learn for the network, leading to missed detection.
- (3)
- In the pursuit of enhanced detection accuracy, many leading detection algorithms often adopt complex models with a substantial number of parameters. However, these models typically incur significant computational costs, making deployment on devices challenging and thereby constraining their widespread applicability.
- (1)
- The method proposed in this paper introduces the latest anchor-free architecture YOLOX tiny as the basic framework and then conducts a series of optimization designs. Extensive tests on the SSDD [25] and the SAR ship dataset [26] demonstrate a noteworthy improvement in the detection accuracy with this method compared to the original YOLOvX tiny network, achieving a 2.32% performance increase on SSDD.
- (2)
- This paper significantly enhances the FPN structure of YOLOvX tiny by introducing the Local Cross-Channel Attention (LCA) module and adaptive weighted fusion (AWF) mechanism. The LCA module strengthens feature extraction in target regions, effectively suppressing the interference of complex backgrounds in the detection results. Additionally, the introduction of the AWF module makes the fusion of multi-scale features more flexible and adaptive, maximizing the complementary information between different scales. These two improvements together form the final AB-FPN structure, significantly enhancing the detection capability across different scales, with a particular boost in the sensitivity of small target detection.
- (3)
- By introducing MGA in the detection head, multi-scale receptive field cascades were employed to capture multi-scale contextual information, effectively expanding the receptive field of the detection head. This enhancement significantly improves the network’s global perception capabilities in complex backgrounds, enabling better differentiation between ships and background clutter.
- (4)
- The proposed model achieves a high inference speed (FPS) while maintaining a low number of parameters. This indicates that our network not only offers significant improvements in detection performance but also remains lightweight, making it suitable for efficient deployment in practical applications. The collaborative effect of the three improved modules results in enhanced performance in small target detection. Furthermore, compared to similar methods, the proposed model demonstrates a distinct advantage in nearshore target detection, making it better equipped to handle challenges in complex scenarios.
2. Related Work
2.1. Two-Stage Method Based on Region Suggestions
2.2. One-Stage Method Based on Target Regression
2.3. Deep Learning-Based SAR Image Object Detection Algorithm
3. Materials and Methods
3.1. The Overall Architecture of YOLOX
3.2. The Overall Architecture of the AFMSFFNet Model
- (1)
- A SAR image is sent to the main feature network for initial feature extraction, generating multiple feature maps from C1 to C5.
- (2)
- Then, the three output maps, C3, C4, and C5, are further enhanced for feature extraction. The feature map undergoes adaptive feature fusion in the AB-FPN to effectively extract multi-scale target information, ultimately generating three output maps, P3_out, P4_out, and P5_out.
- (3)
- Finally, the feature maps, P3_out, P4_out, and P5_out, are passed to the MGAHead for final ship position detection and category classification.
3.3. AB-FPN
3.4. MGA
4. Experiments and Analysis
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Barber, B.C.; Barker, J.E. The use of SAR-ATI for maritime surveillance and difficult target detection. In Proceedings of the 2009 International Radar Conference Surveillance for a Safer World (RADAR 2009), Bordeaux, France, 12–16 October 2009; pp. 1–6. [Google Scholar]
- Friedman, K.; Wackerman, C.; Funk, F.; Schwenzfeier, M.; Pichel, W.; Colon-Clemente, P.; Li, X. Analyzing the dependence between RADARSAT-1 vessel detection and vessel heading using CFAR algorithm for use on fishery management. In Proceedings of the Oceans 2003. Celebrating the Past... Teaming Toward the Future (IEEE Cat. No.03CH37492), San Diego, CA, USA, 22–26 September 2003; Volume 5, pp. P2819–P2823. [Google Scholar]
- Mazzarella, F.; Vespe, M.; Santamaria, C. SAR Ship Detection and Self-Reporting Data Fusion Based on Traffic Knowledge. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1685–1689. [Google Scholar] [CrossRef]
- Rey, M.; Tunaley, J.; Folinsbee, J.; Jahans, P.; Dixon, J.; Vant, M. Application of Radon Transform Techniques to Wake Detection in Seasat-A SAR Images. IEEE Trans. Geosci. Remote Sens. 1990, 28, 553–560. [Google Scholar] [CrossRef]
- Goldstein, G.B. False-Alarm Regulation in Log-Normal and Weibull Clutter. IEEE Trans. Aerosp. Electron. Syst. 1973, AES-9, 84–92. [Google Scholar] [CrossRef]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
- Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
- Zhang, T.; Zhang, X.; Shi, J.; Wei, S. High-Speed Ship Detection in SAR Images by Improved Yolov3. In Proceedings of the 2019 16th International Computer Conference on Wavelet Active Media Technology and Information Processing, Chengdu, China, 14–15 December 2019; pp. 149–152. [Google Scholar]
- Guo, Y.; Chen, S.; Zhan, R.; Wang, W.; Zhang, J. SAR Ship Detection Based on YOLOv5 Using CBAM and BiFPN. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 2147–2150. [Google Scholar]
- Sun, Z.; Dai, M.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. An Anchor-Free Detection Method for Ship Targets in High-Resolution SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7799–7816. [Google Scholar] [CrossRef]
- Ma, X.; Hou, S.; Wang, Y.; Wang, J.; Wang, H. Multiscale and Dense Ship Detection in SAR Images Based on Key-Point Estimation and Attention Mechanism. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
- Rey, M.; Tunaley, J.; Sibbald, T. Use of the Dempster-Shafer algorithm for the detection of SAR ship wakes. IEEE Trans. Geosci. Remote Sens. 1993, 31, 1114–1118. [Google Scholar] [CrossRef]
- Copeland, A.C.; Ravichandran, G.; Trivedi, M.M. Localized Radon transform-based detection of ship wakes in SAR images. IEEE Trans. Geosci. Remote Sens. 1995, 33, 35–45. [Google Scholar] [CrossRef]
- Ringrose, R.; Harris, N. Ship detection using polarimetric SAR data. In Proceedings of the SAR workshop: CEOS Committee on Earth Observation Satellites, Toulouse, France, 26–29 October 1999; Volume 450, p. 687. [Google Scholar]
- Ritcey, J. An Order-Statistics-Based CFAR for SAR Applications; Electrical Engineering Department, University of Washington: Seattle, WA, USA, 1990. [Google Scholar]
- Xing, X.; Chen, Z.; Zou, H.; Zhou, S. A fast algorithm based on two-stage CFAR for detecting ships in SAR images. In Proceedings of the 2009 2nd Asian-Pacific Conference on Synthetic Aperture Radar (APSAR), Xi’an, China, 26–30 October 2009; pp. 506–509. [Google Scholar]
- Long, Z.; Suyuan, W.; Zhongma, C.; Jiaqi, F.; Xiaoting, Y.; Wei, D. Lira-YOLO: A lightweight model for ship detection in radar images. J. Syst. Eng. Electron. 2020, 31, 950–956. [Google Scholar] [CrossRef]
- Cui, Z.; Li, Q.; Cao, Z.; Liu, N. Dense Attention Pyramid Networks for Multi-Scale Ship Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8983–8997. [Google Scholar] [CrossRef]
- Zhao, Y.; Zhao, L.; Xiong, B.; Kuang, G. Attention Receptive Pyramid Network for Ship Detection in SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2738–2756. [Google Scholar] [CrossRef]
- Cui, Z.; Wang, X.; Liu, N.; Cao, Z.; Yang, J. Ship Detection in Large-Scale SAR Images Via Spatial Shuffle-Group Enhance Attention. IEEE Trans. Geosci. Remote Sens. 2020, 59, 379–391. [Google Scholar] [CrossRef]
- Zhang, T.; Zhang, X.; Li, J.; Xu, X.; Wang, B.; Zhan, X.; Xu, Y.; Ke, X.; Zeng, T.; Su, H.; et al. SAR ship detection dataset (SSDD): Official release and comprehensive data analysis. Remote Sens. 2021, 13, 3690. [Google Scholar] [CrossRef]
- Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR dataset of ship detection for deep learning under complex backgrounds. Remote Sens. 2019, 11, 765. [Google Scholar] [CrossRef]
- Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
- Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
- Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
- Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
- Huang, L.; Yang, Y.; Deng, Y.; Yu, Y. Densebox: Unifying landmark localization with end to end object detection. arXiv 2015, arXiv:1509.04874. [Google Scholar]
- Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
- Guo, H.; Yang, X.; Wang, N.; Gao, X. A CenterNet++ model for ship detection in SAR images. Pattern Recognit. 2021, 112, 107787. [Google Scholar] [CrossRef]
- Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
- Fu, H.; Wang, X.; Peng, C.; Che, Z.; Wang, Y. A dual-task algorithm for ship target detection and semantic segmentation based on improved YOLOv5. In Proceedings of the OCEANS 2023—Limerick, Limerick, Ireland, 5–8 June 2023; pp. 1–7. [Google Scholar]
- Guo, Y.; Zhan, R.; Chen, S.; Li, L.; Zhang, J. A lightweight SAR ship detection method based on improved YOLOv8. In Proceedings of the IET International Radar Conference (IRC 2023), Chongqing, China, 3–5 December 2023; pp. 1322–1327. [Google Scholar]
- Tan, X.; Leng, X.; Wang, J.; Ji, K. A ship detection method based on YOLOv7 in range-compressed SAR data. In Proceedings of the IET International Radar Conference (IRC 2023), Chongqing, China, 3–5 December 2023; pp. 948–952. [Google Scholar]
- Yang, Y.; Ju, Y.; Zhou, Z. A Super Lightweight and Efficient SAR Image Ship Detector. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
- Wei, H.; Wang, Z.; Hua, G.; Ni, Y. A Zero-Shot NAS Method for SAR Ship Detection Under Polynomial Search Complexity. IEEE Signal Process. Lett. 2024, 31, 1329–1333. [Google Scholar] [CrossRef]
- Chen, C.; Zeng, W.; Zhang, X.; Zhou, Y. CSnNet: A Remote Sensing Detection Network Breaking the Second-Order Limitation of Transformers with Recursive Convolutions. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4207315. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
- Liu, J.; Ma, F.; Yin, Q.; Zhang, F. An improved deep active learning method for SAR ship detection. In Proceedings of the IET International Radar Conference (IRC 2023), Chongqing, China, 3–5 December 2023; pp. 3336–3341. [Google Scholar]
- Wu, S.; Wang, W.; Ruan, F.; Zhang, H.; Deng, J.; Guo, P.; Fan, H. Inshore ship detection using high-resolution SAR images based on multi-feature fusion. In Proceedings of the IET International Radar Conference (IRC 2023), Chongqing, China, 3–5 December 2023; pp. 1067–1072. [Google Scholar]
- Sun, M.; Li, Y.; Chen, X.; Zhou, Y.; Niu, J.; Zhu, J. A Fast and Accurate Small Target Detection Algorithm Based on Feature Fusion and Cross-Layer Connection Network for the SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 8969–8981. [Google Scholar] [CrossRef]
- Zhou, Y.; Liu, H.; Ma, F.; Pan, Z.; Zhang, F. A Sidelobe-Aware Small Ship Detection Network for Synthetic Aperture Radar Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
- Hu, B.; Miao, H. An Improved Deep Neural Network for Small-Ship Detection in SAR Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sensing 2024, 17, 2596–2609. [Google Scholar] [CrossRef]
- Ge, R.; Mao, Y.; Li, S.; Wei, H. Research on Ship Small Target Detection in SAR Image Based on Improved YOLO-v7. In Proceedings of the 2023 International Applied Computational Electromagnetics Society Symposium (ACES-China), Hangzhou, China, 15–18 August 2023; pp. 1–3. [Google Scholar]
- Zhang, A.; Zhu, X. Research on ship target detection based on improved YOLOv5 algorithm. In Proceedings of the 2023 5th International Conference on Communications, Information System and Computer Engineering (CISCE), Guangzhou, China, 14–16 April 2023; pp. 459–463. [Google Scholar]
- Bai, L.; Yao, C.; Ye, Z.; Xue, D.; Lin, X.; Hui, M. Feature Enhancement Pyramid and Shallow Feature Reconstruction Network for SAR Ship Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1042–1056. [Google Scholar] [CrossRef]
- Zhang, W.; Wu, Q.M.J.; Yang, Y.; Akilan, T.; Zhao, W.G.W.; Li, Q.; Niu, J. Fast Ship Detection with Spatial-Frequency Analysis and ANOVA-Based Feature Fusion. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
Method | Principle | Advantages | Disadvantages | References |
---|---|---|---|---|
Approaches relying on auxiliary features | This approach primarily leverages auxiliary features in the vicinity of the target, such as the wake or oil spills from vessels. By detecting these auxiliary features, the presence of the target is indirectly inferred. | Simple and intuitive, requiring no complex sensors or data processing Effective target detection can be carried out when the target itself is difficult to directly detect but has obvious auxiliary features | There is a certain dependence on the detection environment, and effective detection may be challenging in situations where the target is directly visible but lacks clear auxiliary features | Ref. [4] Ref. [16] correctly classified 86 out of 93 SEASAT ship images and 21 out of 24 ocean scenes Ref. [17] |
Multi-polarization techniques based on polarization data | Enhancing target detection accuracy is achieved by analyzing the response of targets to different polarization modes using multi-polarization data acquired by SAR sensors. | Utilize the electromagnetic scattering characteristics of the target to provide additional information and enhance target identification and detection Enhances the contrast between targets and backgrounds in SAR images | Handling polarized data introduces complexity to the algorithm and processing workflow Imposes greater demands on sensors and data collection | Ref. [18] PMD = 3(10)−3, PFA = 2.5(10)−7(PFA: the probability of a false alarm, PMD: the probability of miss detection) |
CFAR methods | Based on the statistical distribution of background clutter, targets are detected by comparing pixel grayscale values with a threshold within a specific region. | Usually, statistical methods are used, and the calculations are relatively simple By considering local background statistical information to adapt to targets of different sizes, small targets can usually be effectively detected | It is prone to multiple target occlusion and false alarms at the edges of clutter Manual adjustment of parameters, such as window size and threshold, are required to adapt to different scenarios | Ref. [5] Ref. [19] Ref. [20] |
Deep learning-based detection methods | Employing deep learning models, such as convolutional neural networks (CNNs) or Recurrent Neural Networks (RNNs), features representing targets are learned from SAR images for target detection. | Capable of automatically learning complex feature representations without manual feature extraction Independent of specific prior knowledge, applicable to various scenarios Typically exhibits strong generalization capabilities and good robustness | Training the model typically requires a large amount of well-annotated data Models typically require high computational resources, especially for complex deep networks, which may require high-performance computing devices, such as GPUs or TPUs | Ref. [12]: SSDD, mAP = 90.08% Ref. [13]: GF3 SAR images, AP = 92.8% Ref. [14]: (high-resolution SAR images dataset) HRSID, mAP = 96.01% Ref. [15]: SSDD, mAP = 96.3% Ref. [21]: SSDD, mAP = 85.46% Ref. [22]: SSDD, mAP = 89.8% Ref. [23]: SSDD Offshore mAP = 98.2%, inshore mAP = 84.1% Ref. [24]: SAR ship dataset, mAP = 94.7% |
Method | Parameters/Trainable Params (M) | Training Time | Inference Time | FPS |
---|---|---|---|---|
YOLOX tiny | 5.05/5.03 | 52.21 s | 40.4 ms | 117.89 |
Ours | 8.46/8.42 | 48.96 s | 45.9 ms | 78.26 |
Method | Recall | Precision | F1 | AP50 | Parameters/Trainable Parameters (M) | FPS |
---|---|---|---|---|---|---|
YOLOv3 | 87.79% | 90.91% | 0.89 | 88.19% | 61.58/61.52 | 63.93 |
YOLOv4 tiny | 85.88% | 88.24% | 0.87 | 88.23% | 5.88/5.87 | 96.31 |
YOLOv4 | 91.60% | 93.02% | 0.92 | 92.00% | 64.00/63.94 | 52.54 |
YOLOv5-x | 94.66% | 96.12% | 0.95 | 94.44% | 87.34/87.24 | 25.86 |
YOLOX-x | 95.80% | 95.80% | 0.96 | 95.36% | 99.10/99.00 | 23.58 |
YOLOX tiny | 94.66% | 96.50% | 0.96 | 95.55% | 5.05/5.03 | 117.89 |
YOLOv7 tiny | 93.13% | 94.21% | 0.94 | 93.83% | 6.03/6.01 | 125.81 |
YOLOv8 tiny | 90.84% | 94.44% | 0.93 | 94.23% | 3.01/3.01 | 63.05 |
YOLOv8-s | 92.75% | 95.29% | 0.94 | 94.43% | 11.14/11.14 | 60.17 |
Efficientdet | 67.56% | 94.15% | 0.79 | 91.56% | 3.89/3.84 | 44.19 |
SSD | 48.85% | 93.43% | 0.64 | 85.67% | 23.61/23.61 | 103.90 |
FCOS | 92.37% | 94.90% | 0.94 | 96.23% | 32.17/32.12 | 22.48 |
CenterNet | 91.98% | 97.18% | 0.94 | 94.68% | 191.36/191.24 | 16.66 |
Faster R-CNN | 95.80% | 88.38% | 0.92 | 96.59% | 28.34/28.24 | 11.61 |
Ours | 96.95% | 97.69% | 0.97 | 97.87% | 8.46/8.42 | 78.26 |
Method | AP50 (Inshore) | AP50 (Offshore) | FPS |
---|---|---|---|
DAPN [34] | 67.55% | 95.93% | \ |
ARPN [35] | 84.1% | 98.2% | 13 |
Ours | 93.06% | 98.59% | 78.26 |
Method | AP | AP50 | AP75 |
---|---|---|---|
YOLOX tiny | 54.70% | 95.55% | 61.36% |
YOLOX tiny + LCA | 56.40% | 97.17% | 61.65% |
YOLOX tiny + AWF | 54.60% | 96.91% | 59.96% |
YOLOX tiny + MGAHead | 54.90% | 97.16% | 61.49% |
YOLOX tiny + LCA + AWF | 56.30% | 97.43% | 62.02% |
YOLOX tiny + MGAHead + LCA | 55.80% | 97.23% | 62.52% |
YOLOX tiny + MGAHead + AWF | 54.30% | 97.38% | 61.56% |
Ours | 55.90% | 97.87% | 63.67% |
Method | AP | AP50 | AP75 | APS | APm | APl |
---|---|---|---|---|---|---|
Basic | 54.7% | 95.55% | 61.36% | 47.1% | 65.3% | 62.5% |
SE | 51.6% | 95.82% | 59.51% | 43.1% | 63.5% | 71.2% |
ECA | 55.4% | 96.54% | 62.76% | 48.1% | 65.5% | 70.7% |
CBAM | 55.0% | 96.23% | 62.46% | 46.3% | 67.0% | 67.4% |
LCA | 56.4% | 97.17% | 61.65% | 49.5% | 66.0% | 70.3% |
Method | Recall | Precision | F1 | AP50 |
---|---|---|---|---|
YOLOv3 | 87.78% | 94.43% | 0.91 | 95.62% |
YOLOv4 | 79.83% | 91.97% | 0.86 | 89.72% |
YOLOv5-x | 95.44% | 93.32% | 0.94 | 96.91% |
YOLOX tiny | 93.12% | 93.25% | 0.93 | 96.49% |
YOLOX-x | 92.51% | 93.38% | 0.93 | 96.12% |
YOLOv7 tiny | 90.99% | 93.18% | 0.92 | 95.56% |
Efficientdet | 90.02% | 92.85% | 0.91 | 95.80% |
SSD | 84.66% | 93.34% | 0.89 | 94.72% |
Faster R-CNN | 93.18% | 93.18% | 0.93 | 95.63% |
DETR | 95.56% | 80.48% | 0.87 | 94.97% |
Ours | 95.78% | 92.46% | 0.94 | 97.21% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, Y.; Dong, C.; Guo, L.; Meng, X.; Liu, Y.; Wei, Q. AFMSFFNet: An Anchor-Free-Based Feature Fusion Model for Ship Detection. Remote Sens. 2024, 16, 3465. https://doi.org/10.3390/rs16183465
Zhang Y, Dong C, Guo L, Meng X, Liu Y, Wei Q. AFMSFFNet: An Anchor-Free-Based Feature Fusion Model for Ship Detection. Remote Sensing. 2024; 16(18):3465. https://doi.org/10.3390/rs16183465
Chicago/Turabian StyleZhang, Yuxin, Chunlei Dong, Lixin Guo, Xiao Meng, Yue Liu, and Qihao Wei. 2024. "AFMSFFNet: An Anchor-Free-Based Feature Fusion Model for Ship Detection" Remote Sensing 16, no. 18: 3465. https://doi.org/10.3390/rs16183465
APA StyleZhang, Y., Dong, C., Guo, L., Meng, X., Liu, Y., & Wei, Q. (2024). AFMSFFNet: An Anchor-Free-Based Feature Fusion Model for Ship Detection. Remote Sensing, 16(18), 3465. https://doi.org/10.3390/rs16183465