VHRShips: An Extensive Benchmark Dataset for Scalable Deep Learning-Based Ship Detection Applications
Abstract
:1. Introduction
Paper Contributions
- A database with a hierarchical level of classes which makes it suitable for all DLRI stages,
- A database that consists of a large variety of conditions in terms of geography, weather, and spatial arrangement,
- A method that follows all DLRI stages to present a proper remote sensing solution,
- A process that eliminates negative examples (images without ship), which are the largest percentage of images used in the majority of applications, before getting deeper stages,
- A method that forms a suitable base to adopt various deep learning networks and enables end-to-end and repeatable evaluation metrics which exemplify real scenarios.
2. Related Works
3. Materials and Methods
3.1. Background for Ship Dataset
Dataset | Image Source | Application Purpose | Number of Classes /Number of Ship Classes | Description Assessment |
---|---|---|---|---|
VHR-10_dataset_coco [41] | Optical Satellite (Google Earth) | Object classification | 10/1 | Ships on the images are located yet not classified. |
NWPU RESISC45 [42] | Optical Satellite | Scene classification | 45/1 | Ships on the images are located yet not classified. |
DOTA [43] | Optical Satellite | Object classification | 15/1 | Ships on the images are located yet not classified. |
HRSC2016 [23] | Optical Satellite (Google Earth) | Ship detection-localization & classification | 19/19 | Mostly in-shore images. 19 classes in Level 3. Experimental results only for Level 2 with 4 classes. |
Airbus Sandbox Ship Detection ML [44] | Optical Satellite (SPOT) | Ship detection-localization | 1/1 | Ships on the images are located yet not classified. |
xView [45] | Optical Satellite (WorldView-3) | Object classification | 60/9 | 8 parent classes with 60 child classes (maritime vessel parent. class contains 9 different child classes). |
HRRSD [46] | Optical Satellite (Google Earth and Baidu Map) | Object classification | 13/1 | Ships on the images are located yet not classified. |
FGSD [47] | Optical Satellite (Google Earth) | Ship detection-localization & classification | 44/43 | Mostly in-shore images. There are two classification levels. Level 1 consists of submarine, merchant ship, aircraft carrier and warship classes. |
BCCT200 [48] | Optical Satellite | Ship classification | 4/4 | A broad class definition with only gray-scale ship images from the RAPIER system. |
MASATI v2 [49] | Optical Satellite (Microsoft Bing Maps) | Ship detection-localization | 1/1 | Classes: Ship, multiple ships, ship with no coast, ship(s) with coast, sea with no ship, and land with no sea. |
FGSC-23 [22] | Optical Satellite (Google Earth and Gaofen-1 Satellite) | Ship classification | 23/23 | Three classification levels are defined which are ship-non ship (L1), coarse (L2) and grain (L3). |
- High definition images collected by optical satellites,
- Images with spatial resolution information to ensample a real scenario,
- Image metafile of an existing ship with location and class information,
- Defined recognition and identification levels of each ship in the dataset,
- Images from the inshore and offshore sea,
- Images with rural and urban coasts,
- Images from different locations around the world,
- Some images with clouds and wave clutter to sample a noisy background.
3.2. VHR Ship Dataset
3.3. Class Definitions
- Detection: The discovery of the ship’s existence (s) in the optical satellite image.
- Localization: Determination of the ship’s precise location (s) in the given optical satellite image.
- Recognition: Defining the parent class of each image among the civilian and navy ship groups.
- Identification: Defining the precise child class of each navy ship.
- dredging and bargePontoon classes are grouped into bargePontoon (representing the steady platforms),
- smallPassenger, smallBoat, tug, yacht, and fishing classes are grouped into smallBoat (small size ships),
- oilTanker and tanker classes are grouped into tanker (common tanker group),
- generalCargo, bulkCarrier, and oreCarrier are grouped into generalCargo (cargo ships),
- cruiser, frigate, patrolForce, and destroyer are grouped into destroyer (combat navy ship group).
3.4. Hardware and Software Configuration
3.5. Methodology
3.5.1. Detection
3.5.2. Localization
3.5.3. Recognition
- Analyzing the length-width ratios of the bBox and rotating it to ensure that the long side is horizontally aligned.
- Putting the rotated bBox in the center and overlaying the rotated bBox around the center bBox until the input size of the stage is filled.
- While overlaying the bBox, the reflection in the X and Y directions is applied to prevent the unintended gradient forming.
- The input is gathered through PFMO with 416 × 416 × 3 patch size.
- The output is the probability of 16 parent classes.
- No threshold is applied at the classification level output, the class with the highest score is selected.
3.5.4. Identification
4. Results and Discussion
4.1. Individual Stage Performances
- The coaster class consists of small water tankers, ore carriers, or cargo ships, which contain small ships that can be confused with small boats, as in this study. In particular, relatively low-resolution images do not include precise details and fail to detect compelling features of these small ships.
- Even though the number of ships in the drill ship class is minimal, the accuracy is close to the overall accuracy. Additionally, the drill class features are very similar to the barge-pontoon class. The drilling instruments in a drill ship are generally located on a barge-pontoon.
- The floating dock class is another class with limited sampling, illustrating a mixture with the barge-pontoon due to the similar shape form and the navy classes since floating docks carry navy ships in the test set.
- The offshore ship class has two characteristics: the navigating bridge on the bow side of the ship and a high bow freeboard. It seems that the number of samples in this class and the level of the differential features wasn’t enough to perform over the overall value.
- Some big yachts in the small boat class can be confused with passenger ships. A decline in accuracy regarding the passenger class relative to the overall results from this variety can be noted.
- The “Undefined” class comprises ships with different features from the defined classes or having a pure resolution that makes them impossible to be classified. Most ships in the “undefined” class do not have common features to be learned by the network. Thus, the low accuracy in this class is an expected result.
- The auxiliary class is mainly confused with the destroyer class. Both classes have similar sizes, and some auxiliary classes have similar features such as helicopter decks, sensor masts, and a navigational bridge at the bow.
- The same confusion exists between the coast guard and the destroyer classes. The coast guard ships mainly have the same form as the destroyers (especially with the frigates or the patrol ships). They differ in color (white for the coast guard class and navy gray for the destroyer class).
- The “other” class case is the same as the “undefined” class in the recognition stage.
4.2. End-to-End Performance
- The detection stage has the same results as the individual because it is the first stage to stay out of the error cycle. Only two images are missing, and each of them have one ship. The results of the detection phase are very satisfactory, except for the false positives.
- The localization stage is also not affected by the error cycle. Sixteen false true images detected in the previous stage are eliminated in this stage with true negative labeling. Thus, the results are very close to the individual evaluation. The number of missed ships is a bit high, which lowers the recall values of the following stages. On the other hand, a low false positive value increases the precision value of the next stages.
- The recognition starts with 274 errors from the previous two stages. For this reason, the accuracy and recall values are lower than they are in the individual evaluation.
- The identification evaluation metrics are a bit higher than 50 percent. Even though this may seem low, the relatively low number of ships belonging to navy classes, the compulsory samples in the dataset, and the below-average recognition performance (falseNegative and falsePositive2) of the navy class generated these results.
4.3. Comparison with the State of Art Network
4.4. Model Limitations and Future Work
- An additional sub-stage can be designed to detect and eliminate false positive samples for the localization, recognition, and identification stages, if further performance is needed. This design can also reduce the effects of the error cycle created by false positive outcomes.
- The distribution of unbalanced classes can be improved by implementing different over/under-sampling approaches, and the number of samples can be increased for the classes with limited data.
- Introducing more images with different classes will strengthen the dataset for the future.
- Different CNN networks can be integrated to increase the performance of the stages.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A
No. | Location | No. | Location |
---|---|---|---|
1 | Port of Shanghai | 22 | Port of Hong Kong |
2 | Port of Tianjin | 29 | Port of Shenzhen * |
3 | Port of Guangzhou | 30 | Suez Canal |
4 | Port of Ningbo | 31 | The Panama Canal |
5 | Port of Rotterdam | 32 | Naval Station Norfolk |
6 | Port of Qingdao | 33 | Naval Station Mayport |
7 | Port of Dalian | 34 | Devonport, Plymouth |
8 | Port of Busan | 35 | Portsmouth |
9 | Port of Nassau | 36 | Port of Canaveral |
10 | Port of Barcelona | 37 | Port of Everglades |
11 | Port of Civitavecchia | 38 | Port of Cozumel |
12 | Port of The Balearic Islands | 39 | Port of New York and New Jersey |
13 | Port of Southampton | 40 | Okinawa Naval Base White Beach |
14 | Snezhnogorsk | 41 | Yokosuka Naval Base |
15 | Tartus, Syria | 42 | Sasebo Naval Base |
16 | Venice | 43 | Toulon |
17 | Taranto Navy Base | 44 | Sevastopol |
18 | Navy Augusta | 45 | San Francisco |
19 | Port of Los Angeles | 46 | Dardanelles * |
20 | Port of Haydarpasa | 47 | Bosphorus |
21 | Port of Izmir | 48 | Port of Mersin |
23 | England | 49 | Port of Antalya |
24 | Jordan Agaba | 50 | Hanau Germany |
25 | Kocaeli gulf | 51 | Port of Mersin |
26 | Norway | 52 | Pearl Harbour |
27 | Qatar | 53 | Qatar Port of Raundabout |
28 | San Diego | 54 | Port of Singapore |
Appendix B
References
- Alganci, U.; Soydas, M.; Sertel, E. Comparative Research on Deep Learning Approaches for Airplane Detection from Very High-Resolution Satellite Images. Remote Sens. 2020, 12, 458. [Google Scholar] [CrossRef] [Green Version]
- Beşbinar, B.; Alatan, A.A. Inshore ship detection in high-resolution satellite images: Approximation of harbors using sea-land segmentation. In Image and Signal Processing for Remote Sensing XXI. Int. Soc. Opt. Photonics 2015, 9643, 687–698. [Google Scholar] [CrossRef]
- Bakırman, T.; Komurcu, I.; Sertel, E. Comparative analysis of deep learning based building extraction methods with the new VHR Istanbul dataset. Expert Syst. Appl. 2022, 202, 117346. [Google Scholar] [CrossRef]
- Kurekin, A.A.; Loveday, B.R.; Clements, O.; Quartly, G.D.; Miller, P.I.; Wiafe, G.; Adu Agyekum, K. Operational monitoring of illegal fishing in Ghana through exploitation of satellite earth observation and AIS data. Remote Sens. 2019, 11, 293. [Google Scholar] [CrossRef] [Green Version]
- Page, B.P.; Olmanson, L.G.; Mishra, D.R. A harmonized image processing workflow using Sentinel-2/MSI and Landsat-8/OLI for mapping water clarity in optically variable lake systems. Remote Sens. Environ. 2019, 231, 111284. [Google Scholar] [CrossRef]
- Zollini, S.; Alicandro, M.; Cuevas-González, M.; Baiocchi, V.; Dominici, D.; Buscema, P.M. Shoreline extraction based on an active connection matrix (ACM) image enhancement strategy. J. Mar. Sci. Eng. 2019, 8, 9. [Google Scholar] [CrossRef] [Green Version]
- Mishra, D.; Narumalani, S.; Rundquist, D.; Lawson, M. Benthic habitat mapping in tropical marine environments using QuickBird multispectral data. Photogramm. Eng. Remote Sens. 2006, 72, 1037–1048. [Google Scholar] [CrossRef]
- Corbane, C.; Najman, L.; Pecoul, E.; Demagistri, L.; Petit, M. A complete processing chain for ship detection using optical satellite imagery. Int. J. Remote Sens. 2010, 31, 5837–5854. [Google Scholar] [CrossRef]
- Early, B.R.; Gartzke, E. Spying from space: Reconnaissance satellites and interstate disputes. J. Confl. Resolut. 2021, 65, 1551–1575. [Google Scholar] [CrossRef]
- Norris, P. Developments in high resolution imaging satellites for the military. Space Policy 2011, 27, 44–47. [Google Scholar] [CrossRef]
- Li, B.; Xie, X.; Wei, X.; Tang, W. Ship detection and classification from optical remote sensing images: A survey. Chin. J. Aeronaut. 2021, 34, 145–163. [Google Scholar] [CrossRef]
- Kanjir, U.; Greidanus, H.; Oštir, K. Vessel detection and classification from spaceborne optical images: A literature survey. Remote Sens. Environ. 2018, 207, 1–26. [Google Scholar] [CrossRef]
- McDonnell, M.J.; Lewis, A.J. Ship detection from LANDSAT imagery. Photogramm. Eng. Remote Sens. 1978, 44, 297–301. [Google Scholar]
- Jin, T.; Zhang, J. Ship detection from high-resolution imagery based on land masking and cloud filtering. In Seventh International Conference on Graphic and Image Processing (ICGIP 2015); SPIE: Bellingham WA, USA, 2015; Volume 9817, p. 981716. [Google Scholar] [CrossRef]
- You, X.; Li, W. A sea-land segmentation scheme based on statistical model of sea. In Proceedings of the 2011 4th International Congress on Image and Signal Processing, Shanghai, China, 15–17 October 2011; Volume 3, pp. 1155–1159. [Google Scholar] [CrossRef]
- Li, R.; Liu, W.; Yang, L.; Sun, S.; Hu, W.; Zhang, F.; Li, W. DeepUNet: A deep fully convolutional network for pixel-level sea-land segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3954–3962. [Google Scholar] [CrossRef] [Green Version]
- Yang, G.; Li, B.; Ji, S.; Gao, F.; Xu, Q. Ship detection from optical satellite images based on sea surface analysis. IEEE Geosci. Remote Sens. 2013, 11, 641–645. [Google Scholar] [CrossRef]
- Xie, X.; Xu, Q.; Hu, L. Fast ship detection from optical satellite images based on ship distribution probability analysis. In Proceedings of the 4th International Workshop on Earth Observation and Remote Sensing Applications (EORSA), Guangzhou, China, 4–6 July 2016; pp. 97–101. [Google Scholar] [CrossRef]
- Li, S.; Zhou, Z.; Wang, B.; Wu, F. A novel inshore ship detection via ship head classification and body boundary determination. IEEE Geosci. Remote Sens. 2016, 13, 1920–1924. [Google Scholar] [CrossRef]
- Li, W.; Fu, K.; Sun, H.; Sun, X.; Guo, Z.; Yan, M.; Zheng, X. Integrated localization and recognition for inshore ships in large scene remote sensing images. IEEE Geosci. Remote Sens. 2017, 14, 936–940. [Google Scholar] [CrossRef]
- Li, X.; Li, Z.; Lv, S.; Cao, J.; Pan, M.; Ma, Q.; Yu, H. Ship detection of optical remote sensing image in multiple scenes. Int. J. Remote Sens. 2021, 1–29. [Google Scholar] [CrossRef]
- Zhang, X.; Lv, Y.; Yao, L.; Xiong, W.; Fu, C. A New Benchmark and an Attribute-Guided Multilevel Feature Representation Network for Fine-Grained Ship Classification in Optical Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1271–1285. [Google Scholar] [CrossRef]
- Liu, Y.; Cui, H.Y.; Kuang, Z.; Li, G.Q. Ship detection and classification on optical remote sensing images using deep learning. ITM Web Conf. EDP Sci. 2017, 12, 05012. [Google Scholar] [CrossRef] [Green Version]
- Liu, Z.; Yuan, L.; Weng, L.; Yang, Y. A high resolution optical satellite image dataset for ship recognition and some new baselines. In Proceedings of the International Conference on Pattern Recognition Applications and Method, Porto, Portugal, 24–26 February 2017; SciTePress: Setúbal, Portugal, 2017; Volume 2, pp. 324–331. [Google Scholar] [CrossRef]
- Heiselberg, P.; Heiselberg, H. Ship-Iceberg discrimination in Sentinel-2 multispectral imagery by supervised classification. Remote Sens. 2017, 9, 1156. [Google Scholar] [CrossRef] [Green Version]
- Kanjir, U. Detecting migrant vessels in the Mediterranean Sea: Using Sentinel-2 images to aid humanitarian actions. Acta Astronaut. 2019, 155, 45–50. [Google Scholar] [CrossRef]
- Zhang, R.; Yao, J.; Zhang, K.; Feng, C.; Zhang, J. S-CNN-based ship detection from high-resolution remote sensing images. Int. Arch. Photogramm. 2016, 41, 423–430. [Google Scholar] [CrossRef]
- Zhang, S.; Wu, R.; Xu, K.; Wang, J.; Sun, W. R-CNN-based ship detection from high resolution remote sensing imagery. Remote Sens. 2019, 11, 631. [Google Scholar] [CrossRef] [Green Version]
- Topputo, F.; Massari, M.; Lombardi, R.; Gianinetto, M.; Marchesi, A.; Aiello, M.; Banda, F. Space shepherd: Search and rescue of illegal immigrants in the mediterranean sea through satellite imagery. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4852–4855. [Google Scholar] [CrossRef]
- Xing, Q.; Meng, R.; Lou, M.; Bing, L.; Liu, X. Remote sensing of ships and offshore oil platforms and mapping the marine oil spill risk source in the Bohai Sea. Aquat. Pract. 2015, 3, 127–132. [Google Scholar] [CrossRef]
- Liu, Y.; Yao, L.; Xiong, W.; Zhou, Z. Fusion detection of ship targets in low resolution multi-spectral images. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 6545–6548. [Google Scholar] [CrossRef]
- Tang, J.; Deng, C.; Huang, G.B.; Zhao, B. Compressed-domain ship detection on spaceborne optical image using deep neural network and extreme learning machine. IEEE Trans. Geosci. Remote Sens. 2014, 53, 1174–1185. [Google Scholar] [CrossRef]
- Zou, Z.; Shi, Z. Ship detection in space borne optical image with SVD networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 5832–5845. [Google Scholar] [CrossRef]
- Chen, L.; Shi, W.; Fan, C.; Zou, L.; Deng, D. A novel coarse-to-fine method of ship detection in optical remote sensing images based on a deep residual dense network. Remote Sens. 2020, 12, 3115. [Google Scholar] [CrossRef]
- Shi, Q.; Li, W.; Tao, R.; Sun, X.; Gao, L. Ship classification based on multi feature ensemble with convolutional neural network. Remote Sens. 2019, 11, 419. [Google Scholar] [CrossRef] [Green Version]
- Zhuang, Y.; Li, L.; Chen, H. Small sample set inshore ship detection from VHR optical remote sensing images based on structured sparse representation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2145–2160. [Google Scholar] [CrossRef]
- Gao, L.; He, Y.; Sun, X.; Jia, X.; Zhang, B. Incorporating Negative Sample Training for Ship Detection Based on Deep Learning. Sensors 2019, 19, 684. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Feng, Y.; Diao, W.; Sun, X.; Yan, M.; Gao, X. Towards Automated Ship Detection and Category Recognition from High-Resolution Aerial Images. Remote Sens. 2019, 11, 1901. [Google Scholar] [CrossRef] [Green Version]
- Sasikala, J. Ship detection and recognition for offshore and inshore applications: A survey. Int. J. Intell. Unman. Syst. 2019, 7, 177–188. [Google Scholar] [CrossRef]
- Ekim, B.; Sertel, E. Deep neural network ensembles for remote sensing land cover and land use classification. Int. J. Digit. Earth 2021, 14, 1868–1881. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J.; Zhou, P.; Guo, L. Multi-class geospatial object detection and geographic image classification based on collection of part detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef] [Green Version]
- Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
- Kaggle. Airbus Ship Detection Challenge. Available online: https://www.kaggle.com/c/airbus-ship-detection/data (accessed on 6 October 2021).
- Lam, D.; Kuzma, R.; McGee, K.; Dooley, S.; Laielli, M.; Klaric, M.; McCord, B. xview: Objects in context in overhead imagery. arXiv 2018, arXiv:1802.07856. [Google Scholar]
- Zhang, Y.; Yuan, Y.; Feng, Y.; Lu, X. Hierarchical and robust convolutional neural network for very high-resolution remote sensing object detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5535–5548. [Google Scholar] [CrossRef]
- Chen, K.; Wu, M.; Liu, J.; Zhang, C. FGSD: A dataset for fine-grained ship detection in high resolution satellite images. arXiv 2020, arXiv:2003.06832. [Google Scholar]
- Rainey, K.; Parameswaran, S.; Harguess, J.; Stastny, J. Vessel classification in overhead satellite imagery using learned dictionaries. In Applications of Digital Image Processing XXXV. Int. Soc. Opt. Photonics 2012, 8499, 84992F. [Google Scholar] [CrossRef]
- Gallego, A.J.; Pertusa, A.; Gil, P. Automatic ship classification from optical aerial images with convolutional neural networks. Remote Sens. 2018, 10, 511. [Google Scholar] [CrossRef] [Green Version]
- Dataset, the Codes and VHRShips Test Data to Evaluate HieD Approach. 2022. Available online: https://github.com/radres333/VHRShips (accessed on 7 June 2022).
- STANAG 3769. In Minimum Resolved Object Sizes and Scales for Imagery Interpretation, AIR STD 80/15, 2nd ed.; Air Standards: Washington, DC, USA, 1970; pp. 20330–25058.
- Register, L. Rules and Regulations for the Classification of Ships; Lloyd’s Register: London, UK, 2018. [Google Scholar]
- Saunders, S. (Ed.) Jane’s Fighting Ships, 2003–2004; Jane’s Information Group: Coulsdon, Surrey, UK; Alexandria, VA, USA, 2002. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar] [CrossRef] [Green Version]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
- Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar] [CrossRef] [Green Version]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
Property | Value/Description |
---|---|
Number of images | 6312 |
Number of images with ship(s) | 5312 |
Number of images without ship | 1000 |
No. of ships | 11,179 |
No. of classes | 24 parent, 11 child |
Image spectral channel | red-green-blue |
Image spatial resolution | 1280 × 720 |
Image source | Google Earth Professional |
Bounding box (bBox) type | Horizontal-rectangle |
bBox statistics (min–mean–max) | Length: 7–183–991 Width: 6–47–428 bBox area: 42–13,640–282,226 |
Class Distribution-Number of Ships | ||
---|---|---|
Parent classes (23 civilian classes and 1 navy class) and non-ship images | ||
Lpg: 33 | drill: 33 | floatingDock: 51 |
Ferry: 99 | Roro: 102 | Offshore: 122 |
Passenger: 143 | Dredging: 194 | bargePontoon: 714 |
dredgerReclamation: 268 | Coaster: 344 | Undefined: 417 |
smallPassenger: 419 | smallBoat: 891 | Tug: 846 |
Yacht: 1581 | Fishing: 37 | Container: 580 |
oilTanker: 594 | Tanker: 777 | generalCargo: 655 |
bulkCarrier: 677 | oreCarrier: 771 | Navy: 831 |
nonShip: 1000 | ||
Child classes (navy): | ||
Other: 27 | Aircraft: 35 | Landing: 34 |
coastGuard: 40 | Submarine: 50 | Cruiser: 68 |
Frigate: 71 | patrolForce: 101 | Destroyer: 90 |
serviceCraft: 151 | Auxilary: 164 |
Hardware Parameters | |
Parameter | Value/Type |
Processor | Intel Core i7-6700 CPU 3.4 GHz |
RAM | 32 GB |
System | 64 bit Windows 10 |
GPU Name | NVIDIA GeForce GTX 1060 3 GB |
GPU Compute Capability | 6.1 |
GPU Total Memory | 3.22 GB (2.5 GB available for computing) 1152 CUDA cores 128 bit |
Software Parameters | |
Software Name | MATLAB 2021a Academic Version Google Colaboratory |
Software Libraries | MATLAB Computer Vision Toolbox 10.0 MATLAB GPU Coder Support Package |
MATLAB Parallel Computing Toolbox 7.4 MATLAB Deep Learning Toolbox 14.2 MATLAB Embedded Coder 7.6 MATLAB Coder 5.2 MATLAB Image Processing Toolbox 11.3 CUDA 10.1 cuDNN 9.0 |
Parameter | Value/Type |
---|---|
Network | Xception (MATLAB version 21.1.1) pre-trained with ImageNet database |
Network input | 416 × 416 × 3 (original images are resized) |
Network output | ship/non-ship (binary classification) |
Training optimizer | Stochastic Gradient Descent with Momentum (SGDM) |
Loss function | Cross-entropy |
Data augmentation | Random rotation [0, 360] & random X and Y reflection |
Initial learning rate (LR) | 0.001 |
LR drop frequency | Every 10 epochs |
LR drop factor | 0.1 |
Number of epoch | 50 |
Mini-batch size | 8 |
Detection threshold | 0.2 |
Parameter | Value/Type |
---|---|
Network | YOLO v4 pre-trained with Coco dataset |
Backbone | Cross Stage Partial Network Darknet-53 |
Network input | 608 × 608 × 3 (original images are resized) |
Network output | [x, y, width, height] bBox with the score |
Loss function | Complete Intersection Over Union (CIoU) |
Data augmentation | Cutmix |
Initial learning rate (LR) | 0.001 |
LR decay factor | 0.0005 |
Number of epochs | 160 |
Batch size | 64 |
Mini batch size | 2 |
bBox score threshold | 0.4 |
Ground truth overlap threshold | 0.05 |
Non-maximum suppression threshold | 0.75 |
Stage | Accuracy | Recall | Precision | F1-Score |
---|---|---|---|---|
Detection | 98.59 | 99.81 | 98.53 | 99.17 |
Localization | 89.04 | 91.04 | 97.60 | 94.20 |
Recognition | 84.31 | 72.57 1 | 78.33 1 | 75.34 1 |
84.30 2 | 83.85 2 | 84.08 2 | ||
Identification | 80.99 | 83.67 1 | 86.63 1 | 85.12 1 |
80.98 2 | 83.31 2 | 82.13 2 |
Predicted Class | |||
---|---|---|---|
Non-Ship | Ship | ||
True Class | Non-ship | 92 | 8 |
Ship | 0.2 | 99.8 |
Predicted Class | Recall | |||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
bargePontoon | Coaster | Container | dredgerRec | Drill | Ferry | floatingDock | generalCargo | Lpg | Navy | Offshore | Passenger | Roro | smallBoat | Tanker | Undefined | |||
True Class | bargePontoon | 81.7 | 0.6 | 1.1 | 0.6 | 3.9 | 5.0 | 0.6 | 4.4 | 1.7 | 0.6 | 82 | ||||||
Coaster | 3.7 | 31.5 | 1.9 | 7.4 | 1.9 | 29.6 | 13.0 | 11.1 | 31 | |||||||||
Container | 0.9 | 78.4 | 20.7 | 78 | ||||||||||||||
dredgerRec | 3.7 | 94.4 | 1.9 | 94 | ||||||||||||||
Drill | 16.7 | 66.7 | 16.7 | 67 | ||||||||||||||
Ferry | 11.8 | 47.1 | 5.9 | 5.9 | 17.6 | 5.9 | 5.9 | 47 | ||||||||||
floatingDock | 25 | 50 | 8.3 | 16.7 | 50 | |||||||||||||
generalCargo | 1.7 | 0.7 | 1.0 | 90.4 | 0.5 | 0.2 | 1.2 | 2.4 | 1.9 | 90 | ||||||||
Lpg | 100 | 100 | ||||||||||||||||
Navy | 2.1 | 0.7 | 1.4 | 0.7 | 1.4 | 79.6 | 0.7 | 7.0 | 2.1 | 4.2 | 80 | |||||||
Offshore | 8.0 | 4.0 | 4.0 | 68.0 | 4.0 | 4.0 | 8.0 | 68 | ||||||||||
Passenger | 3.7 | 3.7 | 70.4 | 14.8 | 7.4 | 70 | ||||||||||||
Roro | 5.0 | 5.0 | 5.0 | 85 | 85 | |||||||||||||
smallBoat | 1.2 | 1.2 | 0.4 | 3.6 | 0.1 | 0.3 | 91.7 | 0.5 | 0.9 | 92 | ||||||||
Tanker | 0.8 | 1.5 | 0.4 | 3.4 | 0.4 | 0.4 | 2.3 | 91.0 | 91 | |||||||||
Undefined | 15.2 | 2.5 | 1.3 | 1.3 | 7.6 | 6.3 | 3.8 | 1.3 | 22.8 | 2.5 | 35.4 | 35 | ||||||
Precision | 76 | 46 | 92 | 100 | 80 | 62 | 86 | 86 | 88 | 69 | 74 | 76 | 94 | 91 | 89 | 46 |
Predicted Class | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
Aircraft | Auxiliary | coastGuard | Destroyer | Landing | Other | serviceCraft | Submarine | Recall | ||
True Class | Aircraft | 100 | 100 | |||||||
Auxiliary | 5 | 60 | 30 | 5 | 60 | |||||
coastGuard | 70 | 30 | 70 | |||||||
Destroyer | 1.8 | 3.6 | 1.8 | 92.7 | 93 | |||||
Landing | 100 | 100 | ||||||||
Other | 33.3 | 66.7 | 67 | |||||||
serviceCraft | 10 | 10 | 80 | 80 | ||||||
Submarine | 100 | 100 | ||||||||
Precision | 67 | 89 | 88 | 75 | 75 | 100 | 100 | 100 |
Metric | Detection | Localization | Recognition | Identification |
---|---|---|---|---|
falseNegative | 2 (No. of images labeled as without ship(s) even with ship(s)) | 226 (No. of ships which are missed) | 228 (No. of ships which couldn’t be detected & localized, so couldn’t be recognized) | 29 (No. of ships couldn’t be recognized as Navy) |
falseNegative2 | NA (invalid metric for this stage) | 2 (No. of ships couldn’t be detected, so couldn’t be localized) | NA (invalid metric for this stage) | 7 (No. of ships which couldn’t be localized, so couldn’t be identified) |
falsePositive | 16 (No. of images labeled as with ship(s) even without ship (s)) | 46 (bBoxes labeled as ship even not) | 405 (No. of ships wrongly recognized) | 21 (No. of Navy ships wrongly identified) |
falsePositive2 | NA (invalid metric for this stage) | NA (invalid metric for this stage) | 46 (No. of bBox labeled as ship even not, so incorrectly recognized) | 50 (No. of ships recognized as Navy even not, so incorrectly identified) |
truePositive | 1073 (No. of images with ship(s)s) | 1948 (No. of true localized ships) | 1543 GTruth (No. of ships truly recognized) | 85 (No. of ships truly identified) |
trueNegative | 184 (No. of images without ship(s)) | 16 (No. of images labeled as with ship(s) but none of ship localized) | NA (invalid metric for this stage) | NA (invalid metric for this stage) |
Accuracy | 98.59% | 87.76% | 69.44% | 44.27% |
Recall | 99.81% | 89.52% | 70.91% | 59.86% |
Precision | 98.53% | 97.69% | 77.38% | 54.49% |
F1-Score | 99.17% | 93.43% | 74.00% | 57.05% |
Stage/Metric | Proposed HieD | YOLO v4 | ||||||
---|---|---|---|---|---|---|---|---|
Accuracy | Recall | Precision | F1-Score | Accuracy | Recall | Precision | F1-Score | |
Detection | 98.59% | 99.81% | 98.53% | 99.17% | 98.59% | 99.44% | 98.89% | 99.17% |
Localization | 87.76% | 89.52% | 97.69% | 93.43% | 76.35% | 86.41% | 86.77% | 86.59% |
Recognition | 69.44% | 70.91% | 77.38% | 74.00% | 60.72% | 68.72% | 69.01% | 68.87% |
Identification | 44.27% | 59.86% | 54.49% | 57.05% | 43.05% | 45.77% | 73.03% | 56.28% |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kızılkaya, S.; Alganci, U.; Sertel, E. VHRShips: An Extensive Benchmark Dataset for Scalable Deep Learning-Based Ship Detection Applications. ISPRS Int. J. Geo-Inf. 2022, 11, 445. https://doi.org/10.3390/ijgi11080445
Kızılkaya S, Alganci U, Sertel E. VHRShips: An Extensive Benchmark Dataset for Scalable Deep Learning-Based Ship Detection Applications. ISPRS International Journal of Geo-Information. 2022; 11(8):445. https://doi.org/10.3390/ijgi11080445
Chicago/Turabian StyleKızılkaya, Serdar, Ugur Alganci, and Elif Sertel. 2022. "VHRShips: An Extensive Benchmark Dataset for Scalable Deep Learning-Based Ship Detection Applications" ISPRS International Journal of Geo-Information 11, no. 8: 445. https://doi.org/10.3390/ijgi11080445
APA StyleKızılkaya, S., Alganci, U., & Sertel, E. (2022). VHRShips: An Extensive Benchmark Dataset for Scalable Deep Learning-Based Ship Detection Applications. ISPRS International Journal of Geo-Information, 11(8), 445. https://doi.org/10.3390/ijgi11080445