Next Article in Journal
Himawari-8/AHI Aerosol Optical Depth Detection Based on Machine Learning Algorithm
Previous Article in Journal
Performance Assessment of BDS-3 PPP-B2b/INS Loosely Coupled Integration
Previous Article in Special Issue
Analysis of Environmental and Atmospheric Influences in the Use of SAR and Optical Imagery from Sentinel-1, Landsat-8, and Sentinel-2 in the Operational Monitoring of Reservoir Water Level
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessment of Machine Learning Techniques for Oil Rig Classification in C-Band SAR Images

1
Department of Telecommunications, Aeronautics Institute of Technology (ITA), São José dos Campos 12228-900, Brazil
2
Department of Mathematics and Natural Sciences, Blekinge Institute of Technology (BTH), 371 79 Karlskrona, Sweden
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(13), 2966; https://doi.org/10.3390/rs14132966
Submission received: 1 May 2022 / Revised: 14 June 2022 / Accepted: 17 June 2022 / Published: 21 June 2022
(This article belongs to the Special Issue Advances in Spaceborne SAR – Technology and Applications)

Abstract

:
This article aims at performing maritime target classification in SAR images using machine learning (ML) and deep learning (DL) techniques. In particular, the targets of interest are oil platforms and ships located in the Campos Basin, Brazil. Two convolutional neural networks (CNNs), VGG-16 and VGG-19, were used for attribute extraction. The logistic regression (LR), random forest (RF), support vector machine (SVM), k-nearest neighbours (kNN), decision tree (DT), naive Bayes (NB), neural networks (NET), and AdaBoost (ADBST) schemes were considered for classification. The target classification methods were evaluated using polarimetric images obtained from the C-band synthetic aperture radar (SAR) system Sentinel-1. Classifiers are assessed by the accuracy indicator. The LR, SVM, NET, and stacking results indicate better performance, with accuracy ranging from 84.1% to 85.5%. The Kruskal–Wallis test shows a significant difference with the tested classifier, indicating that some classifiers present different accuracy results. The optimizations provide results with more significant accuracy gains, making them competitive with those shown in the literature. There is no exact combination of methods for SAR image classification that will always guarantee the best accuracy. The optimizations performed in this article were for the specific data set of the Campos Basin, and results may change depending on the data set format and the number of images.

1. Introduction

The seas and oceans are natural resources of oil, natural gas, fauna, and flora, rich in biodiversity and ecosystems. Because of those attributes, the Brazilian marine area is known as the “Blue Amazon” [1]. The Campos Basin, one of the largest oil and gas producers in Brazil, is located on the northern coast of Rio de Janeiro, between the cities of Vitória (ES) and Cabo Frio (RJ), covering an area of 120,000 km2 [2].
The production capacity of oil and natural gas at sea can be evidenced by the results for July 2021 in the Monthly Bulletin number 131 [3], highlighting the Campos and Santos basins. For this period, the offshore oil and natural gas production was at about 97.1% and 82.5% capacity, respectively [3]. In particular, the Campos and Santos Basins were responsible for 23.74% and 69.07% of national oil and natural gas production, respectively. Consequently, the sea is of strategic importance to the Brazilian economy. Its constant surveillance is necessary to avoid illegal exploitation, such as fishing, oil spills, maritime traffic, and piracy [4].
Remote sensing allows the surveillance of large maritime areas [5]. The US Seasat mission provided the first earth-imaging synthetic aperture radar (SAR) from space, launched in 1978. The European Space Agency developed and launched the first microwave SAR systems, ERS-1 (launched on 17 July 1991) and ERS-2 (launched on 20 April 1995) [6]. SAR systems have been applied in earth remote sensing applications for more than 30 years due to their ability to provide high-resolution images, independent of the weather conditions and sunlight illumination [7]. Furthermore, SAR images play an important role in maritime surveillance [8,9], being useful for civil and military applications [10], such as detecting environmental accidents related to oil spills [11,12].
SAR images can be difficult to interpret with human vision. To overcome this limitation, automatic target recognition (ATR) can be applied to extract features that contain unique identifying information about the targets. ATR consists of three stages, known as detection, recognition, and classification [13]. This article will focus on the classification step [14], and more precisely, on the classification of two types of targets, oil rigs and ships.
Recently, artificial intelligence (AI) techniques such as machine learning (ML) [15] and deep learning (DL) [13] have been widely used for target classification in different SAR image applications [16,17,18]. As discussed in [16], a hybrid neural network was employed in Sentinel-1 and RADARSAT-2 SAR images for classification purposes. The hybrid algorithm consisted of a convolutional neural network (CNN) and a multilayer perceptron (MLP). As a result, the classification performance increased. ML is a subset area of AI that enables computer systems to learn from past experiences and improve their behavior for specific tasks. Typical examples of ML schemes are support vector machines (SVM), decision tree (DT), naive Bayes (NB), and k-means clustering [19,20]. Neural networks (NET) are a subset of ML inspired by biological neural networks, being represented by artificial neurons connected in layers [20,21], while DL is an NET technique that organizes neurons into several layers [20,22], for example, deep neural network (DNN) and CNN [20]. DL has become popular in object detection due to its ability to learn how to discriminate features automatically [23]. Furthermore, ML techniques, such as logistic regression (LR) and NET, have shown to be useful for oil slick detection in SAR images from Sentinel-1 orbital systems [11]. For instance, in [24], a CNN was used to increase the training and test data sets in ship detection in Gaofen-3 SAR images. A SAR ATR CNN-based application was performed with the public MSTAR data set in [25] to classify military vehicles in SAR images, displaying competitive results with the literature in terms of accuracy.
This article compares ML techniques, such as random forest (RF), LR, NET, SVM, AdaBoost (ADBST), k-nearest-neighbours (kNN), NB, and DT, for classifying maritime targets in high-resolution SAR images. For attribute extraction, two visual geometry group deep learning techniques (VGG-16 and VGG-19) are considered. In particular, we aim to classify oil rigs and ships in the Campos Basin, on the coast of Rio de Janeiro and Espírito Santo, Brazil, considering vertical-horizontal (VH) and vertical-vertical (VV) polarimetric images. For that, we considered the following methods: (M1) reproduction of results already presented in the literature, specifically the ones discussed in [26], which, to the best of our knowledge, is the only study related to the Campos Basin data set; (M2) evaluation of the sensitivity of the classifiers; (M3) expansion of the training data set concatenation considering the whole of the VH and VV polarization samples; (M4) expansion of the training data set concatenating half of the VH and VV polarization samples; and (M5) combining classifiers to obtain better accuracy results (a technique named stacked generalization).
To achieve this objective, the aims of this article are as follows:
(i)
Evaluate the combination of machine learning methods associated with classification with deep learning for feature extraction;
(ii)
Evaluate ML methods in the classification of SAR images in the C-Band and DL for attribute extraction;
(iii)
Reproduce the results of [26] to validate the methodology;
(iv)
Identify parameter adjustments with more significant gains than those achieved in the literature;
(v)
Combine data sets to increase the variability of training samples;
(vi)
Combine classification techniques;
(vii)
Analyze the performance of VGG-16 and VGG-19 networks as feature generators for classification algorithms;
(viii)
Evaluate the significance of the results through non-parametric statistics.
The remainder of the paper is organized as follows. Section 2.1 presents the Sentinel-1 mission and the employed data set. In Section 2.2, the used classification tools are presented. The classification setup applied in this study is described in Section 2.3. The numerical analyses are included in Section 3. Finally, Section 4 concludes the paper.

2. Materials and Methods

This section presents the characteristics of the Sentinel-1 mission, the classification techniques, and the methodology used in this article.

2.1. Sentinel-1 Mission

The data set employed in this studied was obtained through the Sentinel-1 mission, which is composed of two satellites, namely Sentinel-1A and Sentinel-1B, launched in 3 April 2014 and 25 April 2016, respectively [27]. They are in a sun-synchronized near-polar orbit, operating day and night, with a 12-day repetition cycle and an altitude of 693 km, and they perform C-band SAR imaging [28,29]. This satellite has an SAR sensor capable of generating medium- and high-resolution measurements [28]. In Table 1, some of the Sentinel-1 system characteristics, such as the operating band, bandwidth, antenna size, antenna weight, and pulse repetition frequency, are presented [27]. The Sentinel-1 systems support single- (HH or VV) and dual-polarization (HH + HV or VV + VH) operations, implemented by a transmit chain (switchable between H or V) and two parallel receive chains for H and V polarization. Additionally, the stripmap (SM), interferometric wide swath (IW), and extra-wide swath (EW) products are available with single or dual-polarization. However, the waver product is only available with single polarization [27].
According to [27], Sentinel-1 can acquire data in four modes, which are described in the following and shown in Figure 1. First, SM is a standard SAR stripmap imaging mode. A continuous sequence of pulses with a fixed elevation angle illuminates a strip of ground. Second, in IW mode, the data are acquired in three bands using the terrain observation with progressive scanning SAR (TOPSAR) imaging technique. Third, the data are acquired on five swaths using the TOPSAR imaging technique. The EW mode provides extensive swath coverage at the expense of spatial resolution. Finally, in WV mode, the data are obtained in small stripmap scenes called “vignettes”, located at regular 100 km intervals along the swath.
The four acquisition modes (SM, IW, EW, WV) can generate SAR level 0, level 1 SLC, level 1 GRD, and level 2 OCN products [27], as shown in Figure 2. The product used in our research is level 1 GRD with high-resolution SAR images, in IW mode, as shown in Figure 3. It consists of focused SAR data detected, multi-looked, and projected to the ground range using an ellipsoid model. Table 2 shows some examples of applications separated by operating modes.

Sentinel-1 Image Data Set

For this study, the Sentinel-1 SAR image data set contains 400 images (patches) in VH and VV polarization with maritime targets (platforms and ships), equally distributed (i.e., 200 patches with platforms and 200 patches with ships). Image patches were acquired at different times. There are targets with more than one patch. Despite being from the same target, the patches can be considered distinct because of the SAR image formation process, which is influenced by backscatter and sea currents that cause displacement on the platforms.
Following the methodology employed in [26], the original amplitude-type images were transformed into sigma-zero (dB) images. Figure 4 presents an optical image and its respective SAR image for the following targets: (i) Floating Production Storage and Offloading (FPSO) platforms P-48; (ii) Floating and Production Unit (FPU) P-53; (iii) Tension Leg Wellhead Platform (TLWP) P-61; (iv) Fixed Platform (FIX) PCH-1; and (v) Semisubmersible (SS) P-65. These images were collected with the ground-range-detected product; interferometric wide swath mode; high spatial resolution ( 20 m × 22 m —range × azimuth); pixel spacing equal to 10 m × 10 m in range and azimuth, respectively; and 5 × 1 number of looks (the equivalent number of looks is 4.4 [30]).
The legends ( υ 1 , υ 2 , υ 3 , υ 4 , υ 5 , υ 6 , υ 7 , υ 8 ) represent the images with VH polarization and the legends ( ι 1 , ι 2 , ι 3 , ι 4 , ι 5 , ι 6 , ι 7 , ι 8 ) represent the images with VV polarization, taken on the dates mentioned in Table 3.

2.2. Classification Tools

This section presents the classifier methods employed in this article. In particular, a classifier can be defined as a function f that maps the input vectors of features, x χ , into the output class labels, y 1 , 2 , , C , where χ is the attribute space and C is the number of classes. Usually, it is assumed that χ = R D or χ = 0 , 1 D , that is, that the attribute vector is a vector of D real numbers or binary bits [32]. Classification is one of the most important topics in data mining, especially for large amounts of data (big data) applications. The main task of classification is to predict the labels of the test data based on the training data [33]. In the following, the employed classifiers in this study are presented. The first considered method is the SVM scheme, which is a class of statistical models first developed in the 1960s by Vladimir Vapnik [34] that can be used for classification [35]. SVM has become popular due to its applicability in a variety of contexts, such as extreme learning machines [34], automatic target recognition for CNN-based in SAR images [36], SAR ATR, and independent component analysis [37].
The second scheme applied in our study is the DT, which is a nonparametric supervised learning method used for classification and regression [38].
The main idea of this algorithm is that the trees learn how to approximate a sine curve with a set of decision rules. They are visualized using a graph, which makes them easy to interpret. In addition, they require little previous information and can handle both numerical and categorical data. On the other hand, instability due to small variations in the data can cause changes in the tree [39,40].
Another employed method is the RF, which is a hybrid of the bagging algorithm and random subspace method and uses DT as a basis in the classification process [41]. In other words, RF is a combination of tree predictors, where each tree depends on the values of a random vector sampled independently with the same distribution for all the trees in the forest [42], that is, each tree is built from a sample, which is taken with replacement from the training set. Individual DTs have high variance and tend to overfit. However, the randomness injected into forests produces DTs with reduced prediction errors. Furthermore, increasing the number of trees can produce better accuracy results and limit the generalization error [42].
The NB and kNN methods were investigated in this study. The NB is one of the most efficient algorithms used in ML, classification, pattern recognition, and data mining, and it is based on Bayes’ theorem [43,44,45]. The kNN is a nonparametric classification method that has been used in different real-world applications due to its simplicity and efficiency [33,46]. The main idea of the kNN method is to predict the label of a test data point by the majority rule. In other words, the test data label is predicted with the main class with its k most similar training data points in the attribute space [33]. To avoid inaccurate prediction results, it is necessary to choose an appropriate value of k. A simple way to choose k is to load the algorithm several times with different values of k and select the one with the best result [47].
We also considered the LR scheme, which is a linear model useful for classification tasks. The sigmoid function is the basis for LR. Particularly, the logistic sigmoid function is expressed as
σ x = 1 1 + exp { x } ,
where the input x ( , ) produces results in the range of 0 , 1 . LR adds an exponential function at the linear regression, bounding the output y i 0 , 1 , and i = 1 , 2 , , n , where n is the total number of training samples [48]. The relationship between the input and the predicted output for LR is presented as
y ^ i = σ j = 1 n x i j w j + w 0 ,
where x i j is the input given by an n-dimensional vector belonging to reals; y i is the current output value that is given by a one-dimensional array; y ^ i is the predicted output value that is given by an array; w j is the weight parameters; and w 0 a bias term. Since the output is limited to the interval 0 , 1 , it can be interpreted as a probabilistic measure, that is, the LR is a variation of the linear regression [48].
This article also considered the AdaBoost (ADBST) classifier. ADBST is an adaptive boosting algorithm proposed by Freund and Schapire in 1999 [49], developed for binary classifications. The purpose of the classifier is to train predictors sequentially, trying to correct previous predictors and focusing on the most difficult cases. The algorithm increases the weight for training samples that have been misclassified, that is, the classifier learns from previous prediction errors [46]. The weight is associated with the degree of difficulty in getting it right. It builds a stronger classifier from a combination of weaker classifiers. If there are correct answers, then the classifier is rewarded. The process is repeated in T rounds, for t = 0 , 1 , , T , and n training samples. In each iteration of the algorithm, the weights are adjusted, and the samples are trained [50]. The final model is defined by the weighted majority of weak T learners, where their weights are adjusted during the training [46,50]. Initially, all training samples must have the same weight, so w i = 1 / n , x i , i = 1 , 2 , , n . Then, the algorithm considers all possible classifiers and identifies the f i ( x ) that minimizes ϵ t , which is the sum of the weights of the misclassified points. The weight α t of the new classifier is expressed as
α t = 1 2 ln 1 ϵ t ϵ t ,
which depends on the accuracy with respect to the current set of measured points. The weights are then normalized as i = 1 n = w i = 1 , and as a result, we have a classifier with an error ϵ t 0.5 3 . In the next round, incorrect classifications have their weights adjusted to make them more significant. Let y ^ t ( x i ) be a class—assuming values of 1 or 1 —predicted for x i , and y i be the correct value of the class. For the situation where y ^ t (predicted) value is equal to y i (observed) value, the y ^ t ( x i ) . y signal is a positive value; otherwise, it will assume a negative. The adjusted weights are expressed as w i , t + 1 = w i , t exp { y i α t y ^ t x i } , before renormalizing them all, so that they continue to sum to 1, that is, C = i = 1 n w i , t + 1 , and w i , t + 1 = w i , t + 1 / C [50].
We also employed a neural network as a classification tool. A NET is a very complex technology that requires a large amount of data for the training process, which is based on how human neurons work, receiving a set of inputs that are used to predict one or more outputs [51]. One of the main uses of NETs is in grouping data into two or more classes. Neural networks can be trained in two ways: (i) supervised learning, where each training input vector is paired with a target vector or desired output, and (ii) unsupervised learning, where the net self-organizes to extract patterns from data with no target information. In the n-dimensional space, the input vectors are represented as x 1 , x 2 , , x n or x, and the coefficients or weights are represented as v 1 , v 2 , , v n or v, i.e., x . v = y [52].
Finally, the stacking or stacked generalization technique is an ensemble method that combines multiple models to achieve better classification results [53,54], and this was also used in our study. This type of scheme can be more accurate than an individual classifier [55]. For instance, [56] demonstrates the efficiency of the technique by combining three different algorithms, DT, NB, and IB1 (a variation of the lazy algorithm).

2.3. Classification Setup

This section describes the methodology employed in this article, introducing the training and test groups and the steps used in the classification tools. To obtain the training and test groups, we adopted the methodology proposed in [26], where 50 groups were created randomly, resulting in 320 training images (80% of the total samples) and 80 test images (20% of the total samples). Before starting the classification steps, the image attributes are extracted using CNN VGG-16 and VGG-19 [57]. A VGG is a CNN with a convolutional layer stacking with different levels of depth.
The difference between VGG-16 and VGG-19 is in the number of convolutional layers. More precisely, VGG-16 comprises 13 convolutional layers [58,59], while VGG-19 is made up of 16 convolutional layers [60,61]. This difference can be seen in Figure 5 and Figure 6. The VGG-16 is composed of: (i) conv 1—two convolutional layers with 64 channels; (ii) conv 2—two convolutional layers with 128 channels; (iii) conv 3—three convolutional layers with 256 channels; (iv) conv 4—three convolutional layers with 512 channels; (v) conv 5—three convolutional layers with 512 channels; (vi) fc6—4096 channels; (vii) fc7—4096 channels; and (viii) fc8—1000 channels. On the other hand, VGG-19 is composed of: (i) conv 1—two convolutional layers of 64 channels; (ii) conv 2—two convolutional layers with 128 channels; (iii) conv 3—four convolutional layers with 256 channels; (iv) conv 4—four convolutional layers with 512 channels; (v) conv 5—four convolutional layers with 512 channels; (vi) fc6—4096 channels; (vii) fc7—4096 channels; and (viii) fc8—1000 channels. In short, the difference between the two occurs in conv 3–5. The VGG architecture used in the research ends in the FC7 layer. Therefore, the FC7 layer with the embedding is the input for the classification algorithms, as shown in Figure 7.
After extracting attributes with the CNNs, four different data sets are created, df-16vh, df–16vv, df-19vh, and df-19vv, which are the results of the combinations of the two CNNs, VGG-16/VGG-19, and the VH/VV polarizations.
The bootstrap technique was used to ensure reproducibility and to make sure that the classifiers are evaluated under the same conditions. Bootstrap is a random resampling technique with replacement from the primary dataset [39,62]. This technique makes it possible to estimate the empirical distribution of statistics [39,63]. In this work, each data set (i.e., df-16vh, df-16vv, df-19vh, df-19vv) is resampled 50 times, as described in Figure 8. Similar to [26], each resampling consists of 320 training samples and 80 test samples.
The methodology of this work is described in the items below and shown in Figure 9.
(1)
Data set
The data set consists of eight Sentinel-1 SAR images in VH polarization and eight in VV polarization, GRD product, IW mode, obtained through ESA’s Copernicus project [64] in the periods listed in Table 3.
(2)
Preprocessing
The original data set went through a calibration process using SNAP software (Sentinel Application Platform) to transform the amplitude image into zero-sigma. Then, the cutouts of the targets were made manually through SNAP. The dimensions of the images are displayed in Table 4. The identification of oil platforms is through geolocation (latitude × longitude) provided by the ANP [65]. Targets without geolocation are considered ships. The number of targets extracted in each SAR image is presented in Table 3. Each patch is individually exported as a TIFF image. The types of platforms in the images are listed in Table 5. TIFF image patches form the VH and VV data sets from platforms and ships.
(3)
Extraction of attributes
The VH and VV image patches are the input of two CNNs, VGG-16 and VGG-19, pre-trained in the ImageNet data set, which extract features and generate four data sets: df-16vh, df-16vv, df-19vh, and df-19vv. Figure 5 presents an example of the application of VGG-16. It is noticed that the VGG-16 of our research uses the FC7 layer to provide the attributes to the classification algorithms. In this article, VGG-19 also uses the FC7 layer to provide the attributes to the classification algorithms.
(4)
Formation of train and test samples
The proportion of 80% (training) and 20% (test) is considered. Training and testing samples are generated randomly. This was the methodology applied by [26].
(5)
Classification with the M1 method
In this method, ML techniques are applied in order to reproduce the results of [26].
(6)
Bootstrap formation
The samples vectored in the four data sets (df-16vh, df-16vv, df-19vh, df-19vv) are randomly distributed with replacement in 50 bootstrap groups and saved to submit the classifiers to the same reproducibility conditions. Table 5 presents the distribution of platforms between training and test samples in a bootstrap group.
(7)
Bootstrap and formation of train and test samples
Each of the 50 bootstrap groups comprises subsets of the original data sets (df-16vh, df-16vv, df-19vh, df-19vv).
(8)
Classification with the M2-M5 method
ML techniques are applied to the M2–M5 methods considering the kNN, SVM, LR, DT, RF, NB, NET, and ADBST algorithms.
(9)
Statistical analysis
Statistical analysis is performed using the Shapiro–Wilk methods (normality analysis); Kruskal–Wallis (significant difference analysis); and Dunn (identifies who owns differs). For brevity purposes, the two best results in each method (M1–M5) were considered.
To perform the classification, we used the following five methods, namely M1, M2, M3, M4, and M5, defined as follows:
(M1)
Aiming at reproducing the results obtained by [26], the LR, SVM, RF, kNN, DT, and NB classifiers were used with parameters in the default setting. To the best of our knowledge, [26] is the only study available in the literature for maritime target classification in Sentinel-1 SAR data based on ML techniques. For comparison purposes, in this method, the samples were randomly generated only at the time of classification and were not saved, and the parameters of the applied classification tool are the ones predefined as default in the Orange Canvas software [66], which are described in Table 6.
(M2)
The number of algorithms used in M1 was increased with the addition of the NET and ADBST methods. In this step, an extensive computational search was performed, varying the parameters of the considered classification algorithms, aiming to maximize their performance. The employed parameters are presented in Table 7. The basis for parameter adjustment is empirical and was optimized to improve the results presented by [26]. For example, [67] shows that 500 trees are a good choice for constituting the RF. However, this number can be increased to approximately 3000 to evaluate the results. For SVM and ADBST, [67] shows that the proper adjustment is made by gradually changing the parameter values. Indeed, there is no analytical methodology to reach optimal parameter values because the optimization depends on the data. This was evident when [68] optimized the parameters of SVM, kNN, and DT, demonstrating that the ideal values of the parameters can vary with the size of the training data.
(M3)
In this method, the training data set was expanded with the concatenation of all samples from the VH and VV data sets. The test data set remained unchanged.
(M4)
The training data set was extended with the concatenation of half of the samples of the VH and VV image data set. The test set remained with the same samples.
(M5)
The stacked generalization technique consists of combining several classifiers, aiming to obtain better classification results [53,56,69]. Since supervised classification is performed in all steps, the distribution of the training and test sets is done according to Table 8.
In this section, the numerical results are presented and discussed. To perform the maritime target classification in the Sentinel SAR data based on ML techniques, we extracted 4096 features from the images using CNNs VGG-16 and VGG-19, available in the Orange Canvas software [66]. For this architecture, the number of filters doubles after each max pool layer [20]. Consequently, from the data set generated by feature extraction, 50 distinct groups were defined, separated by networks VGG-16 and VGG-19, and polarizations VH and VV. Additionally, to perform the classification, we employed the tools described in Section 2.2 and the steps presented in Section 2.3.
To evaluate the performance of the employed methods, we considered the following metrics: area under the curve (AUC), accuracy (Acc), F1 score, precision, and recall. The mean for each metric was computed considering the 50 groups of data (randomly created, as mentioned in Section 2.3). Since all the results lead to the same conclusion, we decided only to discuss the Acc results in this section. The remaining results are detailed in Appendix A.
Finally, statistical analyses, such as the Shapiro–Wilk and Kruskal–Wallis tests, are presented to assess the overall performance of the methods. A flowchart with the stages of image acquisition, attribute extraction with DL algorithm, image classification methods, and statistical analysis of classification accuracy is presented in Figure 9.
Table 9 presents the setups that optimize the performance of the tested methods. The parameters are the number of CNN layers and the polarization channel. The NB and ADBST methods parameters are not presented since NB has no configuration parameters, and ADBST presented the same results, regardless of the parameter setup.
Considering the parameters displayed in Table 9, Table 10 shows the Acc mean values of 50 classification results; the best results are highlighted in bold. The classifiers that excelled in classification results were LR, NET, SVM, RF, and kNN, presenting a classification gain of 32.5%, 32.5%, 17.5%, 15.5%, and 2.5%, respectively.
Comparing our results with [26], the accuracy of the CNN VGG-16 was increased by 7.7% and 4.2% for the M4-kNN (VV polarization) and M2-SVM (VH polarization), respectively. For the CNN VGG-19, the gains are about 7.0% and 3% for the M4-kNN (VV polarization) and M2-SVM (VH polarization), respectively. In methods M3 and M4, there is an upgrade in the variability of the samples and the accuracy of the classification results. The stacking technique presents accuracy results ranging from 76.1% to 84.1%. Compared with the results shown in [26], the stacking technique does not excel only for the LR scheme.
Therefore, the RF, SVM, and NET classifiers excel in all the evaluated scenarios, and the LR had poor performance in method M4. In summary, considering all the classification techniques, the following ones stand out: LR, SVM, NET, STACK, and RF, with the highest accuracy results ranging between 80.5% and 85.45% for all the tested scenarios.
The VH polarization presents better results in detecting targets, mainly on oil platforms formed by large metallic structures of complex geometry. In general, the brightness of the targets is more intense in the VH polarization, and the background (sea) in the VV polarization. Therefore, feature extraction is best represented in VH polarization in the VGG-16 and VGG-19 networks.
To emphasize the results highlighted in Table 10, Figure 10 demonstrates the top two classification results in each method with VGG-16VH. The other graphical results for VGG-19VH, VGG-16VV, and VGG-19VV are presented in Appendix A.
Complementing the results of Table 10, Table 11 displays the average of the classification results for the M1 method with the CNN VGG-16/VGG-19 and the classification metrics (AUC, F1 Score, Precision, and Recall). In addition, it reproduces the results of [26], using the same training/testing and classification data generation approach. Each metric is calculated for the six classifiers (kNN, LR, NB, RF, SVM, and DT). As in [26], LR is the classifier with the best performance. Furthermore, it is observed that the results with VH polarization are superior to those with VV polarization. Methods M2 to M5 present result tables with the same structure as M1 for the metrics (AUC, F1 Score, Precision, and Recall). Therefore, tables are included in Appendix A.

3. Results and Discussion

To further evaluate the performance of the tested methods, we considered the Kruskal–Wallis and post hoc Dunn’s tests to identify if the employed methods present significantly different mean behavior in terms of accuracy in comparison with the approaches described in [26]. Both tests are widely explored in several non-Gaussian signal processing applications for comparison purposes of machine learning tools, such as in [70,71,72,73,74]. To verify the normality of the data, we performed the Shapiro–Wilk test, which indicates that 70%, 10%, 30%, and 30% of the VGG-16VH, VGG-16VV, VGG-19VH, and VGG-19VV data cannot be modeled by the normal distribution, respectively. For all the employed tests, we set the significance level equal to 0.05, which is a convenient cutoff level to reject the null hypothesis [75].
Table 12 shows the p-values of the Kruskal–Wallis and Dunn’s tests for the results reproduced from [26] (M1) and the proposed tools; for brevity, just the significant results are displayed. Among them, SVM results in an accuracy gain of 4.17%, 3.94%, and 3.03% for VGG-16VH (M2), VGG-19VV (M4), and VGG -19VH (M2), respectively, in comparison with the results presented in [26].
Another analysis to consider is comparing the methods that present the best performances. For VGG-16VH (M2), the LR classifier presents gains of 0.35%, 1.76%, 0.97%, and 1.24% about M1, M3, M4,and M5, respectively. For VGG-19VH (M1), the LR classifier presents gains of 0.18%, 0.77%, 0.65%, 1.13% over M2, M3, M4, and M5, respectively.
The performed analysis and statistical tests highlight that the applied schemes presented competitive performance when compared with [26], which, to the best of our knowledge, is the only study available in the literature related to detecting maritime targets in Sentinel-1 SAR images from the Campos Basin.

4. Conclusions

This article applied machine learning algorithms to classify maritime targets in Pol-SAR images (VH and VV) obtained with the Sentinel-1 system. The classifiers were evaluated considering five different methods (M1, M2, M3, M4, M5). As a pre-stage, the features were extracted using two CNN algorithms: VGG-16 and VGG-19. The classifiers were assessed in terms of accuracy. The RF, SVM, and NET classifiers excelled in all the evaluated scenarios over the reference methods, and the LR classifier performed poorly in M3. The classification results for all the tested classifiers except LR presented mean accuracy values above 80%, which were 1.44 times better than the baseline results in VGG-16VH and 1.55 better in VGG-19VH. VH polarization stands out in the classification of maritime targets, as oil platforms and ships have higher brightness (i.e., higher backscattering) because of their geometries formed by large metallic structures. With the parameters optimization, the tested classifiers showed more accurate classification results. The stacking technique also showed satisfactory accuracy results.

Author Contributions

Conceptualization, F.G.d.S., B.G.P. and R.M.; methodology, F.G.d.S., L.P.R., B.G.P. and R.M.; software, F.G.d.S.; validation, F.G.d.S.; formal analysis, F.G.d.S.; writing—original draft preparation, F.G.d.S.; writing—review and editing, L.P.R., B.G.P. and R.M.; supervision, B.G.P. and R.M. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financed in part by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior—Brazil (CAPES-Brazil)—Finance Code 001 (Pró-Defesa IV), and by Brazilian National Council for Scientific and Technological Development (CNPq-Brazil). The authors also thank the Brazilian Institute of Data Science (BI0S), grant 2020/09838-0, São Paulo Research Foundation (FAPESP), Fundação de Amparo à Pesquisa do Estado do Rio Grande do Sul (FAPERGS), Brazil, and the Swedish–Brazilian Research and Innovation Centre (CISB).

Data Availability Statement

A publicly available data set was analysed in this study.

Acknowledgments

The authors would like to thank the European Space Agency (ESA) for the Sentinel-1 Open Data Set provided by Copernicus Open Access Hub.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADBSTAdaBoost
AIArtificial Intelligence
AISAutomatic Identification System
ATRAutomatic Target Recognition
AccClassification Accuracy
AUCArea Under the Curve
CNNConvolutional Neural Network
DLDeep Learning
DNNDeep Neural Network
DTDecison Tree
EWExtra-Wide
GANGenerative Adversarial Network
GRDGround Range Detected
IWInterferometric Wide
kNNk-Nearest Neighbor
LRLogistic Regression
MLMachine Learning
MLPMultilayer Perceptron
MSTARMoving and Stationary Target Acquisition and Recognition
NBNaive Bayes
NETNeural Networks
NLPNatural Language Processing
OCNOcean
RFRandom Forest
RNNRecurrent Neural Network
SARSynthetic Aperture Radar
SLCSingle Look Complex
SMStripmap
SVMSupport Vector Machine
VGGVisual Geometry Group

Appendix A. Numerical Analysis

In this appendix, the mean of the 50 results in all classification methods in terms of AUC, F1 Score, Precision, and Recall are presented. Table A1, Table A2, Table A3 and Table A4 show the numerical results for M2, M3, M4, and M5, respectively. The two best ranking results by methods are summarized in Figure A1, Figure A2 and Figure A3.
Table A1. Overall summary of the average of all metrics employed in the classification methods—M2.
Table A1. Overall summary of the average of all metrics employed in the classification methods—M2.
Classifier
MethodCNN-PolMetricADBSTkNNLRNBNETRFSVMDT
M2VGG-16VHAUC0.7320.8650.9270.7240.9030.9050.9100.754
F10.7310.7910.8540.6910.8370.8180.8370.737
Precision0.7340.7980.8580.7070.8420.8220.8440.742
Recall0.7320.7920.8550.6960.8380.8180.8380.738
VGG-16VVAUC0.6500.7520.8680.6740.8510.8250.8350.654
F10.6490.7070.7820.6500.7830.7380.7590.652
Precision0.6520.7100.7860.6690.7880.7440.7650.656
Recall0.6500.7080.7830.6570.7840.7400.7600.654
VGG-19VHAUC0.7680.8610.9170.7410.9030.9070.9110.732
F10.7670.8020.8410.6990.8350.8330.8480.764
Precision0.7700.8050.8440.7140.8380.8370.8540.768
Recall0.7680.8020.8420.7040.8360.8330.8490.765
VGG-19VVAUC0.6400.7720.8450.6830.8170.8300.8300.669
F10.6370.7240.7670.6440.7580.7500.7530.661
Precision0.6420.7370.7700.6640.7610.7530.7560.665
Recall0.6400.7280.7680.6520.7610.7510.7540.663
Table A2. Overall summary of the average of all metrics employed in the classification methods—M3.
Table A2. Overall summary of the average of all metrics employed in the classification methods—M3.
Classifier
MethodCNN-PolMetricADBSTkNNLRNBNETRFSVMDT
M3VGG-16VHAUC0.7390.8290.7470.7180.9130.8800.8780.724
F10.7380.7670.5780.6860.8390.7980.7840.756
Precision0.7410.7730.5730.7030.8430.8020.7890.761
Recall0.7390.7680.6270.6910.8400.7990.7850.757
VGG-16VVAUC0.6880.8050.7170.6720.8560.8080.8160.643
F10.6870.7470.5600.6500.7750.7270.7250.675
Precision0.6890.7500.5490.6670.7780.7310.7390.677
Recall0.6880.7480.6080.6560.7760.7280.7290.676
VGG-19VHAUC0.7700.8550.7850.7360.9240.8930.8840.733
F10.7690.7930.6000.7030.8440.8100.7870.774
Precision0.7740.7960.5950.7190.8470.8130.7970.778
Recall0.7700.7930.6410.7070.8440.8100.7890.775
VGG-19VVAUC0.6980.8270.7380.6880.8650.8190.8330.668
F10.6970.7630.5820.6530.7770.7410.7640.705
Precision0.7000.7650.5770.6670.7790.7440.7760.708
Recall0.6980.7630.6240.6580.7770.7420.7660.706
Table A3. Overall summary of the average of all metrics employed in the classification methods—M4.
Table A3. Overall summary of the average of all metrics employed in the classification methods—M4.
Classifier
MethodCNN-PolMetricADBSTkNNLRNBNETRFSVMDT
M4VGG-16VHAUC0.7260.8220.9190.7220.9080.8630.8880.699
F10.7240.7710.8460.6820.8300.7790.7910.723
Precision0.7290.7780.8510.7000.8360.7830.8000.727
Recall0.7260.7720.8460.6870.8300.7800.7920.724
VGG-16VVAUC0.6670.7540.8620.6790.8290.7780.8100.647
F10.6660.6990.7820.6520.7560.7010.7200.666
Precision0.6700.7020.7850.6680.7590.7050.7280.671
Recall0.6670.7000.7830.6580.7570.7030.7220.668
VGG-19VHAUC0.7520.8400.9240.7350.8910.8750.8940.724
F10.7510.7870.8440.6930.8320.8040.7970.757
Precision0.7550.7920.8480.7090.8360.8070.8050.761
Recall0.7520.7880.8450.6980.8330.8040.7990.758
VGG-19VVAUC0.6800.7560.8500.6880.8340.8040.8300.668
F10.6780.7080.7720.6510.7600.7250.7450.691
Precision0.6830.7130.7760.6660.7640.7300.7490.694
Recall0.6800.7100.7730.6570.7610.7270.7460.692
Table A4. Overall summary of the average of all metrics employed in the classification methods—M5.
Table A4. Overall summary of the average of all metrics employed in the classification methods—M5.
MethodCNN-PolAUCF1PrecisionRecall
M5VGG-16VH0.9170.8420.8560.844
VGG-16VV0.8440.7560.7790.763
VGG-19VH0.9130.8400.8460.841
VGG-19VV0.8410.7570.7720.761
Figure A1. Summary of the two best classification results obtained in each method with CNN VGG-19 and VH polarization.
Figure A1. Summary of the two best classification results obtained in each method with CNN VGG-19 and VH polarization.
Remotesensing 14 02966 g0a1
Figure A2. Summary of the two best classification results obtained in each method with CNN VGG-16 and VV polarization.
Figure A2. Summary of the two best classification results obtained in each method with CNN VGG-16 and VV polarization.
Remotesensing 14 02966 g0a2
Figure A3. Summary of the two best classification results obtained in each method with CNN VGG-19 and VV polarization.
Figure A3. Summary of the two best classification results obtained in each method with CNN VGG-19 and VV polarization.
Remotesensing 14 02966 g0a3

References

  1. de Oliveira Soares, M.; da Cruz Lotufo, T.M.; Vieira, L.M.; Salani, S.; Hadju, E.; Matthews-Cascon, H.; Leão, Z.M.; Kenji, R.; de Kikuchi, P. Brazilian marine animal forests: A new world to discover in the southwestern Atlantic. In Marine Animal Forests: The Ecology of Benthic Biodiversity Hotspots; Springer International Publishing: Cham, Switzerland, 2017; pp. 73–110. [Google Scholar] [CrossRef]
  2. Armelenti, G.; Goldberg, K.; Kuchle, J.; de Ros, L. Deposition, diagenesis and reservoir potential of non-carbonate sedimentary rocks from the rift section of Campos Basin, Brazil. Pet. Geosci. 2016, 22, 223–239. [Google Scholar] [CrossRef]
  3. ANP. Boletim Mensal da Produção de Petróleo e Gás Natural; ANP: Rio de Janeiro, Brazil, 2021.
  4. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR dataset of ship detection for deep learning under complex backgrounds. Remote. Sens. 2019, 11, 765. [Google Scholar] [CrossRef] [Green Version]
  5. Jiang, Y.; Li, W.; Liu, L. R-CenterNet+: Anchor-free detector for ship detection in SAR images. Sensors 2021, 21, 693. [Google Scholar] [CrossRef]
  6. Snoeij, P.; Attema, E.; Davidson, M.; Duesmann, B.; Floury, N.; Levrini, G.; Rommen, B.; Rosich, B. The Sentinel-1 radar mission: Status and performance. In Proceedings of the 2009 International Radar Conference “Surveillance for a Safer World” (RADAR 2009), Bordeaux, France, 12–16 October 2009; IEEE: New York, NY, USA, 2009; pp. 1–6. [Google Scholar]
  7. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K. A tutorial on synthetic aperture radar. IEEE Geosci. Remote. Sens. Mag. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
  8. Brusch, S.; Lehner, S.; Fritz, T.; Soccorsi, M.; Soloviev, A.; Van Schie, B. Ship surveillance with TerraSAR-X. IEEE Trans. Geosci. Remote. Sens. 2011, 49, 1092–1103. [Google Scholar] [CrossRef]
  9. Rane, A.; Sangili, V. Implementation of improved ship-iceberg classifier using deep learning. J. Intell. Syst. 2020, 29, 1514–1522. [Google Scholar] [CrossRef]
  10. El-Darymli, K.; McGuire, P.; Power, D.; Moloney, C. Target detection in synthetic aperture radar imagery: A state-of-the-art survey. J. Appl. Remote. Sens. 2013, 7, 071598. [Google Scholar] [CrossRef] [Green Version]
  11. Cantorna, D.; Dafonte, C.; Iglesias, A.; Arcay, B. Oil spill segmentation in SAR images using convolutional neural networks. A comparative analysis with clustering and logistic regression algorithms. Appl. Soft Comput. J. 2019, 84, 105716. [Google Scholar] [CrossRef]
  12. Solberg, A. Remote sensing of ocean oil-spill pollution. Proc. IEEE 2012, 100, 2931–2945. [Google Scholar] [CrossRef]
  13. Gao, F.; Huang, T.; Sun, J.; Wang, J.; Hussain, A.; Yang, E. A new algorithm for SAR image target recognition based on an improved deep convolutional neural network. Cogn. Comput. 2019, 11, 809–824. [Google Scholar] [CrossRef] [Green Version]
  14. Kanjir, U.; Greidanus, H.; Oštir, K. Vessel detection and classification from spaceborne optical images: A literature survey. Remote. Sens. Environ. 2018, 207, 1–26. [Google Scholar] [CrossRef]
  15. Liang, X.; Zhen, Z.; Song, Y.; Jian, L.; Song, D. Pol-SAR based oil spillage classification with various scenarios of prior knowledge. IEEE Access 2019, 7, 66895–66909. [Google Scholar] [CrossRef]
  16. Sharifzadeh, F.; Akbarizadeh, G.; Seifi Kavian, Y. Ship classification in SAR images using a new hybrid CNN–MLP classifier. J. Indian Soc. Remote. Sens. 2019, 47, 551–562. [Google Scholar] [CrossRef]
  17. Bentes, C.; Velotto, D.; Tings, B. Ship Classification in TerraSAR-X images with convolutional neural networks. IEEE J. Ocean. Eng. 2018, 43, 258–266. [Google Scholar] [CrossRef] [Green Version]
  18. Chen, S.; Wang, H.; Xu, F.; Jin, Y.Q. Target classification using the deep convolutional networks for SAR images. IEEE Trans. Geosci. Remote. Sens. 2016, 54, 4806–4817. [Google Scholar] [CrossRef]
  19. Igual, L.; Seguí, S. Introduction to data science. In Introduction to Data Science: A Python Approach to Concepts, Techniques and Applications; Springer International Publishing: Cham, Switzerland, 2017; pp. 1–4. [Google Scholar] [CrossRef]
  20. Nguyen, G.; Dlugolinsky, S.; Bobák, M.; Tran, V.; García, Á.L.; Heredia, I.; Malík, P.; Hluchỳ, L. Machine Learning and deep learning frameworks and libraries for large-scale data mining: A survey. Artif. Intell. Rev. 2019, 52, 77–124. [Google Scholar] [CrossRef] [Green Version]
  21. Shalev-Shwartz, S.; Ben-David, S. Understanding Machine Learning: From Theory to Algorithms; Cambridge University Press: New York, NY, USA, 2014. [Google Scholar] [CrossRef]
  22. Kubat, M. An Introduction to Machine Learning; Springer: Coral Gables, FL, USA, 2017. [Google Scholar] [CrossRef]
  23. Wang, Y.; Wang, C.; Zhang, H. Combining a single shot multibox detector with transfer learning for ship detection using Sentinel-1 SAR images. Remote. Sens. Lett. 2018, 9, 780–788. [Google Scholar] [CrossRef]
  24. Wang, Y.; Wang, C.; Zhang, H.; Zhang, C.; Fu, Q. Combing single shot multibox detector with transfer learning for ship detection using Chinese Gaofen-3 images. In Proceedings of the 2017 Progress in Electromagnetics Research Symposium-Fall (PIERS-FALL), Singapore, 19–22 November 2017; Volume 2017, pp. 712–716. [Google Scholar] [CrossRef]
  25. Morgan, D. Deep convolutional neural networks for ATR from SAR imagery. In Algorithms for Synthetic Aperture Radar Imagery; Society of Photo Optical: Bellingham, DC, USA, 2015; Volume 22, p. 9475. [Google Scholar] [CrossRef]
  26. Falqueto, L.; Sa, J.; Paes, R.; Passaro, A. Oil rig recognition using convolutional neural network on Sentinel-1 SAR images. IEEE Geosci. Remote. Sens. Lett. 2019, 16, 1329–1333. [Google Scholar] [CrossRef]
  27. Torres, R.; Snoeij, P.; Geudtner, D.; Bibby, D.; Davidson, M.; Attema, E.; Potin, P.; Rommen, B.; Floury, N.; Brown, M.; et al. GMES Sentinel-1 mission. Remote. Sens. Environ. 2012, 120, 9–24. [Google Scholar] [CrossRef]
  28. Potin, P.; Rosich, B.; Miranda, N.; Grimont, P.; Shurmer, I.; O’Connell, A.; Krassenburg, M.; Gratadour, J.B. Copernicus Sentinel-1 constellation mission operations status. In Proceedings of the IGARSS 2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 5385–5388. [Google Scholar] [CrossRef]
  29. Geudtner, D.; Gebert, N.; Tossaint, M.; Davidson, M.; Heliere, F.; Navas Traver, I.; Furnell, R.; Torres, R. Copernicus and ESA SAR missions. In Proceedings of the 2021 IEEE Radar Conference (RadarConf21), Atlanta, GA, USA, 8–14 May 2021. [Google Scholar] [CrossRef]
  30. Schubert, A.; Small, D.; Miranda, N.; Geudtner, D.; Meier, E. Sentinel-1A product geolocation accuracy: Commissioning phase results. Remote. Sens. 2015, 7, 9431–9449. [Google Scholar] [CrossRef] [Green Version]
  31. Navy, B. Directorate of Ports and Coasts—DPC. 2021. Available online: https://www.marinha.mil.br/dpc/helideques (accessed on 27 December 2021).
  32. Murphy, K.P. Naive Bayes Classifiers; University of British Columbia: Vancouver, BC, Canada, 2006; Volume 18, pp. 1–8. [Google Scholar]
  33. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Cheng, D. Learning k for kNN classification. ACM Trans. Intell. Syst. Technol. 2017, 8, 43. [Google Scholar] [CrossRef] [Green Version]
  34. Huang, G.B.; Wang, D.H.; Lan, Y. Extreme learning machines: A survey. Int. J. Mach. Learn. Cybern. 2011, 2, 107–122. [Google Scholar] [CrossRef]
  35. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
  36. Wu, T.D.; Yen, Y.; Wang, J.H.; Huang, R.; Lee, H.W.; Wang, H.F. Automatic target recognition in SAR images based on a combination of CNN and SVM. In Proceedings of the 2020 International Workshop on Electromagnetics: Applications and Student Innovation Competition (iWEM), Penghu, Taiwan, 26–28 August 2020; pp. 1–2. [Google Scholar] [CrossRef]
  37. Maokuan, L.; Jian, G.; Hui, D.; Xin, G. SAR ATR based on support vector machines and independent component analysis. In Proceedings of the 2006 CIE International Conference on Radar, Shanghai, China, 16–19 October 2006; pp. 1–3. [Google Scholar] [CrossRef]
  38. Quinlan, J.R. C4.5: Programs for Machine Learning; Elsevier: San Mateo, CA, USA, 2014. [Google Scholar]
  39. Friedman, J.; Hastie, T.; Tibshirani, R. The Elements of Statistical Learning; Springer: New York, NY, USA, 2001. [Google Scholar]
  40. Breiman, L.; Friedman, J.; Olshen, R.; Stone, C. Classification and Regression Trees; Thomson Wadsworth: Belmont, CA, USA, 1984; ISBN 9780412048418. [Google Scholar]
  41. Sammut, C.; Webb, G.I. Random Forests. In Encyclopedia of Machine Learning; Springer: Boston, MA, USA, 2010; p. 828. [Google Scholar] [CrossRef]
  42. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  43. Zhang, H. The optimality of Naive Bayes. In Proceedings of the Seventeenth International Florida Artificial Intelligence Research Society Conference, Miami Beach, FL, USA, 12–14 May 2004; Volume 2, pp. 562–597. [Google Scholar]
  44. Yan, X.; Li, W.; Wu, Q.; Sheng, V.S. A double weighted naive bayes for multi-label classification. Computational Intelligence and Intelligent Systems; Li, K., Li, J., Liu, Y., Castiglione, A., Eds.; Springer: Singapore, 2016; pp. 382–389. [Google Scholar] [CrossRef]
  45. Langley, P.; Iba, W.; Thompson, K. An analysis of Bayesian classifiers. AAAI 1992, 90, 223–228. [Google Scholar]
  46. Wu, X.; Kumar, V.; Quinlan, J.R.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.J.; Ng, A.; Liu, B.; Philip, S.Y.; et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar] [CrossRef] [Green Version]
  47. Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. kNN model-based approach in classification. Lect. Notes Comput. Sci. 2003, 2888, 986–996. [Google Scholar] [CrossRef]
  48. Joshi, A.V. Machine Learning and Artificial Intelligence; Springer: Cham, Switzerland, 2020. [Google Scholar]
  49. Freund, Y.; Schapire, R. A short introduction to boosting. J. Jpn. Soc. Artif. Intell. 1999, 14, 1612. [Google Scholar]
  50. Skiena, S.S. The Data Science Design Manual; Springer: Cham, Switzerland, 2017. [Google Scholar] [CrossRef]
  51. Bramer, M. Principles of Data Mining; Springer: London, UK, 2007; Volume 180. [Google Scholar] [CrossRef]
  52. Daumé, H., 3rd. A Course in Machine Learning; University of Maryland: College Park, MA, USA, 2017. [Google Scholar]
  53. Wolpert, D.H. Stacked generalization. Neural Netw. 1992, 5, 241–259. [Google Scholar] [CrossRef]
  54. Chatzimparmpas, A.; Martins, R.M.; Kucher, K.; Kerren, A. Stackgenvis: Alignment of data, algorithms, and models for stacking ensemble learning using performance metrics. IEEE Trans. Vis. Comput. Graph. 2020, 27, 1547–1557. [Google Scholar] [CrossRef]
  55. Dietterich, T.G. Machine Learning. Annu. Rev. Comput. Sci. 1990, 4, 255–306. [Google Scholar] [CrossRef]
  56. Ting, K.M.; Witten, I.H. Stacked Generalization: When Does it Work? University of Waikato, Department of Computer Science: Hamilton, New Zealand, 1997. [Google Scholar]
  57. Sujatha, R.; Chatterjee, J.; Jhanjhi, N.; Brohi, S. Performance of deep learning vs machine learning in plant leaf disease detection. Microprocess. Microsyst. 2021, 80, 103615. [Google Scholar] [CrossRef]
  58. Bharati, S.; Podder, P.; Mondal, M.R.H. Hybrid deep learning for detecting lung diseases from X-ray images. Informatics Med. Unlocked 2020, 20, 100391. [Google Scholar] [CrossRef] [PubMed]
  59. Behzadi-Khormouji, H.; Rostami, H.; Salehi, S.; Derakhshande-Rishehri, T.; Masoumi, M.; Salemi, S.; Keshavarz, A.; Gholamrezanezhad, A.; Assadi, M.; Batouli, A. Deep learning, reusable and problem-based architectures for detection of consolidation on chest X-ray images. Comput. Methods Programs Biomed. 2020, 185, 105162. [Google Scholar] [CrossRef]
  60. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  61. Dong, Y.; Zhang, H.; Wang, C.; Wang, Y. Fine-grained ship classification based on deep residual learning for high-resolution SAR images. Remote. Sens. Lett. 2019, 10, 1095–1104. [Google Scholar] [CrossRef]
  62. Numbisi, F.N.; Van Coillie, F. Does Sentinel-1A Backscatter Capture the Spatial Variability in Canopy Gaps of Tropical Agroforests? A Proof-of-Concept in Cocoa Landscapes in Cameroon. Remote. Sens. 2020, 12, 4163. [Google Scholar] [CrossRef]
  63. Martínez-Agüero, S.; Marques, A.G.; Mora-Jiménez, I.; Álvarez Rodríguez, J.; Soguero-Ruiz, C. Data and Network Analytics for COVID-19 ICU Patients: A Case Study for a Spanish Hospital. IEEE J. Biomed. Health Inform. 2021, 25, 4340–4353. [Google Scholar] [CrossRef] [PubMed]
  64. ESA. Copernicus Open Access Hub. 2021. Available online: https://scihub.copernicus.eu/ (accessed on 28 December 2021).
  65. National Petroleum Agency (ANP). List of Platforms in Operation. 2021. Available online: https://www.gov.br/anp/pt-br/centrais-de-conteudo/dados-abertos/lista-de-plataformas-em-operacao (accessed on 28 December 2021).
  66. Demšar, J.; Curk, T.; Erjavec, A.; Gorup, C.; Hočevar, T.; Milutinovič, M.; Možina, M.; Polajnar, M.; Toplak, M.; Starič, A.; et al. Orange: Data mining toolbox in python. J. Mach. Learn. Res. 2013, 14, 2349–2353. [Google Scholar]
  67. Berk, R.A. Statistical Learning from a Regression Perspective; Springer: New York, NY, USA, 2008. [Google Scholar]
  68. Qian, Y.; Zhou, W.; Yan, J.; Li, W.; Han, L. Comparing machine learning classifiers for object-based land cover classification using very high resolution imagery. Remote. Sens. 2014, 7, 153–168. [Google Scholar] [CrossRef]
  69. Ting, K.M.; Witten, I.H. Issues in stacked generalization. J. Artif. Intell. Res. 1999, 10, 271–289. [Google Scholar] [CrossRef] [Green Version]
  70. Murata, T.; Yanagisawa, T.; Kurihara, T.; Kaneko, M.; Ota, S.; Enomoto, A.; Tomita, M.; Sugimoto, M.; Sunamura, M.; Hayashida, T.; et al. Salivary metabolomics with alternative decision tree-based machine learning methods for breast cancer discrimination. Breast Cancer Res. Treat. 2019, 177, 591–601. [Google Scholar] [CrossRef] [PubMed]
  71. Emami, A.; Kunii, N.; Matsuo, T.; Shinozaki, T.; Kawai, K.; Takahashi, H. Seizure detection by convolutional neural network-based analysis of scalp electroencephalography plot images. Neuroimage Clin. 2019, 22, 101684. [Google Scholar] [CrossRef] [PubMed]
  72. Sutton, E.J.; Dashevsky, B.Z.; Oh, J.H.; Veeraraghavan, H.; Apte, A.P.; Thakur, S.B.; Morris, E.A.; Deasy, J.O. Breast cancer molecular subtype classifier that incorporates MRI features. J. Magn. Reson. Imaging 2016, 44, 122–129. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Paiva, F.D.; Cardoso, R.T.N.; Hanaoka, G.P.; Duarte, W.M. Decision-making for financial trading: A fusion approach of machine learning and portfolio selection. Expert Syst. Appl. 2019, 115, 635–655. [Google Scholar] [CrossRef]
  74. Taghavi, M.; Trebeschi, S.; Simões, R.; Meek, D.; Beckers, R.; Lambregts, D.; Verhoef, C.; Houwers, J.; van der Heide, U.; Beets-Tan, R.; et al. Machine learning-based analysis of CT radiomics model for prediction of colorectal metachronous liver metastases. Abdom. Radiol. 2021, 46, 249–256. [Google Scholar] [CrossRef] [PubMed]
  75. Fisher, R.A. Statistical methods for research workers. In Breakthroughs in Statistics; Springer: New York, NY, USA, 1992; pp. 66–70. [Google Scholar]
Figure 1. Data acquisition modes from Sentinel-1 imaging during its orbital shift.
Figure 1. Data acquisition modes from Sentinel-1 imaging during its orbital shift.
Remotesensing 14 02966 g001
Figure 2. Composition diagram of Sentinel-1 operating modes (SW, IW, EW, and WV) in the three product levels (L0, L1, and L2).
Figure 2. Composition diagram of Sentinel-1 operating modes (SW, IW, EW, and WV) in the three product levels (L0, L1, and L2).
Remotesensing 14 02966 g002
Figure 3. Level 1 product (GRD) of the IW mode used in the research.
Figure 3. Level 1 product (GRD) of the IW mode used in the research.
Remotesensing 14 02966 g003
Figure 4. Examples of oil platforms with optical and SAR (VH and VV polarizations) images in the Campos Basin, Brazil. Optical and SAR images extracted from [31] and [26], respectively.
Figure 4. Examples of oil platforms with optical and SAR (VH and VV polarizations) images in the Campos Basin, Brazil. Optical and SAR images extracted from [31] and [26], respectively.
Remotesensing 14 02966 g004
Figure 5. VGG-16 formation diagram.
Figure 5. VGG-16 formation diagram.
Remotesensing 14 02966 g005
Figure 6. VGG-19 formation diagram.
Figure 6. VGG-19 formation diagram.
Remotesensing 14 02966 g006
Figure 7. VGG-16 formation diagram.
Figure 7. VGG-16 formation diagram.
Remotesensing 14 02966 g007
Figure 8. Formation of the 50 training and test groups using the bootstrap resampling technique.
Figure 8. Formation of the 50 training and test groups using the bootstrap resampling technique.
Remotesensing 14 02966 g008
Figure 9. Flowchart with image acquisition, attribute extraction with DL algorithm, image classification (five methods), and statistical analysis of classification accuracy.
Figure 9. Flowchart with image acquisition, attribute extraction with DL algorithm, image classification (five methods), and statistical analysis of classification accuracy.
Remotesensing 14 02966 g009
Figure 10. Summary of the two best classification results obtained in each method with CNN VGG-16 and VH polarization.
Figure 10. Summary of the two best classification results obtained in each method with CNN VGG-16 and VH polarization.
Remotesensing 14 02966 g010
Table 1. Summary of the Sentinel-1 system parameters and characteristics.
Table 1. Summary of the Sentinel-1 system parameters and characteristics.
NameSentinel-1
BandC
Bandwidth0–100 MHz (programmable)
Centre frequency5.405 GHz
Storage capacity1410 Gb
PolarizationHH+HV, VV+VH
VV, HH
Incidence angle range20–46°
Look directionright
Antenna typeSlotted wave-guide radiators
Antenna size12.3 m × 0.821 m
Antenna mass880 kg
Azimuth beam width0.23°
PRF (pulse repetition frequency)1–3 kHz (programmable)
Data quantization10 bit
Total instrument mass (including antenna)945 kg
Font: [27].
Table 2. Typical Sentinel-1 applications distributed by modes.
Table 2. Typical Sentinel-1 applications distributed by modes.
ApplicationMode
SMIWEWWV
Arctic and sea-ice XX
Open ocean ship surveillance XX
Oil pollution monitoring XX
Marine winds XXX
Forestry X
Agriculture X
Urban deformation mapping X
Flood monitoringXX
Earthquake analysisXX
Landslide and volcano monitoringXX
Table 3. Number of targets extracted in each SAR image.
Table 3. Number of targets extracted in each SAR image.
Polarization
VHVV
SAR Image DateLegendOil RigShipLegendOil RigShip
20180430 at 08:04 a.m. υ 1 4250 ι 1 4250
20180605 at 08:05 a.m. υ 2 4049 ι 2 4049
20180512 at 08:05 a.m. υ 3 4048 ι 3 4048
20171118 at 08:13 a.m. υ 4 1229 ι 4 1229
20180524 at 08:05 a.m. υ 5 3811 ι 5 3811
20180617 at 08:04 a.m. υ 6 80 ι 6 80
20180622 at 08:13 a.m. υ 7 130 ι 7 130
20180430 at 08:05 a.m. υ 8 713 ι 8 713
Table 4. Size of image patches.
Table 4. Size of image patches.
Dimension: Range × Azimuth
912 × 596 pixels
968 × 596 pixels
939 × 596 pixels
901 × 596 pixels
930 × 596 pixels
944 × 596 pixels
985 × 596 pixels
1057 × 596 pixels
Table 5. Distribution of platform patches between training and testing samples.
Table 5. Distribution of platform patches between training and testing samples.
PolarizationTotal
VGG-16VHVGG-16VVVGG-19VHVGG-19VV
IdOil RigTestTrainTestTrainTestTrainTestTrain
1FIX PCA1 2 2 2118
2FIX PCA211 2 2 28
3FIX PCH12323231420
4FIX PCP11212 3 312
5FIX PCP2 3 3 31212
6FIX PEREGRINOA 1 1 1 14
7FIX PNA11423 52320
8FIX PNA214 5 51420
9FIX POLVOA 1 1 1 14
10FIX PPG114 514 520
11FIX PPM1 211 2 28
12FIX PRA113 4221316
13FIX PVM11414 5 520
14FIX PVM22314 51420
15FIX PVM314 514 520
16FPSO CAPX112 11118
17FPSO CDAN12 3211212
18FPSO DYNAMIC 1 1 11 4
19FPSO ESPST 2 2 2 28
20FPSO FPF3214 5 520
21FPSO FPNIT 312 31212
22FPSO FPRJ 422 41316
23FPSO FPRO2213132216
24FPSO FRADE11 2 2118
25FPSO OSX3 1 1 1 14
26FPSO P31 4 4 4 416
27FPSO P331313133116
28FPSO P35 413131316
29FPSO P37 4 4 4 416
30FPSO P4313 4 42216
31FPSO P47 312 32112
32FPSO P481212 31212
33FPSO P49 11 1 14
34FPSO P50 413131316
35FPSO P54 3 312 312
36FPSO P57 3 312 312
37FPSO P58 3 3121212
38FPSO P62 532142320
39FPSO P63 1 1 11 4
40FPSO P64 11 1 14
41FPSO PEREGRINO 11 1 14
42FPSO POLVO 11 1 14
43FPU P5312 321 312
44FSO MACAE 4 4 41316
45FSO P3211 2 2 28
46FSO P38 312 31212
47SS P0723 523 520
48SS P081414321420
49SS P15 52314 520
50SS P1822 4222216
51SS P191313 4 416
52SS P203214 52320
53SS P25132222 416
54SS P261313131316
55SS P401313221316
56SS P51131313 416
57SS P52 2 22 28
58SS P55 4 4 4 416
59SS P56131313 416
60SS P651313311316
61TWLP P61 1 1 1 14
Total40160401604016040160800
Table 6. Default parameters available of the employed classification methods in Orange Canvas software [66].
Table 6. Default parameters available of the employed classification methods in Orange Canvas software [66].
ClassifierParameterValue
kNNNumber of neighbors3
DistanceEuclidean
WeightUniform
DTTree typeBinary
Minimum instances per sheet2
Minimum instances for splitting a node5
Tree depth limit100 node levels
Stop criteria based on majority95%
RFNumber of trees10
Minimum instances for splitting a node5
SVMCost1
Regression loss epsilon0.1
KernelRBF
Numerical tolerance0.001
Number of interactions100
NBNo parameters-
LRRegularization methodRidge (L2)
Force1
Table 7. Parameters employed in the classification tools.
Table 7. Parameters employed in the classification tools.
ClassifierParameterValue
RFNumber of trees10, 50, 100, 200, 300, 400
500, 600, 700, 800, 900
1000, 1100, 1200, 1300
kNNNumber of neighbors3, 5, 7, 10
Distance usedEuclidean and Manhattan
ADBSTNumber of estimators10, 50, 100
LRStrength0.6, 1, 10, 50, 100, 200
NETNumber of neurons10, 50, 100, 200
SVMCost1, 2, 3, 4, 5
KernelRBF and SIGM
Number of interactions100 and 200
DTMinimum instances per sheet2, 5, 7, 10, and 20
NBNo parameters
Table 8. Distribution of the training and test samples.
Table 8. Distribution of the training and test samples.
Samples
CNNPolTotalTrainingTest
M1, M2,andM5M3M4
VGG-16VH400320640160 (VH) + 160 (VV)80
VV40032080
VGG-19VH400320640160 (VH) + 160 (VV)80
VV40032080
Table 9. Optimized parameters in step M2.
Table 9. Optimized parameters in step M2.
Parameter
Classifier
CNNPolkNNLRRFSVMDTNET
VGG-16VH7-Manhattan1504-SIGM-200710
VV3-Manhattan12003-SIGM-200550
VGG-19VH5-Manhattan112001-SIGM-100250
VV5-Euclidean0.61002-SIGM-2005100
Table 10. Overall summary of accuracy mean results of the employed classification methods for the VGG—16VH, 16VV, 19VH, and 19VV SAR images. The best results are highlighted in bold. The symbol (i) “-” indicates that the classification result is absent, and (ii) “x” indicates which classifiers are combined to obtain the final result with the stacked generalization technique in step M5.
Table 10. Overall summary of accuracy mean results of the employed classification methods for the VGG—16VH, 16VV, 19VH, and 19VV SAR images. The best results are highlighted in bold. The symbol (i) “-” indicates that the classification result is absent, and (ii) “x” indicates which classifiers are combined to obtain the final result with the stacked generalization technique in step M5.
ClassifierMrefM1M2M3M4M5
VGG - 16VHkNN0.772 0.778 ± 0.04 0.792 ± 0.04 0.768 ± 0.03 0.772 ± 0.04
LR0.864 0.852 ± 0.03 0.855 ± 0.03 0.627 ± 0.08 0.846 ± 0.04 x
NB0.698 0.711 ± 0.05 0.696 ± 0.04 0.691 ± 0.04 0.687 ± 0.05
RF0.802 0.792 ± 0.04 0.818 ± 0.04 0.799 ± 0.04 0.780 ± 0.04 x
SVM0.804 0.806 ± 0.04 0.838 ± 0.03 0.785 ± 0.04 0.792 ± 0.04 x
DT0.750 0.744 ± 0.04 0.738 ± 0.04 0.757 ± 0.05 0.724 ± 0.05 x
ADBST 0.732 ± 0.05 0.739 ± 0.05 0.726 ± 0.04
NET 0.838 ± 0.03 0.840 ± 0.03 0.830 ± 0.03
Stack 0.844 ± 0.03
VGG - 16VVkNN0.694 0.701 ± 0.04 0.708 ± 0.04 0.748 ± 0.04 0.700 ± 0.05
LR0.781 0.782 ± 0.04 0.783 ± 0.03 0.608 ± 0.07 0.783 ± 0.04
NB0.660 0.645 ± 0.04 0.657 ± 0.04 0.656 ± 0.04 0.658 ± 0.05
RF0.704 0.695 ± 0.05 0.740 ± 0.04 0.728 ± 0.05 0.703 ± 0.04
SVM0.723 0.714 ± 0.04 0.760 ± 0.03 0.729 ± 0.04 0.722 ± 0.04
DT0.659 0.642 ± 0.04 0.654 ± 0.04 0.676 ± 0.05 0.668 ± 0.04
ADBST 0.650 ± 0.05 0.688 ± 0.05 0.667 ± 0.05
NET 0.784 ± 0.03 0.776 ± 0.04 0.757 ± 0.04
Stack 0.763 ± 0.03
VGG - 19VHkNN0.801 0.801 ± 0.04 0.802 ± 0.04 0.793 ± 0.04 0.788 ± 0.04
LR0.841 0.851 ± 0.03 0.842 ± 0.03 0.641 ± 0.09 0.845 ± 0.03 x
NB0.707 0.699 ± 0.05 0.704 ± 0.04 0.707 ± 0.04 0.698 ± 0.04
RF0.815 0.806 ± 0.03 0.833 ± 0.05 0.810 ± 0.04 0.804 ± 0.03 x
SVM0.824 0.816 ± 0.04 0.849 ± 0.03 0.789 ± 0.04 0.799 ± 0.04 x
DT0.776 0.762 ± 0.04 0.765 ± 0.04 0.775 ± 0.04 0.758 ± 0.04 x
ADBST 0.768 ± 0.05 0.770 ± 0.05 0.752 ± 0.04
NET 0.836 ± 0.04 0.844 ± 0.04 0.833 ± 0.03
Stack 0.841 ± 0.03
VGG - 19VVkNN0.713 0.713 ± 0.04 0.724 ± 0.04 0.763 ± 0.04 0.710 ± 0.04
LR0.774 0.766 ± 0.04 0.768 ± 0.06 0.624 ± 0.07 0.773 ± 0.05 x
NB0.643 0.653 ± 0.04 0.652 ± 0.04 0.658 ± 0.04 0.657 ± 0.04
RF0.719 0.717 ± 0.05 0.751 ± 0.05 0.742 ± 0.05 0.727 ± 0.05 x
SVM0.737 0.731 ± 0.05 0.754 ± 0.06 0.766 ± 0.04 0.746 ± 0.05 x
DT0.670 0.667 ± 0.05 0.663 ± 0.05 0.706 ± 0.04 0.692 ± 0.05 x
ADBST 0.640 ± 0.05 0.698 ± 0.04 0.680 ± 0.04
NET 0.761 ± 0.06 0.777 ± 0.04 0.761 ± 0.05
Stack 0.761 ± 0.04
Table 11. Overall summary of the average of all metrics employed in the classification method—method M1.
Table 11. Overall summary of the average of all metrics employed in the classification method—method M1.
Classifier
MethodCNN-PolMetrickNNLRNBRFSVMDT
M1VGG-16VHAUC0.8370.9280.7380.8760.8970.719
F10.7760.8510.7070.7910.8030.743
Precision0.7820.8550.7210.7950.8260.747
Recall0.7780.8520.7110.7920.8060.744
VGG-16VVAUC0.7460.8730.6630.7670.8020.633
F10.7000.7810.6380.6940.7120.641
Precision0.7030.7850.6570.6970.7220.645
Recall0.7010.7820.6450.6950.7140.642
VGG-19VHAUC0.8420.9230.7370.8860.9110.723
F10.8000.8500.6950.8050.8140.761
Precision0.8030.8520.7080.8090.8250.766
Recall0.8010.8510.6990.8060.8160.762
VGG-19VVAUC0.7610.8500.6790.7930.8090.652
F10.7110.7650.6470.7160.7300.666
Precision0.7180.7690.6660.7200.7340.670
Recall0.7130.7660.6530.7170.7310.667
Table 12. Kruskal–Wallis and Dunn’s test results, comparing the proposed methods with the approaches presented in [26]. For brevity, only the cases with a significant difference in the mean behavior are displayed.
Table 12. Kruskal–Wallis and Dunn’s test results, comparing the proposed methods with the approaches presented in [26]. For brevity, only the cases with a significant difference in the mean behavior are displayed.
MClassifierMClassifierp-ValueMClassifierMClassifierp-Value
VGG-16VHVGG-19VH
p-value <0.001p-value < 0.001
1SVM2LR<0.0011SVM2SVM0.003
1SVM2SVM0.0271LR3RF<0.001
1SVM3NET0.0081NET3RF0.006
1LR3RF<0.0011SVM4LR0.024
1SVM4LR<0.001
1LR4NET0.041
VGG-16VVVGG-19VV
p-value < 0.001p-value < 0.001
1SVM2LR<0.0011SVM2LR0.009
1SVM2NET<0.0011SVM3NET<0.001
1LR3kNN0.0211SVM3SVM0.047
1SVM3kNN0.0261SVM4LR0.008
1SVM3NET<0.001
1SVM4NET<0.001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

da Silva, F.G.; Ramos, L.P.; Palm, B.G.; Machado, R. Assessment of Machine Learning Techniques for Oil Rig Classification in C-Band SAR Images. Remote Sens. 2022, 14, 2966. https://doi.org/10.3390/rs14132966

AMA Style

da Silva FG, Ramos LP, Palm BG, Machado R. Assessment of Machine Learning Techniques for Oil Rig Classification in C-Band SAR Images. Remote Sensing. 2022; 14(13):2966. https://doi.org/10.3390/rs14132966

Chicago/Turabian Style

da Silva, Fabiano G., Lucas P. Ramos, Bruna G. Palm, and Renato Machado. 2022. "Assessment of Machine Learning Techniques for Oil Rig Classification in C-Band SAR Images" Remote Sensing 14, no. 13: 2966. https://doi.org/10.3390/rs14132966

APA Style

da Silva, F. G., Ramos, L. P., Palm, B. G., & Machado, R. (2022). Assessment of Machine Learning Techniques for Oil Rig Classification in C-Band SAR Images. Remote Sensing, 14(13), 2966. https://doi.org/10.3390/rs14132966

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop