Next Article in Journal
Evaluation of Sentinel-3A Wave Height Observations Near the Coast of Southwest England
Previous Article in Journal
Exposure of Marine Turtle Nesting Grounds to Named Storms Along the Continental USA
Previous Article in Special Issue
Inter-Calibration of the OSIRIS-REx NavCams with Earth-Viewing Imagers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ship Identification and Characterization in Sentinel-1 SAR Images with Multi-Task Deep Learning

1
IMT Atlantique—Lab-STICC, UMR CNRS 6285, 29238 Brest, France
2
Univ. Bretagne Sud—IRISA, UMR CNRS 6074, 56017 Vannes, France
3
Collecte Localisation Satellites, 29280 Brest, France
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(24), 2997; https://doi.org/10.3390/rs11242997
Submission received: 2 November 2019 / Revised: 8 December 2019 / Accepted: 11 December 2019 / Published: 13 December 2019
(This article belongs to the Special Issue Remote Sensing: 10th Anniversary)

Abstract

:
The monitoring and surveillance of maritime activities are critical issues in both military and civilian fields, including among others fisheries’ monitoring, maritime traffic surveillance, coastal and at-sea safety operations, and tactical situations. In operational contexts, ship detection and identification is traditionally performed by a human observer who identifies all kinds of ships from a visual analysis of remotely sensed images. Such a task is very time consuming and cannot be conducted at a very large scale, while Sentinel-1 SAR data now provide a regular and worldwide coverage. Meanwhile, with the emergence of GPUs, deep learning methods are now established as state-of-the-art solutions for computer vision, replacing human intervention in many contexts. They have been shown to be adapted for ship detection, most often with very high resolution SAR or optical imagery. In this paper, we go one step further and investigate a deep neural network for the joint classification and characterization of ships from SAR Sentinel-1 data. We benefit from the synergies between AIS (Automatic Identification System) and Sentinel-1 data to build significant training datasets. We design a multi-task neural network architecture composed of one joint convolutional network connected to three task specific networks, namely for ship detection, classification, and length estimation. The experimental assessment shows that our network provides promising results, with accurate classification and length performance (classification overall accuracy: 97.25%, mean length error: 4.65 m ± 8.55 m).

Graphical Abstract

1. Introduction

Deep learning is considered as one of the major breakthroughs related to big data and computer vision [1]. It has become very popular and successful in many fields including remote sensing [2]. Deep learning is a paradigm for representation learning and is based on multiple levels of information. When applied on visual data such as images, it is usually achieved by means of convolutional neural networks. These networks consist of multiple layers (such as convolution, pooling, fully connected, and normalization layers) aiming to transform original data (raw input) into higher level semantic representation. With the composition of enough such elementary operations, very complex functions can be learned. For classification tasks, higher level representation layers amplify aspects of the input that are important for discrimination and discard irrelevant variations. For humans, it is simple through visual inspection to know what objects are in an image, where they are, and how they interact in a very fast and accurate way, allowing performing complex tasks. Fast and accurate algorithms for object detection are thus sought to allow computers to perform such tasks, at a much larger scale than humans can achieve.
Ship detection and classification have been extensively addressed with traditional pattern recognition techniques for optical images. Zhu et al. [3] and Antelo et al. [4] extracted handcrafted features from images such as shapes, textures, and physical properties, while Chen et al. [5] and Wang et al. [6] exploited dynamic Bayesian networks to classify different kinds of ships. Such extracted features are known for their lack of robustness, which can raise challenges in practical applications (e.g., they may lead to poor performances when the images are corrupted by blur, distortion, or illumination, which are common artifacts in remote sensing). Furthermore, they cannot overcome the issues raised by big data such as image variabilities (i.e., ships of the same type may have different shapes, colors, sizes, etc.) and data volume. Recently, following the emergence of deep learning, an autoencoder based deep neural network combined with extreme learning machine was proposed [7] and outperformed some other methods using SPOT-5 spaceborne optical images for ship detection.
Compared with optical remote sensing, satellite SAR imaging appears more suited for maritime traffic surveillance in operational contexts, as it is not critically affected by weather conditions and day-night cycles. In this context, open-source Sentinel-1 SAR data are particularly appealing. Almost all coastal zones and shipping routes are covered by Interferometric Wide Swath Mode (IW), while the Extra-Wide Swath Mode (EW) acquires data over open oceans, providing a global coverage for sea oriented applications. Such images, combined with the Automatic Identification System (AIS), represent a large amount of data that can be employed for training deep learning models [8]. AIS provides meaningful and relevant information about ships (such as position, type, length, rate of turn, speed over ground, etc.). The combination of these two data sources could leverage new applications to the detection and estimation of ship parameters from SAR images, which remains a very challenging task. Indeed, detecting inshore and offshore ships is critical in both military and civilian fields (e.g., for monitoring of fisheries, management of maritime traffic, safety of coast and sea, etc.). In operational contexts, the approaches used so far still rely on manual visual interpretations that are time consuming, possibly error prone, and definitely irrelevant to scale up to the available data streams. On the contrary, the availability of satellite data such as Sentinel-1 SAR makes possible the exploration of efficient and accurate learning based schemes.
One may however consider with care AIS data as they involve specific features. AIS is mandatory for large vessels (e.g., >500 GT, passenger vessels). As such, it provides representative vessel datasets for international maritime traffic, but may not cover some maritime activities (e.g., small fishing vessels). Though not authorized, ships can easily turn off their AIS and/or spoof their identity. While AIS tracking strategies [9] may be considered to address missing track segments, the evaluation of spoofing behavior is a complex task. Iphar et al. [10] evaluated that amongst ships with AIS, about 6% have no specified type, and 3% are only described as “vessels”. Besides, respectively 47% and 18% of the vessels may involve uncertain length and beam data. These points should be considered with care in the analysis of AIS datasets, especially when considering learning strategies as addressed in this work.
Among existing methods for ship detection in SAR images, Constant False Alarm Rate (CFAR) based methods have been widely used [11,12]. The advantage of such methods is their reliability and high efficiency. Using AIS information along with SAR images significantly improves ship detection performance [13]. As the choice of features has an impact on the performance of discrimination, deep neural networks have recently taken the lead thanks to their ability to extract (or learn) features that are richer than handcrafted (or expert) features. In [14], a framework named Sea-Land Segmentation based Convolutional Neural Network (SLS-CNN) was proposed for ship detection, combined with the use of saliency computation. A modified Faster R-CNN based on the CFAR algorithm for SAR ship detection was proposed in [15] with good detection performance. In [16], Texture Features extracted from SAR images were fed into Artificial Neural Networks (TF-ANN) to discriminate ship pixels from sea ones. Schwegmann et al. [17] employed a highway network for ship detection in SAR images and achieved good results, especially in reducing the false detection rate. These state-of-the-art approaches focused on ship detection in SAR images. In this paper, we aim to go beyond ship detection and investigate higher level tasks, namely the identification of ship types (also known as classification) and their length estimation, which to our knowledge remain poorly addressed using learning based frameworks.
The problem of ship length estimation from SAR images was briefly discussed in [18,19]. In [18], the best shape of a ship was extracted from a SAR image using inertia tensors. The estimated shape allowed obtaining the ship length. However, the absence of a ground truth does not allow validating the accuracy of this method. In [19], a three step method was proposed in order to extract a rectangle that would be the reference model for ship length estimation. The method produced good results (mean absolute error: 30 m ± 36.6 m). However, the results were presented on a limited dataset (only 127 ships), and their generalization may be questioned.
In this paper, we propose a method based on deep learning for ship identification and characterization with the synergetic use of Sentinel-1 SAR images and AIS data.

2. Material and Methods

The proposed framework combined the creation of a reference ground truthed dataset using AIS-SAR synergies and the design of a multi-task deep learning model. In this section, we first introduce the proposed multi-task neural network architecture, which jointly addresses ship detection, classification, and length estimation. Second, we describe the training framework in terms of the considered training losses and of the implemented optimization scheme. Third, we detail the creation of the considered reference datasets, including how we tackled data augmentation and class imbalance issues, which have been shown to be critical for the learning process.

2.1. Proposed Framework

The proposed multi-task framework was based on two stages, with a first common part and then three task oriented branches for ship detection, classification, and length estimation, respectively (see Figure 1). The first part was a convolutional network made of 5 layers. It was followed by the task oriented branches. All these branches were made of convolutional layers followed by fully connected layers (the number of which depended on the complexity of the task). For the detection task, the output consisted of a pixel-wise probability map of the presence of ships. It only required 1 fully connected layer after the 4 convolutional layers. For the classification task, we considered 4 or 5 ship classes (cargo, tanker, fishing, passenger, and optionally tug). The branch also required 4 convolutional layers and 2 fully connected layers. The last task was related to the length estimation. This branch was composed of 4 convolutional layers and 5 fully connected layers.
This architecture was inspired by state-of-the-art architectures [20,21,22]. The number of layers was chosen to be similar to the first layers of the VGG network [22]. All the activations of the convolutional layers and fully connected layers were ReLU [23]. Other activation functions were employed for the output layers: a sigmoid for the detection, a softmax activation for the classification, and a linear activation employed for the length estimation; further details are presented in Section 2.2. We may emphasize that our model included a detection component. Though this was not a targeted operational objective in our context, it was shown to improve the performance for the other tasks (see Section 3.3).

2.2. Training Procedure

We describe below the considered training strategy, especially the training losses considered for each task specific component. The proposed end-to-end learning scheme combined task specific training losses as follows:
  • Detection loss: The detection output was the ship presence probability. We employed a binary cross-entropy loss, which is defined by:
    L d e t = 1 N n = 1 N k I ( y k log ( p ( k ) ) + ( 1 y k ) log ( 1 p ( k ) ) ,
    where N is the number of samples, k is a pixel of the output detection image I, y k is the ground truth of ship presence (0 or 1), and p ( k ) is the predicted probability of ship presence. It is a usual loss function for binary classification tasks [24].
  • Classification loss: The output for the last classification layer was the probability that the input image corresponded to one of the considered ship types. We used here the categorical cross-entropy loss:
    L c l a s s = 1 N n = 1 N c = 1 n c ( y o , c log ( p o , c ) ) ,
    where N is the number of samples, n c is the number of classes (here, n c = 4 or n c = 5 ), y o , c is a binary indicator (0 or 1) if class label c is the correct classification for observation o, and p o , c is the predicted probability for the observation o to belong to class c. It is a widely used loss function for multiclass classification tasks [25,26].
  • Length estimation loss: In the length estimation network, the 4 fully connected layers of shape (64 × 1 × 1) were connected to each other (see Figure 1). The idea was to propagate the difference between the first layer and the current layer and was related to residual learning [27]. We used here the mean squared error, defined as:
    L l e n g t h = 1 N n = 1 N ( l p r e d l t r u e ) 2 ,
    where N is the number of samples, l p r e d is the predicted length, and l t r u e is the true length.
Overall, we define the loss function of the whole network as:
L = L d e t + L c l a s s + L l e n g t h .
Each specific loss employed to design the loss of the whole network could have been weighted. Nevertheless, we observed no significant effect of such a weighting scheme. Thus, we decided to rely on a simple combination through adding the different task dedicated losses, giving the same importance to each task. Our network was trained end-to-end using the RMSProp optimizer [28]. The weights of the network were updated by using a learning rate of 1 × 10−4 and a learning rate decay over each update of 1 × 10−6 over the 500 iterations. Such parameterization showed good results for our characterization tasks.

2.3. Creation of Reference Datasets

With a view toward implementing deep learning strategies, we first address the creation of reference datasets from the synergy between AIS data and Sentinel-1 SAR data. AIS transceiver sent data every 2 to 10 s. These data mainly consisted of the positional accuracy (up to 0.0001 min precision) and the course over ground (relative to True North to 0.1°). For a given SAR image, one could interpolate AIS data from the associated acquisition time. Thus, it was possible to know the precise location of the ships in the SAR image and the related information (in our case, length and type). The footprint of the ship was obtained by thresholding the SAR image in the area where it was located (the brightest pixel of the image). Since the database was very unbalanced in terms of class distribution, a strategy was also proposed in order to enlarge the training set with translations and rotations, which is a standard procedure for database enlargement (also known as data augmentation). Concurrently to our work, a similar database was proposed in [29]. We also evaluated our framework with this dataset (see Section 3.5).
In our experiments, we considered a dataset composed of 18,894 raw SAR images of size 400 × 400 pixels with a 10 m resolution. The polarization of the images were either HH (proportion of Horizontally transmitted waves that return Horizontally) or VV (proportion of Vertically transmitted waves that return Vertically). Polarization had a significant effect on SAR backscatter. However, our goal was to allow us to process any Sentinel-1 SAR images. We thus considered any HH and VV polarized image without prior information on the type of polarization. Each image was accompanied with the incidence angle since it impacted the backscatter intensity of the signal. For the proposed architecture, the input was a 2 band image (backscatter intensity and incidence angle). Thus, we did not use any pre-trained network since we assumed that they could not handle such input data. We relied on the Automatic Identification System (AIS) to extract images that contained a ship in their center. AIS also provided us with information about the ship type and length. As stated before, AIS may have been corrupted (e.g., with spoofing), so when creating the database, we only considered ships that responded to the two following criteria: (i) their type was clearly defined (i.e., they belonged to the retained classes); (ii) their length was greater than 0 and smaller than 400 m (the largest ship in the world). Besides, the SAR images we selected were acquired over European waters, where we expected AIS data to be of higher quality compared with other maritime areas.
The dataset was strongly imbalanced, amongst the 5 classes (tanker, cargo, fishing, passenger and tug), cargo was the most represented (10,196 instances), while tug was the least represented (only 444 instances). The class distribution is detailed in Figure 2 and Table 1. The length distribution showed that tanker, cargo, and passenger ships had similar length distributions. fishing ships had relatively small lengths, while tug ships’ lengths were intermediate.
To account for class imbalance [30], we applied data augmentation with translations and rotations. We first performed a rotation of a random angle centered on the brightest pixel of the SAR image (the center of the ship) and then performed a random translation. The same transformation was applied to the incidence angle image. The images employed to train the networks were of size 80 × 80 pixels. They contained ships (not necessarily in their center; see Figure 3). The ship footprint ground truth was generated by thresholding the SAR image since we precisely knew the location of the ship (i.e., it was the brightest pixel of the SAR image; see Figure 4). The obtained footprint was not perfect (see Figure 4b), but was shown to be sufficient to train the network. Let us note that a CFAR approach could have been employed in order to extract more precisely the ship footprint [11]. However, since our goal was not to detect ships, a coarse ship footprint was sufficient. We considered 2 configurations for the databases; a 4 class database, employed to compare our baseline to other state-of-the-art approaches (namely MLP and R-CNN), and a 5 class database in order to evaluate how our network responded with more classes. Each database was composed of 20,000 images of 80 × 80 pixels, with the same amount of samples per class (5000 per class for the 4 class database and 4000 per class for the 5 class database). The networks were trained with 16,000 images, and the remaining 4000 were used for validation. Throughout the data augmentation process, we ensured that images could be seen either in the training or validation set, but not in both. Ships with no AIS signal were not considered in our dataset (neither to train nor evaluate our model), since our strategy to build the dataset relied on matching the AIS signal with SAR imagery. However, once a model was trained, it could perform in an operational settings to detect ships with no AIS (this is indeed one of our long term goals).

3. Results

We ran all numerical experiments on a PC with an NVidia GTX 1080 Ti, an Intel Xeon W-2145 CPU 3.70 GHz, and 64 GB RAM (with a Keras [31] implementation). We evaluated the proposed framework with respect to other popular deep learning based solutions. We first considered a Multi-Layer Perceptron (MLP) [32] with only one hidden layer with 128 hidden units. The MLP is the most simple network that can be proposed for the desired task and could be a good basis in order to evaluate the performance of our network. We also designed a R-CNN (Regions with CNN features) [33] network in order to extract ship bounding boxes along with classification. Even if the R-CNN based bounding boxes did not allow precisely measuring the ship length, they could provide a good basis for its estimation. R-CNN is a state-of-the-art algorithm for object detection and classification [33]. Thus, it is worth being compared with our proposed model. The R-CNN had a very simple architecture, as presented in Figure 5. The networks were trained using 16,000 images from the augmented dataset, and the remaining 4000 images were used for validation.
The evaluation of the models was performed using several metrics. The classification task was assessed through the confusion matrix, giving, for each class and overall, several metrics. The Intersection over Union (IoU or Jaccard index) [34] measures similarity between finite sample sets and is defined as the size of the intersection divided by the size of the union of the sample sets. It was designed for the evaluation of object detection. The F-score is the harmonic mean of precision and recall, and it reaches its best value at one and the worst at zero. The Kappa coefficient [35] ( κ ) is generated from a statistical test to evaluate the accuracy of a classification. Kappa essentially evaluates how well the classification performs as compared to just randomly assigning values (i.e., did the classification do better than randomness?). The Kappa coefficient can range from −1 to one. A value of zero (respectively −1 or one) indicates that the classification is no better (respectively worse or better) than a random classification. For the length estimation task, the mean error (and its standard deviation) were employed. For a ship k, the length error is defined as e k = l k , p r e d l k , t r u e , where l k , p r e d is the predicted length and l k , t r u e is the actual length. The mean error m e r r _ l e n g t h (respectively the standard deviation s t d e v e r r _ l e n g t h ) is the mean (respectively the standard deviation) of all the e k . We further refer to the mean error as m e r r _ l e n g t h ± s t d e v e r r _ l e n g t h .

3.1. MLP Model

For an 80 × 80 image, the MLP ran at 2000 frames per second. The whole training took about one hour. The testing took less than a minute. It produced very poor results. Indeed, the overall accuracy for classification was 25%, which means that the classifier assigned the same class to all the images (see Table 2). The length estimation was also rather inaccurate, the ship length being underestimated with a very large standard deviation (mean error: −7.5 m ± 128 m).

3.2. R-CNN Model

For an 80 × 80 image, the R-CNN ran at 333 frames per second. The whole training took about 6.5 h and the testing about a minute. It produced better results than the MLP. The network estimated the 4 corners of the bounding box. As the ground truth for bounding boxes was obtained from the ship footprint extracted by thresholding the SAR image, it might not be well defined (see Figure 6c,d). In Figure 6c, the bounding box is well centered on the ship, but has the wrong size. In Figure 6d, the bounding box is also not well sized and accounts for the brightest part of the ship. We recall that the detection task was not our main objective, but rather regarded as a means to better constrain the training of the models. The R-CNN had a classification overall accuracy of 89.29%. Several other metrics are presented in Table 3.

3.3. Our Network

For an 80 × 80 image, our method could run at 250 frames per second. The whole training took about 9 h and the testing about a minute. With an overall accuracy and a mean F-score of 97.2%, the proposed multi-task architecture significantly outperformed the benchmarked MLP and R-CNN models. We report in Table 4 the confusion matrix and additional accuracy metrics. Interestingly, the classification performances were relatively homogeneous across ship types (mean accuracy above 92% for all classes). Tankers involved a greater misclassification rate with some confusion with cargo.
Regarding length estimation performance, our framework achieved very promising results. The length was slightly over-estimated (mean error: 4.65 m ± 8.55 m), which is very good regarding the spatial resolution of the Sentinel-1 SAR data (10 m/pixel). To our knowledge, this was the first demonstration that reasonably accurate ship length estimates could be derived from SAR images using learning based schemes, whereas previous attempts using model driven approaches led to much poorer performance. Overall, the results of the classification and length estimation tasks for all the tested architectures are summarized in Table 5.
We also trained our model with five classes, and it confirmed that our framework performed well. The length was slightly over-estimated (mean error: 1.93 m ± 8.8 m), and the classification was also very good (see Table 6). Here, we still report some light confusion for the tanker and cargo classes. The accuracy metrics were slightly worse than the four class model, but we still report an overall accuracy and a mean F-score of 97.4%.
We further analyze the proposed scheme and the relevance of the multi-task setting, compared with task specific architectures. To this end, we performed an ablation study and trained the proposed architecture using (i) length estimation loss only, (ii) classification loss only, and (iii) the combination of length estimation and classification losses (i.e., without the detection loss). We report in Table 7 the resulting performances compared to those of the proposed end-to-end learning strategy. Regarding the classification issue, combined losses resulted in an improvement of about 1.3% (above 25% in terms of relative gain). The improvement was even more significant for length estimation with a relative gain in the mean error of about 36%. Interestingly, we noted that the additional use of the detection loss also greatly contributed to the improvement of length estimation performance (mean error 2.85 m without using the detection loss during training vs. 1.93 m when using jointly detection, classification, and length estimation losses). As an illustration of the detection component of the proposed architecture, we illustrate in Figure 7a the detection result. As mentioned above, the thorough evaluation of this detection model was not the main objective of this study. Furthermore, without any precise ship footprint ground truth, it was impossible to evaluate quantitatively the performance of the network for this specific task. Let us recall that the detection task has been widely addressed in the literature [14,15,16]. Overall, this complementary evaluation supported the idea that neural network architectures for SAR image analysis may share some low level task independent layers, whose training can highly benefit from the existence of multi-task datasets.

3.4. Application to a Full SAR Image

We illustrate here an application of the proposed approach to a real SAR image acquired on 4 April 2017 in Western Brittany, France. We proceeded in several steps as follows. First, a CFAR based ship detector was applied. Then, for each detected ship, we applied the trained deep network model to predict the ship category and its length. For illustration purposes, we report in Figure 8 the detected ships that could be matched to AIS signals.
For the considered SAR image, among the 98 ships detected by the CFAR based ship detector, 66 ships had their length documented, and 69 ships belonged to one of the five proposed classes after AIS matching. We may point out that the tug class is not represented. We report classification and length estimation performance in Table 8. Ship classification performance was in line with the performance reported above. Regarding length estimation, the mean error 14.56 m ± 39.98 m was larger than that reported for the ground truthed dataset. Still, this error level was satisfactory given the pixel resolution of 10m of the SAR image. Let us note that, given the limited samples available, the standard deviation was not fully relevant here. While a special care was undertaken for the creation of our SAR-AIS dataset, this application to a single SAR image exploited the raw AIS data. AIS data may be significantly corrupted, which may partially explain these differences.

3.5. Application to the OpenSARShip Dataset

The OpenSARShip dataset [29] has recently been made available to the community. We report here the results obtained with our framework when applied on this dataset. This dataset comprises SAR data with different polarization characteristics and also includes ship categories. With a view toward easing the comparison with the previous results, we focused on SAR images that were in VV polarization and the ship categories considered, which led to considering the following four categories; tanker, cargo, fishing, and passenger. Overall, we considered a dataset of 5225 ships (80% were employed for training and 20% for testing). In the OpenSARShip dataset, the classes are not equally represented. We report classification and length estimation performance in Table 9. We also evaluated the performance of the model trained on our dataset and applied on the OpenSARShip dataset and conversely.
The results showed that our model produced good results when trained and tested on the same database. However, the results did not transfer from one dataset to another. We suggest that this may relate to differences in the maritime traffic and environment between Europe (our dataset) and Asia (OpenSARShip dataset). The comparison to previous work on the OpenSARShip dataset was not straightforward. For instance, the work in [36] considered only a three class dataset (tanker, cargo, and other). The reported accuracy score (76%) was lower than our 87.7% accuracy score for the considered five class dataset. We may also emphasize that [36] did not address ship length estimation.

4. Discussion

The reported results showed that a dedicated architecture was necessary for ship classification and length estimation, while state-of-the art architectures failed to achieve satisfying performances. The MLP was sufficient for ship detection on SAR images (from a visual assessment). However, this should not be considered as a good result since we only had (positive) examples of ships in our database (no negative samples, so we could not assess the false positives). Thus, the network only learned a thresholding and could not discard a ship from other floating objects (e.g., icebergs). Indeed, iceberg detection and discrimination between iceberg and ship are specific research questions [37,38]. Overall, the performance of the MLP stressed the complexity of the classification and length estimation tasks. In terms of classification accuracy, the R-CNN performed better than the MLP, with an overall accuracy of 88.57%. These results support the proposed architecture with three task specific networks, which shared a common low level network. The latter was interpreted as a feature extraction unit that the task specific networks relied on.
Compared to the state-of-the art architectures (MLP and R-CNN), our model produced better results for ship classification and length estimation from Sentinel-1 SAR images with only a few confusions between classes. A multi-task architecture was well adapted for simultaneous ship classification and length estimation. Our model also performed well when a new class was added (e.g., tug). Furthermore, adding a detection task (even with a coarse ground truth) tended to improve the length estimation. Our experiments also showed that the trained models did not transfer well from one dataset to another. We suggest that this may relate to differences in the characteristics of the maritime traffic and/or marine environment. Future work should further explore these aspects for the application of the proposed model worldwide.

5. Conclusions

In this paper, a multi-task neural network approach was introduced. It jointly addressed the detection, classification, and length estimation of ships in Sentinel-1 SAR images. We exploited synergies between AIS and Sentinel-1 to build reference datasets automatically for training and evaluation purposes, with the ultimate goal of relying solely on SAR imagery to counter a lack or corruption of AIS information that corresponds to illegal activities. While the polarization type had a significant effect on SAR backscatter, we were able to train a model that jointly processed HH or VV polarization without prior information on the type of polarization. Our results supported the assumption that HH and VV polarizations shared common image features and that differences in backscatter distributions could be handled through an appropriate parameterization of the network.
Regarding the considered architecture, a mutual convolutional branch transformed raw inputs into meaningful information. Such information was fed into three task specific branches. Experimental evaluation showed improvement over standard MLP or R-CNN. Ship detection cannot be totally assessed, but a visual inspection supported the relevance of this detection stage. Besides, it was shown to significantly contribute to improved performance of the classification and length estimation components. Overall, we report the promising performance for ship classification (above 90% of correct classification) and length estimation (relative bias below 10%). Considering a residual architecture appeared as a critical feature to reach good length estimation performance, this would require further investigation.
Future work may further investigate the training and evaluation of the detection stage. The automation of the matching process between AIS data and SAR images has the potential for significantly increasing the size and diversity of the training and evaluation datasets. This may provide new avenues to address generalization and transfer issues between geographic areas pointed out in our results. Furthermore, while SAR imagery is less affected by weather conditions than optical imagery, a specific analysis of the impact of weather conditions on the identification performance would also be of interest. Finally, the specificity of the SAR imagery would call for dedicated operations, while our network relied on standard techniques issuing from computer vision.

Author Contributions

C.D. carried out the experiment and wrote the manuscript with support from S.L. and R.F., R.V. and G.H. provided the SAR data. All authors contributed to the final manuscript.

Funding

This research was funded by ANR (Agence Nationale de la Recherche) SESAME (ANR-16-ASTR-0026).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  3. Zhu, C.; Zhou, H.; Wang, R.; Guo, J. A novel hierarchical method of ship detection from spaceborne optical image based on shape and texture features. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3446–3456. [Google Scholar] [CrossRef]
  4. Antelo, J.; Ambrosio, G.; Gonzalez, J.; Galindo, C. Ship detection and recognition in high-resolution satellite images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Cape Town, South Africa, 12–17 July 2009. [Google Scholar]
  5. Chen, H.; Gao, X. Ship recognition based on improved forwards-backwards algorithm. In Proceedings of the International Conference on Fuzzy Systems and Knowledge Discovery, Tianjin, China, 14–16 August 2009; pp. 509–513. [Google Scholar]
  6. Wang, Q.; Gao, X.; Chen, D. Pattern recognition for ship based on Bayesian networks. In Proceedings of the International Conference on Fuzzy Systems and Knowledge Discovery, Haikou, China, 24–27 August 2007; pp. 684–688. [Google Scholar]
  7. Tang, J.; Deng, C.; Huang, G.B.; Zhao, B. Compressed-domain ship detection on spaceborne optical image using deep neural network and extreme learning machine. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1174–1185. [Google Scholar] [CrossRef]
  8. Fablet, R.; Bellec, N.; Chapel, L.; Friguet, C.; Garello, R.; Gloaguen, P.; Hajduch, G.; Lefèvre, S.; Merciol, F.; Morillon, P.; et al. Next Step for Big Data Infrastructure and Analytics for the Surveillance of the Maritime Traffic from AIS & Sentinel Satellite Data Streams. In Proceedings of the BiDS’2017-Conference on Big Data from Space, Toulouse, France, 28–30 November 2017; pp. 1–4. [Google Scholar]
  9. Nguyen, D.; Vadaine, R.; Hajduch, G.; Garello, R.; Fablet, R. Multi-task Learning for Maritime Traffic Surveillance from AIS Data Streams. In Proceedings of the IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA), Turin, Italy, 1–3 October 2018; pp. 331–340. [Google Scholar]
  10. Iphar, C.; Napoli, A.; Ray, C. Data quality assessment for maritime situation awareness. In Proceedings of the ISSDQ 2015-The 9th International Symposium on Spatial Data Quality, Montpellier, France, 29–30 September 2015; Volume 2. [Google Scholar]
  11. Liao, M.; Wang, C.; Wang, Y.; Jiang, L. Using SAR images to detect ships from sea clutter. IEEE Geosci. Remote Sens. Lett. 2008, 5, 194–198. [Google Scholar] [CrossRef]
  12. An, W.; Xie, C.; Yuan, X. An improved iterative censoring scheme for CFAR ship detection with SAR imagery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4585–4595. [Google Scholar]
  13. Pelich, R.; Chini, M.; Hostache, R.; Matgen, P.; Lopez-Martinez, C.; Nuevo, M.; Ries, P.; Eiden, G. Large-Scale Automatic Vessel Monitoring Based on Dual-Polarization Sentinel-1 and AIS Data. Remote Sens. 2019, 11, 1078. [Google Scholar] [CrossRef] [Green Version]
  14. Liu, Y.; Zhang, M.H.; Xu, P.; Guo, Z.W. SAR ship detection using sea-land segmentation based convolutional neural network. In Proceedings of the International Workshop on Remote Sensing with Intelligent Processing, Shanghai, China, 18–21 May 2017; pp. 1–4. [Google Scholar]
  15. Kang, M.; Leng, X.; Lin, Z.; Ji, K. A modified faster R-CNN based on CFAR algorithm for SAR ship detection. In Proceedings of the International Workshop on Remote Sensing with Intelligent Processing, Shanghai, China, 18–21 May 2017; pp. 1–4. [Google Scholar]
  16. Khesali, E.; Enayati, H.; Modiri, M.; Aref, M.M. Automatic ship detection in Single-Pol SAR Images using texture features in artificial neural networks. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2015, 40, 395. [Google Scholar] [CrossRef] [Green Version]
  17. Schwegmann, C.P.; Kleynhans, W.; Salmon, B.P.; Mdakane, L.W.; Meyer, R.G. Very deep learning for ship discrimination in synthetic aperture radar imagery. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 104–107. [Google Scholar]
  18. Bedini, L.; Righi, M.; Salerno, E. Size and Heading of SAR-Detected Ships through the Inertia Tensor. Proceedings 2018, 2, 97. [Google Scholar] [CrossRef] [Green Version]
  19. Stasolla, M.; Greidanus, H. The exploitation of Sentinel-1 images for vessel size estimation. Remote Sens. Lett. 2016, 7, 1219–1228. [Google Scholar] [CrossRef] [Green Version]
  20. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in neural information processing systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  22. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  23. Klambauer, G.; Unterthiner, T.; Mayr, A.; Hochreiter, S. Self-normalizing neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 971–980. [Google Scholar]
  24. Drozdzal, M.; Vorontsov, E.; Chartrand, G.; Kadoury, S.; Pal, C. The importance of skip connections in biomedical image segmentation. In Deep Learning and Data Labeling for Medical Applications; Springer: Berlin, Germany, 2016; pp. 179–187. [Google Scholar]
  25. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  26. Geng, J.; Fan, J.; Wang, H.; Ma, X.; Li, B.; Chen, F. High-resolution SAR image classification via deep convolutional autoencoders. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2351–2355. [Google Scholar] [CrossRef]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  28. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the Gradient by a Running Average of Its Recent Magnitude. In COURSERA: Neural Networks for Machine Learning; Coursera Inc.: Mountain View, CA, USA, 2012. [Google Scholar]
  29. Huang, L.; Liu, B.; Li, B.; Guo, W.; Yu, W.; Zhang, Z.; Yu, W. OpenSARShip: A dataset dedicated to Sentinel-1 ship interpretation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 11, 195–208. [Google Scholar] [CrossRef]
  30. Kellenberger, B.; Marcos, D.; Tuia, D. Detecting mammals in UAV images: Best practices to address a substantially imbalanced dataset with deep learning. Remote Sens. Environ. 2018, 216, 139–153. [Google Scholar] [CrossRef] [Green Version]
  31. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 12 December 2019).
  32. Kruse, R.; Borgelt, C.; Klawonn, F.; Moewes, C.; Steinbrecher, M.; Held, P. Multi-layer perceptrons. In Computational Intelligence; Springer: Berlin, Germany, 2013; pp. 47–81. [Google Scholar]
  33. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  34. Jaccard, P. The distribution of the flora in the alpine zone. 1. New Phytol. 1912, 11, 37–50. [Google Scholar] [CrossRef]
  35. Cohen, J. A coefficient of agreement for nominal scales. Educ. Psychol. Meas. 1960, 20, 37–46. [Google Scholar] [CrossRef]
  36. Seungryong, K.; Jeongju, B.; Chan-Su, Y. Satellite image based ship classification method with sentinel-1 IW mode data. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Yokohama, Japan, 28 July–2 August 2019; pp. 1300–1301. [Google Scholar]
  37. Power, D.; Youden, J.; Lane, K.; Randell, C.; Flett, D. Iceberg detection capabilities of RADARSAT synthetic aperture radar. Can. J. Remote Sens. 2001, 27, 476–486. [Google Scholar] [CrossRef]
  38. Bentes, C.; Frost, A.; Velotto, D.; Tings, B. Ship-iceberg discrimination with convolutional neural networks in high resolution SAR images. In Proceedings of the EUSAR 2016: 11th European Conference on Synthetic Aperture Radar, Hamburg, Germany, 6–9 June 2016; pp. 1–4. [Google Scholar]
Figure 1. Proposed multi-task architecture for ship detection, classification (4 classes) and length estimation from a Sentinel-1 SAR image.
Figure 1. Proposed multi-task architecture for ship detection, classification (4 classes) and length estimation from a Sentinel-1 SAR image.
Remotesensing 11 02997 g001
Figure 2. Boxplot of the length distribution for each class in our dataset. Length data are given in meters.
Figure 2. Boxplot of the length distribution for each class in our dataset. Length data are given in meters.
Remotesensing 11 02997 g002
Figure 3. Examples of SAR image (with backscatter intensity) for each ship type of the collected database.
Figure 3. Examples of SAR image (with backscatter intensity) for each ship type of the collected database.
Remotesensing 11 02997 g003
Figure 4. Example of a SAR image (with backscatter intensity) and associated ship footprint (best viewed in color).
Figure 4. Example of a SAR image (with backscatter intensity) and associated ship footprint (best viewed in color).
Remotesensing 11 02997 g004
Figure 5. R-CNN architecture considered for ship classification.
Figure 5. R-CNN architecture considered for ship classification.
Remotesensing 11 02997 g005
Figure 6. Illustration of the detection and classification performance of the evaluated R-CNN model: each subpanel depicts a SAR image with the superimposed detected bounding box. (a) Accurate bounding box superimposed on the SAR image. The ship type (passenger) is well predicted. (b) Accurate bounding box superimposed on the SAR image. The ship type (tanker) is well predicted. (c) Inaccurate bounding box superimposed on the SAR image. The ship type predicted (fishing) is not the right one (cargo). (d) Inaccurate bounding box superimposed on the SAR image. The ship type (tanker) is well predicted.
Figure 6. Illustration of the detection and classification performance of the evaluated R-CNN model: each subpanel depicts a SAR image with the superimposed detected bounding box. (a) Accurate bounding box superimposed on the SAR image. The ship type (passenger) is well predicted. (b) Accurate bounding box superimposed on the SAR image. The ship type (tanker) is well predicted. (c) Inaccurate bounding box superimposed on the SAR image. The ship type predicted (fishing) is not the right one (cargo). (d) Inaccurate bounding box superimposed on the SAR image. The ship type (tanker) is well predicted.
Remotesensing 11 02997 g006
Figure 7. Example of detection output for the considered multi-task architecture: (left), SAR image (with backscatter intensity) used as the input; (right), the output of the detection module of the considered architecture.
Figure 7. Example of detection output for the considered multi-task architecture: (left), SAR image (with backscatter intensity) used as the input; (right), the output of the detection module of the considered architecture.
Remotesensing 11 02997 g007
Figure 8. SAR image acquired on 4 April 2017 in Western Brittany, France. (a) Image location. (b) SAR image with ships identified from the matching of CFAR based detection and the Automatic Identification System (AIS). tanker, cargo, fishing, passenger (best viewed in color).
Figure 8. SAR image acquired on 4 April 2017 in Western Brittany, France. (a) Image location. (b) SAR image with ships identified from the matching of CFAR based detection and the Automatic Identification System (AIS). tanker, cargo, fishing, passenger (best viewed in color).
Remotesensing 11 02997 g008
Table 1. Length distribution and number of samples for each class in our dataset.
Table 1. Length distribution and number of samples for each class in our dataset.
TankerCargoFishingPassengerTug
Number of
samples
47371019626641071444
Length mean (m)168.4146.526.6153.547.3
Length standard
deviation (m)
64.660.812.268.926.0
Number of augmented
samples (4 classes)
263023363929-
Number of augmented
samples (5 classes)
00133629293556
Table 2. Confusion matrix and accuracy metrics for the MLP with 4 classes.
Table 2. Confusion matrix and accuracy metrics for the MLP with 4 classes.
Confusion Matrix
PredictionTankerCargoFishingPassengerPrecision
Ground Truth
Tanker1000000100.0
Cargo10000000.0
Fishing10000000
Passenger10000000.0
Recall25.0---
Accuracy metrics
LabelTankerCargoFishingPassengerOverall
IoU25.00.00.00.06.25
F-Score40.0---10.00
Accuracy25.075.075.075.025.00
κ 0.00.00.00.00.25
Table 3. Confusion matrix and accuracy metrics for the R-CNN with 4 classes.
Table 3. Confusion matrix and accuracy metrics for the R-CNN with 4 classes.
Confusion Matrix
PredictionTankerCargoFishingPassengerPrecision
Ground Truth
Tanker8459733386.40
Cargo98787247779.82
Fishing298915193.49
Passenger815196197.56
Recall88.6786.6796.9585.65
Accuracy metrics
LabelTankerCargoFishingPassengerOverall
IoU77.8171.0990.8383.8680.90
F-Score87.5283.1095.1991.2289.26
Accuracy93.8291.8097.6995.2689.29
κ 0.830.780.940.880.88
Table 4. Confusion matrix and accuracy metrics for the proposed network with 4 classes.
Table 4. Confusion matrix and accuracy metrics for the proposed network with 4 classes.
Confusion Matrix
PredictionTankerCargoFishingPassengerPrecision (%)
Ground Truth
Tanker985110498.5
Cargo65907121690.7
Fishing02998099.8
Passenger0001000100.0
Recall(%)93.8198.5998.8198.04
Accuracy metrics
LabelTankerCargoFishingPassengerOverall
IoU(%)92.4989.5498.6298.0494.67
F-Score(%)96.1094.4899.3099.0197.22
Accuracy(%)98.0097.3599.6599.5097.25
κ 0.950.930.990.990.97
Table 5. Results of all the tested architectures for the classification (4 classes) and length estimation.
Table 5. Results of all the tested architectures for the classification (4 classes) and length estimation.
ArchitectureLength Mean Error (m)Classification Overall Accuracy (%)
MLP−7.50 ± 12825.00
R-CNN-88.57
Our network4.65 ± 8.5597.25
Table 6. Confusion matrix and accuracy metrics for the proposed network with 5 classes.
Table 6. Confusion matrix and accuracy metrics for the proposed network with 5 classes.
Confusion Matrix
PredictionTankerCargoFishingPassengerTugPrecision (%)
Ground Truth
Tanker7712801096.38
Cargo6073233291.50
Fishing017990099.88
Passenger310796099.50
Tug0000800.0100.00
Recall(%)92.4596.0699.6399.5099.75
Accuracy metrics
LabelTankerCargoFishingPassengerTugOverall
IoU(%)89.3488.1999.5099.0099.7595.16
F-Score(%)94.3793.7399.7599.5099.8897.44
Accuracy(%)97.7097.5599.9099.8099.9597.45
κ 0.930.921.000.991.000.97
Table 7. Ablation study; performance of the network for different scenarios: (i) only length estimation, (ii) only classification, and (iii) length estimation and classification without detection.
Table 7. Ablation study; performance of the network for different scenarios: (i) only length estimation, (ii) only classification, and (iii) length estimation and classification without detection.
Length Mean Error (m)Classification Overall Accuracy (%)
(i) 3.07 ± 9.0 -
(ii)-96.10
(iii) 2.85 ± 8.9 97.50
Full network 1.93 ± 8.8 97.45
Table 8. Classification scores of the proposed network on small patches extracted from a SAR scene.
Table 8. Classification scores of the proposed network on small patches extracted from a SAR scene.
Confusion Matrix
PredictionTankerCargoFishingPassengerPrecision (%)
Ground Truth
Tanker1330272.22
Cargo5300085.71
Fishing00140100.00
Passenger0001100.00
Recall72.2290.91100.0033.33
Accuracy metrics
LabelTankerCargoFishingPassengerOverall(%)
IoU56.5278.95100.0033.3367.20
F-Score72.2288.24100.0050.0077.61
Accuracy85.2988.24100.0097.0685.29
κ 0.620.761.000.490.85
Table 9. Comparison of the results of the network on our database and on the OpenSARShip database.
Table 9. Comparison of the results of the network on our database and on the OpenSARShip database.
TestTrain
OursOpenSARShip OursOpenSARShip
Ours97.4522.18Ours1.93 ± 8.856.78 ± 314.78
Ours34.0587.71Ours−102.51 ± 123.94−0.23 ± 11.04
Classification overall
accuracy (%)
Length mean
error (m)

Share and Cite

MDPI and ACS Style

Dechesne, C.; Lefèvre, S.; Vadaine, R.; Hajduch, G.; Fablet, R. Ship Identification and Characterization in Sentinel-1 SAR Images with Multi-Task Deep Learning. Remote Sens. 2019, 11, 2997. https://doi.org/10.3390/rs11242997

AMA Style

Dechesne C, Lefèvre S, Vadaine R, Hajduch G, Fablet R. Ship Identification and Characterization in Sentinel-1 SAR Images with Multi-Task Deep Learning. Remote Sensing. 2019; 11(24):2997. https://doi.org/10.3390/rs11242997

Chicago/Turabian Style

Dechesne, Clément, Sébastien Lefèvre, Rodolphe Vadaine, Guillaume Hajduch, and Ronan Fablet. 2019. "Ship Identification and Characterization in Sentinel-1 SAR Images with Multi-Task Deep Learning" Remote Sensing 11, no. 24: 2997. https://doi.org/10.3390/rs11242997

APA Style

Dechesne, C., Lefèvre, S., Vadaine, R., Hajduch, G., & Fablet, R. (2019). Ship Identification and Characterization in Sentinel-1 SAR Images with Multi-Task Deep Learning. Remote Sensing, 11(24), 2997. https://doi.org/10.3390/rs11242997

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop