Next Article in Journal
Reducing Emissions in the Maritime Sector: Offshore Wind Energy as a Key Factor
Next Article in Special Issue
Data-Driven Analysis of Ocean Fronts’ Impact on Acoustic Propagation: Process Understanding and Machine Learning Applications, Focusing on the Kuroshio Extension Front
Previous Article in Journal
Long-Term Impacts of Runoff and Coastal Reclamation on Tidal Bore Variations in the Qiantang River Estuary, China
Previous Article in Special Issue
Range-Domain Subspace Detector in the Presence of Direct Blast for Forward Scattering Detection in Shallow-Water Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Acoustic Imaging Learning-Based Approaches for Marine Litter Detection and Classification

1
INESCTEC—Institute for Systems and Computer Engineering, Technology and Science, Rua Dr. Roberto Frias, 4200-465 Porto, Portugal
2
Imperial College London, South Kensington Campus, London SW7 2AZ, UK
3
ISEP—School of Engineering, Polytechnic Institute of Porto, Rua Dr. António Bernardino de Almeida 431, 4249-015 Porto, Portugal
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(11), 1984; https://doi.org/10.3390/jmse12111984
Submission received: 29 July 2024 / Revised: 26 September 2024 / Accepted: 1 November 2024 / Published: 3 November 2024
(This article belongs to the Special Issue Applications of Underwater Acoustics in Ocean Engineering)

Abstract

:
This paper introduces an advanced acoustic imaging system leveraging multibeam water column data at various frequencies to detect and classify marine litter. This study encompasses (i) the acquisition of test tank data for diverse types of marine litter at multiple acoustic frequencies; (ii) the creation of a comprehensive acoustic image dataset with meticulous labelling and formatting; (iii) the implementation of sophisticated classification algorithms, namely support vector machine (SVM) and convolutional neural network (CNN), alongside cutting-edge detection algorithms based on transfer learning, including single-shot multibox detector (SSD) and You Only Look once (YOLO), specifically YOLOv8. The findings reveal discrimination between different classes of marine litter across the implemented algorithms for both detection and classification. Furthermore, cross-frequency studies were conducted to assess model generalisation, evaluating the performance of models trained on one acoustic frequency when tested with acoustic images based on different frequencies. This approach underscores the potential of multibeam data in the detection and classification of marine litter in the water column, paving the way for developing novel research methods in real-life environments.

1. Introduction

Higher standards and improved living conditions have increased the demand for more materials, leading to greater waste generation [1]. Billions of metric tons are discarded annually, and without proper waste disposal, many environments are being damaged [2]. Recycling policies are failing to adequately address the disposal of plastic products, primarily composed of single-use plastics. Many companies engage in greenwashing, providing misinformation about their environmental actions to enhance their reputations and gain competitive advantages among environmentally conscious consumers [3,4].
The United Nations Environment Programme (UNEP) defines marine litter as any persistent, manufactured, or processed solid material discarded, disposed of, or abandoned in the marine and coastal environment [5,6,7], with plastic being the most prominent. The accumulation of plastic in the water is rapidly increasing [6]. It is estimated that 60% of all plastics ever made have been discarded in landfills or the natural environment. Positively buoyant plastic objects, commonly called floating plastic debris, illustrated in Figure 1, are influenced by a wide range of physical transport processes [7,8,9,10].
Submerged plastic debris can originate from low-density plastics combined with heavier materials, resulting in negatively buoyant masses that sink below the water’s surface and biofouling.
Microplastics can originate from fragmented parts of larger objects [6]. Detecting submerged macroplastics in the water column is necessary to capture them before they generate microplastics, reducing potential global damage. Most marine litter detection methods are related to debris on the surface of a body of water, with data essentially acquired by satellites [11]. Some solutions use drones or cameras in aircraft to acquire data to survey the amount of marine debris that covers the surface of rivers and seas. More recently, novel techniques are being applied to data extracted by hyperspectral cameras to better detect and classify marine litter [12].
Robotics can dramatically improve the detection and prediction of risks related to water pollution, providing new tools for the global management of water resources [13]. The multibeam echosounder (MBES) acquires acoustic data, and depending on the application, its raw data can be applied differently. MBES acoustic data, although composed of backscatter data, can be expressed as bathymetric data and as water column data (WCD) in the form of acoustic images.
While acoustic images can detect objects in the water column, it is crucial to test the ability to classify the footprint of a target in such images. Most methods use classical classifiers, such as support vector machine (SVM), Bayesian classifiers, and random forest classifiers. Other approaches explore deep neural networks for forward-looking sonar image classification [14]. There is limited research on classification and detection tasks performed with WCD, largely due to a low number of available datasets.
This paper contributes to the aforementioned topics and uses the imaging system setup and the foundations of the qualitative characterisation study that was made in [15] by leveraging multibeam water column data at various frequencies to detect and classify marine litter. The study encompasses:
  • Using the previously acquired test tank data with different types of marine litter at multiple acoustic frequencies.
  • Creation of an extensive acoustic image dataset with detailed labelling and formatting.
  • Implementation of two novel classification algorithms: support vector machine (SVM) and convolutional neural network (CNN).
  • Utilisation of two detection algorithms based on transfer learning: single-shot multibox detector (SSD), and You Only Look Once (YOLO), specifically YOLOv8 tuned for marine litter detection.
The results demonstrate the system’s effectiveness in distinguishing between different classes of marine litter using varied algorithms for both detection and classification. Cross-frequency studies were conducted to evaluate model generalisation, assessing performance when models trained on one acoustic frequency were tested with images based on different frequencies. This innovative characterisation of multibeam data underscores its potential in enhancing the detection and classification of marine litter within the water column.
This paper is organised as follows: Section 2 has an overview of the related work in marine litter detection, focusing on water column litter detection. Section 3 describes the acoustic acquisition system setup and the solution’s high-level architecture. Section 4 details the acoustic images dataset and the different acoustic image representations. Section 5 details the training procedures that were made for all the different models and a discussion of the results. Finally, Section 6 draws some conclusions on the work carried out and describes future work.

2. Related Work

This section reviews MBES acoustic imaging methods for marine litter detection and underwater image classification. Section 2.1 covers techniques like aerial surveys, hyperspectral imaging, and acoustic imaging. Section 2.2 explores classical and deep learning methods for classifying and detecting underwater acoustic data.

2.1. Marine Litter Detection

Marine litter detection mostly relies on remote sensing methods, often using unmanned aerial vehicles (UAV) [14,16]. Other methods include using satellite data, aircraft cameras, or remote hyperspectral imaging to assess debris on water surfaces [12]. Artificial intelligence (AI) is employed to detect and classify marine litter using surface imaging datasets (RGB and hyperspectral cameras), bathymetry studies with acoustic sensors, and in situ observations at the seabed. However, the latter often occurs without sonar imaging [14].
A literature review by Politikos et al. [14] summarises studies on marine litter research, focusing on detection-based algorithms for the automatic visual recognition and identification of macroplastics. Most marine debris data are acquired from the surface. Of the eighty scientific articles surveyed by Politikos et al. [14], only eight involved underwater data collection. This underwater data collection is primarily conducted using remotely operated vehicles (ROVs) and autonomous underwater vehicles (AUV) equipped with echosounders and cameras [14,17]. Only three of these publications addressed floating debris in the water column, with the remainder focusing on seafloor studies, as do most studies that rely on echosounder data. The author specified different types of data collected, including optical and acoustic (sonar) images.
Table 1 is based on a review on marine litter surveying and contains the eight publications on underwater marine debris studies [14]. These studies evaluated how the adopted algorithms recognised targets from the collected data.
From the perspective of underwater submerged litter detection, only two studies in the survey focused on this topic using sonar imaging [19,25]; however, these studies focused on a small-scale dataset collected from the bottom of a water tank. There is a lack of field measurements of underwater plastic debris, especially when concerned with submerged plastic [26]. Some studies used echosounders to detect macroplastics in the water column. Tests by Broere et al. [17] using a low-cost commercial fish finder identified plastic objects within the extracted acoustic images in controlled and uncontrolled environments. This study demonstrated that plastics can be detected in the water column with acoustic sensors and that different macroplastics have specific signatures related to their backscatter intensity, allowing for the estimation of litter size or even classification. It was confirmed that backscatter is affected by the orientation and deformation of an object. Water flow velocity significantly impacts detection, as increased velocity decreases the exposure time to the acoustic beams, reducing detection rates. High flow velocity can deter small macroplastics, which may not produce significant reflections. Additionally, the variability of reflections from the same object can impact object classification.

2.2. Underwater Acoustic Image Classification and Detection

Since most underwater marine litter detection applications do not rely on acoustic data, most of the work in this subsection, although varying in application, comprises solutions based on acoustic images and classification methods.
Object detection similar to that performed with multibeam echosounders is often facilitated by radars, which provide comparable information and share similar working principles. Deep neural networks were employed to recognise objects in 300 GHz radar images using returned power data, akin to multibeam backscatter intensity [27]. This study examined how power data varies with range, orientation, and different receivers in a laboratory environment. Due to the limited data available from this type of sensor, transfer learning was utilised. The study explored deep learning methods for detection and classification in scenes with multiple objects.
The detection and classification of marine litter using multibeam echosounders face similar challenges to underwater object recognition tasks performed with optical and sonar images. Deep convolutional neural networks have proven effective in these tasks, but they often require extensive datasets to generalise well to unseen examples. Acquiring and labelling such large volumes of data is costly and time-consuming, particularly for rare objects or in real-time operations. Few-shot learning (FSL) methods can be promising in addressing low data availability. Recent research has compared several supervised and semi-supervised FSL methods using underwater optical and side-scan sonar imagery, demonstrating that FSL significantly outperforms traditional transfer learning methods. These insights can be applied to improve the detection and classification of marine litter with multibeam echosounders, leveraging FSL to enhance model performance despite limited data [28].
A real-time underwater object detection algorithm using forward-looking imaging sonar [29] utilised Haar-like features as a weak classifier combined with AdaBoost to make a strong classifier, improving target detection performance in echosounder images. These features can reflect changes in the grey level, allowing the identification of edges, bars, and other simple image structures [29,30]. An algorithm for automatic detection and segmentation of gas plumes from water column images (WCIs) was proposed by Zhao et al. [31]. This author also combined Haar features with local binary patterns (LBP), which are highly discriminating and invariant to monotonic grey-level images, as are WCIs. Histogram of oriented gradients (HOG) was considered for feature extraction in underwater coloured acoustic images [32]. The HOG descriptor focuses on the structure or shape of an object, generating histograms using the magnitude and orientations of the gradient for each image region. Due to the large dimension of the HOG feature vector, a large amount of redundant information was reduced. An SVM classifier trained the extracted features and chose the radial basis function as the kernel function of the SVM, which defined the optimal classification surface.
A back propagation neural network (BPNN) was developed as a classifier for seabed sediments with MBES backscatter data in [33]. Of the thirty-four dimensions extracted that made up the initial feature vector, eight features with high classification were selected, reducing the workload of the classifier and improving classification efficiency and accuracy. Particle swarm optimisation was applied to increase the global optimisation ability of the model as well as to achieve an optimal initial weight for the parameters. Another optimisation was AdaBoost, which uses weak classifiers in the training set and assigns weights based on the weak classifier’s error. The author proposed the PSO-BP-AdaBoost classification algorithm and compared it with a one-level decision tree, a PSO-Backpropagation algorithm, and an SVM, achieving better accuracy than the three.
According to Valdenegro-Toro [34], classic computer vision methods like AdaBoost only work well in objects that produce large echosounder shadows, as the Haar features previously mentioned. The author proposed to detect new objects without class information, which applies to detecting hard-to-model objects such as marine debris. That was achieved by using a CNN to estimate and rank the objectness maps from sonar images, being a data-driven approach. Using forward-looking sonar (FLS) images, different types of marine litter, such as plastic bottles, cans, glass bottles, tyres, and more, were successfully detected at the bottom of a test tank by multiple authors [34,35]. Later, an adapted faster region-based convolutional neural network (R-CNN) algorithm was developed [19]. Transfer learning was applied for feature extraction to address data scarcity, incorporating the Residual Network 50 (ResNet-50) into the algorithm.
Transfer learning was used by many authors, with trained networks on large datasets such as the COCO and ImageNet datasets. A multi-branch shuttle network embedded in You Only Look Once 5 (YOLO5) was proposed to detect fish with an FLS [36], and a real-time object detection was proposed with a YOLO5 model for an obstacle avoidance algorithm in the underwater environment [37]. Transfer learning has been used to create an automatic, multi-class object classifier using data from a side-scan sonar to detect sunken shipwrecks and drowning victims, among others [38,39,40]. A CNN model was developed based on transfer learning, where a visual geometry group (VGG) model was pre-trained with ImageNet data and all of its trained layers. This model’s last fully connected layers were transferred to a new CNN, which was fine-tuned with a semi-synthetic sonar dataset [40].
The related work on underwater acoustic image classification and detection is summarised in Table 2. The table details the related work, considering the following items: the author, the year of publication, whether the acquired data are in the water column, developed algorithms, the application and the type of sensor used.
MBES data, previously acquired in [15], were necessary due to the scarcity of sonar images for object detection and classification in the scientific community compared to optical datasets. Existing sonar datasets typically focus on specific target objects and exhibit varied representations. Additionally, most sonar images do not capture the water column or are not for marine litter detection. The acquired test tank dataset, previously characterised qualitatively, was augmented with synthetic data, including multiple targets in a single acoustic image for numerous detections. This work extends earlier efforts by developing, training, and testing a multi-label SVM classifier and a shallow CNN multi-label classification model. Transfer learning was also applied to a single-shot detector (SSD) and the YOLO8 model to detect and label marine litter in the test tank data, validating the effectiveness of marine litter detection using MBES.

3. Multibeam Echosounder Test Tank Acquisition Setup

This section describes the setup of the acoustic imaging system, including the characteristics of the multibeam echosounder (MBES) used, and the test tank setup to perceive marine debris targets.

3.1. System Setup

Water column data were collected using the Kongsberg M3 high-frequency model, which provides high-resolution underwater acoustic imaging composed of beams and reflections. A sonar beam is defined by its shape, with horizontal and vertical angles that vary based on the type of sonar used. This array is divided into bins that associate intensity values with their respective distances and the reflections in the environment [43]. Each acoustic image from this MBES consists of 256 beams, each with 1573 reflections. A Windows processing unit manages the data acquisition process. The Kongsberg application programming interface (API) interfaces with the MBES, handling communication with the sonar head and processing raw data, specifically for MBES beamforming. This communication occurs over a standard ethernet cable using TCP/IP commands.
A custom software module was developed within the Robot Operating System 1 (ROS1) framework. This module acquires sensor message data packages. A dedicated subscriber was implemented to capture these messages and extract essential MBES data, including range, bearing, and normalised intensity.
The MBES operates at four distinct frequencies. The field of view (FOV) changes with each frequency, affecting the angular separation between the 256 beams. Higher frequencies enhance spatial resolution but may reduce detections because of a narrower FOV. The MBES uses maximum pulse power for imaging applications, with pulse duration adjustments to increase long-range detection. This results in a lower ping rate to extend the detectable range and manage power consumption. The sonar head was configured for image-enhanced mode, using a time delay fast Fourier transform beamformer for high-quality image generation. Table 3 depicts the main features of the Kongsberg M3 multibeam high-frequency echosounder model characteristics in image enhanced mode.
Data acquisition was carried out in a test tank, as illustrated in Figure 2a. The data acquired in [15] lacked variety and quantity for each class of selected marine debris. Tests were conducted in a controlled tank environment, where limitations allowed data acquisition at only two distances and three directions of arrival relative to the sonar head. These constraints were caused by the tank’s walls, bottom, and possible surface backscatter. Range limitations were addressed by setting a maximum range based on the target’s position. The MBES auto-range feature was disabled, ensuring all 1573 reflections per beam were confined within the defined range to enhance image resolution.
During the data acquisition in [15], targets were placed approximately 2.8 m away, directly in front of the sonar head. In the latest tests, data were collected at approximately 1.66 m and 3.3 m from the sonar head at a depth of around 2.5 m. For each range, targets were positioned at angles of 0° (directly in front), −18° (left), and 18° (right). Targets were also attached to ropes in varying ways to alter their appearance in the water column.
The MBES parameters were adjusted during the tests, while time-variant and image gains remained constant across different frequencies and acquisition ranges.
The initial MBES mounting was designed for test tank conditions but later upgraded to a versatile mounting pole suitable for tank and surface vehicle applications, as illustrated in Figure 2b. This mount pole allows depth adjustment and pitch configuration for various mounting angles. Different angles were tested by adjusting the MBES pitch to evaluate their effects on the results, considering the MBES azimuth. The mounting point was designed for future compatibility with an AUV.
Targets were placed at depths that prevented the detection of the tank’s bottom and walls, minimising high backscatter intensities from hard surfaces. This arrangement ensured that the highest backscatter intensities, normalised by the MBES, originated from the targets, simulating real-world conditions. The depth placements varied for each target, depending on its size, to maintain this setup. Data acquisition involved varying the frequency of the acoustic signals to generate acoustic images.

3.2. Marine Debris Selection

This study selected a collection of marine debris objects typically composed of plastic materials, such as a square made of polyvinyl chloride (PVC), for experimental testing. Various objects were chosen, some with the same squared shape, so it would be possible to study if algorithms could distinguish objects without relying solely on their shape.
The objects were selected for macroplastics detection and the NetTag+ project, with the primary objective of identifying concentrations of fish nets. Consequently, target sizes ranged from a PVC square measuring 0.5 m × 0.5 m to a smaller PVC square of 0.15 m × 0.15 m. Specific details of these objects are outlined in Table 4. Although these objects can occur naturally in the water column, their buoyancy was artificially modified for the experiments since the materials were new and had not been affected by wear or biofouling. Positively buoyant targets were weighted while negatively buoyant targets were tethered using buoys or a crane. The crane facilitated placing the targets at different fields of view (FOVs) and ranges and allowed for easy adjustment of the target’s depth. These objects compose the current test tank dataset, with each object corresponding to a class.
The PVC square class contains more occurrences due to the additional data collected with the smaller square of the same material. This enabled testing whether the material could still be detected at varying ranges and directions, as well as assessing the potential for false positives.
Each type of debris in Table 4 has a number that can be mapped in Figure 3.

3.3. High Level Architecture

The acoustic imaging for detection and classification problems of high-level architecture is summarised in Figure 4. After raw data acquisition, data are processed into three different acoustic imaging representations, which will be detailed in Section 4. Labelling and annotations were performed and tailored for the specific tasks and model architectures, enabling the training and validation of different types of networks.

4. Acoustic Images Dataset

This section outlines water column image processing. Section 4.1 describes three representations of acoustic images. Section 4.2 details the labelling and formatting of datasets for different models, ensuring readiness for training and validation.

4.1. Acoustic Images Representation

The experimental setup described in Section 3 enabled the extraction of raw MBES data, represented in the PointCloud2 format. This format includes the following fields:
  • x , y : Acoustic image pixel coordinates.
  • h e i g h t : Acoustic image height representing 1573 backscatter points.
  • w i d t h : Acoustic image width representing 256 beams.
  • i n t e n s i t y : Normalised backscatter intensity.
Three types of acoustic image representations were generated: raw, polar, and Cartesian. The raw and polar representations were tested as inputs for multiple models to evaluate, which led to better results. The raw representation was designed for real-time visualisation of MBES acoustic images due to its low storage requirements and fast generation. The polar representation was created to provide a format that is more interpretable for human operators. The following subsections detail each type of representation.

4.1.1. Raw Acoustic Image

The raw acoustic image representation, depicted in Figure 5, is a direct mapping of the normalised intensity of each pixel to its corresponding x and y coordinates, being composed of a single channel (greyscale). This representation has a fixed resolution of (256 × 1573) and ensures consistent pixel spacing. However, the varying field of view (FOV) associated with different acoustic frequencies and the changing range of targets relative to the MBES induce distortions in target shapes, as shown in Figure 5. This illustration kept the range constant, yet the resulting acoustic images differ significantly due to the varying field of view (FOV). Although the range, angle, and intensity data are preserved, this representation distorts target shapes, making it more challenging to distinguish targets, especially with varying ranges. This deformation makes it harder to use the objects’ shape as a feature for classification and detection models. Nonetheless, this method enables fast image generation due to the simplicity of its creation, which could be used for real-time visualisation.

4.1.2. Cartesian Acoustic Image

The Cartesian acoustic image shown in Figure 6 facilitates the identification of target locations, and sonar operators typically use it. In this representation, the range is associated with the y axis, with lower y values indicating closer targets. The x axis represents the bearing relative to the sonar head, with the centre corresponding to a target directly in front, having a direction of arrival (DoA) of 0°. The entire field of view is divided in half, each representing left and right bearings, providing spatial orientation information. This representation is based on variable pixel sizes influenced by the minimum distance between acoustic backscatter points, which affect the FOV and range parameters. Consequently, the image size varies due to these pixel size differences. To mitigate gaps in the data, a kernel is applied to the image to interpolate empty pixels. While this enhances visualisation clarity, many interpolated pixels lack precise intensity data. This Cartesian representation was developed as an alternative to the rviz ROS package, enhancing the visualisation of captured targets, particularly with adjustable kernel sizes. Additionally, post-processing was made, where a jet colour map was used on the normalised intensity data. This representation was not utilised for classification or detection models, and the image generation by this method is slower than the other two methods due to the kernel application.

4.1.3. Polar Acoustic Image

The Polar acoustic image illustrated in Figure 7 addresses the issues present in the raw representation by offering a consistent and interpretable format.
The x and y coordinates of points, along with their intensity values, are used as input to generate a two-dimensional (2D) image representing these points in polar space. The maximum accepted field of view (FOV) and range are considered. A fixed FOV is applied for all frequencies to ensure all acoustic images have the same fixed width without pixels lacking intensity values. The algorithm computes the range (r) and angle ( θ ) for each point (i). The range is computed using the Euclidean distance formula:
r a n g e i = x i 2 + y i 2
The angle ( θ ) is computed using the arctangent function:
t h e t a i = a r c t a n 2 ( y i , x i )
The range resolution is defined as 0.002 m/pixel based on specifications, and the theta resolution is set to 0.1°, given by:
t h e t a r e s o l u t i o n = m i n i m u m f o v b e a m s n u m b e r
The dimensions of the output image are determined based on these resolutions and the maximum values, ensuring consistent image resolution without compromising scale:
i m a g e h e i g h t = r a n g e m a x r a n g e r e s o l u t i o n + 1
i m a g e w i d t h = f o v m a x t h e t a r e s o l u t i o n + 1
Two mapping operations convert polar coordinates to polar image indices. Equation (6) converts the angular position ( θ ) into a column index (x) within the polar image:
x = i m a g e w i d t h 1 f o v m a x θ + i m a g e w i d t h 1 2
The current angle ( θ ) is scaled to fit within the image width, defined as the first member of Equation (6), and then centred to ensure symmetrical angular positions across the image width.
Equation (7) converts the radial distance (r) into a row index (y) within the polar image:
y = i m a g e h e i g h t 1 r a n g e m a x r
The range (r) is scaled to fit within the image height. The mapped indices assign each point’s normalised intensity to its corresponding polar image position. This process iterates over all points represented in polar coordinates, effectively populating the polar image with intensity values based on their spatial distribution.
This method allows for generating acoustic images that facilitate the computation of a target’s range and bearing relative to the sonar head. The varying FOV with different acoustic frequencies does not affect the resulting image dimensions or scale due to the fixed range and theta resolutions. Consistent image dimensions are beneficial for input into classification and detection models.

4.2. Dataset Labelling and Format

The dataset comprises five classes, as detailed in Table 4. Three acoustic frequencies were selected for the dataset: 950 kHz, 1200 kHz, and 1400 kHz. The 700 kHz frequency was excluded due to excessive noise in the MBES backscatter data, likely caused by the large field of view (FOV) and side-lobe interference from the test tank walls. The number of images per class at each frequency is presented in Table 5. The dataset is primarily balanced, except for the PVC square, which has higher occurrences.
For classification tasks, images were stored in directories named after their respective classes, facilitating organised dataset management and straightforward class identification during training. Simple data augmentation techniques, including horizontal flips and brightness adjustments, were applied to enhance the dataset. Since the data consist of sonar images, horizontal flips were used to simulate different fields of view (FOVs). Brightness adjustments were made to mimic variations in backscatter intensity. These transformations aimed to improve generalisation and have been discussed by other authors [44,45]. Although synthetic data augmentation methods, such as generative adversarial networks (GANs), were considered, they were not implemented in the current scope of development [46].
Labelling and annotation were necessary for detection tasks due to the presence of bounding boxes. This was accomplished using the Supervisely platform, where the bounding boxes were annotated for the entire dataset, including augmented data. The augmented dataset included images with multiple objects and some without targets. Data augmentation techniques were also applied to this dataset, including horizontal and vertical flips, zoom, and brightness adjustments. The Common Objects in Context (COCO) format was used for detection tasks, and a specific COCO format adaptation was made for the You Only Look Once 8 (YOLOv8) model.
This structured approach to dataset labelling and annotation, combined with data augmentation, ensures that both classification and detection models are trained on diverse and comprehensive datasets, enhancing their performance and robustness.

5. Detection and Classification

This section details the detection and classification algorithms implemented and tested to detect marine debris objects in the water column acoustic images. Section 5.1 describes the chosen algorithms based on the state-of-the-art approaches. Section 5.2 details how the algorithms were implemented and their results.

5.1. Algorithms for Detection and Classification Problems

The complexity and variability of marine litter, along with the characteristics of sonar images, present significant challenges for developing accurate and efficient classification and detection methods. The selection of algorithms was based on the nature of the acoustic images, their footprints, and the review of the state-of-the-art techniques and where they were commonly applied. Machine learning and deep learning methods, typically used for this type of data, can offer promising solutions. Selecting appropriate algorithms is crucial for achieving high accuracy and efficiency. Testing these algorithms with the acquired data was necessary to determine if it was possible to accurately discriminate the selected marine debris in the test tank dataset.
SVM and CNN were chosen for the multi-class classification of sonar acoustic images. The decision to use SVM is based on its robustness in high-dimensional feature spaces, which is advantageous for the complex features in sonar images. Although SVMs can be slower to train with large datasets, efficient kernels, such as the Radial Basis Function (RBF), mitigate this issue [47].
CNNs were also selected due to their ability to capture spatial hierarchies in images through convolutional layers. This capability allows CNNs to automatically learn and extract relevant features from images, reducing the need for manual feature engineering. CNNs have demonstrated superior performance in image classification tasks due to their proficiency in modelling complex patterns. Despite requiring substantial computational resources and time for training, CNNs generally achieve higher accuracy than traditional methods like SVM [48].
YOLO (You Only Look Once) version 8 and the single-shot multibox detector (SSD) were chosen for object detection in sonar acoustic images. YOLOv8 is designed for real-time detection, treating detection as a single regression problem that directly predicts bounding boxes and class probabilities. This design facilitates efficient processing, maintaining a good balance between speed and accuracy, essential for real-time applications. While YOLOv8 may not achieve the highest accuracy among detection models, its efficiency makes it a practical choice for applications requiring real-time processing [49].
The SSD balances speed and accuracy, making it suitable for applications where both factors are essential. SSD uses feature maps at different scales to detect objects of various sizes, enhancing detection performance. Like YOLO, SSD requires only a single forward pass through the network to detect objects, ensuring efficiency. Although SSD is more complex due to its use of multiple feature maps and default boxes, it allows for greater flexibility in detecting objects of different sizes [50].
Three acquisitions were conducted, each corresponding to one of the three operating frequencies: 950 kHz, 1200 kHz, and 1400 kHz. Consequently, three models were trained for each frequency. The training was performed on an NVIDIA GeForce RTX 3060 Mobile GPU. The development environment primarily utilised TensorFlow 2, with specific APIs employed for different models. The SVM model was implemented using the sklearn library, and the CNN model was built with keras. The detector models were trained using transfer learning from initial weights derived from the COCO dataset, including the SSD and the YOLOv8. The SSD model was trained using the TensorFlow 2 Detection API, the YOLO8 model was trained using the Supervisely platform, and the inference was performed using the ultralytics library.
The machine and deep learning algorithms used in this study are widely applied in state-of-the-art research across various domains. The focus of this work was not on developing new algorithms but rather on applying proven learning-based methods to this specific application. The successful characterisation and detection of the proposed targets in a controlled environment serve as essential groundwork for future real-world experiments.

5.2. Training and Results

5.2.1. Support Vector Machine (SVM)

An algorithm was developed to optimise the performance of the support vector machine (SVM) model through grid search. The process involves systematically exploring and evaluating hyperparameter combinations to identify the best-performing configurations. The aim was to ensure that the SVM classifier is robust and generalises unseen data well.
The algorithm considers three critical hyperparameters for optimisation: the regularisation parameter (C), the kernel coefficient (gamma), and the kernel types (that were chosen taking into account that it is a multi-class classification problem):
  • C—varies from 0.2 to 1.6 with a step of 0.2.
  • gamma—varies from 0.25 to 2.0 with a step of 0.25.
  • Kernel types—polynomial (poly), radial basis function (RBF), and sigmoid.
The combined dataset, made of the original and augmented data, was split into training, validation, and test sets. This step ensures that the model has sufficient data to learn from while also providing separate sets for validation and testing to evaluate model performance.
The grid search algorithm evaluated all possible combinations of the specified hyperparameters using 2-fold cross-validation (CV) for computational efficiency. The mean performance scores across the cross-validation folds are recorded for each combination. These scores are analysed to identify the best hyperparameters for each performance metric. The algorithm also evaluates the model performance based on accuracy, precision, and recall. Custom scoring functions are defined to assess precision and recall with weighted averaging, which accounts for class imbalances. The optimal hyperparameters for accuracy, precision, and recall are identified, defining an optimal classification surface. If different hyperparameters are optimal for each metric, multiple models are trained, each optimised for a specific metric. This ensures that the best possible model is available for different evaluation criteria. If the optimal parameters are the same across all metrics, a single model is trained with these parameters.
Inference was applied to the dataset corresponding to the frequency the model was trained on and the data of the other frequencies. This cross-frequency evaluation helps determine the models’ robustness and adaptability when applied to data acquired under different acoustic conditions. The results of each model from the other acoustic frequencies data are available in Table 6. The models trained on a specific frequency demonstrated overfitting when tested with acoustic images of the same frequency. This overfitting can be attributed to the low variability of the images obtained in a controlled environment. In such settings, the images did not capture objects from different angles, resulting in limited variations in the object footprints, underlining the current dataset limitations. This is notorious across all trained models. This will be addressed in the near future, where MBES data will be acquired from an unmanned surface vehicle (USV) in a real-life controlled environment where targets will be placed in a harbour within known locations.

5.2.2. Convolutional Neural Network (CNN)

A convolutional neural network (CNN) algorithm was developed, and class activation maps (CAM) were employed to identify the most informative regions in the images used by the model. The images and labels were split into training, validation, and test sets. The labels are one-hot encoded to facilitate multi-class classification.
The CNN architecture comprises several layers: an input layer and two convolutional layers with 32 and 64 filters, each with a kernel size of 3 × 3 and rectified linear unit (ReLU) activation functions. Max-pooling layers follow each convolutional layer to reduce the spatial dimensions of the feature maps. The output from these layers is then flattened and passed through a dense layer with 64 units and ReLU activation. The final output layer uses a softmax activation function to produce a probability distribution over the classes. The model is compiled using the Adam optimizer, categorical cross-entropy as the loss function, and metrics including accuracy, precision, and recall.
The model’s performance is evaluated on the test set, generating key metrics such as test loss, accuracy, precision, and recall. Class activation maps (CAM) were generated to visualize the regions in the input images that CNN focused on when making its predictions. This helps in understanding which parts of the pictures are most influential in the model’s decision-making process, providing insight into the model’s interpretability.
Each trained model was evaluated on datasets corresponding to different acoustic frequencies to assess generalisation. The models were trained using both raw and polar acoustic image representations. The results for each acoustic image representation are summarised in Table 7 and Table 8, respectively. This cross-frequency evaluation, similar to that applied in the support vector machine (SVM) models, aimed to determine the robustness and adaptability of the CNN models when applied to data acquired under different acoustic conditions.
The models trained on polar acoustic images were less prone to overfitting compared to those trained on raw images, a trend observed across all frequencies. While these models still showed signs of overfitting, the effect was less pronounced than with the SVM model. The generation of CAMs provided insight into the models’ focus areas, revealing whether they concentrated on pixels and features associated with backscatter data from marine debris targets. Upon generating the CAMs, it was confirmed that the marine debris objects and their backscatter significantly influenced the model’s decision-making process, as illustrated in Figure 8. This finding underscored the importance of the marine debris and their backscatter in the model’s predictions, affirming that the models primarily relied on these features despite the overfitting issue.

5.2.3. Single Shot Detector (SSD)

A single-shot multiBox detector (SSD) algorithm was implemented to detect objects in images across different acoustic frequencies. The workflow encompasses data preprocessing, model training, evaluation, and visualisation of the results using bounding boxes.
The data were loaded and pre-processed using annotations in COCO format. The data were divided into training and validation sets, ensuring the data distribution was maintained across sets. Annotations and images were processed to create TFRecord files to use the TensorFlow 2 Object Detection API. This step involved reading image files, encoding them, and normalising the bounding box coordinates. The annotations included bounding box coordinates and class labels.
The model architecture included a pre-trained backbone, RetinaNet50, followed by multiple convolutional layers to predict bounding boxes and class probabilities. The training involved optimising a multi-task loss function, a weighted sum of localisation loss and confidence loss. The Adam optimiser was used to minimise this loss function, facilitating efficient convergence.
The SSD model was loaded, and inference was performed on test images. The inference process involved passing images through the model and extracting bounding box predictions, class labels, and confidence scores. The COCO evaluation toolkit was integrated to compare the predicted bounding boxes with ground truth annotations, facilitating a detailed performance analysis. The results are summarised in Table 9.
It is noticeable from the results that the average precision (AP) for the intersection over union (IoU) for both thresholds is higher when the models that were trained with a specific acoustic frequency detect targets within images from the same frequency. The same goes for all the other metrics collected and showcased in Table 9. The results are visualised using bounding boxes drawn on the images, as illustrated in Figure 9, with scores above the 0.5 thresholds considered as valid detections. Cross-frequency detection seems unlikely for detection models.

5.2.4. You Only Look Once 8 (YOLOv8)

The YOLOv8 model, pre-trained using the Supervisely platform, is loaded from a checkpoint file containing weights for the previously labelled dataset. This model is configured to detect objects in images with a confidence threshold of 0.5 to filter predictions. Inference begins by loading and iterating through a set of photos from a predefined directory. The YOLOv8 model processes each image to detect objects, producing bounding boxes and c for each detected object. The detection results are saved as images with highlighted objects, and the bounding boxes are extracted and stored for further analysis. The YOLOv8 model’s performance is evaluated using metrics such as precision, recall, and F1 score metrics to assess accuracy and reliability.
During training, the YOLOv8 model was fine-tuned for 100 epochs with a patience of 50 epochs and a batch size of 16. The input image size was 640 pixels. The AdamW optimiser was used with an initial learning rate of 0.01, a final learning rate of 0.01, a momentum of 0.937, and a weight decay of 0.0005. The training utilised automatic mixed precision (AMP). Data augmentations included HSV-Hue augmentation (0.015), HSV-Saturation augmentation (0.7), and HSV-Value augmentation (0.4). Eight worker threads were used for data loading to ensure efficient training.
The results are summarised in Table 10.
Bounding boxes were drawn on the images, as illustrated in Figure 10, with scores above the 0.5 thresholds considered valid detections. As with the SSD models, it is noticeable that the mean average precision (mAP) and the fitness metric are higher when the models trained with a specific acoustic frequency detect targets within images from the same frequency, as with the other metrics. The results are visualised using bounding boxes drawn on the images, as illustrated in Figure 10, with scores above the 0.5 thresholds considered valid detections.

5.3. Results Discussion

The original dataset from [15] consisted of targets captured directly in front of the sonar head at multiple frequencies. However, the dataset lacked variation in fields of view and ranges relative to the sonar head. As a result, most models overfitted when tested on acoustic images using the same frequency as the training set, particularly in classification tasks, where results of 1.0 were observed. To mitigate overfitting and improve model generalisation, additional data were collected in the test tank using the same three acoustic frequencies, incorporating new data for all object classes, including the addition of a smaller PVC square. This process involved acquiring data at two additional distances and from three fields of view (FOV) relative to the sonar head, with targets arranged in different ways to create varied footprints in the sonar images. The ranges were selected to ensure an equal number of occurrences for each frequency, considering the minimum field of view (FOV) of the multibeam echosounder (MBES). Varying the range altered the appearance of targets across the water column, allowing assessment of whether the models could still detect and classify the targets accurately. Due to the test tank’s physical constraits, such as its walls, bottom, and surface reflections, more distances could not be tested. While the expanded dataset reduced overfitting in most models, SVM models continued to struggle with generalisation and required significantly longer training times compared to other models, even if there were changes in the cross-frequency test results.
CNN models demonstrated that the polar representation of the acoustic images outperformed the raw image representation, showing better generalisation. However, despite the mostly balanced test tank dataset, the lack of environmental and range variability still led to overfitting, even with simple data augmentation and early stopping techniques in place. These issues were primarily due to test tank limitations. Class activation maps revealed that the CNN models focused on pixels and features linked to the backscatter from marine debris targets. CNN models also trained faster than SVM models.
The SSD model produced good results with acoustic images at the same frequency as the training set, outperforming the YOLOv8 model in this respect. However, SSD models required more time and resources for training and were only effective for detecting objects at the same acoustic frequency they were trained on.
The YOLOv8 models, while not as effective as SSD models in same-frequency tests, performed better in cross-frequency tests. The training was faster and more optimised. YOLOv8 might yield better results with more exhaustive hyperparameter tuning.
Overall, the classification and detection of different types of marine debris were achieved, despite some selected marine debris sharing similar shapes. This suggests that features other than shape were key to distinguishing between the selected materials, allowing the material’s characterisation with the features available within the dataset. Data were gathered at different acoustic frequencies to assess whether there was greater discriminative power for each type of object and material at specific frequencies and to verify if the objects could be reliably detected at those frequencies. Additionally, a cross-frequency study was conducted, confirming the expected outcome that models would underperform when trained and tested across different frequencies. This was a preliminary investigation into the potential of multi-frequency approaches for target detection.
Across the models, the cross-frequency tests generally yielded better results in those trained with 1400 kHz data. Higher acoustic frequencies tended to produce better metrics, such as accuracy, likely due to the improved spatial resolution, which helps distinguish between targets based on different types of acoustic data. This improvement is primarily linked to the reduced distance between beams at higher frequencies. However, these tests are preliminary and inconclusive, as they do not provide sufficient information to conduct a comprehensive multi-frequency study, which would require the use of different hardware. However, a multi-frequency MBES is not yet available for data acquisition, which could otherwise facilitate techniques similar to those used in hyperspectral imaging systems [12]. Another approach worth exploring involves using multiple single-frequency models to detect and classify the same target, potentially increasing its detectability. Although this method does not achieve the same results as a multi-frequency or adaptive-frequency multibeam echosounder, it presents a feasible line of investigation when the required hardware is unavailable. It is not possible to conclude if the models can generalise with the current test tank dataset. To achieve generalisation and enable the application of new algorithms, a more extensive dataset obtained using the MBES in real-world scenarios across all frequencies and various ranges and poses relative to placed targets is required.

6. Conclusions and Future Work

Addressing the growing threat of marine litter in the oceans is crucial. This study obtained acoustic images using a multibeam echosounder setup at various frequencies to facilitate classification and detection. Data were collected in a test tank scenario with multiple targets positioned in the water column, enabling their identification through acoustic imagery.
Supervised machine learning methods were employed for automated target classification. Two algorithms, CNN and SVM, were implemented, and a model was trained to categorise different object classes at different acoustic frequencies. Deep learning approaches based on transfer learning were also trained with new weights, specifically for SSD and YOLO8 detection models. These methods demonstrated the capability to classify and detect marine debris targets at different acoustic frequencies from water column imaging data, distinguishing different debris even if their shape was similar. However, while mostly balanced, the data lacked sufficient variety for the models to generalise across multiple scenarios. The test results revealed that models trained on data with the same acoustic frequency exhibited overfitting, revealing a dataset limitation. Cross-frequency inference yielded poorer results, particularly in models based on frequencies with better spatial resolution. It highlights their difficulty in detecting or classifying targets in images with lower spatial resolution, such as those based on 950 kHz frequency, being more notorious in the detection algorithms. Tests with different algorithms indicated that the polar representation of acoustic data led to better results than the raw image representation, as demonstrated in the CNN model study.
The successful characterisation and detection of the proposed targets in a controlled environment serve as essential groundwork for future real-world experiments that are already planned.
Future work will focus on acquiring more data in real-world scenarios and placing various marine debris targets at known locations to develop more generalised models. Plans include creating a dataset with multiple acoustic frequencies and varying MBES orientations relative to the targets at different ranges. A data acquisition campaign is being organised within the scope of the NetTag+ project to gather data from the same marine debris classes used in this study, along with additional debris types, to expand the dataset and improve the models’ generalisation capabilities. This dataset will enable testing of current state-of-the-art networks and support the development of new techniques for detecting and classifying marine litter. The successful characterisation and detection proposed by this work laid the foundation for the planned dataset campaign.
Moreover, further experiments with multiple MBES systems will be necessary to advance marine litter detection and classification. However, this study has already demonstrated that acoustic data collected at different single frequencies using a MBES can differentiate between various objects in the water column in test tank conditions. Future tests will investigate whether detecting the same target at multiple frequencies, even if multi-band information is not available simultaneously, can enhance the discriminative power of marine debris detection algorithms.

Author Contributions

Conceptualisation, H.M.S.; methodology, H.M.S.; software, P.A.G.; validation, H.M.S. and S.W.; formal analysis, H.M.S. and S.W.; investigation, P.A.G.; data curation, P.A.G.; writing—original draft preparation, P.A.G.; writing—review and editing, H.M.S., S.W. and P.A.G.; supervision, H.M.S. and S.W.; project administration, H.M.S., A.M., J.A. and E.S.; funding acquisition, H.M.S., A.M., J.A. and E.S. All authors have read and agreed to the published version of the manuscript.

Funding

National Funds finance this work through the Portuguese funding agency, FCT—Fundação para a Ciência e a Tecnologia, within the project UIDB/50014/2020 with DOI (https://doi.org/10.54499/uidb/50014/2020), the Studentship for Doctoral Research Funding Programme, with the DOI (https://doi.org/10.54499/SFRH/BD/151460/2021) and by the European Union under the Horizon Europe Program, Grant No. 101112812 (NETTAG+).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fauziah, S.H.; Rizman-Idid, M.; Cheah, W.; Loh, K.H.; Sharma, S.; NoorMaiza, M.; Bordt, M.; Praphotjanaporn, T.; Samah, A.A.; bin Sabaruddin, J.S.; et al. Marine debris in Malaysia: A review on the pollution intensity and mitigating measures. Mar. Pollut. Bull. 2021, 167, 112258. [Google Scholar] [CrossRef]
  2. Alizadeh, L.; Liscio, M.C.; Sospiro, P. The phenomenon of greenwashing in the fashion industry: A conceptual framework. Sustain. Chem. Pharm. 2024, 37, 101416. [Google Scholar] [CrossRef]
  3. Li, M.; Trencher, G.; Asuka, J. The clean energy claims of BP, Chevron, ExxonMobil and Shell: A mismatch between discourse, actions and investments. PLoS ONE 2022, 17, e0263596. [Google Scholar] [CrossRef]
  4. Yu, E.P.y.; Van Luu, B.; Chen, C.H. Greenwashing in environmental, social and governance disclosures. Res. Int. Bus. Financ. 2020, 52, 101192. [Google Scholar] [CrossRef]
  5. Galgani, F.; Michela, A.; Gérigny, O.; Maes, T.; Tambutté, E.; Harris, P.T. Marine Litter, Plastic, and Microplastics on the Seafloor. Plastics and the Ocean: Origin, Characterization, Fate, and Impacts; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2022; pp. 151–197. [Google Scholar]
  6. Löhr, A.; Savelli, H.; Beunen, R.; Kalz, M.; Ragas, A.; Van Belleghem, F. Solutions for global marine litter pollution. Curr. Opin. Environ. Sustain. 2017, 28, 90–99. [Google Scholar] [CrossRef]
  7. Sivadas, S.K.; Mishra, P.; Kaviarasan, T.; Sambandam, M.; Dhineka, K.; Murthy, M.R.; Nayak, S.; Sivyer, D.; Hoehn, D. Litter and plastic monitoring in the Indian marine environment: A review of current research, policies, waste management, and a roadmap for multidisciplinary action. Mar. Pollut. Bull. 2022, 176, 113424. [Google Scholar] [CrossRef]
  8. Egger, M.; Quiros, L.; Leone, G.; Ferrari, F.; Boerger, C.M.; Tishler, M. Relative abundance of floating plastic debris and neuston in the eastern North Pacific Ocean. Front. Mar. Sci. 2021, 8, 626026. [Google Scholar] [CrossRef]
  9. Corcoran, P.L. Benthic plastic debris in marine and fresh water environments. Environ. Sci. Process. Impacts 2015, 17, 1363–1369. [Google Scholar] [CrossRef]
  10. Soto-Navarro, J.; Jordá, G.; Compa, M.; Alomar, C.; Fossi, M.; Deudero, S. Impact of the marine litter pollution on the Mediterranean biodiversity: A risk assessment study with focus on the marine protected areas. Mar. Pollut. Bull. 2021, 165, 112169. [Google Scholar] [CrossRef]
  11. Topouzelis, K.; Papageorgiou, D.; Karagaitanakis, A.; Papakonstantinou, A.; Arias Ballesteros, M. Remote sensing of sea surface artificial floating plastic targets with Sentinel-2 and unmanned aerial systems (plastic litter project 2019). Remote Sens. 2020, 12, 2013. [Google Scholar] [CrossRef]
  12. Freitas, S.; Silva, H.; Silva, E. Hyperspectral Imaging Zero-Shot Learning for Remote Marine Litter Detection and Classification. Remote Sens. 2022, 14, 5516. [Google Scholar] [CrossRef]
  13. Ribotti, A.; Magni, P.; Mireno, B.; Schroeder, K.; Barton, J.; McCaul, M.; Diamond, D. New cost-effective, interoperable sensors tested on existing ocean observing platforms in application of European directives: The COMMON SENSE European project. In Proceedings of the OCEANS 2015-Genova, Genova, Italy, 18–21 May 2015; pp. 1–9. [Google Scholar]
  14. Politikos, D.V.; Adamopoulou, A.; Petasis, G.; Galgani, F. Using artificial intelligence to support marine macrolitter research: A content analysis and an online database. Ocean. Coast. Manag. 2023, 233, 106466. [Google Scholar] [CrossRef]
  15. Guedes, P.; Silva, H.; Wang, S.; Martins, A.; Almeida, J.; Silva, E. Multibeam Multi-Frequency Characterization of Water Column Litter. In Proceedings of the OCEANS 2024-Singapore, Singapore, 15–18 April 2024; pp. 1–6. [Google Scholar]
  16. Garaba, S.P.; Park, Y.J. Riverine litter monitoring from multispectral fine pixel satellite images. Environ. Adv. 2024, 15, 100451. [Google Scholar] [CrossRef]
  17. Broere, S.; van Emmerik, T.; González-Fernández, D.; Luxemburg, W.; de Schipper, M.; Cózar, A.; van de Giesen, N. Towards underwater macroplastic monitoring using echo sounding. Front. Earth Sci. 2021, 9, 628704. [Google Scholar] [CrossRef]
  18. Politikos, D.V.; Fakiris, E.; Davvetas, A.; Klampanos, I.A.; Papatheodorou, G. Automatic detection of seafloor marine litter using towed camera images and deep learning. Mar. Pollut. Bull. 2021, 164, 111974. [Google Scholar] [CrossRef]
  19. Aleem, A.; Tehsin, S.; Kausar, S.; Jameel, A. Target Classification of Marine Debris Using Deep Learning. Intell. Autom. Soft Comput. 2022, 32, 73–85. [Google Scholar] [CrossRef]
  20. Bajaj, R.; Garg, S.; Kulkarni, N.; Raut, R. Sea debris detection using deep learning: Diving deep into the sea. In Proceedings of the 2021 IEEE 4th International Conference on Computing, Power and Communication Technologies (GUCON), Kuala Lumpur, Malaysia, 24–26 September 2021; pp. 1–6. [Google Scholar]
  21. Deng, H.; Ergu, D.; Liu, F.; Ma, B.; Cai, Y. An embeddable algorithm for automatic garbage detection based on complex marine environment. Sensors 2021, 21, 6391. [Google Scholar] [CrossRef]
  22. Fossum, T.O.; Sture, Ø.; Norgren-Aamot, P.; Hansen, I.M.; Kvisvik, B.C.; Knag, A.C. Underwater autonomous mapping and characterization of marine debris in urban water bodies. arXiv 2022, arXiv:2208.00802. [Google Scholar]
  23. Fulton, M.; Hong, J.; Islam, M.J.; Sattar, J. Robotic detection of marine litter using deep visual detection models. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 5752–5758. [Google Scholar]
  24. Hong, J.; Fulton, M.; Sattar, J. Trashcan: A semantically-segmented dataset towards visual detection of marine debris. arXiv 2020, arXiv:2007.08097. [Google Scholar]
  25. Valdenegro-Toro, M. Submerged marine debris detection with autonomous underwater vehicles. In Proceedings of the 2016 International Conference on Robotics and Automation for Humanitarian Applications (RAHA), Amritapuri, India, 18–20 December 2016; pp. 1–7. [Google Scholar]
  26. Van Emmerik, T.; Kieu-Le, T.C.; Loozen, M.; Van Oeveren, K.; Strady, E.; Bui, X.T.; Egger, M.; Gasperi, J.; Lebreton, L.; Nguyen, P.D.; et al. A methodology to characterize riverine macroplastic emission into the ocean. Front. Mar. Sci. 2018, 5, 372. [Google Scholar] [CrossRef]
  27. Sheeny, M.; Wallace, A.; Wang, S. 300 GHz radar object recognition based on deep neural networks and transfer learning. IET Radar, Sonar Navig. 2020, 14, 1483–1493. [Google Scholar] [CrossRef]
  28. Ochal, M.; Vazquez, J.; Petillot, Y.; Wang, S. A comparison of few-shot learning methods for underwater optical and sonar image classification. In Proceedings of the Global Oceans 2020: Singapore–US Gulf Coast, Biloxi, MS, USA, 5–30 October 2020; pp. 1–10. [Google Scholar]
  29. Kim, B.; Yu, S.C. Imaging sonar based real-time underwater object detection utilizing AdaBoost method. In Proceedings of the 2017 IEEE Underwater Technology (UT), Busan, Republic of Korea, 21–24 February 2017; pp. 1–5. [Google Scholar]
  30. Zhao, J.; Mai, D.; Zhang, H.; Wang, S. Automatic Detection and Segmentation on Gas Plumes from Multibeam Water Column Images. Remote Sens. 2020, 12, 3085. [Google Scholar] [CrossRef]
  31. Zhao, S.; Wang, C.; Bai, B.; Jin, H.; Wei, W. Study on the polystyrene plastic degradation in supercritical water/CO2 mixed environment and carbon fixation of polystyrene plastic in CO2 environment. J. Hazard. Mater. 2022, 421, 126763. [Google Scholar] [CrossRef]
  32. Wang, X.; Wang, J.; Yang, F.; Zeng, G. Target detection in colorful imaging sonar based on HOG. In Proceedings of the 2018 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Qingdao, China, 14–16 September 2018; pp. 1–5. [Google Scholar]
  33. Ji, X.; Yang, B.; Tang, Q. Acoustic seabed classification based on multibeam echosounder backscatter data using the PSO-BP-AdaBoost algorithm: A case study from jiaozhou bay, China. IEEE J. Ocean. Eng. 2020, 46, 509–519. [Google Scholar] [CrossRef]
  34. Valdenegro-Toro, M. Learning objectness from sonar images for class-independent object detection. In Proceedings of the 2019 European Conference on Mobile Robots (ECMR), Prague, Czech Republic, 4–6 September 2019; pp. 1–6. [Google Scholar]
  35. Singh, D.; Valdenegro-Toro, M. The marine debris dataset for forward-looking sonar semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3741–3749. [Google Scholar]
  36. Wang, J.; Feng, C.; Wang, L.; Li, G.; He, B. Detection of Weak and Small Targets in Forward-Looking Sonar Image Using Multi-Branch Shuttle Neural Network. IEEE Sens. J. 2022, 22, 6772–6783. [Google Scholar] [CrossRef]
  37. Yu, Y.; Zhao, J.; Gong, Q.; Huang, C.; Zheng, G.; Ma, J. Real-time underwater maritime object detection in side-scan sonar images based on transformer-YOLOv5. Remote Sens. 2021, 13, 3555. [Google Scholar] [CrossRef]
  38. Huo, G.; Wu, Z.; Li, J. Underwater object classification in sidescan sonar images using deep transfer learning and semisynthetic training data. IEEE Access 2020, 8, 47407–47418. [Google Scholar] [CrossRef]
  39. Fuchs, L.R.; Gällström, A.; Folkesson, J. Object recognition in forward looking sonar images using transfer learning. In Proceedings of the 2018 IEEE/OES Autonomous Underwater Vehicle Workshop (AUV), Porto, Portugal, 6–9 November 2018; pp. 1–6. [Google Scholar]
  40. Ge, Q.; Ruan, F.; Qiao, B.; Zhang, Q.; Zuo, X.; Dang, L. Side-scan sonar image classification based on style transfer and pre-trained convolutional neural networks. Electronics 2021, 10, 1823. [Google Scholar] [CrossRef]
  41. Gaida, T. Acoustic Mapping and Monitoring of the Seabed: From Single-Frequency to Multispectral Multibeam Backscatter. Ph.D. Thesis, TU Delft, Delft, The Netherlands, 2020. [Google Scholar]
  42. Janowski, L.; Trzcinska, K.; Tegowski, J.; Kruss, A.; Rucinska-Zjadacz, M.; Pocwiardowski, P. Nearshore benthic habitat mapping based on multi-frequency, multibeam echosounder data using a combined object-based approach: A case study from the Rowy Site in the Southern Baltic Sea. Remote Sens. 2018, 10, 1983. [Google Scholar] [CrossRef]
  43. Gonçalves, P.M.; Ferreira, B.M.; Alves, J.C.; Cruz, N.A. Image segmentation and mapping in an underwater environment using an imaging sonar. In Proceedings of the OCEANS 2022, Hampton Roads, VA, USA, 17–20 October 2022; pp. 1–8. [Google Scholar]
  44. Chandrashekar, G.; Raaza, A.; Rajendran, V.; Ravikumar, D. Side scan sonar image augmentation for sediment classification using deep learning based transfer learning approach. Mater. Today Proc. 2023, 80, 3263–3273. [Google Scholar] [CrossRef]
  45. Qin, X.; Luo, X.; Wu, Z.; Shang, J. Optimizing the Sediment Classification of Small Side-Scan Sonar Images Based on Deep Learning. IEEE Access 2021, 9, 29416–29428. [Google Scholar] [CrossRef]
  46. Zhu, J.; Li, H.; Qing, P.; Hou, J.; Peng, Y. Side-Scan Sonar Image Augmentation Method Based on CC-WGAN. Appl. Sci. 2024, 14, 8031. [Google Scholar] [CrossRef]
  47. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  48. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6999–7019. [Google Scholar] [CrossRef]
  49. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  50. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  51. Guedes, P. MBES M3 Kongsberg HF Model Test Tank Marine Debris Dataset. 2024. Available online: https://zenodo.org/records/13759505 (accessed on 13 September 2024).
Figure 1. Marine litter in the water column. Courtesy of Unsplash by Naja Jensen.
Figure 1. Marine litter in the water column. Courtesy of Unsplash by Naja Jensen.
Jmse 12 01984 g001
Figure 2. Kongsberg M3 Multibeam High-Frequency Echosounder system setup in the test tank. (a) Test tank setup, (b) MBES capturing the Wooden deck in the water column.
Figure 2. Kongsberg M3 Multibeam High-Frequency Echosounder system setup in the test tank. (a) Test tank setup, (b) MBES capturing the Wooden deck in the water column.
Jmse 12 01984 g002
Figure 3. Marine debris used for the test tank dataset. PVC Squares (1); PVC traffic cone (2); Wooden deck (3); vinyl sheet (4); fish net (5).
Figure 3. Marine debris used for the test tank dataset. PVC Squares (1); PVC traffic cone (2); Wooden deck (3); vinyl sheet (4); fish net (5).
Jmse 12 01984 g003
Figure 4. High-level architecture for the MBES sensor and acoustic imaging for detection and classification problems.
Figure 4. High-level architecture for the MBES sensor and acoustic imaging for detection and classification problems.
Jmse 12 01984 g004
Figure 5. Raw acoustic images of a PVC square at the same range, with varying FOV due to the different acoustic frequencies. (a) Raw acoustic image of 1200 kHz, (b) Raw acoustic image of 1400 kHz.
Figure 5. Raw acoustic images of a PVC square at the same range, with varying FOV due to the different acoustic frequencies. (a) Raw acoustic image of 1200 kHz, (b) Raw acoustic image of 1400 kHz.
Jmse 12 01984 g005
Figure 6. Cartesian acoustic image of a PVC square in the water column.
Figure 6. Cartesian acoustic image of a PVC square in the water column.
Jmse 12 01984 g006
Figure 7. Polar acoustic image of a PVC square in the water column.
Figure 7. Polar acoustic image of a PVC square in the water column.
Jmse 12 01984 g007
Figure 8. Class Activation Map applied to the CNN with a polar image of a PVC square as an input.
Figure 8. Class Activation Map applied to the CNN with a polar image of a PVC square as an input.
Jmse 12 01984 g008
Figure 9. SSD model inference in two polar acoustic images with multiple targets with the target detection confidence.
Figure 9. SSD model inference in two polar acoustic images with multiple targets with the target detection confidence.
Jmse 12 01984 g009
Figure 10. YOLO8 model inference in polar acoustic images with multiple targets with the target detection confidence.
Figure 10. YOLO8 model inference in polar acoustic images with multiple targets with the target detection confidence.
Jmse 12 01984 g010
Table 1. Submerged Marine Litter Surveying, based on [14,18].
Table 1. Submerged Marine Litter Surveying, based on [14,18].
ReferencesSampling SystemDataset TypeLitter DomainTask
Aleem et al. [19]SonarSonar imageFloating, SeafloorClassification, Detection
Bajaj et al. [20]AUV/ROVOptical imageSeafloorDetection
Deng et al. [21]AUV/ROVOptical imageSeafloorClassification, Detection
Fossum et al. [22]AUV/ROVOptical imageSeafloorClassification, Detection
Fulton et al. [23]AUV/ROVOptical imageSeafloorClassification, Detection
Hong et al. [24]AUV/ROVOptical imageFloating, SeafloorClassification
Politikos et al. [18]AUV/ROVOptical imageSeafloorClassification, Detection, Segmentation
Valdenegro-Toro [25]SonarSonar imageFloatingClassification, Detection
Table 2. MBES acoustic data classification and detection related work summary.
Table 2. MBES acoustic data classification and detection related work summary.
ReferenceYearWCDTask/TechniqueSensor TypeApplication
Gaida [41]2020YesStatistical ClassificationMBESAcoustic sediment classification
Janowski et al. [42]2018YesSVM ClassificationMBESBenthic habitat classification
Aleem et al. [19]2022NoFaster-RCNN classification model with pre-trained ResNet-50 ModelAdaptive Image Resolution Sonar (ARIS) in FLS configurationLitter classification in test tank bottom
Yu et al. [37]2021NoPre-trained YOLO5 detection modelSide-Scan SonarShipwreck and submerged container detector
Ge et al. [40]2021NoPre-trained CNN classification modelSide-Scan and synthetic dataDetection of acoustic targets within synthetic data
Fuchs et al. [39]2018NoPre-trained CNN modelForward-looking imaging sonarDetection of targets for obstacle avoidance
Wang et al. [36]2022NoPre-trained YOLO5 detection model and YOLO5 adaptionForward-looking imaging sonarDetection of weak and small litter
Valdenegro-Toro [34]2019NoSupervised CNN detection modelARIS in FLS configurationDetect litter without class information
Wang et al. [32]2018NoSVM detection with HOG featuresColourful imaging sonarWood stakes detection
Zhao et al. [30]2020NoAdaBoost cascade detectorMBESDetection of gas plumes
Kim and Yu [29]2017NoAdaBoost cascade detectorForward-looking imaging sonarObject detection
Ji et al. [33]2020NoPSO-BP-AdaBoost classifierMBESAcoustic sediment classification
Ochal et al. [28]2020NoFew shot learningOptical and Side-scan sonarUnderwater image classification
Table 3. Kongsberg M3 Multibeam High-Frequency Echosounder model characteristics in Image Enhanced mode.
Table 3. Kongsberg M3 Multibeam High-Frequency Echosounder model characteristics in Image Enhanced mode.
Operating Frequency
Specifications700 kHz950 kHz1200 kHz1400 kHz
Angular Resolution140° × 30°140° × 27°75° × 21°45° × 18°
Range Interval (m)0.2–150 (m)
Beams256
Reflections per beam1573
Pulse TypeContinuous Wave and Linear Frequency Modulation (Chirp)
Table 4. Marine Litter objects in the dataset.
Table 4. Marine Litter objects in the dataset.
ObjectDescription
PVC Square (1)Window PVC square of 0.5 m × 0.5 m
and small PVC square of 0.15 m × 0.15 m
PVC traffic cone (2)Traffic cone lying vertically
Wooden deck (3)Square wooden deck with slats
Vinyl sheet (4)Thin squared shaped vinyl sheet
Fish net (5)Agglomerate with fish nets and buoys
Table 5. Test tank dataset with its classes and number of occurrences at each operating frequency.
Table 5. Test tank dataset with its classes and number of occurrences at each operating frequency.
Operating Frequency
Class950 kHz1200 kHz1400 kHz
PVC Square551549453
PVC traffic cone317356305
Wooden deck356355344
Vinyl sheet356355425
Fish net301330313
Table 6. Performance Metrics for SVM Models with different multibeam acoustic frequencies.
Table 6. Performance Metrics for SVM Models with different multibeam acoustic frequencies.
ClassPrecisionRecallF1-ScoreAccuracy
950 kHz SVM model
950 kHz data
Fish net1.001.001.00
Wooden deck1.001.001.00
PVC traffic cone1.001.001.001.00
Vinyl1.001.001.00
PVC square1.001.001.00
1200 kHz data
Fish net0.000.000.00
Wooden deck0.191.000.32
PVC traffic cone0.000.000.000.19
Vinyl0.000.000.00
PVC square0.000.000.00
1400 kHz data
Fish net0.000.000.00
Wooden deck0.191.000.32
PVC traffic cone0.000.000.000.19
Vinyl0.000.000.00
PVC square0.000.000.00
1200 kHz SVM model
950 kHz data
Fish net0.720.670.69
Wooden deck0.460.740.56
PVC traffic cone0.310.350.330.50
Vinyl0.510.510.51
PVC square0.560.320.41
1200 kHz data
Fish net1.001.001.00
Wooden deck1.001.001.00
PVC traffic cone1.001.001.001.00
Vinyl1.001.001.00
PVC square1.001.001.00
1400 kHz data
Fish net1.000.160.28
Wooden deck0.231.000.38
PVC traffic cone0.540.110.190.26
Vinyl0.000.000.00
PVC square0.400.100.16
1400 kHz SVM model
950 kHz data
Fish net0.380.810.52
Wooden deck0.660.360.46
PVC traffic cone0.260.400.320.42
Vinyl0.440.450.44
PVC square1.000.180.31
1200 kHz data
Fish net0.970.180.30
Wooden deck0.220.650.32
PVC traffic cone0.350.250.290.35
Vinyl0.500.120.20
PVC square0.560.540.55
1400 kHz data
Fish net1.001.001.00
Wooden deck1.001.001.00
PVC traffic cone1.001.001.001.00
Vinyl1.001.001.00
PVC square1.001.001.00
Table 7. Performance Metrics for CNN Models with different multibeam acoustic frequencies with the raw acoustic image representation.
Table 7. Performance Metrics for CNN Models with different multibeam acoustic frequencies with the raw acoustic image representation.
ClassPrecisionRecallF1-ScoreAccuracy
950 kHz raw CNN model
950 kHz data
Fish net1.000.700.82
Wooden deck0.451.000.62
PVC traffic cone1.000.510.680.77
Vinyl sheet1.000.960.98
PVC Square1.000.690.82
1200 kHz data
Fish net0.520.810.63
Wooden deck0.700.770.74
PVC traffic cone0.480.330.390.56
Vinyl sheet0.410.540.47
PVC Square0.820.390.53
1400 kHz data
Fish net0.360.870.51
Wooden deck0.600.830.69
PVC traffic cone0.120.020.030.46
Vinyl sheet0.430.340.38
PVC Square0.620.290.39
1200 kHz raw CNN model
950 kHz data
Fish net1.000.470.64
Wooden deck1.000.530.69
PVC traffic cone0.300.280.290.46
Vinyl sheet0.360.620.45
PVC Square0.340.410.37
1200 kHz data
Fish net0.881.000.94
Wooden deck0.990.830.90
PVC traffic cone0.960.880.920.94
Vinyl sheet0.890.990.93
PVC Square1.001.001.00
1400 kHz data
Fish net0.650.940.77
Wooden deck0.600.420.49
PVC traffic cone0.040.020.020.51
Vinyl sheet0.330.710.45
PVC Square1.000.450.62
1400 kHz raw CNN model
950 kHz data
Fish net1.000.190.32
Wooden deck0.510.310.39
PVC traffic cone0.030.000.010.33
Vinyl sheet0.490.310.38
PVC Square0.250.740.38
1200 kHz data
Fish net1.000.210.34
Wooden deck0.720.450.55
PVC traffic cone0.610.520.560.55
Vinyl sheet0.420.920.57
PVC Square0.580.610.60
1400 kHz data
Fish net1.000.960.98
Wooden deck1.000.980.99
PVC traffic cone0.831.000.910.96
Vinyl sheet0.971.000.98
PVC Square1.000.880.94
Table 8. Performance Metrics for CNN Models with different multibeam acoustic frequencies with the polar acoustic image representation.
Table 8. Performance Metrics for CNN Models with different multibeam acoustic frequencies with the polar acoustic image representation.
ClassPrecisionRecallF1-ScoreAccuracy
950 kHz polar CNN model
950 kHz data
Fish net1.001.001.00
Wooden deck1.000.990.99
PVC traffic cone0.931.000.960.98
Vinyl sheet0.971.000.99
PVC Square1.000.950.97
1200 kHz data
Fish net0.850.830.84
Wooden deck0.920.760.83
PVC traffic cone1.000.420.590.65
Vinyl sheet0.361.000.53
PVC Square0.960.410.57
1400 kHz data
Fish net0.570.710.63
Wooden deck0.670.620.64
PVC traffic cone0.550.470.500.59
Vinyl sheet0.490.830.61
PVC Square1.000.350.52
1200 kHz polar CNN model
950 kHz data
Fish net0.850.750.85
Wooden deck0.950.660.80
PVC traffic cone0.680.840.750.76
Vinyl sheet0.860.550.71
PVC Square0.610.930.73
1200 kHz data
Fish net0.991.000.99
Wooden deck1.000.830.91
PVC traffic cone0.901.000.950.97
Vinyl sheet0.971.001.00
PVC Square0.981.000.99
1400 kHz data
Fish net0.670.990.80
Wooden deck0.700.370.48
PVC traffic cone0.480.980.640.67
Vinyl sheet1.000.470.64
PVC Square0.820.630.71
1400 kHz polar CNN model
950 kHz data
Fish net0.660.660.66
Wooden deck0.770.550.64
PVC traffic cone0.530.820.640.63
Vinyl sheet0.640.790.71
PVC Square0.620.440.51
1200 kHz data
Fish net0.900.880.89
Wooden deck0.660.180.28
PVC traffic cone0.800.750.780.75
Vinyl sheet0.511.000.68
PVC Square0.970.880.92
1400 kHz data
Fish net0.951.000.98
Wooden deck0.910.910.91
PVC traffic cone0.850.980.910.90
Vinyl sheet0.840.930.88
PVC Square0.910.740.85
Table 9. Performance Metrics for SSD Models with different multibeam acoustic frequencies with the polar acoustic image representation.
Table 9. Performance Metrics for SSD Models with different multibeam acoustic frequencies with the polar acoustic image representation.
ClassPrecisionRecallmAPAP for IoUAP for IoU
Threshold at 0.6Threshold at 0.75
950 kHz SSD model
950 kHz data
PVC traffic cone0.86570.87120.86800.96930.8680
PVC Square0.84400.84890.84940.53470.8494
Fish net0.71040.72220.71860.71230.7186
Wooden deck0.97040.97690.97800.98020.9780
Vinyl sheet0.98310.99000.98720.9580.9872
1200 kHz data
PVC traffic cone0.17250.33540.28430.32480.2843
PVC Square0.00120.00390.00120.00.0012
Fish net0.23530.35180.26310.00.2631
Wooden deck0.00.00.00.00.0
Vinyl sheet0.02450.08200.02950.65050.0295
1400 kHz data
PVC traffic cone0.00160.00470.01950.02970.0195
PVC Square0.00.00.00.00.0
Fish net0.00.00.00.00.0
Wooden deck0.00490.01260.01720.07130.0172
Vinyl sheet0.02090.03230.02950.00.0295
1200 kHz SSD model
950 kHz data
PVC traffic cone0.27990.35090.28890.50490.2889
PVC Square0.00420.00760.02700.00000.0270
Fish net0.31530.42660.35280.00.3528
Wooden deck0.00000.00000.00000.00000.0000
Vinyl sheet0.00000.00000.00000.00000.0000
1200 kHz data
Fish net0.95010.98320.9720-1.00.9720
PVC traffic cone0.97160.98010.97890.98220.9789
PVC Square0.72560.74230.73320.57430.7332
Wooden deck0.87300.88100.88121.00000.8812
Vinyl sheet0.96330.99000.97620.00000.9762
1400 kHz data
Fish net0.00.00.00.00.0
PVC traffic cone0.00.00.00.00.0
PVC Square0.00040.00080.01430.02180.0143
Wooden deck0.00230.00290.04560.00000.0456
Vinyl sheet0.19630.27660.26950.00000.2695
1400 kHz SSD model
950 kHz data
PVC traffic cone0.00090.00170.01380.00490.0138
PVC Square0.22270.31130.31110.00000.3111
Wooden deck0.00210.00640.03840.00000.0384
Fish net0.00290.00480.01590.01780.0159
Vinyl sheet0.00050.00140.02420.00000.0242
1200 kHz data
PVC traffic cone0.00030.00030.01600.01980.0160
PVC Square0.01290.03550.13010.33070.1301
Wooden deck0.00670.00700.12250.00000.1225
Fish net0.03910.04720.17140.17720.1714
Vinyl sheet0.50670.62780.54260.80000.5426
1400 kHz data
Fish net0.98470.99010.99300.98710.9930
PVC traffic cone0.97910.99000.98620.98510.9862
PVC Square0.94250.95010.94790.87520.9479
Wooden deck0.87000.87130.87390.49500.8739
Vinyl sheet0.88960.90920.90170.90000.9017
Table 10. Performance Metrics for YOLOv8 Models with different multibeam acoustic frequencies with the polar acoustic image representation.
Table 10. Performance Metrics for YOLOv8 Models with different multibeam acoustic frequencies with the polar acoustic image representation.
ClassPrecisionRecallF1mAPFitness
950 kHz YOLOv8 model
950 kHz data
Fish net1.001.001.000.9921
PVC traffic cone0.93030.99660.96230.9186
Wooden deck0.99251.000.99620.97570.9603
Vinyl sheet0.99151.000.99570.9911
PVC Square0.99250.92130.95560.9100
1200 kHz data
Fish net0.74900.63160.68530.2702
PVC traffic cone0.01570.01200.01360.0031
Wooden deck0.10160.10000.10080.02600.0695
Vinyl sheet0.000.000.000.00
PVC Square0.000.000.000.00
1400 kHz data
Fish net0.01340.96300.02640.0119
PVC traffic cone0.00250.500.00500.0012
Wooden deck0.000.000.000.000.0031
Vinyl sheet0.000.000.000.00
PVC Square0.00150.07140.00300.0002
1200 kHz YOLOv8 model
950 kHz data
Fish net0.93331.000.96550.6900
PVC traffic cone0.02680.95450.05200.0488
Wooden deck0.02380.95650.04640.01330.1599
Vinyl sheet0.00970.92590.01930.0010
PVC Square0.00.00.00.0
1200 kHz data
Fish net0.99061.000.99530.9950
PVC traffic cone1.000.99580.99790.9279
Wooden deck0.99431.000.99710.98310.9789
Vinyl sheet0.99801.000.99900.9905
PVC Square0.99451.000.99720.9887
1400 kHz data
Fish net1.000.000.000.00
PVC traffic cone1.000.000.000.00
Wooden deck0.000.000.000.07510.1024
Vinyl sheet0.74450.90480.81690.3754
PVC Square0.000.000.000.00
1400 kHz YOLOv8 model
950 kHz data
Fish net0.000.000.000.0047
PVC traffic cone1.000.000.000.00
Wooden deck0.000.000.000.000.1109
Vinyl sheet0.000.000.000.00
PVC Square0.94940.91300.93090.5055
1200 kHz data
Fish net1.000.000.000.00
PVC traffic cone1.000.000.000.04
Wooden deck1.000.000.000.000.0891
Vinyl sheet0.91080.91670.91370.3401
PVC Square1.000.000.000.00
1400 kHz data
Fish net0.77561.000.87360.8568
PVC traffic cone0.87070.79210.82950.5482
Wooden deck0.98571.000.99280.82500.7822
Vinyl sheet0.79380.95290.86610.7661
PVC Square1.000.84760.91750.8106
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guedes, P.A.; Silva, H.M.; Wang, S.; Martins, A.; Almeida, J.; Silva, E. Acoustic Imaging Learning-Based Approaches for Marine Litter Detection and Classification. J. Mar. Sci. Eng. 2024, 12, 1984. https://doi.org/10.3390/jmse12111984

AMA Style

Guedes PA, Silva HM, Wang S, Martins A, Almeida J, Silva E. Acoustic Imaging Learning-Based Approaches for Marine Litter Detection and Classification. Journal of Marine Science and Engineering. 2024; 12(11):1984. https://doi.org/10.3390/jmse12111984

Chicago/Turabian Style

Guedes, Pedro Alves, Hugo Miguel Silva, Sen Wang, Alfredo Martins, José Almeida, and Eduardo Silva. 2024. "Acoustic Imaging Learning-Based Approaches for Marine Litter Detection and Classification" Journal of Marine Science and Engineering 12, no. 11: 1984. https://doi.org/10.3390/jmse12111984

APA Style

Guedes, P. A., Silva, H. M., Wang, S., Martins, A., Almeida, J., & Silva, E. (2024). Acoustic Imaging Learning-Based Approaches for Marine Litter Detection and Classification. Journal of Marine Science and Engineering, 12(11), 1984. https://doi.org/10.3390/jmse12111984

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop