Next Article in Journal
Using a Neural Network to Model the Incidence Angle Dependency of Backscatter to Produce Seamless, Analysis-Ready Backscatter Composites over Land
Previous Article in Journal
AquaPile-YOLO: Pioneering Underwater Pile Foundation Detection with Forward-Looking Sonar Image Processing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DBCE-Net: A Novel Deep Learning Framework for Annual Mapping of Coastal Aquaculture Ponds in China with Sentinel-2 Data

1
State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China
2
College of Computer Science, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(3), 362; https://doi.org/10.3390/rs17030362
Submission received: 2 December 2024 / Revised: 18 January 2025 / Accepted: 19 January 2025 / Published: 22 January 2025

Abstract

:
Despite the promising advancements of deep learning techniques in coastal aquaculture pond extraction, their capacity for large-scale mapping tasks remains relatively limited. To address this challenge, this study developed a novel deep learning framework, Dual-Branch Enhanced Network (DBCE-Net), for mapping the annual aquaculture ponds at the national scale using Sentinel-2 imagery. The DBCE-Net framework effectively mitigates the contextual information loss inherent in traditional methods and reduces classification errors by processing both down-sampled large-scale images and block images at their original resolution. The architecture comprises local feature extraction and global feature extraction, along with feature fusion and decoding. The pivotal Multi-scale Dynamic Feature Fusion (DFF) module synthesizes local and global features while incorporating complementary information, demonstrating strong robustness with smaller training areas, compared to previous methods that required a larger number of samples distributed across different regions. By applying the DBCE-Net to Sentinel-2 imagery from 2017 to 2023, we mapped the spatiotemporal distribution of coastal aquaculture ponds across all coastal counties in China, achieving an overall classification accuracy approximately 93%. The results demonstrate substantial changes in the area of coastal aquaculture ponds in China from 2017 to 2023, with the total area declining from 8970.25 km2 to 8261.17 km2, representing a notable decrease of 7.90%. The most pronounced reduction was observed in Shanghai, with a decrease of 38.92%, followed by Zhejiang (31.57%) and Jiangsu (19.07%). These reductions are primarily attributed to policies converting aquaculture ponds into natural wetlands. In contrast, the area of coastal aquaculture ponds in Liaoning Province slightly increased by 5.75%. This DBCE-Net demonstrates good accuracy and generalizability and is promising to further expand its application to the extraction of coastal aquaculture areas worldwide, providing important scientific value and practical significance for the global coastal aquaculture industry.

1. Introduction

Aquaculture, an essential pillar of the global food supply chain, serves a crucial function in maintaining biodiversity, ecological equilibrium, and ensuring food security. With the continual growth of the global population and the progressive depletion of marine resources, the aquaculture sector is becoming ever more vital [1]. According to the Food and Agriculture Organization of the United Nations (FAO), global fishery and aquaculture production has reached an unprecedented high, with the industry projected to play an increasingly significant role in providing food and nutrition in the future. China is the largest aquaculture nation globally, supplying over 60% of the world’s aquaculture products [2,3,4]. However, the rapid expansion of coastal pond aquaculture has precipitated a series of environmental issues, including disrupting ecological balance, damaging crucial ecosystems such as coastal wetlands, coral reefs, and seagrass beds, and adversely affecting local biodiversity and habitats [5,6]. Interactions between aquaculture ponds and surrounding marine ecosystems, such as the discharge of aquaculture wastewater, can threaten the survival and reproduction of marine organisms, resulting in marine pollution and other issues [7,8,9,10]. As the scale of aquaculture continues to change, traditional methods of manual field monitoring and statistical surveys are increasingly insufficient to meet the demands for efficient and precise management, creating an urgent need for developing efficient and reliable aquaculture monitoring methods.
In recent years, an escalating number of researchers have leveraged machine learning and deep learning techniques for the automatic extraction and mapping of aquaculture ponds. For mapping coastal aquaculture ponds, random forest algorithms and decision tree models are prevalently employed. These methods are integrated with the analysis of various features of aquaculture pond water bodies, such as spectral indices (e.g., Normalized Difference Water Index [NDWI] [11] and Modified Normalized Difference Water Index [MNDWI] [12]), morphology, and texture [13,14,15]. Existing studies have extensively applied these methods using Sentinel-2 satellite imagery in China and other large-scale coastal regions [16,17]. However, this method necessitates manually setting distinct thresholds and parameters for various regions, requiring multiple parameters nationwide, and its effectiveness is contingent upon the support of other datasets [16,17,18,19,20,21]. Additionally, methods combining edge detection and morphological processing can identify and analyze pond edges and morphological features to extract aquaculture pond information. However, these methods are limited in distinguishing coastal aquaculture ponds from similar surface water bodies [22]. Due to their high similarity in size, appearance, and spectral characteristics, these water bodies often lead to classification errors. In addition to these challenges, traditional approaches are further hindered by their limited capacity to effectively integrate spatial information, which impairs their performance in areas characterized by complex environmental conditions. Moreover, the reliance on external datasets and the necessity of adjusting thresholds for different regions at large scales introduce additional complexities, thereby diminishing both the efficiency and accuracy of the mapping process.
Deep learning methods, with their powerful feature extraction capabilities, have demonstrated exceptional performance in various remote sensing applications, including the precise identification and classification of aquaculture ponds. However, their application for large-scale extraction of aquaculture ponds on a national level remains limited. This highlights an urgent need for a highly generalizable and automated approach to accurately extract large-scale coastal aquaculture ponds. In recent years, several studies have attempted to address these challenges. Gao et al. proposed a semantic segmentation model based on D-ResUnet for high-precision extraction of marine floating raft aquaculture (FRA) areas [23]. Deng et al. introduced a coastal aquaculture network to resolve issues of boundary accuracy and complex backgrounds, successfully extracting aquaculture areas from high-resolution remote sensing images [24]. Zou et al. developed a U2-Net-based model for coastal aquaculture pond extraction and validated it in the Zhoushan Islands [25]. Ai et al. proposed the SAMALNet model, which utilizes high-resolution satellite imagery along with a self-attention mechanism and auxiliary loss network to enhance the accuracy of boundary extraction in aquaculture areas [26]. Furthermore, Liang et al. innovatively introduced the DIAS index and proposed the MAFU-Net model, achieving high-precision extraction of aquaculture ponds [27]. Despite these advancements, traditional single-branch structures in remote sensing deep learning still face inherent limitations in capturing contextual and boundary information. To address this, existing studies have incorporated pyramid structures or dilated convolutions to enhance the model’s ability to process multi-scale information [28,29]. However, these approaches often struggle with the precise differentiation of features in regions with complex geographical characteristics, such as multi-branch rivers and aquaculture ponds, due to their limited ability to handle small-scale perspectives. Additionally, edge information loss further reduces the accuracy of classification and identification.
In this study, we developed DBCE-Net (Dual-Branch Enhanced Network), an innovative deep learning framework specifically tailored for the mapping of remote sensing imagery of coastal aquaculture ponds. This framework employs a dual-branch structure: the main branch concentrates on accurately capturing local features of high-resolution images, while the Context Edge Supplement Branch (CESB) processes down-sampled images, supplementing global contextual and boundary information, thereby enhancing overall scene understanding. Through the Multi-scale Dynamic Feature Fusion (DFF) module, we achieved efficient integration of local details with global information, markedly enhancing the model’s accuracy in identification and classification within complex backgrounds.
The primary contributions of this study are
(1)
Design a dual-branch DBCE-Net deep learning framework that combines high-resolution small image features with low-resolution large image features, effectively addressing the information loss problem inherent in traditional single-branch networks.
(2)
Propose an innovative Multi-scale Dynamic Feature Fusion (DFF) module to effectively integrate complementary information from the main branch and CESB, capturing and utilizing multi-scale features.
(3)
Selecting typical aquaculture pond regions in China and applying the small sample to large-scale coastal areas, achieving extensive extraction and mapping of aquaculture ponds across China.
Through these innovations, DBCE-Net effectively addresses key challenges in analyzing remote sensing imagery of coastal aquaculture environments, such as improving classification accuracy, enhancing boundary identification accuracy, and reducing spectral similarity interference.

2. Study Areas and Data

2.1. Study Areas

China’s coastline begins at the Yalu River estuary and extends to the Beilun River estuary, spanning a total length of 18,000 km and characterized by its winding paths and numerous islands [30]. These unique natural conditions provide an exceptionally favorable environment for the thriving development of the marine aquaculture industry. Coastal marine aquaculture primarily involves the cultivation of fish, shellfish, seaweed, shrimp, and crabs. These aquaculture activities play a pivotal role in promoting local economic development and are also essential for maintaining regional ecological balance [31].
In this study, we selected Chinese coastal counties and a 10 km buffer zone extending seaward from the coastline as the study area (Figure 1). This choice was motivated by the significant annual changes in coastline position and coastal land use caused by natural and human factors, which render static coastline data insufficient to capture the dynamic characteristics of aquaculture ponds. The 10 km buffer zone provides better coverage of these spatial variations while using counties as the basic statistical unit, offering more accurate and targeted decision making and resource management support to local governments.

2.2. Data and Preprocessing

The Sentinel-2 satellite series constitutes a pivotal element of the European Space Agency’s (ESA) remote sensing observation program, encompassing two key satellites: Sentinel-2A and Sentinel-2B. This dual-satellite configuration achieves a revisit period of at least once every 5 days and offers a wide field of view of up to 290 km. Equipped with Multispectral Instruments (MSI) featuring 13 spectral bands spanning visible to near-infrared wavelengths and spatial resolutions ranging from 10 to 60 m, Sentinel-2 provides crucial data for diverse applications [5,32,33]. With its 10 m spatial resolution, rich spectral information across 13 bands, and high-frequency revisit period, Sentinel-2 significantly reduces cloud interference, supporting large-scale and long-term aquaculture monitoring [6,34,35]. The L1C level data comprise raw, georeferenced multispectral images widely used for agricultural monitoring, urban planning, and environmental protection [36]. Sentinel-2’s wide coverage, high temporal resolution, and spectral diversity make it invaluable for land cover and dynamic change detection, particularly in remote sensing research.
It is important to note that many aquaculture ponds undergo cleaning and drying operations from December to February, which may result in imagery from this period not accurately reflecting typical conditions in these areas. To enhance data quality, this study focuses on remote sensing imagery from March to November each year, synthesizing annual median images for subsequent monitoring. Cloud masking is performed using the QA60 band, with a filtering criterion set to CLOUDY_PIXEL_PERCENTAGE less than 20% to effectively identify and exclude cloudy pixels.

3. Method

3.1. Creation of Sample Dataset

The primary objective of this study is to utilize semantic segmentation techniques for the binary classification of surface water bodies in coastal areas, distinctly categorizing them into aquaculture ponds and background. By accurately identifying aquaculture ponds within coastal water bodies, we aim to conduct a detailed analysis and research on these key areas, while classifying all other non-aquaculture surface water bodies as background to concentrate on the identification and study of aquaculture ponds.
The Multispectral Imaging Instrument (MSI) on Sentinel-2 satellite is capable of capturing image data in 13 spectral bands (Table 1), covering spectra from visible light to near-infrared and shortwave infrared light. Among them, B2 (blue), B3 (green), B4 (red), B8 (near-infrared), and B11 (shortwave infrared) exhibit significant spectral differences, making them effective for distinguishing aquaculture ponds, salt fields, and dikes. In this study, we selected bands B2, B3, B4, B8, and B11. The spectral differences of these bands are crucial for extracting key land features in coastal areas, particularly for distinguishing aquaculture ponds from other water bodies or background features.
The Sentinel-2 satellite series constitutes a pivotal element of the European Space Agency’s (ESA) remote sensing observation program, encompassing two key satellites: Sentinel-2A and Sentinel-2B. This dual-satellite configuration achieves a revisit period of at least once every 5 days and offers a wide field of view of up to 290 km. These satellites are outfitted with Multispectral Instruments (MSI), featuring 13 spectral bands spanning wavelengths from visible to near-infrared, with spatial resolutions ranging from 10 to 60 m [5,32,33].
Sentinel-2 satellite imagery, with its 10 m median spatial resolution and rich spectral information across 13 bands, coupled with a high-frequency revisit period of 5 days, significantly reduces cloud interference, thereby supporting large-scale and long-term aquaculture dynamic monitoring needs [6,34,35]. The L1C level data provided by the Sentinel-2 series consist of raw, georeferenced multispectral images that are widely recognized for their broad applicability [36]. These data play a pivotal role in various remote sensing analyses, encompassing agricultural monitoring, urban planning, and environmental protection. The primary features of satellite data include wide coverage, high temporal resolution, and a rich selection of spectral bands, rendering Sentinel-2 extremely valuable for land cover and dynamic change monitoring. Particularly, its comprehensive optical imaging capabilities make it an indispensable tool in the field of remote sensing research.
Accurately distinguishing between salt pans, aquaculture ponds, embankments, and paddy fields is a critical step in the extraction and classification of coastal aquaculture ponds. To address this challenge, this study selected three regions along China’s coastline, spanning from north to south: Bohai Bay, Yancheng, and Shantou, as the study areas for spectral band analysis. A total of 161 embankment samples, 251 aquaculture pond samples, 92 salt pan samples, and 159 paddy field samples were collected.
Through an in-depth analysis of the spectral reflectance characteristics of these sample points, it was found that Sentinel-2 satellite imagery bands B2 (blue), B3 (green), B4 (red), B8 (near-infrared), and B11 (shortwave infrared) exhibited significant spectral reflectance differences among the four land cover types (Figure 2). Notably, paddy fields showed high spectral similarity to aquaculture ponds in the B2, B3, and B4 bands, making them difficult to differentiate. However, significant differences were observed in the B8 and B11 bands, especially during the vegetative growth stage of paddy fields. By incorporating these two bands into the analysis, combined with the robust feature extraction capabilities of deep learning models, the classification model was able to reduce confusion between paddy fields and aquaculture ponds. Consequently, these five bands were selected for subsequent analysis to enhance classification accuracy and improve the reliability and applicability of the classification model.
It is important to note that many aquaculture ponds undergo cleaning and drying operations from December to February [37], which may result in imagery from this period not accurately reflecting typical conditions in these areas. To enhance data quality, this study focuses on remote sensing imagery from March to November each year, synthesizing annual median images for subsequent monitoring. Cloud masking is performed using the QA60 band, with a filtering criterion CLOUDY_PIXEL_PERCENTAGE set to less than 20% to effectively identify and exclude cloudy pixels.
The characteristics of aquaculture ponds in China exhibit significant regional diversity, particularly the stark contrast between the northern and southern regions. The north is characterized by large-scale aquaculture ponds and salt fields, while the south features smaller, more complex ponds with diverse attributes. To capture this diversity, we selected Sentinel-2 satellite imagery from 2021 and chose Bohai Bay (Figure 1a) in the north, along with Hangzhou Bay and Sanmen Bay in Zhejiang Province (Figure 1b) in the south, as regions for training sample selection. These areas not only represent typical features of aquaculture ponds in northern and southern China, but also include land features such as salt pan in the north and rice fields in the south that are easily confused with aquaculture ponds. The image information for each area is presented in Table 2. These regions encapsulate the representative features of aquaculture ponds across China, enabling the model to generalize effectively for nationwide and multi-year applications. The sample data consist of high-resolution images provided by Google Earth [38], which were annotated through visual interpretation methods. These annotations classified the data into aquaculture ponds and background categories to ensure the precision and practical applicability of the training dataset.
During this study, we generated a total of 802 images, each with a pixel size of 512 × 512. These images encompass samples from various geographical environments in coastal areas, including lakes, rivers, aquaculture ponds, and salt pans. For the semantic segmentation task, we explicitly categorized these images into aquaculture ponds and background to meet the requirements of our study.
In this study, an 80/20 split was employed to partition the training and validation datasets. Specifically, 628 images were used for training the model, while 174 images were allocated to the validation set. This ratio is widely acknowledged as a standard practice in deep learning tasks, particularly for medium-sized datasets [39], due to its balance between efficiency and representativeness. The chosen partitioning strategy ensures that a sufficient number of samples are available for training, allowing the model to effectively learn the underlying data features, while preserving a representative subset for validation. This approach facilitates a robust evaluation of the model’s generalization ability and mitigates the risk of performance degradation caused by overfitting.

3.2. Data Augmentation

Compared to traditional sliding window methods, the sample augmentation strategy proposed in this study effectively avoids over-training on repetitive regions, thereby mitigating the risk of model overfitting. Given the limitation of training sample size, we designed and implemented an online data augmentation strategy to enhance sample diversity in real-time during the training process. Specifically, during the construction of the Dataloader, we introduced random parameters that allow, with a certain probability, operations such as rotation, cropping, and scaling to be applied to samples before training each batch, thus generating augmented samples dynamically. This online augmentation approach eliminates the need for additional storage of augmented data, enabling the optimal utilization of limited samples while significantly improving the model’s generalization capability across diverse scenarios.

3.3. Model Structure

In the design of DBCE-Net (Dual-Input Context Enhanced Network), the network architecture primarily comprises the main branch, the Context Edge Supplement Branch (CESB), and the feature fusion and decoding modules (Figure 3), where DBCE-Net outputs the final segmentation map exclusively from the main branch, and with CESB enhancing the network’s performance by supplementing contextual information.
In this study, we selected ResNeSt [40] as the encoder for our model due to its substantial performance improvement and efficient capture of complex data features. ResNeSt, a significant evolution of ResNet [41], represents the latest generation of deep convolutional neural networks. By introducing the Split-Attention mechanism, it maintains controllability over network depth and parameter scale while markedly enhancing the network’s feature representation capabilities. This mechanism allows the network to perform adaptive recalibration of features within each block, thereby strengthening the model’s ability to capture key features and improving overall performance.
During the feature extraction stage, DBCE-Net’s main branch employs ResNeSt200e as the encoder, concentrating on extracting detailed information from high-resolution remote sensing images. Meanwhile, CESB processes images with fixed center points and expanded sizes, primarily to supplement the contextual and boundary information missing due to cropping in the main branch. This structural design enables the Network to comprehensively address contextual loss due to cropping in remote sensing images and also enhances the network’s multi-scale capabilities.
In the feature fusion stage, the Dynamic Feature Fusion (DFF) module intelligently integrates features from the main branch and CESB. In the decoding phase, DBCE-Net gradually restores compressed feature maps to their original size, ultimately outputting precise aquaculture pond segmentation maps exclusively from the main branch.
Through this multi-branch structure and innovative loss function design, DBCE-Net fully exploits the rich information provided by remote sensing imagery, thereby enhancing the accuracy and efficiency of aquaculture pond extraction. The workflow for aquaculture pond classification based on DBCE-Net (Figure 4) is shown below.

3.3.1. Context-Enhanced Supplementary Branch

When extracting aquaculture ponds, traditional networks often struggle due to the lack of global contextual information and boundary features. These limitations result in frequent misclassifications, such as mistakenly identifying complex, multi-branch rivers or lakes as aquaculture ponds, or failing to extract the ponds completely. This issue primarily stems from the inherent constraints of single-branch networks, whose feature extraction capabilities are limited by their local receptive fields, making it difficult to capture the broader spatial relationships and global background information within the image.
The fundamental principle of the Context-Enhanced Supplementary Branch (CESB) lies in its ability to process larger-scale input images and extract expanded contextual information to compensate for the main branch’s inability to independently capture spatial relationships and boundary details. In the task of aquaculture pond extraction, CESB effectively recognizes the global relationships between ponds and surrounding geographical features, such as rivers, lakes, and vegetation. This capability aids in distinguishing visually similar but fundamentally different objects, while also enhancing the completeness and accuracy of pond boundary delineation. Unlike traditional single-branch networks, which primarily focus on local feature extraction, CESB emphasizes macro-level feature representation, significantly improving the network’s understanding of complex scenes.
Specifically, CESB dynamically integrates the localized features extracted by the main branch with its own global contextual information. This integration not only prevents the misclassification of rivers or lakes as aquaculture ponds but also reduces the risk of incomplete pond extraction caused by insufficient local information. By supplementing the main branch with expanded contextual features, CESB addresses challenges such as boundary ambiguity and classification errors, leading to substantial improvements in segmentation accuracy and robustness.
To achieve this, CESB incorporates ResNeSt50 as its lightweight encoder, striking an optimal balance between computational efficiency and feature extraction capacity. The lightweight design of ResNeSt50 minimizes the computational overhead introduced by CESB, while ensuring the effective capture of additional contextual features essential for enhancing segmentation performance. This design consideration makes CESB particularly suitable for computationally demanding tasks requiring high efficiency.
The outputs generated by CESB do not replace the feature representations of the main branch. Instead, they undergo fine-grained integration through a Dynamic Feature Fusion module, which ensures seamless incorporation of the contextual information extracted by CESB into the localized feature representations of the main branch. By combining global contextual information with fine-grained local details, the fusion module enables the network to produce unified, hierarchical feature representations. This approach enhances overall segmentation accuracy and robustness, particularly for aquaculture pond extraction. It effectively differentiates aquaculture ponds from other water bodies and ensures accurate delineation of pond boundaries.

3.3.2. Dynamic Feature Fusion Module

To effectively integrate the backbone features and contextual boundary features extracted by the dual-branch network, we designed the Dynamic Feature Fusion (DFF) module. The DFF module is designed to learn the complementary relationship between high-resolution features and contextual boundary features, resulting in enhanced feature representations that encompass both local details and global semantics. This is particularly beneficial in the context of aquaculture pond extraction, where it demonstrates effective effectiveness.
In the extraction of aquaculture ponds, the DFF module improves segmentation performance by enhancing the precise capture of boundaries and the overall perception of pond regions. Aquaculture ponds are often located in complex geographical environments, such as areas with branching rivers or adjacent water bodies. Their boundaries are susceptible to interference, and the internal texture information is relatively sparse, making it challenging for traditional segmentation methods to distinguish ponds from other water bodies. By dynamically adjusting the weight distribution between backbone features and contextual boundary features, the DFF module effectively mitigates issues such as ambiguous boundaries, incomplete feature representation, and misclassification, enabling accurate and comprehensive extraction of pond regions.
In the design of the DFF module (Figure 5), three dilated convolutions with different dilation rates (2, 3 and 4) are first applied to the contextual boundary features. These operations aggregate rich global contextual information through multi-scale dilated convolutions [28], enlarging the receptive field and capturing long-range relationships within the image, thereby enhancing the semantic representation of features. For aquaculture pond segmentation, this global perception capability enables the network to better distinguish ponds from surrounding environments, reducing the likelihood of misclassification.
The DFF module further incorporates two feature processing and fusion branches. The first branch applies a 1 × 1 convolution followed by a Sigmoid activation function and performs element-wise multiplication with the original feature map. This process enhances the significant regions of the feature map while suppressing less important areas, achieving feature selection and weight allocation. The second branch begins with a pooling operation to reduce spatial dimensions and aggregate feature information. It then applies two sequential 1 × 1 convolutions followed by Sigmoid activation functions, culminating in element-wise multiplication with the original feature map. This design refines the information within the feature map, ensuring the accurate extraction of features relevant to aquaculture ponds.
By combining the outputs of the two branches, the DFF module achieves efficient feature fusion and dynamic weighting, ensuring precise extraction of critical features. Through the introduction of a dynamic weight learning mechanism, the DFF module not only adjusts weight distributions across feature channels but also optimizes weights in the spatial dimension. This flexibility allows the module to adapt to various image scenes. In aquaculture pond segmentation, the DFF module significantly enhances global perception, contextual boundary information processing, and local detail capture, resulting in improved accuracy and robustness.

3.3.3. Training Parameter Settings and Post-Processing

In this study, to meet the high computational demands of deep learning, we selected Python as the development language and PyTorch as the primary framework. CUDA (version 11.2) was employed for GPU acceleration in the experiments, running on a system equipped with an NVIDIA GeForce RTX 3090 GPU and an Intel(R) Xeon(R) Gold 6226R CPU @ 2.90 GHz. The Python version used in development was 3.8.5. Additionally, the GDAL library was utilized for handling geospatial data. After a series of tests, we determined the optimal training configuration: 200 training epochs, a batch size of 24, the Adam optimizer, an initial learning rate set to 1 × 10−4, and the ReLU activation function. We chose Focal Loss [42] and Dice Loss [43] as the loss functions. The training loss curve is illustrated (Figure 6).
In addition, it is essential to differentiate aquaculture ponds from independent reservoirs, which, although often possessing similar shapes and sizes, are distinct entities. These independent reservoirs cannot be reliably distinguished from aquaculture ponds based solely on spectral and geometric characteristics. Consequently, to account for the spatial distribution patterns of aquaculture ponds, the distance between the nearest neighboring objects of potential aquaculture ponds was calculated. A threshold was then established to effectively exclude independently distributed reservoirs. The results from the training process indicate that a minimum distance threshold of 50 m is optimal; if no other water bodies are located within 50 m of a given water body, it is classified as an independent reservoir.
FocalLoss = α t ( 1 p t ) γ log ( p t )
DiceLoss = 1 2 T P 2 T P + F P + F N
TotalLoss = FocalLoss + DiceLoss
p t is the model’s estimated probability for the true class, α t is a balancing factor for class imbalance, and γ is a focusing parameter used to reduce the impact of easy examples. T P is the number of true positives, F P is the number of false positives, F N is the number of false negatives.

3.4. Evaluation Metrics

We used the F1 score [44], precision, recall, and OA as metrics to evaluate the performance of the model. We also calculated confusion matrices for accuracy assessment. The formulas for the evaluation metrics are as follows:
Precision = T P T P + F P
Recall = T P T P + F N
F 1 Score = 2 · Precision · Recall Precision + Recall
OA = T P + T N T P + T N + F P + F N
where T P is the number of true positive samples (i.e., samples correctly predicted as aquaculture ponds), T N is the number of true negative samples (i.e., samples correctly predicted as non-aquaculture ponds), F N is the number of false negative samples (i.e., aquaculture pond samples incorrectly predicted as non-aquaculture ponds), and F P is the number of false positive samples (i.e., non-aquaculture pond samples incorrectly predicted as aquaculture ponds).
Our sampling points were initially generated based on the 2021 ESA dataset [45], covering various water bodies such as rivers, reservoirs, lakes, farmland, seasonal wetlands, and rice fields (Figure 7). To address the issue of annual variations in the distribution of aquaculture ponds, we also rely on high-resolution Google Earth images manually annotated each year to migrate samples to different years.

4. Results

4.1. Accuracy Assessment

The results (Table 3) indicate that the overall accuracy evaluation metric for the aquaculture pond map produced in this study averages around 93%. Specifically, the precision and F1 score for aquaculture ponds both exceed 90%. The confusion matrix results demonstrate a high level of agreement between the aquaculture pond map generated in this study and the validation sample points. The training model established based on 2019 samples maintains comparable accuracy in other years, indicating that the model exhibits good temporal consistency and stability. The model’s ability to sustain high accuracy across different years underscores its reliability in temporal detection tasks.
To further verify the accuracy of the contours and shapes of aquaculture ponds predicted by this method, we compared the ponds generated using this method with polygons manually delineated from high-resolution Google Earth images.
The comparison results indicate that the extracted contours and shapes are highly consistent with the manually delineated high-precision results, as illustrated in the figure. This comparison validates the reliability of the method proposed in this study and demonstrates its effectiveness in extracting aquaculture pond information (Figure 8).
We observed that the aquaculture pond extraction results obtained using this method are highly consistent with visual interpretation results in most cases, exhibiting significant overlap and extraction performance.
However, we also noticed some errors in extracting pond edges, resulting in an overall extracted aquaculture pond area that is slightly smaller than the actual area. This phenomenon is primarily due to the limitations of spatial resolution, which affect the precise extraction of aquaculture pond boundaries. To further improve the accuracy of boundary extraction, future research could consider using higher resolution imagery or developing more advanced image processing algorithms to mitigate the impact of spatial resolution on the results.

4.2. Verification of Typicality of Selected Data

To verify the representativeness of the training data and the generalization ability of the model, we conducted statistical analysis and comparative analysis on the provincial sample points from all years.
(Table 4) shows that the accuracy between provinces with training samples and provinces without training samples is basically the same. This indicates that our training samples have high representativeness and exhibit strong generalization ability.

4.3. Ablation Experiments

To validate the impact of the proposed DFF (Dynamic Feature Fusion) module on model performance, we conducted a series of ablation experiments, comparing the classification performance of the model with and without the DFF module. We selected samples from the year 2023 to ensure the precision and reliability of the experimental results.
(Table 5) indicates that the introduction of the DFF module increased the model’s overall accuracy from 91.9% to 93.1%, an improvement of 1.2%, demonstrating the effectiveness of DFF in enhancing the overall performance of the model. Specifically, the DFF module improved the precision of the AP category from 87.6% to 92.6%, although the recall slightly decreased (from 91.4% to 88.6%). Nevertheless, the F1 score increased by 1.1%, indicating that the DFF module provides significant benefits in extracting aquaculture pond features.

4.4. Comparative Experiments with Single-Branch Structure

To verify the effectiveness of improvements made by DBCE-Net compared to traditional single-branch structure networks, we selected validation sample points from 2023 for comparative experiments. This comparative experiment aims to demonstrate the performance enhancements achieved by DBCE-Net. By comparing the performance of DBCE-Net and single-branch structure networks at the same validation sample points, we can directly evaluate the advantages of the improved network structure in terms of processing efficiency and accuracy.
(Table 6) below presents the comparison results of DBCE-Net and other single-branch networks, revealing a significant increase in accuracy.
It is evident that the single-branch structure of networks results in misclassification of rivers and lakes due to insufficient multi-scale processing capability and loss of edge information, leading to decreased accuracy (Figure 9). These results indicate that DBCE-Net possesses significant advantages over traditional single-branch structures in handling boundary information and complex scenes, thereby enhancing overall segmentation performance.

5. Discussion

5.1. Comparison with Other Datasets

We further compared the results of this study with other 10 m resolution land cover datasets, including the FROM-GLC10 dataset [46], the CWaC dataset [47] the ESA land cover dataset [45], and the China Aquaculture Ponds V1 [48]. Although these datasets do not specifically classify aquaculture ponds, they are generally categorized as a type of water body in land cover classification systems. Therefore, by comparing the extraction results of water body categories in these datasets, we can indirectly evaluate their performance in identifying aquaculture ponds.
The comparison (Figure 10) reveals that while these datasets can generally extract the locations of aquaculture ponds, their accuracy is relatively low, with prevalent issues of misclassification and omission. Specifically, the ESA dataset exhibits blurred boundaries of aquaculture ponds, with closely connected ponds making it difficult to discern the shape and boundaries of individual ponds, thereby failing to effectively distinguish between aquaculture ponds and surrounding land, which is inconsistent with the expected performance of 10 m high-resolution data. The FROM-GLC10 dataset performs better in extracting aquaculture ponds, effectively distinguishing between aquaculture ponds and land, but fails to differentiate between dikes and aquaculture ponds. The CWaC dataset is relatively superior in pond boundary recognition but exhibits lower accuracy in extracting small aquaculture ponds. The China Aquaculture Ponds V1 dataset demonstrates an enhanced ability to effectively distinguish between salt fields and aquaculture ponds, achieving improved overall accuracy. Nevertheless, for certain small-scale aquaculture ponds, the dataset only managed to extract a minimal portion of these features, resulting in incomplete and less comprehensive extraction outcomes. In this study, the total area of aquaculture ponds in 2022 was calculated to be 8592 square kilometers, exceeding the area covered by the CAV1 dataset (7919 square kilometers). This larger area includes a more comprehensive extraction of small aquaculture ponds, demonstrating the reliability of our study. However, unlike the CAV1 dataset, our study did not exclude abandoned aquaculture ponds. Moreover, most of the aforementioned datasets effectively distinguish between salt pans and aquaculture ponds. Overall, compared to existing land cover datasets, the method proposed in this paper more accurately identifies the boundaries of aquaculture ponds, more thoroughly extracts small aquaculture ponds, and successfully distinguishes between salt pans and aquaculture ponds. These results indicate that the method proposed in this paper exhibits higher accuracy and discriminative capability in identifying aquaculture ponds.

5.2. Area and Spatial Distribution of Coastal Aquaculture Ponds in China

The spatial distribution of coastal aquaculture ponds, along with their national and provincial regional delineations for the year 2023 (Figure 11A). These aquaculture ponds are extensively dispersed along the Chinese coastline, covering a total area of 8290.52 km2. Coastal aquaculture ponds are densely distributed in the following four regions: Bohai Bay, coastal plains of Jiangsu Province, Beilin Bay of Guangdong Province, and the Pearl River Delta (Figure 11A). These areas account for over 70% of the total coastal aquaculture ponds in China. Shandong has the largest area of coastal aquaculture ponds (2271.30 km2, 27.4%), followed by Guangdong (1205.18 km2, 14.5%), Jiangsu (1094.28 km2, 13.2%), and Liaoning (1092.78 km2, 13.2%) (Figure 11B,C). Collectively, these four provinces account for 68.3% of the total area of coastal aquaculture ponds in China.

5.3. Area and Spatial Distribution of Coastal Aquaculture Ponds in China

This study detailed the spatial changes of coastal aquaculture ponds in China from 2017 to 2023, offering an important perspective on the spatial dynamics of coastal aquaculture. During this period, the total area of coastal aquaculture ponds exhibited a continuous downward trend, decreasing from 8970.25 km2 to 8261.17 km2 (Figure 12C), resulting in a net loss of 708.83 km2, a decline of 7.90%. Additionally, existing studies reveal that before 2016, the area of coastal aquaculture ponds in China generally exhibited an expansion trend. Nevertheless, this study reveals a significant reversal in the trend, starting around 2017, wherein the area dedicated to aquaculture ponds has begun to decline. This finding represents a critical inflection point in the spatial dynamics of coastal aquaculture development in China.
We calculated the annual change rate (ACR) for each province in the entire coastal region (Figure 11A). The changes in the aquaculture area varied significantly across different provinces, highlighting the differences in aquaculture development strategies, natural conditions, and policy orientations among regions. The reduction in the aquaculture area was most significant in Shanghai Province, decreasing by 124.41 square kilometers, with a reduction rate of 38.92%. This was followed by Zhejiang and Jiangsu, which decreased by 31.57 square kilometers and 219.60 square kilometers, with reduction rates of 31.57% and 19.07% (Figure 11B), respectively. The changes observed in Jiangsu’s coastal areas are primarily attributed to natural coastal erosion and the implementation of the “Three Zones” policy along with land reclamation initiatives, with the most severely affected regions experiencing an average annual shoreline retreat of approximately 80 m, in Zhejiang Province, ecological protection efforts have led to the establishment of several wetland protection areas and large-scale projects to convert high-pollution, illegally constructed aquaculture ponds back to farmland. According to the Shandong Fisheries Statistical Yearbook released by the Shandong Provincial Department of Agriculture in recent years (http://nync.shandong.gov.cn/, accessed on 1 November 2024), the trends in the aquaculture pond area align closely with those calculated in our study. The area of aquaculture ponds are basically consistent with the trend we have calculated. In contrast, the aquaculture area in Liaoning exhibited an increasing trend, which is also reflected in the Liaoning Province National Economic and Social Development Statistical Bulletin (https://tjj.ln.gov.cn/, accessed on 1 November 2024), indicating a yearly increase in the area of aquaculture ponds.
As reported in the Fisheries Statistics Bulletin issued by the Ministry of Agriculture and Rural Affairs of the People’s Republic of China since 2017 (http://www.moa.gov.cn/, accessed on 20 October 2024), the area of aquaculture ponds in China has been declining in recent years. This observed trend confirms our research findings, thus confirming the reliability of our study. In addition, data show that although the aquaculture area decreased, the decline in aquatic product production in coastal areas was relatively small. This indicates that the structure of aquaculture ponds in coastal areas of China has been optimized, resulting in a significant increase in yield per unit area.

5.4. Limitations of the Method and Future Directions for Improvement

This study employed a pixel-based semantic segmentation approach [49], which was extensively utilized in aquaculture pond extraction due to its capability to directly classify each pixel into specific categories. However, compared to object-based methods, pixel-based segmentation is often prone to challenges such as over-segmentation and under-segmentation. In this study, under-segmentation is particularly prevalent due to the complex shapes and spatial distributions of aquaculture ponds, resulting in multiple adjacent ponds being erroneously merged into a single entity.
In contrast, object-based methods leverage spectral [27,50,51], spatial, and contextual information to group pixels into meaningful objects, effectively mitigating issues of over-segmentation and under-segmentation. Nevertheless, these methods, including object-based image analysis (OBIA), often require additional preprocessing steps and manual parameter tuning, which increases computational complexity and limits their scalability for large-scale automated applications.
Despite these limitations, this study adopted a pixel-based semantic segmentation method due to its simplicity and its suitability for large-scale automated analyses, particularly when applied to moderate spatial resolution datasets, such as Sentinel-2 imagery. Future research directions could explore the integration of object-based methods or hybrid approaches to address the segmentation challenges encountered. Future research could address these challenges by incorporating higher-resolution remote sensing data, such as PlanetScope [52], WorldView imagery [53], and the Gaofen series [54], to enhance the precision of pond boundary delineation, thereby overcoming the limitations imposed by spatial resolution on segmentation performance.

6. Conclusions

This study proposes a dual-branch deep learning model to improve the accuracy of identifying coastal aquaculture ponds in China. By integrating a Context-Enhanced Supplementary Branch (CESB), the model enhances the recognition of ponds with varying shapes and sizes while reducing misclassification of rivers and lakes. Since 2017, aquaculture pond areas have declined by 10.05%, reflecting policy adjustments and ecological protection efforts. The model demonstrates strong generalization capabilities, effectively adapting to diverse temporal and spatial data with limited training samples. Compared to traditional single-branch models, CESB more accurately captures contextual and boundary features in broader scenes. Beyond its applicability to China’s coastal ecological monitoring, this method has global potential, enabling detailed aquaculture pond mapping and spatiotemporal change analysis worldwide. This research offers new insights into sustainable coastal ecosystem development, emphasizes the environmental impact of aquaculture pond changes, and provides valuable references for policy making, ecological conservation, and interdisciplinary studies, advancing resource management and environmental protection.

Author Contributions

Conceptualization, L.Z., H.Z. and W.C.; Methodology, Y.L. and H.Z.; Software, Y.L.; Writing—original draft, Y.L.; Writing—review & editing, L.Z., H.Z. and W.C.; Visualization, Y.L. and L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 42106174) and the open fund of State Key Laboratory of Satellite Ocean Envi-ronment Dynamics, Second, Institute of Oceanography, MNR (No. QNHX2202).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

We sincerely thank the European Space Agency (ESA) for distributing the Sentinel-2 data and Google Earth Engine (GEE) for distributing the platform.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Edwards, P. Aquaculture environment interactions: Past, present and likely future trends. Aquaculture 2015, 447, 2–14. [Google Scholar] [CrossRef]
  2. FAO. Fisheries Department. In The State of World Fisheries and Aquaculture; FAO: Rome, Italy, 2022. [Google Scholar]
  3. Su, S.; Tang, Y.; Chang, B.; Zhu, W.; Chen, Y. Evolution of marine fisheries management in China from 1949 to 2019: How did China get here and where does China go next? Fish Fish. 2020, 21, 435–452. [Google Scholar] [CrossRef]
  4. Zhao, K.; Zhang, M.; Wang, K.; Zhu, K.; Xu, C.; Xie, J.; Xu, J. Aquaculture impacts on china’s marine wild fisheries over the past 30 years. Front. Mar. Sci. 2021, 8, 710124. [Google Scholar] [CrossRef]
  5. Nagaraju, T.V.; Malegole, S.B.; Chaudhary, B.; Ravindran, G. Assessment of environmental impact of aquaculture ponds in the western delta region of Andhra Pradesh. Sustainability 2022, 14, 13035. [Google Scholar] [CrossRef]
  6. Xu, P.; Zhang, X.; Zhang, F.; Bempah, G.; Lu, C.; Lv, S.; Zhang, W.; Cui, P. Use of aquaculture ponds by globally endangered red-crowned crane (Grus japonensis) during the wintering period in the Yancheng National Nature Reserve, a Ramsar wetland. Glob. Ecol. Conserv. 2020, 23, e01123. [Google Scholar] [CrossRef]
  7. Angel, D.; Jokumsen, A.; Lembo, G. Aquaculture production systems and environmental interactions. In Organic Aquaculture: Impacts and Future Developments; Springer: Berlin/Heidelberg, Germany, 2019; pp. 103–118. [Google Scholar]
  8. Li, X.; Bellerby, R.; Craft, C.; Widney, S.E. Coastal wetland loss, consequences, and challenges for restoration. Anthr. Coasts 2018, 1, 1–15. [Google Scholar] [CrossRef]
  9. Mavraganis, T.; Constantina, C.; Kolygas, M.; Vidalis, K.; Nathanailides, C. Environmental issues of Aquaculture development. Egypt. J. Aquat. Biol. Fish. 2020, 24, 441–450. [Google Scholar] [CrossRef]
  10. Salin, K.R.; Arome Ataguba, G. Aquaculture and the environment: Towards sustainability. Sustain. Aquac. 2018, 1–62. [Google Scholar] [CrossRef]
  11. Gao, B.C. NDWI—A normalized difference water index for remote sensing of vegetation liquid water from space. Remote Sens. Environ. 1996, 58, 257–266. [Google Scholar] [CrossRef]
  12. Xu, H. Modification of normalised difference water index (NDWI) to enhance open water features in remotely sensed imagery. Int. J. Remote Sens. 2006, 27, 3025–3033. [Google Scholar] [CrossRef]
  13. Wang, X.; Xiao, X.; Zou, Z.; Hou, L.; Qin, Y.; Dong, J.; Doughty, R.B.; Chen, B.; Zhang, X.; Chen, Y.; et al. Mapping coastal wetlands of China using time series Landsat images in 2018 and Google Earth Engine. ISPRS J. Photogramm. Remote Sens. 2020, 163, 312–326. [Google Scholar] [CrossRef] [PubMed]
  14. Binh, N.A.; Hauser, L.T.; Salinero-Delgado, M.; Viet Hoa, P.; Thi Phuong Thao, G.; Verrelst, J. Monitoring mangrove traits through optical Earth observation: Towards spatio-temporal scalability using cloud-based Sentinel-2 continuous time series. ISPRS J. Photogramm. Remote Sens. 2024, 214, 135–152. [Google Scholar] [CrossRef]
  15. Liu, Y.; Xiao, X.; Li, J.; Wang, X.; Chen, B.; Sun, C.; Wang, J.; Tian, P.; Zhang, H. Tracking changes in coastal land cover in the Yellow Sea, East Asia, using Sentinel-1 and Sentinel-2 time-series images and Google Earth Engine. ISPRS J. Photogramm. Remote Sens. 2023, 196, 429–444. [Google Scholar] [CrossRef]
  16. Sun, W.; Hou, T.; Chen, C.; Yang, G.; Chen, B.; Meng, X.; Ren, K. Mapping China’s coastal aquaculture ponds expansion with sentinel-2 images during 2017–2021. Int. J. Digit. Earth 2024, 17, 2297943. [Google Scholar] [CrossRef]
  17. Wang, M.; Mao, D.; Xiao, X.; Song, K.; Jia, M.; Ren, C.; Wang, Z. Interannual changes of coastal aquaculture ponds in China at 10-m spatial resolution during 2016–2021. Remote Sens. Environ. 2023, 284, 113347. [Google Scholar] [CrossRef]
  18. Ren, C.; Wang, Z.; Zhang, Y.; Zhang, B.; Chen, L.; Xi, Y.; Xiao, X.; Doughty, R.B.; Liu, M.; Jia, M.; et al. Rapid expansion of coastal aquaculture ponds in China from Landsat observations during 1984–2016. Int. J. Appl. Earth Obs. Geoinf. 2019, 82, 101902. [Google Scholar] [CrossRef]
  19. Xia, Z.; Guo, X.; Chen, R. Automatic extraction of aquaculture ponds based on Google Earth Engine. Ocean. Coast. Manag. 2020, 198, 105348. [Google Scholar] [CrossRef]
  20. Ottinger, M.; Bachofer, F.; Huth, J.; Kuenzer, C. Mapping aquaculture ponds for the coastal zone of Asia with Sentinel-1 and Sentinel-2 time series. Remote Sens. 2021, 14, 153. [Google Scholar] [CrossRef]
  21. Kolli, M.K.; Opp, C.; Karthe, D.; Pradhan, B. Automatic extraction of large-scale aquaculture encroachment areas using Canny Edge Otsu algorithm in Google Earth Engine—The case study of Kolleru Lake, South India. Geocarto Int. 2022, 37, 11173–11189. [Google Scholar] [CrossRef]
  22. Wang, Z.; Zhang, J.; Yang, X.; Huang, C.; Su, F.; Liu, X.; Liu, Y.; Zhang, Y. Global mapping of the landside clustering of aquaculture ponds from dense time-series 10 m Sentinel-2 images on Google Earth Engine. Int. J. Appl. Earth Obs. Geoinf. 2022, 115, 103100. [Google Scholar] [CrossRef]
  23. Gao, L.; Wang, C.; Liu, K.; Chen, S.; Dong, G.; Su, H. Extraction of floating raft aquaculture areas from Sentinel-1 SAR images by a dense residual U-Net model with pre-trained ResNet34 as the encoder. Remote Sens. 2022, 14, 3003. [Google Scholar] [CrossRef]
  24. Deng, J.; Bai, Y.; Chen, Z.; Shen, T.; Li, C.; Yang, X. A convolutional neural network for coastal aquaculture extraction from high-resolution remote sensing imagery. Sustainability 2023, 15, 5332. [Google Scholar] [CrossRef]
  25. Zou, Z.; Chen, C.; Liu, Z.; Zhang, Z.; Liang, J.; Chen, H.; Wang, L. Extraction of aquaculture ponds along coastal region using u2-net deep learning model from remote sensing images. Remote Sens. 2022, 14, 4001. [Google Scholar] [CrossRef]
  26. Ai, B.; Xiao, H.; Xu, H.; Yuan, F.; Ling, M. Coastal aquaculture area extraction based on self-attention mechanism and auxiliary loss. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 2250–2261. [Google Scholar] [CrossRef]
  27. Liang, Z.; Wang, F.; Zhu, J.; Li, P.; Xie, F.; Zhao, Y. Autonomous Extraction Technology for Aquaculture Ponds in Complex Geological Environments Based on Multispectral Feature Fusion of Medium-Resolution Remote Sensing Imagery. Remote Sens. 2024, 16, 4130. [Google Scholar] [CrossRef]
  28. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar] [CrossRef]
  29. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar] [CrossRef]
  30. Fan, Q.; Liang, L.; Liang, F.; Sun, X. Research progress on coastline change in China. J. Coast. Res. 2020, 99, 289–295. [Google Scholar] [CrossRef]
  31. Li, K.; Zhang, L.; Chen, B.; Zuo, J.; Yang, F.; Li, L. Analysis of China’s coastline changes during 1990–2020. Remote Sens. 2023, 15, 981. [Google Scholar] [CrossRef]
  32. Isbaex, C.; Coelho, A.M. The Potential of Sentinel-2 Satellite Images for Land-Cover/Land-Use and Forest Biomass Estimation: A Review; IntechOpen: Rijeka, Croatia, 2021. [Google Scholar] [CrossRef]
  33. Transon, J.; d’Andrimont, R.; Maugnard, A.; Defourny, P. Survey of hyperspectral earth observation applications from space in the sentinel-2 context. Remote Sens. 2018, 10, 157. [Google Scholar] [CrossRef]
  34. Phiri, D.; Simwanda, M.; Salekin, S.; Nyirenda, V.R.; Murayama, Y.; Ranagalage, M. Sentinel-2 data for land cover/use mapping: A review. Remote Sens. 2020, 12, 2291. [Google Scholar] [CrossRef]
  35. Segarra, J.; Buchaillot, M.L.; Araus, J.L.; Kefauver, S.C. Remote sensing for precision agriculture: Sentinel-2 improved features and applications. Agronomy 2020, 10, 641. [Google Scholar] [CrossRef]
  36. Coluzzi, R.; Imbrenda, V.; Lanfredi, M.; Simoniello, T. A first assessment of the Sentinel-2 Level 1-C cloud mask product to support informed surface analyses. Remote Sens. Environ. 2018, 217, 426–443. [Google Scholar] [CrossRef]
  37. Liu, C.; Xu, J.; Xu, K.; Yu, J. Mapping large-scale aquaculture ponds in Jiangsu Province, China: An automatic extraction framework based on Sentinel-1 time-series imagery. Aquaculture 2024, 581, 740441. [Google Scholar] [CrossRef]
  38. Yu, L.; Gong, P. Google Earth as a virtual globe tool for Earth science applications at the global scale: Progress and perspectives. Int. J. Remote Sens. 2012, 33, 3966–3986. [Google Scholar] [CrossRef]
  39. Ko, H.; Chung, H.; Kang, W.S.; Kim, K.W.; Shin, Y.; Kang, S.J.; Lee, J.H.; Kim, Y.J.; Kim, N.Y.; Jung, H.; et al. COVID-19 pneumonia diagnosis using a simple 2D deep learning framework with a single chest CT image: Model development and validation. J. Med. Internet Res. 2020, 22, e19569. [Google Scholar] [CrossRef]
  40. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. Resnest: Split-attention networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2736–2746. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Ross, T.Y.; Dollár, G. Focal loss for dense object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar]
  43. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: Third International Workshop, DLMIA 2017, and 7th International Workshop, ML-CDS 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, 14 September 2017; Proceedings 3. Springer: Berlin/Heidelberg, Germany, 2017; pp. 240–248. [Google Scholar]
  44. Chen, T.Y.; Kuo, F.C.; Merkel, R. On the statistical properties of the f-measure. In Proceedings of the Fourth International Conference onQuality Software, Braunschweig, Germany, 8–9 September 2004; QSIC 2004 Proceedings. IEEE: Piscataway, NJ, USA, 2004; pp. 146–153. [Google Scholar] [CrossRef]
  45. Zanaga, D.; Van De Kerchove, R.; Daems, D.; De Keersmaecker, W.; Brockmann, C.; Kirches, G.; Wevers, J.; Cartus, O.; Santoro, M.; Fritz, S.; et al. ESA WorldCover 10 m 2021 v200. 2022. Available online: https://pure.iiasa.ac.at/18478 (accessed on 20 January 2024).
  46. Gong, P.; Liu, H.; Zhang, M.; Li, C.; Wang, J.; Huang, H.; Clinton, N.; Ji, L.; Li, W.; Bai, Y.; et al. Stable classification with limited sample: Transferring a 30-m resolution sample set collected in 2015 to mapping 10-m resolution global land cover in 2017. Sci. Bull. 2019, 64, 370–373. [Google Scholar] [CrossRef]
  47. Li, Y.; Niu, Z. Systematic method for mapping fine-resolution water cover types in China based on time series Sentinel-1 and 2 images. Int. J. Appl. Earth Obs. Geoinf. 2022, 106, 102656. [Google Scholar] [CrossRef]
  48. Wang, D.; Mao, D.; Wang, M.; Xiao, X.; Choi, C.Y.; Huang, C.; Wang, Z. Identify and map coastal aquaculture ponds and their drainage and impoundment dynamics. Int. J. Appl. Earth Obs. Geoinf. 2024, 134, 104246. [Google Scholar] [CrossRef]
  49. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  50. Hu, Y.; Zhang, L.; Chen, B.; Zuo, J. An Object-Based Approach to Extract Aquaculture Ponds with 10-Meter Resolution Sentinel-2 Images: A Case Study of Wenchang City in Hainan Province. Remote Sens. 2024, 16, 1217. [Google Scholar] [CrossRef]
  51. Li, B.; Gong, A.; Chen, Z.; Pan, X.; Li, L.; Li, J.; Bao, W. An object-oriented method for extracting single-object aquaculture ponds from 10 m resolution Sentinel-2 images on Google Earth Engine. Remote Sens. 2023, 15, 856. [Google Scholar] [CrossRef]
  52. Roy, D.P.; Huang, H.; Houborg, R.; Martins, V.S. A global analysis of the temporal availability of PlanetScope high spatial resolution multi-spectral imagery. Remote Sens. Environ. 2021, 264, 112586. [Google Scholar] [CrossRef]
  53. DigitalGlobe. WorldView-3 Data Sheet. 2014. Available online: https://earth.esa.int/eogateway/missions/worldview-3 (accessed on 28 December 2023).
  54. Guo, Q.; Liu, X.; Jin, S.; Wang, L. A review of high-resolution satellite remote sensing data in China: GF-1 to ZY-3 series. J. Remote Sens. 2017, 21, 658–675. [Google Scholar]
Figure 1. Study area information: (A) Geographic location of the study area, where the green color represents the study regions, including (a) Bohai Bay and (b) Hangzhou Bay and Sanmen Bay in Zhejiang Province. (13) Sentinel-2 satellite images, and (46) typical examples of coastal aquaculture ponds manually delineated from Google Earth images in 2021.
Figure 1. Study area information: (A) Geographic location of the study area, where the green color represents the study regions, including (a) Bohai Bay and (b) Hangzhou Bay and Sanmen Bay in Zhejiang Province. (13) Sentinel-2 satellite images, and (46) typical examples of coastal aquaculture ponds manually delineated from Google Earth images in 2021.
Remotesensing 17 00362 g001
Figure 2. Geographical locations of the three study areas along China’s coastline (A): Bohai Bay (a), Yancheng (b), Hangzhou Bay (c) and Shantou (d). Sampling points are classified into three categories based on land cover types: embankments (orange triangles), aquaculture ponds (pink dots), paddy fields (purple plus sign), and salt pans (yellow squares).
Figure 2. Geographical locations of the three study areas along China’s coastline (A): Bohai Bay (a), Yancheng (b), Hangzhou Bay (c) and Shantou (d). Sampling points are classified into three categories based on land cover types: embankments (orange triangles), aquaculture ponds (pink dots), paddy fields (purple plus sign), and salt pans (yellow squares).
Remotesensing 17 00362 g002
Figure 3. DBCE-Net network architecture and DFF module demonstrate dynamic feature fusion module.
Figure 3. DBCE-Net network architecture and DFF module demonstrate dynamic feature fusion module.
Remotesensing 17 00362 g003
Figure 4. Workflow for aquaculture pond classification by developing DBCE-Net.
Figure 4. Workflow for aquaculture pond classification by developing DBCE-Net.
Remotesensing 17 00362 g004
Figure 5. Dynamic Feature Fusion (DFF) module diagram.
Figure 5. Dynamic Feature Fusion (DFF) module diagram.
Remotesensing 17 00362 g005
Figure 6. Loss curve of network training.
Figure 6. Loss curve of network training.
Remotesensing 17 00362 g006
Figure 7. Spatial distribution of validation samples.
Figure 7. Spatial distribution of validation samples.
Remotesensing 17 00362 g007
Figure 8. Comparison between the aquaculture ponds extracted by our model and manually delineated sample polygons from Google Earth in various regions: (A) Bohai Bay, Liaoning Province; (B) Liaodong Bay, Liaoning Province; (C) Yancheng, Jiangsu Province; (D) Pearl River Estuary, Guangdong Province; 1, 4, 7, and 10 represent Sentinel-2 imagery. 2, 5, 8, and 11 are results from Google Earth; 3, 6, 9, and 12 are results from this study.
Figure 8. Comparison between the aquaculture ponds extracted by our model and manually delineated sample polygons from Google Earth in various regions: (A) Bohai Bay, Liaoning Province; (B) Liaodong Bay, Liaoning Province; (C) Yancheng, Jiangsu Province; (D) Pearl River Estuary, Guangdong Province; 1, 4, 7, and 10 represent Sentinel-2 imagery. 2, 5, 8, and 11 are results from Google Earth; 3, 6, 9, and 12 are results from this study.
Remotesensing 17 00362 g008
Figure 9. Comparing DBCE-Net results with UNet and other single-branch networks: (A,B) Sentinel-2 image, (a,f) DBCE-Net results, (b,g) UNet results, (c,h) U2Net results, (d,i) PSPNet results, (e,j) SegNet results. The red dots represent the study regions.
Figure 9. Comparing DBCE-Net results with UNet and other single-branch networks: (A,B) Sentinel-2 image, (a,f) DBCE-Net results, (b,g) UNet results, (c,h) U2Net results, (d,i) PSPNet results, (e,j) SegNet results. The red dots represent the study regions.
Remotesensing 17 00362 g009
Figure 10. Comparing coastal aquaculture ponds extracted by DBCE-Net with other 10 m resolution datasets, including the FROM-GLC10, CWaC, ESA Worldcover, and China Aquaculture Ponds V1. The red dots represent the study regions.
Figure 10. Comparing coastal aquaculture ponds extracted by DBCE-Net with other 10 m resolution datasets, including the FROM-GLC10, CWaC, ESA Worldcover, and China Aquaculture Ponds V1. The red dots represent the study regions.
Remotesensing 17 00362 g010
Figure 11. Distribution and area statistics of coastal aquaculture ponds in China in 2023 (A): Density map of coastal aquaculture ponds in China in 2023. (B): Area proportion of coastal aquaculture ponds by province. (C): Provincial area statistics of coastal aquaculture ponds. Coastal aquaculture ponds in Guangdong Province include Hong Kong and Macau.
Figure 11. Distribution and area statistics of coastal aquaculture ponds in China in 2023 (A): Density map of coastal aquaculture ponds in China in 2023. (B): Area proportion of coastal aquaculture ponds by province. (C): Provincial area statistics of coastal aquaculture ponds. Coastal aquaculture ponds in Guangdong Province include Hong Kong and Macau.
Remotesensing 17 00362 g011
Figure 12. Spatial distribution and area changes of coastal aquaculture ponds in China from 2017 to 2023 (A) spatial distribution of area changes in coastal aquaculture ponds across provinces from 2017 to 2023. Four key wetland areas with significant changes in (a) Liaoning Bay, (b) Yancheng, (c) Sanmen Bay, and (d) Guanghai Bay. Parts (B,C) represent the area changes of coastal aquaculture ponds at the provincial and national levels from 2017 to 2023. Numbers 1–4 refer to the results from 2017, and numbers 5–8 refer to the results from 2023.
Figure 12. Spatial distribution and area changes of coastal aquaculture ponds in China from 2017 to 2023 (A) spatial distribution of area changes in coastal aquaculture ponds across provinces from 2017 to 2023. Four key wetland areas with significant changes in (a) Liaoning Bay, (b) Yancheng, (c) Sanmen Bay, and (d) Guanghai Bay. Parts (B,C) represent the area changes of coastal aquaculture ponds at the provincial and national levels from 2017 to 2023. Numbers 1–4 refer to the results from 2017, and numbers 5–8 refer to the results from 2023.
Remotesensing 17 00362 g012
Table 1. Sentinel-2 bands and their central wavelengths.
Table 1. Sentinel-2 bands and their central wavelengths.
Spectral RangeBand NameCentral Wavelength (nm)
VisibleBlue (B2)490
Green (B3)560
Red (B4)665
Near-InfraredNIR (B8)842
Shortwave InfraredSWIR1 (B11)1610
SWIR2 (B12)2190
Other BandsB1443
B5705
B6740
B7783
B8A865
B9935
B101375
Table 2. Image information for different areas.
Table 2. Image information for different areas.
AreaImage SizeImage Data (Date Range)
Bohai Bay16,516 × 16,6362021-03-01–2021-12-01
Hangzhou Bay14,978 × 44642021-03-01–2021-12-01
Sanmen Bay7436 × 31362021-03-01–2021-12-01
Table 3. Confusion matrix and F1 scores for the aquaculture pond map generated by DBCE-Net in China from 2017 to 2023.
Table 3. Confusion matrix and F1 scores for the aquaculture pond map generated by DBCE-Net in China from 2017 to 2023.
YearClassAPNon-APPrecisionRecallF1 ScoreOA
2017AP7286891.5%89.7%90.5%92.7%
Non-AP84120993.5%94.8%94.1%
2018AP7456791.7%89.8%90.7%92.8%
Non-AP85121093.4%94.9%94.1%
2019AP7455892.8%89.8%91.2%93.2%
Non-AP85122093.5%95.5%94.5%
2020AP7264993.7%89.9%91.7%93.7%
Non-AP82122893.7%96.2%94.9%
2021AP7736192.7%88.7%90.7%92.5%
Non-AP98117992.3%95.1%93.7%
2022AP7136092.2%89.5%90.9%93.1%
Non-AP84121793.5%95.3%94.4%
2023AP6875592.6%88.6%90.6%93.1%
Non-AP86123093.3%95.7%94.5%
Table 4. Accuracy in different provinces.
Table 4. Accuracy in different provinces.
TypeProvinceClassAPNon-APPrecisionRecallF1 ScoreOA
With TrainingHebeiAP831193.30%88.30%90.70%92.7%
Non-AP613492.40%95.70%94.00%
ShandongAP87610391.10%89.50%90.20%92.3%
Non-AP86141293.20%94.30%93.70%
ZhejiangAP2012791.40%88.20%89.70%95.8%
Non-AP1984296.90%97.80%97.30%
Without TrainingLiaoningAP122611497.20%91.50%94.30%92.8%
Non-AP3570486.10%95.30%90.40%
JiangsuAP8459390.00%90.10%90.00%89.8%
Non-AP9480789.70%89.60%89.60%
ShanghaiAP471488.70%77.00%82.50%90.0%
Non-AP613490.50%95.70%93.10%
FujianAP3118292.60%79.10%85.30%91.0%
Non-AP2577390.40%96.90%93.50%
TaiwanAP3153690.80%89.70%90.30%92.8%
Non-AP3256694.00%94.60%94.30%
GuangdongAP9219692.10%90.60%91.30%94.1%
Non-AP79187695.10%96.00%95.50%
GuangxiAP1581892.40%89.80%91.10%93.3%
Non-AP1327493.80%95.50%94.60%
HainanAP1251187.40%91.90%89.60%96.1%
Non-AP1859898.20%97.10%97.60%
Table 5. Performance of DBCE-Net with and without DFF module.
Table 5. Performance of DBCE-Net with and without DFF module.
TypeClassAPNon-APPrecisionRecallF1 ScoreOA
DBCE-Net (with DFF)AP6875592.6%88.6%90.6%93.1%
Non-AP88123093.3%95.7%94.5%
DBCE-Net (without DFF)AP70810087.6%91.4%89.5%91.9%
Non-AP67118594.6%92.2%93.4%
Table 6. Confusion matrix of DBCE-Net and other single-branch networks.
Table 6. Confusion matrix of DBCE-Net and other single-branch networks.
TypeClassAPNon-APPrecisionRecallF1 ScoreOA
DBCE-NetAP6875592.6%88.6%90.6%93.1%
Non-AP88123093.3%95.7%94.5%
UNetAP68310187.1%88.1%87.6%90.6%
Non-AP92118492.8%92.1%92.5%
U2NetAP6746691.1%87.0%89.0%91.9%
Non-AP101121992.3%94.9%93.6%
PSPNetAP6247189.8%80.5%84.9%89.2%
Non-AP151121488.9%94.5%91.6%
SegNetAP59011583.7%76.1%79.7%85.4%
Non-AP185117086.3%91.1%88.6%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Zhao, L.; Zhang, H.; Cao, W. DBCE-Net: A Novel Deep Learning Framework for Annual Mapping of Coastal Aquaculture Ponds in China with Sentinel-2 Data. Remote Sens. 2025, 17, 362. https://doi.org/10.3390/rs17030362

AMA Style

Li Y, Zhao L, Zhang H, Cao W. DBCE-Net: A Novel Deep Learning Framework for Annual Mapping of Coastal Aquaculture Ponds in China with Sentinel-2 Data. Remote Sensing. 2025; 17(3):362. https://doi.org/10.3390/rs17030362

Chicago/Turabian Style

Li, Yin, Liaoying Zhao, Huaguo Zhang, and Wenting Cao. 2025. "DBCE-Net: A Novel Deep Learning Framework for Annual Mapping of Coastal Aquaculture Ponds in China with Sentinel-2 Data" Remote Sensing 17, no. 3: 362. https://doi.org/10.3390/rs17030362

APA Style

Li, Y., Zhao, L., Zhang, H., & Cao, W. (2025). DBCE-Net: A Novel Deep Learning Framework for Annual Mapping of Coastal Aquaculture Ponds in China with Sentinel-2 Data. Remote Sensing, 17(3), 362. https://doi.org/10.3390/rs17030362

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop