1. Introduction
Small-scale reservoirs are a critical component of water resource management and play an essential role in flood control, irrigation, and ecological preservation, particularly in regions susceptible to climatic and hydrological variability [
1]. In China alone, there are over 98,000 reservoirs, of which 95% are classified as small-scale, reflecting their pivotal importance in the national context. Among these, Jiangxi Province is particularly notable, with over 10,300 small-scale reservoirs—accounting for 11.4% of the country’s total [
2]. These reservoirs significantly contribute to the socio-economic development of the region by mitigating floods, ensuring agricultural productivity, and maintaining ecological balance. However, the safety and stability of these small-scale reservoirs remain a significant challenge due to several factors, including inadequate design, incomplete hydrological and geological data, and substandard construction practices. These challenges render small-scale reservoirs more susceptible to failures compared to their larger counterparts [
3,
4]. Since 2010, 1031 reservoirs in Jiangxi Province have experienced major hazards, including severe dam deformation in 127 cases, leakage damage in 382 cases, and scattered flooding in 170 cases. These hidden dangers can easily lead to dam landslides, collapses, and breaches. As of now, more than 10 dam collapses have occurred, causing casualties and damaging over 5000 acres of farmland, resulting in a direct economic loss of 50 million yuan. The remaining 600 reservoirs have become dangerous and unable to operate normally, with a reinforcement cost of about 200 million yuan and indirect economic losses of 160 million yuan due to the inability to realize their functional benefits.
The monitoring and assessment of small-scale reservoirs have traditionally relied on manual inspections and conventional surveillance methods. These approaches are not only labor-intensive and time-consuming but also prone to human error, which can compromise the timely detection of hazards [
5,
6,
7]. Consequently, there is an increasing demand for automated and accurate techniques that can enhance the safety management of small-scale reservoirs. One promising solution is the application of image processing and machine learning techniques, particularly deep learning, to the detection and characterization of structural defects in reservoirs [
8].
The conventional approaches to reservoir safety monitoring, such as visual inspections, geotechnical instrumentation, and satellite remote sensing, have provided essential baseline data for hazard detection and management. However, these methods exhibit several limitations. Visual inspections and manual monitoring are labor-intensive, limited by human capability, and can be subjective and inconsistent [
9]. Geotechnical instrumentation provides detailed point-based measurements but lacks spatial coverage and can be affected by external environmental factors [
10]. Satellite remote sensing offers wide area coverage but is often hindered by coarse resolution, low revisit frequency, and sensitivity to atmospheric conditions [
11]. In recent years, several studies have explored the potential of machine learning techniques, such as support vector machines (SVM) and random forests, for detecting and classifying structural anomalies in large dams [
12]. These methods, while promising, have not yet been extensively applied to small-scale reservoirs. The specific challenges of small-scale reservoirs, such as their varied construction standards, diverse environmental settings, and unique hydrological conditions, necessitate the development of more tailored and sophisticated monitoring techniques [
13,
14].
Deep learning, particularly convolutional neural networks (CNNs), has emerged as a powerful tool for image-based structural defect detection due to its ability to automatically learn and extract hierarchical features from complex datasets [
15]. CNN-based methods have been successfully applied in various fields, including urban infrastructure monitoring [
16], crack detection in concrete surfaces [
17], and landslide mapping [
18]. These studies demonstrate that deep learning models, especially when combined with high-resolution imagery, can provide detailed and accurate assessments of structural conditions. Semantic segmentation, a subfield of deep learning, has proven particularly effective in detecting and categorizing structural defects in various settings [
19]. Unlike traditional image classification, which assigns a single label to an entire image, semantic segmentation provides pixel-wise classification, enabling precise localization and characterization of defects [
20]. This capability is crucial for identifying and analyzing multiple types of hazards in small-scale reservoirs, where defects may vary significantly in size, shape, and context [
21]. While there have been significant advancements in the application of deep learning for infrastructure monitoring, most studies have focused on large-scale structures or urban environments. There is a noticeable gap in the application of these technologies to small-scale reservoirs, which are characterized by unique hazard patterns that differ significantly in spatial scale and temporal dynamics from those of larger infrastructure [
22]. Furthermore, the integration of deep learning models with multi-source data, such as hydrological, geological, and meteorological data, for comprehensive risk assessment and prediction remains underexplored [
23]. Then, devices equipped with high-resolution visual sensors, such as industrial cameras, underwater robots, and unmanned aerial vehicles (UAVs), are used to collect a large number of reservoir surface images. Feng et al. [
24] proposed a method for classifying reservoir damage based on convolutional neural networks (CNNs) and transfer learning. Huang et al. [
25] improved region-based CNNs to classify and locate multiple reservoir damage regions and achieved accurate results. Zhao et al. [
26] proposed excellent reservoir damage detection methods using the improved You Only Look Once v5 (YOLOv5) and a three-dimensional (3D) photogrammetric reconstruction method. Semantic segmentation networks, which can classify each pixel in an image, have been studied to achieve accurate pixel-level dam crack detection [
27,
28,
29]. Li et al. [
30] proposed a digital twin method for the structural health monitoring of underground pipelines. Underwater structural defects, including cracks and tunnel defects, were detected pixel-wise using DL techniques [
31,
32,
33].
This study aims to bridge this gap by developing a novel semantic segmentation-based framework for small-scale reservoir safety monitoring. The proposed framework utilizes a combination of deep learning techniques and multi-source data fusion to achieve quantitative extraction of key warning indicators, such as cracks, seepage areas, and surface deformation. This approach integrates image data with conventional monitoring datasets to establish a robust model for real-time hazard evaluation and prediction, enhancing early warning capabilities and operational safety [
34]. Our approach involves the development of a deep learning model based on semantic segmentation to identify and categorize various types of structural hazards in small-scale reservoirs. This model leverages multi-source data, including UAV imagery, multispectral and infrared data, and environmental sensor data, to capture the diverse range of hazard indicators present in reservoir environments [
35]. We utilize U-Net architecture, a popular choice in semantic segmentation tasks due to its ability to retain fine spatial details while capturing global contextual information [
36]. To enhance the robustness and accuracy of the model, we integrate domain adaptation techniques to handle the variability in data quality and environmental conditions. Furthermore, we employ data augmentation strategies, such as random cropping, flipping, and brightness adjustment, to increase the diversity of the training data and improve model generalization.
The proposed framework offers several significant contributions to the field of reservoir safety monitoring. First, it provides a comprehensive methodology for the automatic detection and classification of structural hazards in small-scale reservoirs, addressing a critical gap in current research. Second, by integrating multi-source data, our approach enables a more holistic assessment of reservoir safety, capturing a broader range of hazard indicators and providing more accurate predictions of potential failures. Moreover, the development of a real-time hazard evaluation and prediction model has practical implications for improving the operational safety of small-scale reservoirs. The ability to detect and respond to hazards promptly can significantly reduce the risk of failure, minimize damage, and ensure the continued provision of essential water services. This is particularly relevant in regions such as Jiangxi Province, where small-scale reservoirs play a critical role in supporting socio-economic development and maintaining ecological balance. By leveraging semantic segmentation techniques and multi-source data fusion, we aim to provide a more accurate and comprehensive assessment of reservoir safety, enhancing early warning capabilities and contributing to the sustainable management of water resources. This research not only addresses existing gaps in the literature but also offers a practical solution to the challenges faced by small-scale reservoir engineering in China and beyond.
5. Discussion and Application
Based on the image semantic segmentation method framework, we extracted hidden hazard image features and quantified them. The features were as follows:
Find the boundary of the target object in the segmentation result (edge detection); Use the bounding box or fitting ellipse to fit the target object and obtain its length and width.
Use the height and width of the bounding box to represent the length and width of the target object.
Calculate the number of pixels within the target object area (the area of the segmentation region).
Convert pixel area to actual area (given the spatial resolution of the image, i.e., the actual physical size corresponding to each pixel).
Area = number of pixels in the segmented area multiplied by the actual area per pixel.
In the context of transfer learning or model adaptation for the HHSN-25 architecture, the process involves initially training the model on a large, general dataset to learn common features that are transferable across tasks. This is followed by fine-tuning on a specific reservoir engineering dataset to adapt the model to the domain-specific characteristics of hidden hazards in small-scale reservoirs. During this fine-tuning phase, the model selectively updates the weights of certain layers to refine its understanding of unique hazard features, while preserving foundational patterns acquired from the initial training.
To ensure compatibility with reservoir-specific image data, transfer learning also involves adapting feature extraction layers, such as ResNet or ConvNeXt, through incremental training. This helps the model to better capture the localized textures and anomalies associated with different hazards. By leveraging transfer learning, the model can effectively generalize from limited labeled data and reduce training time, as the model starts with a base of pre-learned representations. This approach is particularly useful when training data for certain hazards are scarce, as it enables the model to make informed predictions based on previously learned knowledge from related tasks.
5.1. Integration of Hidden Hazard Feature Extraction Algorithm
The trained image hidden feature extraction algorithm has been integrated into edge devices and carried out demonstration applications. During actual operation, the collected images need to be clear enough to identify small structural changes such as cracks or leaks. The camera equipment needs to be able to capture the target area video clearly within a certain distance, with a distance of less than 50 m. As such, a high-definition (higher than 1080p resolution) camera has been used to ensure rich details in the image. This ensures that the data collected under different environmental conditions is of sufficient quality, so that the model can accurately detect and identify engineering hazards.
The HHSN-25 model can be integrated into existing monitoring systems by running on edge devices for real-time hazard detection or cloud-based systems for large-scale monitoring. The HHSN-25 model is not only versatile in terms of its deployment options but also boasts a robust feature set that enhances its capabilities. When integrated into edge devices, it leverages local processing power to promptly identify potential hazards, ensuring a swift response to any emerging threats. This real-time detection feature is crucial for applications that require immediate action, such as in industrial safety monitoring or disaster prevention systems. On the other hand, when deployed in cloud-based systems, the HHSN-25 model can harness the vast computational resources of the cloud to perform large-scale monitoring across wide geographical areas. This scalability allows it to process and analyze vast amounts of data from multiple sources, providing a comprehensive overview of hazard situations and facilitating decision making at a macro level. Moreover, the HHSN-25 model is designed to be highly adaptable, enabling it to integrate seamlessly with various sensor types and data streams. This adaptability ensures that the model can be tailored to specific use cases and requirements, maximizing its effectiveness in different monitoring scenarios. Furthermore, the model incorporates advanced machine learning algorithms that enable it to continuously learn and improve over time. As it is exposed to more data and hazard situations, the HHSN-25 model becomes increasingly proficient in detecting and classifying hazards, enhancing its overall accuracy and reliability. In summary, the HHSN-25 model offers a versatile and powerful solution for hazard detection and monitoring. Whether deployed on edge devices for real-time detection or in cloud-based systems for large-scale monitoring, the model’s robust feature set, adaptability, and continuous learning capabilities make it an invaluable tool for ensuring safety and security in a wide range of applications.
Through the edge computing architecture, this algorithm realizes real-time monitoring and analysis, and it reduces the total time from data acquisition to hidden danger detection and alarm from the original average of 5 min to 40 s (reduced by about 86%). The high-performance embedded processor NVIDIA Jetson AGX Xavier and software configuration (TensorFlow, PyTorch) are specially designed for edge computing, so that it has efficient processing capacity and low energy consumption and is suitable for real-time data processing and analysis. In addition, the edge computing architecture enables data processing to be completed locally, reducing dependence on the cloud, improving response speed and privacy protection. Comparing the adaptive detection model developed in this project with the mainstream method “Edge AI and Machine Learning”,
Table 4 can be obtained. The results indicate that the differences between the two are mainly reflected in the data processing and resource requirements during model training and inference computation. The deep learning method of the detection model developed in this project does not rely on manually annotated data, and, thus, reduces the need for annotated data when training on edge devices. After the model training is completed and integrated, the inference calculation process is relatively fast, and the computation time is relatively short. In addition, the multimodal data fusion and adaptive deep learning model have improved the anti-interference ability of the research and development model in complex environments, enabling it to dynamically adjust and adapt to sudden changes in the environment, and play an advantage in scenarios where data annotation is difficult. The training process of Edge AI and Machine Learning is usually completed in the cloud. When reasoning on edge devices, due to the use of pre trained models and optimization techniques, the performance in dealing with sudden changes in environments (such as night, rain, and fog) is weak and the adaptability is poor.
Precision and recall vary across hazard types. In this study, the model’s precision and recall were evaluated for specific hidden hazard classes: cracks, collapse, leaching, and seepage. Each hazard type demonstrated slightly different characteristics in terms of these metrics, reflecting the model’s capability to differentiate and accurately identify distinct types of anomalies.
Cracks: The model achieved a high precision of 97% in detecting cracks, indicating that false positives were minimal. The recall for cracks was also robust, at approximately 94%, suggesting strong detection capability, although some cracks might not have been detected under certain conditions, such as in images with varying light or debris covering parts of the crack.
Collapse: For collapse hazards, the precision was around 98%, while recall was slightly lower at 93%. This indicates that the model effectively minimized false alarms, though a few instances of collapse might have been missed due to occlusions or the complex visual nature of collapsed structures.
Leaching: The precision for leaching was noted at 99%, with a recall of 95%. This shows that the model is particularly adept at identifying leaching without mistakenly classifying other anomalies as leaching. The high recall value also demonstrates the model’s ability to consistently recognize leaching occurrences, even in cases where staining or discoloration might present challenges.
Seepage: Seepage detection scored an impressive precision of 98% and recall of 96%. These figures reflect the model’s proficiency in accurately distinguishing seepage hazards, with few false negatives or positives. The characteristics of seepage—such as moisture patterns or soil discoloration—are distinct and were well captured by the model.
5.2. Algorithm Promotion and Practical Application
To sum up, the project not only significantly improves the accuracy and recall rate of hidden danger detection of small reservoir projects but also greatly shortens the detection and response time and significantly improves the safety and operation efficiency of reservoirs by integrating edge computing hardware and software configuration, optimizing response speed and adopting advanced multi-source data fusion and image segmentation adaptive models. At the same time, the image hidden danger feature extraction algorithm has been widely applied in multiple typical reservoirs, including 25 small reservoirs in Jiangxi Province, including Chookeng Reservoir, Jutang Tuanjie Reservoir, Xiashitang Reservoir, Gaokeng Reservoir, Dongzhan Reservoir, Linjiang Zoujia Reservoir, Zhangshu City Daqiao Dam, Yangqiao Tientang Reservoir, Lianhua County Reservoir, Changlan Maqing Dam, Xixi Reservoir, Xiaoshankou Reservoir, Mengtang Reservoir, etc. The application effect of a typical reservoir is shown in
Figure 15,
Figure 16,
Figure 17 and
Figure 18.
On the basis of existing research, further exploration and expansion of the application scope of this project will bring more potential improvement space and innovative value to the safety monitoring and management of water conservancy engineering. Firstly, the existing technological framework can be extended to larger scale water conservancy facilities such as reservoirs, rivers, and gates, especially in areas with complex geographical locations and variable climate conditions. By increasing the adaptability to diverse data sources, the robustness and universality of the model can be improved in different environments and conditions. To further improve system performance, future research can optimize in the following areas:
Introducing deep learning and transfer learning techniques: Combining existing multi-source data fusion and adaptive image segmentation models, introducing deep learning frameworks, especially convolutional neural networks (CNN) and structured data processing methods based on graph neural networks (GNN), to improve the depth and accuracy of feature extraction. Meanwhile, transfer learning techniques can optimize existing models, reduce reliance on large-scale annotated data, and accelerate model deployment and application.
Enhance real-time data processing capability: Although edge computing has significantly improved data processing speed, it can further combine 5G technology and distributed computing methods to enhance the real-time data transmission and processing capability of the system, ensure that a large amount of data can be acquired and processed in time when emergencies occur, and improve the timeliness and accuracy of decision responses.
Developing multimodal data fusion methods: Currently, multi-source data fusion mainly focuses on the integration of visual image data and sensor data. In the future, more types of sensors (such as acoustics, radar, optics, etc.) can be introduced, combined with multimodal data sources such as meteorological data and historical disaster records, to construct a more comprehensive hazard monitoring model. Through multimodal information fusion, potential risk factors in complex environments can be better understood, and the ability to identify and warn of hidden dangers in small reservoirs can be improved.
Improved data management and analysis capabilities: In terms of data management, future research can develop and apply more efficient data storage, retrieval, and analysis tools to meet the storage and analysis needs of large-scale, multi type data. By utilizing data lake and data warehouse technology, data from different sources can be integrated into a unified analysis platform, and more valuable hidden danger information and patterns can be extracted from it with the help of artificial intelligence and machine learning algorithms.
Intelligent management and decision support: Through further algorithm optimization and data mining, this system can not only be used for detection and warning, but also provide intelligent management suggestions and decision support for reservoir managers. For example, by combining historical data and real-time monitoring data, the model can generate dynamic risk assessment reports, provide emergency response plans for different levels of hidden dangers, and help decision-makers manage the safe operation of reservoirs more scientifically.
5.3. Model’s Performance with Different Hyperparameters
Hyperparameters play a crucial role in determining both the performance and stability of deep learning models, including our hazard detection framework. Variations in key hyperparameters, such as learning rate and batch size, can significantly affect the model’s ability to generalize, converge, and remain stable during training.
5.3.1. Learning Rate
The learning rate is a critical hyperparameter that directly influences how quickly the model adapts to new data. A high learning rate can lead to rapid learning but might cause instability, resulting in the model overshooting the optimal parameters, failing to converge, or converging prematurely to a suboptimal solution. Conversely, a low learning rate stabilizes the learning process but might prolong training and lead to overfitting as the model adapts too slowly. In our tests, the learning rate critically impacts the convergence speed and the stability of training. During our experiments, a learning rate of 0.001 was found optimal, balancing accuracy and convergence. A lower learning rate (e.g., 0.0001) resulted in slower convergence and prolonged training, requiring around 50 epochs to reach an accuracy plateau. In contrast, a higher learning rate (e.g., 0.01) accelerated convergence but led to unstable training with oscillations in the loss curve, ultimately reducing accuracy by approximately 3–5% due to overshooting. Empirical results show that maintaining the learning rate around 0.001 provided stable convergence while maximizing accuracy, achieving 95% recall and 98% precision across hazard classes.
5.3.2. Batch Size
The batch size determines how many samples are processed before updating the model’s weights. A larger batch size tends to result in more stable gradient estimates but requires more memory, whereas smaller batch sizes may introduce noise into the gradient descent process, potentially destabilizing the training. In our model, smaller batch sizes introduced more variability, but when set too large, it led to slow updates, impacting the convergence speed. Future improvements may involve dynamic batch sizing strategies, where batch sizes adjust according to the training phase. Batch size influences model generalization and computational efficiency. With a batch size of 32, the model achieved the best balance between stability and performance, reaching 96% accuracy after 30 epochs. Smaller batch sizes (e.g., 16) led to more fluctuation in the loss curve, although final accuracy was comparable, at 95%. Larger batch sizes (e.g., 64) reduced training time but led to slightly worse generalization, reducing accuracy by around 2% on unseen data. The selected batch size of 32 provided optimal stability with a consistent learning rate, achieving high recall and precision while maintaining training efficiency.
5.3.3. Optimizer Choice
Optimizers, such as Adam and SGD (Stochastic Gradient Descent), have different impacts on the model’s performance and stability. Adam, with adaptive learning rates, was found to accelerate convergence and stabilize training early on but was occasionally prone to overfitting. On the other hand, SGD provided smoother convergence but required more careful learning rate scheduling. The choice of optimizer, therefore, directly influenced model stability. Optimizer choice affects how effectively the model navigates the loss landscape. We compared three optimizers: Adam, SGD, and RMSprop. Adam achieved the best performance, with 97% accuracy and stable loss reduction across epochs. SGD required a significantly lower learning rate (0.001), resulting in a slower training process and achieving 94% accuracy after 40 epochs. RMSprop offered a balance, with 95% accuracy but a less stable loss curve than Adam. Based on this analysis, Adam was selected for its ability to maintain stability and high accuracy across epochs.
For future work, several additional hyperparameters could be considered for exploration:
Dropout Rate: Dropout is a regularization technique to prevent overfitting by randomly deactivating neurons during training. Future work could experiment with varying dropout rates across layers, especially under conditions where datasets are limited, to enhance stability and prevent overfitting.
Weight Decay: Adjusting weight decay or L2 regularization could be crucial for maintaining model generalization, particularly in complex models prone to overfitting. Fine-tuning this parameter might improve the stability of the model by penalizing excessively large weight values.
Early Stopping Patience: Early stopping halts training once performance ceases to improve. Fine-tuning the patience parameter, which determines how long to wait for improvement before stopping, could help in achieving a balance between underfitting and overfitting, improving model robustness.
Gradient Clipping: This technique helps to prevent exploding gradients, which can lead to instability, especially in deep models. Investigating different thresholds for gradient clipping may ensure smoother and more stable updates to the model parameters.
5.4. Other Significant Hidden Hazards in Small-Scale Reservoir Engineering
In addition to the hazards already included in the dataset, there are several other potential risks in small-scale reservoir engineering that are critical to address for maintaining safety and operational integrity. These hazards include erosion, sedimentation, and issues related to the foundation and surrounding slopes. Each of these hazards presents unique challenges for detection, and future efforts should focus on developing models and methods to address them effectively.
Erosion: Erosion of the reservoir embankments or surrounding areas can lead to gradual weakening of structural integrity. It is crucial to develop techniques that can detect early signs of erosion, such as soil displacement, changes in vegetation, or subtle shifts in the topography. A multi-temporal image analysis using remote sensing data can help identify these patterns. Machine learning models trained on datasets that include erosion-related features can be effective for early warning.
Sedimentation: Sediment buildup in the reservoir reduces water capacity and can lead to operational issues. Identifying sedimentation patterns requires combining bathymetric surveys with surface monitoring. Using underwater sensors and image analysis can assist in detecting sediment accumulation over time. It is important to create a dataset that captures the characteristics of sedimentation processes in various reservoirs.
Foundation Issues: Problems with the foundation, such as subsidence or differential settlement, are harder to detect because they occur below the surface. Geotechnical monitoring, along with ground-penetrating radar or other subsurface imaging techniques, could be combined with visual monitoring data to track any foundation shifts or settlements. Collecting these data requires collaboration with geological surveys and real-time monitoring through embedded sensors.
Surrounding Slopes: Slope instability can lead to landslides or rockfalls, which can severely impact the reservoir structure. Detecting slope hazards involves monitoring changes in vegetation, rock fractures, or moisture levels. UAVs equipped with high-resolution cameras and LiDAR sensors could be used to monitor slope conditions over time. Deep learning models could be trained to identify early signs of slope failure using time-series image data.
Erosion and sedimentation not only affect the structural safety of reservoirs but also have long-term impacts on water capacity and environmental health. Foundation issues can lead to catastrophic failures if left undetected, making real-time monitoring essential for early intervention. Slope instability poses both immediate and long-term risks to reservoir stability, and proactive detection of landslide-prone areas is key to preventing disaster.
To capture these hidden hazards, a combination of historical data (from past reservoir failures), real-time monitoring (using sensors and UAVs), and synthetic data (generated from simulations) is necessary. Engaging with local authorities and field engineers to gather case studies of erosion, sedimentation, and foundation issues will help enrich the dataset. Datasets should be diverse and include images from various weather conditions, seasons, and environmental contexts to improve model robustness. By expanding the dataset and developing hazard-specific models, future systems can better predict and mitigate these risks, ensuring greater safety and operational efficiency for small-scale reservoirs.
6. Conclusions
This study demonstrates the effectiveness of an advanced image processing framework for detecting and assessing safety hazards in small-scale reservoirs using deep learning techniques. By employing a fully convolutional semantic segmentation method with an encoding–decoding structure, the proposed model effectively utilizes convolutional neural networks (CNNs) to enhance detection accuracy and response efficiency. The pyramid structure, validated as optimal for CNNs, was carefully implemented to balance feature extraction and computational resource usage. Experimental results showed that the best network performance was achieved with a specific channel configuration, highlighting the importance of carefully managing channel dimensions during the decoding stage to preserve hidden features without excessive computational costs.
Furthermore, the HHSN-25 network demonstrated superior performance over existing segmentation methods, such as FCN, SegNet, and Deeplabv3+, with a mean Intersection over Union (mIoU) reaching 87.00%. The study also integrated an improved loss function focusing on the similarity of different samples, which proved to be the most effective in enhancing the model’s performance. The chosen evaluation metrics, particularly the F1 score and mPA index, provided a comprehensive assessment of both hidden and background pixel classification, further validating the superiority of the proposed approach.
The convergence analysis of the loss function also highlighted the optimal training rounds required to achieve a balance between model performance and computational efficiency, avoiding issues of underfitting or overfitting. These findings underscore the potential of the proposed approach to provide accurate and real-time hazard detection, significantly contributing to the safety and sustainability of small-scale reservoirs.
The limitations of the proposed method mainly revolve around the model’s dependence on the quality and variety of training data. As small-scale reservoir data are often scarce, the model’s ability to generalize across different conditions (e.g., weather patterns, sedimentation, or vegetation growth) may be limited. Additionally, the computational complexity of deep learning models can be a constraint when deploying on edge devices in real-time applications. Moreover, the accuracy of the model might be affected in extreme environmental conditions, such as heavy rain, fog, or night-time monitoring, where visibility is drastically reduced.
Overall, the proposed framework successfully integrates advanced neural network architectures and data processing strategies to offer a robust solution for reservoir monitoring, with potential applications in other areas of water resource management and infrastructure safety. Further research can explore the integration of transfer learning techniques to improve model accuracy when faced with unseen conditions or locations where reservoir data are limited. Another promising direction is the development of hybrid models that combine deep learning with physical models of reservoir behavior to enhance prediction accuracy. Additionally, research could be directed at improving the model’s performance under extreme conditions, using enhanced image preprocessing techniques like noise reduction, fog removal, and image enhancement.