1. Introduction
Clouds are condensations of water, consisting of tiny liquid droplets, ice crystals, or a combination of both, suspended in the atmosphere and typically not in contact with the ground [
1]. Clouds play a crucial role in the Earth system. They are not only key components of the global water cycle but also participate in biogeochemical cycles and significantly impact the global radiative energy balance [
2,
3]. Furthermore, the presence and variation of clouds have a substantial influence on the evolution of weather systems and changes in the climate system [
4]. Different clouds have distinct radiative characteristics, which reflect the dynamic and thermal processes in the atmosphere [
5]. Accurate cloud detection and classification are crucial for many research efforts, specifically in the following aspects: firstly, they can improve the effectiveness of numerical model parameterization schemes, thereby enhancing model accuracy; secondly, they can enhance the analysis of precipitation intensity, temperature, and humidity fields; and finally, they contribute to increasing the accuracy of weather forecasts, ensuring flight safety [
6,
7,
8,
9].
Cloud classification research often employs retrieval methods based on the principle that different types of clouds exhibit distinct absorption and reflection characteristics across various wavelength ranges. By analyzing radiative data, meteorological parameters such as cloud-top temperature and cloud optical thickness can be retrieved, allowing cloud types to be categorized according to their physical properties [
10]. The data sources for cloud classification rely on various observation technologies and platforms, with satellite remote sensing being the most common and widely used source, encompassing both active and passive remote sensing methods. Active remote sensing systems emit signals and receive information from the reflection or scattering by clouds, offering high-resolution details of cloud structure and characteristics. This method is well suited for detailed cloud analysis within a focused observational range, though it typically covers a more limited area. In contrast, passive remote sensing utilizes natural light to observe clouds, enabling rapid acquisition of cloud distribution and dynamics over broad areas [
11]. While this approach has a wider observational range, it is constrained by weather conditions and time of day and has limited penetration through thick cloud layers.
Currently, mainstream cloud classification methods fall into two major categories: simple statistical approaches and machine learning-based techniques. Yu et al. [
12] proposed a unit feature space classification method (UFSCM), which identifies six types of clouds based on a feature space constructed from visible light reflectance and infrared brightness temperature. Purbantoro et al. [
13] utilized a split-window algorithm (SWA) and classified clouds into nine types by selecting different thresholds for brightness temperature (BT) and brightness temperature difference (BTD) from satellite data for winter and summer seasons. With the rapid advancement of computing technology, complex machine learning methods have proven to be highly effective for cloud classification. Compared to the aforementioned simple statistical approaches, these methods generally achieve significantly higher classification accuracy. Wohlfarth et al. [
14] trained a support vector machine (SVM) classifier using four channels (color + thermal) from Landsat-8 satellite data, achieving an overall classification accuracy of 95.4% for 13 categories. Wang et al. [
15] combined Himawari-8 multi-band radiation information with CloudSat cloud classification products and applied a random forest (RF) algorithm, training daytime and nighttime cloud classifiers with accuracies of 88.4% and 79.1%, respectively, across a dataset of seven cloud types. Gorooh et al. [
16] proposed the Deep Neural Network Cloud-Type Classification (DeepCTC) model, using 16-channel features from the GOES-16 satellite to classify eight cloud types, achieving an average accuracy of 85%. Jiang et al. [
17] introduced the CLP-CNN based on U-net, incorporating channel and spatial attention mechanisms to classify FY-4A L1 data into nine categories on a per-pixel basis, achieving a classification accuracy of 76.8% compared with Himawari-8 cloud classification products. Guo et al. [
18] developed machine learning-based cloud classification models using data from the Advanced Geostationary Radiation Imager (AGRI) on the Fengyun-4A satellite, along with CPR-CALIOP merged-product classification results for training and validation. The models achieved accuracies of 83.4% for daytime and 79.4% for all-time single-layer cloud-type classification, outperforming Fengyun 4A Level-2 cloud products. Shang et al. [
19] developed the CARE algorithm, which combines a threshold-based method and an extra randomized tree (ERT) model, using Himawari-8 full-disk measurements to achieve all-day cloud detection, with hit rates validated against CALIPSO measurements: daytime cloudy HR 76.95% and clear HR 87.97% and nighttime cloudy HR 73.19% and clear HR 81.77%.
Although the above methods have been successfully applied to cloud classification with commendable results, several unresolved issues remain. Firstly, simple statistical methods typically require prior knowledge from experts and are prone to misclassification in special regions, such as high-latitude desert areas. Additionally, these methods require the development of new algorithms tailored to different satellites. Machine learning methods based on two-dimensional images, on the other hand, generally rely on data from passive remote sensing. Both machine learning and statistical methods often involve manual labeling of cloud types, albeit to different extents. Statistical methods typically require labeled data to validate thresholds or refine empirical rules, whereas machine learning methods rely on extensive labeled datasets for data-driven feature extraction. This dependence on large labeled datasets limits dataset size and hampers model generalization. Consequently, these models tend to achieve high classification accuracy only on specific datasets of two-dimensional cloud images, without the capability for fine-grained, pixel-level cloud classification. Finally, single-pixel-based cloud classification combines data from both active and passive remote sensing and performs automatic cloud labeling during the spatiotemporal matching process, enabling pixel-level classification results. However, this approach disregards the spatial correlation among pixels within a cloud cluster, treating each pixel as an isolated entity. A cloud cluster typically spans multiple pixels within a certain area, and the cloud types of these neighboring pixels are often identical. This spatial correlation is crucial for improving the accuracy and effectiveness of cloud classification tasks.
In this study, we introduce the one-dimensional nested U-Net cloud-classification model (1D-CloudNet), which is designed to address the specific limitations of existing cloud classification methods. Our model aims to (1) mitigate the reliance on expert prior knowledge and the need for satellite-specific algorithm development by utilizing a data-driven approach, (2) overcome the challenges of data scarcity and high annotation costs associated with two-dimensional image-based machine learning methods by leveraging spatiotemporal matching of active and passive remote sensing data, and (3) improve upon single-pixel classification methods by considering the spatial correlations among pixels within cloud clusters. While we acknowledge that these advancements do not completely resolve all limitations of the mentioned methods, they represent significant steps towards enhancing the accuracy and efficiency of cloud classification.
For developing and validating 1D-CloudNet, we utilized data from the passive Himawari-8 and the active CloudSat satellites. Himawari-8 provides two-dimensional (longitude and latitude) images in 16 bands, along with solar and satellite angles, available both day and night (except visible and near-infrared bands at night). CloudSat, due to malfunction, operates only during the day, providing 125 vertical layers of cloud type distribution along its flight track. To fully leverage Himawari-8’s wide observational range and CloudSat’s high vertical resolution, we performed spatiotemporal matching, resulting in one-dimensional images with corresponding cloud types for each pixel.
The introduction of 1D-CloudNet offers a novel perspective on cloud classification. While existing methods have made considerable progress, this model is unique in its focus on one-dimensional and multi-channel image analysis. A key innovation is its ability to exploit spatial correlations inherent in cloud structures, a feature underexplored in previous models. By linking continuous pixels from spatiotemporal matching of active and passive remote sensing data, 1D-CloudNet enhances classification accuracy and extends its operational timeframe to include both day and night. This innovation in leveraging one-dimensional spatial correlations advances cloud classification methodology.
To guide this study, the following research questions are formulated: (1) How does the performance of 1D-CloudNet in cloud classification compare to traditional methods in terms of accuracy? And (2) what is the impact of adjusting input channel combinations, training epochs, and output layers on the cloud classification accuracy of 1D-CloudNet?
The research objectives for the 1D-CloudNet model are to (1) improve overall classification accuracy by at least 3% compared to traditional models, (2) reduce misclassification rates in nighttime data, and (3) effectively utilize both active and passive remote sensing data to enhance the model’s robustness across diverse terrains and weather conditions. Achieving these objectives will contribute to more accurate and reliable cloud classification, vital for climate research and weather forecasting.
Accurate cloud classification is essential for improving weather forecasting and climate modeling precision. In forecasting, better classification can enhance predictions of precipitation, temperature, and wind patterns, impacting sectors like agriculture, aviation, and disaster management. For climate modeling, better cloud distribution understanding can refine cloud feedback mechanisms in global models, improving long-term climate predictions. Furthermore, accurate real-time cloud classification aids decision-making in weather-sensitive industries such as renewable energy, where cloud cover affects solar and wind energy production. The 1D-CloudNet’s improved classification capability will bridge theoretical advancements with practical applications.
To provide a clear structure for this study, the paper is organized as follows:
Section 2 details the methodology, including the design of the 1D-CloudNet, the datasets used, and the preprocessing steps.
Section 3 presents the results of the model evaluation, including accuracy metrics and comparisons with traditional methods.
Section 4 discusses the practical applications of the model, with case studies and long-term analysis. Finally,
Section 5 concludes the paper with a summary of the findings and suggestions for future research.
4. Model Application
4.1. Long-Term Analysis
Since 2015, the Himawari-8 satellite has provided data services, switching to Himawari-9 in 2022. A full-year dataset from 2016 to 2023 has now been collected, containing hourly L1 data. When SOZ in an image exceeds 90°, the image is classified as nighttime; otherwise, it is considered daytime. Cloud classification was conducted using daytime and nighttime models to obtain high spatiotemporal resolution cloud type over Southeast Asia, followed by statistical analysis of cloud type and cloud fraction.
Figure 12 shows the temporal variation in cloud fraction across Southeast Asia. Monthly averages reveal a distinct seasonal pattern, with higher cloud fraction observed during the rainy season from May to October, when precipitation is most concentrated, and lower cloud fraction during the dry season from November to April. Annual averages show long-term trends, with peak cloud fraction occurring in 2017 and 2022 and relatively low values in 2019 and 2023. A linear trend analysis indicates an overall downward trend in total cloud fraction over time, particularly in the long term. Although High Cloud shows an increasing trend in monthly averages, it decreases in annual averages, while other cloud types exhibit a consistent decline across both time scales.
Figure 13 presents the spatial distribution of cloud fraction. All Cloud is concentrated between 10°S and 10°N and near 28°N, particularly around the Maoke Mountains in New Guinea. Peak Middle Cloud fraction corresponds with areas of the highest All Cloud fraction, likely due to the high elevation and persistent snow cover in the Maoke Mountains, which the classifier may mistakenly label as Middle Cloud. High Cloud is primarily distributed between 10°S and 10°N, with Sumatra showing peak High Cloud fraction. Middle Cloud is found around the Maoke Mountains, the surrounding seas near Sumatra, and the northwest of the study area. Low Cloud is concentrated over the mainland regions of northern Southeast Asia, while Ns/DC is mainly distributed over low-latitude oceans.
Figure 14 illustrates the seasonal distribution of All Cloud fraction, showing that the Maoke Mountains consistently experience substantial cloud fraction across all four seasons. In summer, an area with elevated cloud fraction appears in the northern Andaman Sea, while in autumn, regions with greater cloud fraction include northern Sumatra and its surrounding seas. In winter, the area with increased cloud fraction shifts to the Java Sea. Overall, the regions with substantial cloud fraction display a north-south migration pattern across spring, summer, autumn, and winter. The broader distribution of areas with substantial cloud fraction in summer further supports the earlier statistical conclusion that cloud fraction is highest during this season.
Figure 15 shows the seasonal distribution of four cloud types. The enhanced cloud fraction area for High Cloud similarly undergoes a north–south shift. Middle Cloud displays a broader area of elevated cloud fraction in summer and autumn, primarily covering the western part of the study area. In winter, Low Cloud exhibits a significant region of increased cloud fraction in central and western China. Throughout the year, Ns/DC shows regions of increased cloud fraction mainly over the ocean, with peak areas located in the Andaman Sea during summer and autumn.
In summary, analysis of cloud fraction based on Himawari-8/9 satellite data from 2016 to 2023 indicates that cloud fraction is higher during the rainy season from May to October and significantly lower during the dry season from November to April, with an overall declining trend in All Cloud fraction. Additionally, elevated cloud fraction regions shift seasonally along the north–south axis, with the broadest distribution in summer and more concentrated coverage in winter. However, due to the complex terrain and climatic conditions in Southeast Asia, there are classification errors, particularly over high-altitude regions like the Maoke Mountains. These errors may result from interference in cloud classification caused by terrain and snow cover, leading to confusion between mid-level clouds and surface features. For instance, the persistent snow cover in the Maoke Mountains can have similar radiative properties to certain cloud types, making it challenging for the model to distinguish between them. Additionally, the complex topography can cause variations in the observed cloud characteristics, further complicating the classification process. In some cases, the model may misclassify high-altitude snow-covered areas as Middle Cloud, especially when the cloud cover is sparse, or the snow cover is extensive. This highlights the need for incorporating more contextual information, such as terrain data, to improve the model’s accuracy in such regions. Future work will focus on enhancing the model’s ability to recognize and differentiate between high-altitude features and actual cloud formations, thereby reducing misclassification in complex environments.
4.2. Short-Term Analysis
To investigate cloud evolution characteristics in short-term weather events over Southeast Asia, we selected Typhoon Mangkhut in September 2018 as a case study. Typhoon Mangkhut formed on 7 September and gradually dissipated on 17 September.
Figure 16 illustrates the evolution of cloud types associated with Typhoon Mangkhut over Southeast Asia from 11 September to 16 September, recorded at six-hour intervals (04:00, 10:00, 16:00, and 22:00). These time-series images enable us to observe the dynamic changes in the typhoon’s cloud structure, capturing the entire process of formation, movement, landfall, and dissipation, thus allowing an analysis of the spatial distribution and evolution characteristics of the typhoon’s cloud system.
On 11 September, Typhoon Mangkhut’s cloud cluster first entered Southeast Asia, displaying a well-defined structure. The central region was dominated by Ns/DC, transitioning to Middle Cloud and High Cloud toward the outer regions. This cloud distribution reflects the typical structure of a typhoon system, with denser cloud, lower cloud-top temperature, and more intense convective activity concentrated at the center. On 12 September, the typhoon’s cloud cluster expanded further, with increased coverage of Ns/DC, indicating intensified typhoon strength and an expanded precipitation zone. Between 13 and 14 September, the cloud system continued moving westward and made its first landfall in the Philippines. On 15 September, Typhoon Mangkhut crossed the Philippines into the South China Sea. After making landfall along the Chinese coast on 16 September, the typhoon gradually dissipated. The coverage of Ns/DC decreases significantly, while High and Middle Clouds progressively become dominant. This transition marks the weakening phase of the typhoon system, as the typhoon cloud clusters gradually dissipate after landfall and eventually vanish.
In summary, these time-series images reveal the characteristic cloud evolution patterns of Typhoon Mangkhut. During the intensification phase, Ns/DC dominates the core region of the typhoon, with Middle Cloud and High Cloud progressively forming the outer layers. Following landfall, Ns/DC gradually diminishes, transitioning into Middle Cloud and High Cloud as the cloud cluster become increasingly sparse, indicating the dissipation of the typhoon system. These observations provide a clear basis for studying the spatial distribution and evolution of cloud types within a typhoon system, which is crucial for understanding the changes in typhoon-associated heavy precipitation zones and cloud structure. Additionally, the images reveal notable differences in cloud type results between day and night. There are inconsistencies in cloud type distribution over the same area at different times, particularly in the distribution range and intensity of Ns/DC. This day-night inversion inconsistency may be attributed to various factors, such as observational conditions, variations in solar radiation, and the optical properties of the atmosphere. For instance, daytime solar radiation may alter the scattering and absorption characteristics of cloud structures within the atmosphere compared to nighttime, thereby affecting cloud-type identification accuracy. Moreover, diurnal variations in cloud-top temperature and cloud optical thickness could introduce biases in cloud classification results. The existence of these day–night inversion inconsistencies indicates that there is room for improvement in enhancing the robustness and consistency of current models in cloud-type recognition.
5. Conclusions
This study introduces a novel cloud classification model, 1D-CloudNet, developed using 2016 and 2017 data from Himawari-8 and CloudSat satellites. Based on modifications to U-Net++, 1D-CloudNet is specifically designed for one-dimensional multi-channel images, enabling effective use of contextual information from neighboring pixels to enhance cloud classification accuracy. Experimental validation determined that the optimal input data for the daytime model includes all spectral and angular data (VIS, NIR, IR, SAZ, SOZ, SAA, SOA), comprising a total of 24 channels, while the optimal input data for the nighttime model consist of infrared spectral data and satellite angular data (IR, SAZ, SAA), comprising 14 channels. For both models, the Level-1 (L1) output of 1D-CloudNet most closely matched the true cloud type.
A comparison with existing models (RF and DeepCTC) showed that 1D-CloudNet achieved an overall accuracy of 88.19% for daytime classification, representing an improvement of approximately 3%, and 87.40% for nighttime classification, representing an improvement of around 4%. Clear classification performed best, with a precision exceeding 96%, while High Cloud and Ns/DC reached precision levels between 70% and 80%. Middle Cloud and Low Cloud exhibited lower classification performance, with precision between 50% and 60%. Compared with the Himawari-8 L2 product, 1D-CloudNet demonstrated the advantage of all-day applicability and reliable retrieval without classification failures, although it had limitations in underestimating scattered Middle Cloud and Low Cloud.
In practical application, 1D-CloudNet was used to analyze long-term cloud fraction and cloud type variations over Southeast Asia, as well as cloud characteristics in short-term weather events. The long-term analysis of cloud fraction from 2016 to 2023 revealed higher cloud fraction during the rainy season from May to October and significantly lower cloud fraction during the dry season from November to April, with an overall downward trend. Spatially, elevated cloud fraction regions migrated seasonally along the north–south axis, with the broadest distribution in summer and more concentrated coverage in winter. For short-term analysis, Typhoon Mangkhut in 2018 was examined, revealing structural features of the typhoon cloud cluster: the innermost ring consisted of Ns/DC, surrounded by Middle Cloud in the middle ring and High Cloud in the outer ring, with cloud optical thickness decreasing outward. After landfall, Ns/DC gradually transitioned to High Cloud and Middle Cloud and eventually dissipated.
In summary, the key findings are as follows: (1) The development of 1D-CloudNet significantly improves cloud classification accuracy using optimized inputs for daytime and nighttime. (2) The model demonstrates competitive performance compared to existing methods, particularly for Middle and Low Cloud. (3) Practical applications reveal insights into long-term cloud fraction trends and the structural evolution of cloud clusters during typhoons.
However, 1D-CloudNet demonstrated certain limitations in practical applications. First, high-altitude snow cover was always misclassified as Middle Cloud. Second, there were notable differences between the daytime and nighttime model results in certain areas, indicating the need for further improvements in the model’s robustness and consistency in cloud type recognition.
Future work will focus on optimizing 1D-CloudNet in the following ways: (1) incorporating terrain data or enhancing the model’s ability to recognize high-altitude features to improve adaptability in complex environments and reduce misclassification; (2) exploring more consistent feature fusion methods or adjusting the model architecture to increase consistency between daytime and nighttime classification results; (3) expanding the training dataset, especially by adding Middle Cloud and Low Cloud samples, or introducing dedicated modules for Middle Cloud and Low Cloud recognition to improve classification performance for these cloud types; (4) extending the dataset to encompass diverse regions and climate conditions, incorporating higher-resolution observational data and integrating data from other satellites that provide nighttime cloud classification labels to enhance the model’s generalization capability. With these improvements, we anticipate that 1D-CloudNet will provide more reliable cloud classification results for regional climate studies and extreme weather monitoring and can be further extended for global cloud classification and climate research.