Next Article in Journal
Investigation of Turbulent Dissipation Rate Profiles from Two Radar Wind Profilers at Plateau and Plain Stations in the North China Plain
Next Article in Special Issue
Evaluation of HY-2C and CFOSAT Satellite Retrieval Offshore Wind Energy Using Weather Research and Forecasting (WRF) Simulations
Previous Article in Journal
Spatial and Temporal Variation in Vegetation Cover and Its Response to Topography in the Selinco Region of the Qinghai-Tibet Plateau
Previous Article in Special Issue
The Sensitivity of Large Eddy Simulations to Grid Resolution in Tropical Cyclone High Wind Area Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SeaMAE: Masked Pre-Training with Meteorological Satellite Imagery for Sea Fog Detection

1
Artificial Intelligence School, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
National Meteorological Center, China Meteorological Administration, Beijing 100081, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2023, 15(16), 4102; https://doi.org/10.3390/rs15164102
Submission received: 29 June 2023 / Revised: 26 July 2023 / Accepted: 17 August 2023 / Published: 21 August 2023
(This article belongs to the Special Issue Remote Sensing and Parameterization of Air-Sea Interaction)

Abstract

:
Sea fog detection (SFD) presents a significant challenge in the field of intelligent Earth observation, particularly in analyzing meteorological satellite imagery. Akin to various vision tasks, ImageNet pre-training is commonly used for pre-training SFD. However, in the context of multi-spectral meteorological satellite imagery, the initial step of deep learning has received limited attention. Recently, pre-training with Very High-Resolution (VHR) satellite imagery has gained increased popularity in remote-sensing vision tasks, showing the potential to replace ImageNet pre-training. However, it is worth noting that the meteorological satellite imagery applied in SFD, despite being an application of computer vision in remote sensing, differs greatly from VHR satellite imagery. To address the limitation of pre-training for SFD, this paper introduces a novel deep-learning paradigm to the meteorological domain driven by Masked Image Modeling (MIM). Our research reveals two key insights: (1) Pre-training with meteorological satellite imagery yields superior SFD performance compared to pre-training with nature imagery and VHR satellite imagery. (2) Incorporating the architectural characteristics of SFD models into a vanilla masked autoencoder (MAE) can augment the effectiveness of meteorological pre-training. To facilitate this research, we curate a pre-training dataset comprising 514,655 temporal multi-spectral meteorological satellite images, covering the Bohai Sea and Yellow Sea regions, which have the most sea fog occurrence. The longitude ranges from 115.00E to 128.75E, and the latitude ranges from 27.60N to 41.35N. Moreover, we introduce SeaMAE, a novel MAE that utilizes a Vision Transformer as the encoder and a convolutional hierarchical decoder, to learn meteorological representations. SeaMAE is pre-trained on this dataset and fine-tuned for SFD, resulting in state-of-the-art performance. For instance, using the ViT-Base as the backbone, SeaMAE pre-training which achieves 64.18 % surpasses from-scratch learning, natural imagery pre-training, and VRH satellite imagery pre-training by 5.53 % , 2.49 % , and 2.21 % , respectively, in terms of Intersection over Union of SFD.

Graphical Abstract

1. Introduction

The precise detection of sea fog (SFD) using meteorological satellite imagery holds substantial implications for a wide range of marine activities. Meteorological bureaus continuously monitor their jurisdictional ocean areas to forecast weather, manage fishing and shipping, and relieve meteorological disasters (Figure 1a) [1,2]. In this study, our focus lies on the Bohai Sea and Yellow Sea regions (Figure 2a), utilizing deep-learning techniques to tackle this problem.
Through an exhaustive review of previous studies on SFD [3,4,5,6,7,8,9,10,11,12,13,14], we observe that the question of pre-training SFD networks remains an open challenge. Some papers rely on ImageNet pre-training weights [3,4,6,8,10,14], while some others even opt to train SFD models from scratch on fine-tuning data [5,7,8,9,11,12,13]. Lately, with the advancements in self-supervised learning [15,16,17,18], which is independent of expensive semantically labeled data, customizing pre-trained models on large-scale raw data has become a prevailing practice, especially in some specialized domains such as medical imagery [19,20] and VHR satellite imagery [21,22]. The existing VHR satellite imagery pre-training model may be deemed as an alternative suitable for SFD since the data in pre-training and fine-tuning are both captured by satellites from aerial angles. However, vast distinctions between the VHR and meteorological satellite imagery, such as the number and types of bands, image resolution, and observation targets, necessitate the exploration of meteorological pre-training for SFD.
This paper leverages the rich data obtained from the Himawari-8 satellite as the foundation for our investigation [23]. Indeed, the Himawari-8 satellite has emerged as a crucial contributor to the rapid progress in SFD. It offers the capability to capture up to 16 Advanced Himawari Imager (AHI) observation bands (Figure 2b) and provides substantial temporal publicly available data (Figure 3c). Hence, we curate a dataset comprising 514,655 samples as the pre-training data. To better differentiate the samples of our meteorological dataset from the counterparts, Figure 3a,b exhibits natural imagery of ImageNet [24] and VHR satellite imagery of Million-AID [22]. Based on these comparisons, it becomes evident that there exist significant domain gaps between meteorological satellite imagery versus both natural imagery and VHR satellite imagery. It is justifiable to argue that pre-training SFD models using meteorological satellite imagery can result in the acquisition of more valuable meteorological representations.
Besides the large-scale meteorological satellite imagery, another pivotal aspect of successful meteorological pre-training lies in the adoption of learning methodologies. Our approach builds upon Masked Image Modeling (MIM), as Masked Autoencoders (MAE) [15], a canonical method of MIM, have been empowering various realms through competently pre-trained Vision Transformers [25]. Given the absence of SFD networks incorporating a Vision Transformer as the backbone, we assert that employing MAE for pre-training can simultaneously fill this gap, making contributions not only to pre-training SFD networks but also to fine-tuning SFD networks.
Furthermore, we tweak the architecture of MAE by replacing its plain decoder with a convolutional hierarchical decoder [26,27]. The motivation stems from the performance of SFD improving a lot when the plain decoder in the SFD network is substituted by a hierarchical decoder [6,10]. The plain decoder indicates the upsampling is intensive [15,28], but the hierarchical decoder indicates the upsampling is progressive [29,30,31], as shown in Figure 1b,c. Regarding meteorological satellite imagery, the main objects it covers, such as clouds and fog, possess diverse and complex forms, along with intricate texture features and fragile edge information. Therefore, iterative upsampling operations are necessary for processing the latent representations to progressively restore spatial dimensions. Such an architectural idea can be defined as an inductive bias highly associated with meteorological satellite imagery, irrespective of the task being SFD or MIM. Due to the introduction of this inductive bias of the SFD network, we term our tweaked MAE as SeaMAE.
To be brief and clear, we enumerate the contributions of this work:
  • This paper proposes meteorological pre-training for sea fog detection (SFD). To this end, we collect 514,655 Himawari-8 satellite multi-channel images and use MAE to learn representations from them. On SFD, this pre-training paradigm outperforms from-scratch learning, ImageNet pre-training, and VHR satellite imagery pre-training. The power of available large-scale meteorological satellite raw data is unparalleledly utilized.
  • We investigate a decoder architecture tailored for meteorological pre-training, which results in a novel variant of MAE, SeaMAE. Specifically, an off-shelf decoder effective in SFD is utilized to model masked patches in pre-training. Meteorological pre-training on the proposed SeaMAE facilitates additional performance gains for SFD. To our knowledge, this paper first pioneers the application of Vision Transformers for SFD in this community.
  • SeaMAE intrinsically introduces an architecturally end-to-end pre-training paradigm, wherein the decoders in both pre-training and fine-tuning share the same architecture, revolutionizing the previous routine that only the encoder of SFD is pre-trained. We manifest that the pre-trained decoder performs better than the from-scratch learning decoder on fine-tuning data. The extension of pre-trained components in the SFD network shows great promise.
  • Finally, training SFD networks typically involves using either all bands or (3, 4, 14) bands. Generally, the former performs better than the latter. Our proposed learning paradigm can adapt to both training settings, and make the latter performance on par with the former because our proposed learning paradigm has learned representations from all bands during the pre-training.

2. Materials and Methods

This section contains three parts. The first part introduces the pre-training dataset as well as the collection thereof, and the fine-tuning dataset on which every learning paradigm will be verified. The second part elaborates on our new MAE with a decoder architecturally consistent with that for sea fog detection (SFD). In the last part, we comprehensively describe the pipeline of the proposed meteorological learning paradigm based on SeaMAE for SFD.

2.1. Himawari-8 Meteorological Satellite Imagery

Himawari-8 is a geostationary weather satellite operated by the Japan Meteorological Agency (JMA) [23]. It was launched in 2014 and is positioned over the western Pacific Ocean, providing continuous observation of the Asia-Pacific region. Himawari-8 captures high-resolution imagery and monitors weather patterns, cloud cover, and other atmospheric states. It has been instrumental in enhancing weather forecasting and disaster management efforts in the region. We use its meteorological satellite imagery to construct the pre-training dataset for self-supervised learning, and the fine-tuning dataset for SFD.

2.1.1. Pre-Training Dataset

The meteorological satellite imagery in the area of Bohai Sea and Yellow Sea (115.00E to 128.75E, 27.60N to 41.35N) from the year 2018 to 2021 serves as our database. To accomplish the meteorological pre-training, we collect all 2018 images within the study region, regardless of whether these images capture any sea fog activity, and crop them with a spatial size of 256 × 256. As a result, the pre-training dataset involves 514,655 samples. Despite the meteorological dataset not being dominant in terms of quantity compared to ImageNet (1.43 million) or Million-AID (1.1 million), it possesses 16 channels and every band has special meteorological, cloud, and marine information, as reported in Table 1 and visualized in Figure 2, which can provide the SFD model more meteorological representations in the ocean region. Furthermore, its coverage area aligns precisely with the scope of SFD in that, generally speaking, SFD is also a geostationary interpretation task on satellite imagery, distinguishing it from numerous other satellite imagery pre-training datasets.

2.1.2. Fine-Tuning Dataset

The fine-tuning dataset comprises a training set and a validation set. The training set includes images from the period of 2019 to 2020, while the validation set consists of images from 2021. Unlike the pre-training dataset that involves all samples of the entire year, fine-tuning focuses merely on the images that capture sea fog.
We organize a group with meteorological expertise in recognizing sea fog from the meteorological satellite imagery, employing them to annotate the sea fog of training and validation set in pixels. Consequently, a pixel-level annotated SFD dataset is contributed to the community by us, including 1128 training image-label pairs and 382 test image-label pairs involving 108 sea fog events, both of which are with spatial sizes of (1024, 1024) covering the ocean region with the longitude from 115.00E to 128.75E as well as the latitude from 27.60N to 41.35N. We believe this dataset will be of great applicability for researchers in this field due to the delicate annotation, sufficient samples, and various sea fog activities. Therefore, another article will be organized to provide a deeper introduction and a wider analysis of this SFD dataset, and the dataset will also be publicly available soon. Before that, we report some state-of-the-art (SOTA) methods’ performance on this fine-tuning dataset to quantify its challenge in advance, as displayed in Table 2.

2.2. SeaMAE

Similar to the majority of Masked Image Modeling (MIM) methods, our approach utilizes the masked autoencoder (MAE) to learn visual representations [15], as illustrated in Figure 4. In line with a standard MAE, the input image in our SeaMAE model is divided into non-overlapping patches, with 75% of the patches being randomly masked. The encoder, which is a plain Vision Transformer, operates solely on the unmasked patches and produces latent representations. The decoder takes these latent representations along with learnable mask tokens as inputs and reconstructs the masked patches at the pixel level.

2.2.1. SFD Driven by ViT

Adopting MAE to pre-train means that Vision Transformer is taken as the backbone of the network for sea fog detection. Since there have been never previous works probing this topic, we must first introduce a simple method to utilize Vision Transformer, particularly ViT [25]. Now that the hourglass architecture, or so-called U-shape architecture, has proved effective in SFD [26,31], the key issue is how to extract the feature-map hierarchy from ViT in that the U-shape network transfers the encoder’s feature to the decoder in a stage-to-stage fashion by building skip connections. Unlike hierarchical networks [32] which can naturally generate feature hierarchy, a common manner to handle the plain ViT is separating the ViT into four stages regularly [33], hooking the last feature-map of each stage, and rescaling them to multiple levels.
We propose another easier method to deal with this. Specifically, only the last feature map is used to generate all-level feature maps because it has the strongest representations. By denoting F as the feature-map output by ViT, the generation of the feature-map corresponding n-th stage on the top of the decoder can be formulated by:
F n = B i c u b i c ( F , r a t i o = 2 3 n ) , n [ 1 , 4 ] ,
where B i c u b i c is a function acting on F through bicubic interpolation with a re-sampling r a t i o . Generally, ViT’s output feature map has an output stride of 16 compared to the input 2D image, hence the values of r a t i o are 0.5, 1, 2, and 4. Given such a group of outputs, most decoders for sea fog detection can be used without any bells and whistles.

2.2.2. Hierarchical Decoder

As aforementioned in Section 1, the downstream task of SFD motivates the utilization of a hierarchical decoder instead of a plain decoder in MIM pre-training. In terms of effortlessness, we select LinkNet [26], a widely favored hierarchical decoder in SFD [3,6], to serve as the decoder during both the pre-training and fine-tuning stages. The decoder has 4 blocks and the n-th decoder block takes its corresponding feature F k of the feature pyramid as input. The structure of one decoder block is visualized in Figure 5c. It upsamples F n and outputs E n which is added to F n 1 , and the summation is input to the ( n 1 ) -th decoder block.

2.2.3. Skip Connection for Masked Input

The skip connection plays a crucial role in the U-shape network, connecting the backbone and the decoder. Although implementing the skip connection is straightforward during fine-tuning, it becomes more challenging in MIM pre-training, where the patch sequence in the backbone is masked. For the masked patch sequence, directly using interpolation to upsample it may associate with potential risks due to the considerable information loss. We hence perform interpolation on the concatenation of masked patches and mask tokens, as shown in Figure 5b. The operation of interpolation is identical to Equation (1).

2.3. Learning Paradigm

Like the conventional “first pre-training then fine-tuning” learning paradigm illustrated in Figure 4, our resulting learning paradigm also contains two steps, as shown in Figure 5a. However, different from the steps of the previous learning paradigm shown in Figure 4 which are conducted on two different architectures with imagery from different domains, our method’s two steps depend on the same architecture except for the output layer. The first step is pre-training when SeaMAE learns representations on 514,655 meteorological satellite images. By denoting the output of hidden layers as F h R H W d i m , the output layer can be defined as W r e c R d i m 3 , and the operation of the output reconstruction layer can be formulated by
O r e c R H W 3 = F h × W r e c .
The objective function is Mean Square Error (MSE):
M S E = 1 n i n ( y i y i ^ ) 2 ,
where n is the number of masked patch tokens, y i R p p 3 is i-th masked patch token of raw image, and y i ^ R p p 3 is i-th reconstructed counterpart (p denotes the patch size) of O r e c .
In the second fine-tuning step, the SFD model driven by ViT is trained on labeled meteorological satellite imagery. For F h R H W d i m , the output layer can be defined as W s f d R d i m 1 , and the operation of the output SFD layer can be formulated by
O s f d R H W 1 = F h × W s f d .
The objective function is binary cross-entropy (BCE) loss with sigmoid activating logits:
B C E = 1 n i n ( t i · l o g ( o i ) + ( 1 t i ) · l o g ( 1 o i ) ) ,
where n is number of pixels in the image, t { 0 , 1 } represents the background or sea fog, and o is the sigmoid activating logits.
In summary, the pre-training is indeed architecturally close to the fine-tuning, as shown in Figure 5a. The distinctions between them only include mask tokens which are indispensable in pre-training but redundant in fine-tuning and the output layer whose output channel numbers are 3 or 1 in pre-training and fine-tuning as evident in Equation (2) and Equation (4), respectively. Therefore, two knowledge-transferring ways, i.e., loading pre-training weights, are proposed to initialize the parameters of the SFD network. (1) Only the knowledge of the SeaMAE encoder is transferred. (2) The entire knowledge of SeaMAE is transferred except for the mask tokens and output layer.

3. Results

In this section, we primarily validate the following hypotheses: (1) Under the standard MAE architecture, pre-training with meteorological satellite imagery yields better representations for sea fog detection (SFD) compared to pre-training with natural imagery or Very High-Resolution (VHR) satellite imagery. (2) The decoder of the pre-training MAE using meteorological satellite imagery should exhibit architectural characteristics that align with the SFD network. (3) In the context of sea fog detection, an architecturally end-to-end transfer of pre-trained knowledge surpasses the knowledge transfer limited to the backbone network alone. (4) Our proposed approach, SeaMAE, enjoys several desirable properties, such as data scalability, model-capacity scalability, fast pre-training convergence, and fine-tuning band adaptability. (5) Such architectural improvements result in a novel SeaMAE that performs better than existing state-of-the-art satellite imagery pre-training methods in terms of pre-training for SFD.

3.1. Implementation and Metric

Our hardware environment is a server with 4 NVIDIA3090 GPUS. We use PyTorch as the deep-learning framework to conduct all experiments. For the pre-training setting, we follow the learning protocols of the standard MAE [15]. However, towards the fast tempo of attaining results, all our pre-training experiments set learning epoch 300 instead of 800. For fine-tuning settings, we follow the learning protocols of [6]. In the training stage of fine-tuning, crops of 512 × 512 are input to the network, and in the evaluation stage, the whole image with a size of 1024 × 1024 is input to the network. The performance of detecting sea fog is evaluated through two metrics, sea fog Intersection over Union (IoU) and sea fog Accuracy (Acc.). The sea fog IoU can be calculated as:
I o U S F = T P S F T P S F + F N S F + F P S F
where S F denotes sea fog, and T P , F N , F P are True Positive, False Negative, and False Positive of detecting sea fog, respectively.

3.2. MAE Pre-Training on Meteorological Satellite Imagery

Table 3 compares pre-training on natural imagery—ImageNet [24]—and VHR satellite imagery—Million-AID [22]—and our collected Himawari-8 dataset. The experiment is conducted on three ViT backbones with different model capacities, ViT-Tiny (Ti), ViT-Small (S), and ViT-Base (B) [25]. First, Table 3 showcases the performance of from-scratch UNet, also displayed in Table 2, to define a baseline. For any variant of ViT, the performance of SFD with pre-trained models on any dataset outperforms the performance of from-scratch SFD by large margins. Except for ViT-Ti, pre-training with Million-AID results in superiority to the ImageNet pre-training model on both ViT-S and ViT-B. On all three backbones, the meteorological pre-training outperforms natural imagery pre-training and VHR satellite imagery pre-training by large margins.

3.3. SeaMAE Pre-Training on Meteorological Satellite Imagery

Table 4 analyzes the introduction of SFD’s architectural characteristics to the meteorological MIM pre-training, based on SeaMAE instead of MAE as the pre-training model. Also, the effectiveness is evaluated on three backbones. According to Table 4, the experimental results are divided into three groups based on different backbones, with the first two results in each group directly comparing the fine-tuning performance of SeaMAE and MAE on meteorological image pre-training. It can be observed that SeaMAE consistently outperforms MAE. As described in Section 2.2, the introduced architectural characteristics are hierarchical decoder and skip connection. Considering that the hierarchical decoder involves more tensor operations compared to the skip connection, Table 4 further probes the impact of retaining only the hierarchical decoder, i.e., removing the skip connection, in the SeaMAE on the fine-tuning performance. It is evident that the skip connection in the U-shape network under MIM learning, where the network takes masked images as input, cannot be overlooked.

3.4. Ablation Study

3.4.1. End-to-End Pre-Training

Integrating architectural ideas from the SFD model into MAE brings consistency between the pre-training architecture with fine-tuning architecture, which allows the whole knowledge learned by pre-training, i.e., pre-training weights, can be transferred to almost every corner of the fine-tuning network. In the SeaMAE experiments of Table 4, the fine-tuning results are obtained with such an end-to-end knowledge transfer by default. Therefore, Table 5 compares using only the encoder pre-trained weights with both encoder and decoder pre-trained weights, suggesting that the SFD’s decoder initialization with pre-trained weights is a promising pre-training methodology.

3.4.2. Data Scalability and Pre-Training Time

The standard MAE learns representations on ImageNet involving 1.43 million images. In contrast, our meteorological pre-training dataset contains 0.51 million images. Indeed, the meteorological MIM pre-training with one third of ImageNet’s sample numbers leads to better SFD performance. We are interested in (1) how much performance drops if under some meteorological scenario where available data are insufficient, and (2) how much performance increases if under some meteorological scenario where more large-scale data are available.
For the first question, we reduce the pre-training dataset to 25% (0.125 m), 50% (0.25 m), and 70% (0.375 m) of total samples, respectively. Figure 6a draws the fine-tuning results of ViT at various sizes under four pre-training data scales and plots the trend curve. According to the trend curve, it can be observed that the fine-tuning performance is linearly correlated with the pre-training data scale.
For the second question, although it is possible to collect data before 2018 to enlarge the pre-training dataset, the large time span may result in a significant inner variance of representations in pre-training, which may impair performance. Thus, we adopt a robust alternative trick to simulate a larger-scale pre-training dataset. The epoch number is extended from 300 to 400, 600, and 800, approximately simulating a scaling factor of 1.33, 2, and 2.66 times the pre-training dataset. Figure 6a demonstrates this trend, showing that increasing the training epochs leads to better fine-tuning performance.
Based on these analyses, further increasing the data scale on SeaMAE is expected to yield performance gains in fog detection. Additionally, Figure 6b also examines the performance with reduced pre-training epochs. By comparing the results in Figure 6a with those in Figure 6b where the epochs are less than 300, we find that while varying the pre-training epochs can simulate data scaling, shorter learning schedule on a larger dataset induces higher performance, which also reflects the fast convergence of SeaMAE.

3.4.3. Band Number of Input

The images captured by the Himawari-8 satellite are multi-channel images with 16 channels. In the conducted experiments both in pre-training and fine-tuning, all channels are input to the network. However, for the annotation of the fine-tuning dataset, we only labeled the fog based on the synthesized images from the three most prominent channels: 3, 4, and 14. Some previous literature has also focused on SFD research only using these three channels, despite that fine-tuning on all bands performs better than three channels [6,10]. To reduce the overhead of processing multi-spectral data, Table 6 explores two learning strategies where only three channels are used for fine-tuning. The first strategy uses only a combination of three channels, namely (3, 4, 14), for both pre-training and fine-tuning. The first line in the second group of results of Table 6 shows this way makes the performance degrades by 0.66. The second is using all bands to pre-train, and three bands to fine-tune. With this strategy, the input channel numbers of the input layer, i.e., patch embedding layer, in pre-training and fine-tuning are different, 16 and 3. So three methods are introduced, R a n d o m , I n d e x , and R e s i z e , to adjust the patch embedding to transfer the knowledge of the input layer. R a n d o m means the input layer in fine-tuning loads no knowledge, whose parameters are randomly initialized. I n d e x means the input layer only loads the parameters of (3, 4, 14) channel. Denoting the weights of patch embedding layer from pre-training as W p t R ( p s × p s × 16 × d i m ) , where p t is short for pre-training, p s is short for patch size, and d i m is short for dimension of ViT, I n d e x makes the patch embedding layer in fine-tuning just load W p t [ : , : , [ 3 , 4 , 14 ] , : ] . R e s i z e means W p t is resized to W f t R ( p s × p s × 3 × d i m ) , which acts on the input channel through bicubic interpolation. Table 6 compares these three methods and shows that R a n d o m and I n d e x both impair the performance apparently but R e s i z e only causes an acceptable performance drop 0.22 I o U . This very slight performance drop indeed improves the performance of SFD using only (3, 4, 14) bands to fine-tune, implying our pre-training approach is well adaptive in various band number settings.

3.5. Comparison to Other Satellite Imagery Pre-Training Methods

SatMAE [21] and RingMo [22] are two outstanding pre-training methods, both of which utilize MIM and learn representations on large-scale satellite imagery. What they employed as pre-training network architecture are improved versions of MAE. SatMAE introduces a temporal encoding for temporal imagery and a new masking strategy for input to MAE. RingMo also proposes a new masking strategy and a lightweight decoder with L1 loss for reconstruction. Table 7 shows the SFD performance using SatMAE and RingMo to learn representations of meteorological imagery and as the pre-training model. On backbone with different model capacities, SatMAE and RingMo show varying performance. Therefore, we calculate the performance difference between SeaMAE and the best-performing method among them, as shown in Table 5 for each group of results. SeaMAE outperforms the current state-of-the-art satellite imagery pre-training methods, yielding the best results in pre-training with meteorological satellite imagery for SFD.

3.6. Qualitative Results of Sea Fog Detection

Figure 7 compares SFD qualitative results of Himawari-8 pre-training SeaMAE with ImageNet pre-training MAE. In terms of friendly visualization, the group of bands (3, 4, 14) is used to show the image. Looking into their SFD results, we find that ours presents more consistent annotations, more complete sea fog objects, and more precise boundaries between fog and background. For example, in the 3rd row of the first column, ImageNet pre-training one outputs two holes in the central area of sea fog, and mistakes a cluster of clouds as sea fog, but our result is very close to the ground truth with good prediction consistency.
Figure 7 also demonstrates that in SFD, the most challenging objects to differentiate from sea fog are various types of clouds, which can be categorized as high clouds and low clouds based on their altitude. Analyzing the visualization from bands (3, 4, 14) of meteorological satellite imagery, sea fog areas appear mostly smooth in texture, continuous in color, and dense in structure. In contrast, most low cloud areas show sparse structures, rough textures, and diverse color tones. However, since low cloud areas can also be extensive during sea fog occurrence, there are some sub-regions of low cloud where they resemble sea fog and exhibit similar representations. Experienced meteorologists can distinguish between them based on their rich experience. Nevertheless, for deep-learning models, it is still prone to making errors, especially False Positive predictions. For instance, in the 1st row and 2nd row of the 1st column, SeaMAE pre-training model shows more accurate predictions compared to ImageNet. Another challenge arises from the presence of high clouds frequently over a large area. High clouds are much sparser and rougher, and sometimes appear in a scattered distribution, unlike sea fog and low clouds that tend to cluster together. High clouds often present in a thin state and may overlay the sea fog, resulting in varying representations that can be easily overlooked by SFD models, leading to False Negative predictions. For example, in the 4th row of the 1st column and the 3rd row of the 2nd column, SeaMAE accurately detects sea fog even in these overlaid circumstances, unlike ImageNet’s incomplete predictions.

3.7. Visualization

Qualitative Results of MIM

Figure 8 shows some reconstruction results. The network is pre-trained SeaMAE with ViT-B as the backbone, and we use the validation data in fine-tuning as input to evaluate its reconstruction performance qualitatively. Due to using all bands are not friendly to visualization, we just select the (3, 4, 14) band group to showcase. The objects in meteorological satellite imagery mainly are cloud and fog. SeaMAE succeeds in the reconstruction of the shape of main objects, even the very tiny and hazy ones. For detailed textures, SeaMAE can also depict them ambiguously but not that vividly because the cloud or fog has very sophisticated forms. In general, the pre-trained SeaMAE model demonstrates the ability to model images that capture representations of clouds and fogs, including their shape, size, region, and quantity, utilizing only a small portion of remaining patch cues.

4. Discussions

The content of this discussion includes aspects: (1) Explanation of Results. (2) Validation over 2021 sea fog events. (3) Literature review and comparison to ours. (4) Future directions.

4.1. Explanation of Results

The initial step in applying deep-learning algorithms is to select a pre-training model and load its weights as the initialization of the deep network. Appropriate pre-training methods can lead to performance improvements and more stable optimization. Therefore, when using deep networks to interpret meteorological satellite imagery for sea fog detection (SFD), it is essential to utilize a pre-training model. Our results confirm that applying the most commonly used ImageNet pre-training model yields better results than training the SFD from scratch, i.e., with random parameter initialization.
However, ideally, the distribution of knowledge learned during pre-training should align with the knowledge that the deep-learning model needs to learn during the SFD learning phase (hereinafter referred to as fine-tuning). For instance, recent studies have shown that when interpreting very high-resolution (VHR) satellite imagery, pre-training the model with VHR satellite imagery (using self-supervised methods) outperforms pre-training on natural images (e.g., the ImageNet dataset), leading to better interpretation performance [21,22]. As both VHR satellite imagery and the Himawari-8 meteorological satellite imagery learned by the SFD model have some similarities, such as (1) bird’s-eye camera angle and (2) densely arranged small objects, we compared the performance of the SFD model pre-trained on both natural imagery and VHR satellite imagery, and the results showed that the latter slightly outperformed the former, which meant closing the distribution gap between the pre-training data and the fine-tuning data is promising.
We continue to make efforts to bring the pre-training data representation closer to the fine-tuning data because there are still big gaps between VHR satellite imagery and meteorological satellite imagery, such as (1) channel numbers and types, (2) stationary background in SFD and non-stationary background in VHR satellite imagery, and (3) disparity of image resolution. To achieve this, we collected a large-scale pre-training dataset consisting of images captured by the Himawari-8 meteorological satellite with ample meteorologically marine representations, covering the same marine region and spectrum as the fine-tuning data for SFD. The meteorological pre-training approach largely outperformed the two previously mentioned pre-training methods and utilized only one third of their magnitudes. In the context of deep learning for SFD, our results revealed, for the first time, the importance of pre-training and pre-training data representation, indicating that pre-training the deep-learning model with meteorological satellite imagery can lead to marked SFD performance gains.
On the other hand, besides the data for pre-training, we delve into the network architecture for pre-training. By effectively leveraging Masked Image Modeling (MIM), masked autoencoders (MAE) [15] have emerged as state-of-the-art pre-training architecture. When comparing it with our designed SFD network architecture, it becomes evident that apart from the output layer, which differs due to the different prediction tasks (MIM and SFD), the main distinction lies in the decoder structure. Therefore, we hypothesized that the decoder structure of the SFD network (including hierarchical decoders and skip connections) may be inherently equipped with suitable inductive biases for processing meteorological satellite imagery (as introduced in Section 1), consequently, during the meteorological pre-training, if the decoder of MAE is replaced with the SFD network’s decoder, it could result in further improvements in SFD. The results validated this hypothesis, as incorporating the architectural characteristics of the SFD network’s decoder into the pre-training architecture MAE, be it hierarchical decoders or skip connections, further augmented SFD performance. In addition, the existing state-of-the-art endeavors on pre-training architecture for satellite imagery were compared to ours. The results clarify that for SFD with meteorological satellite imagery, our architecture is the best for pre-training.
Furthermore, we uncovered another advantage of such architectural improvement, which is the complete utilization of the pre-training model. Typically, due to the differences (learning targets) between pre-training and fine-tuning, the decoder structures of the pre-training and fine-tuning models also differ. As a result, the utilization of the pre-training model involves loading only the encoder’s pre-training weights. However, our improvement keeps the consistency of decoders between pre-training and fine-tuning, enabling the complete utilization of the pre-training model, not only loading the encoder’s weights but also the decoder’s to the SFD network. The results further analyzed the contributions of such architectural improvement in pre-training, revealing that the complete utilization of the pre-training model indeed elevates SFD performance.
Although our efforts in pre-training, whether in terms of data or network architecture, were initially aimed at improving SFD’s performance, the analysis of results motivates us to assume that the proposed method has generalizability for most meteorological satellite imagery interpretation tasks, especially in marine regions. In summary, collecting a large-scale dataset of meteorological satellite imagery that resembles the downstream learning task’s regions, meteorological activities, and satellite characteristics, and employing a decoder structure that exhibits excellent performance in that task as the pre-training network’s decoder, holds promising potential as a pre-training methodology.

4.2. Validation over 2021 Sea Fog Events

The validation dataset involves all sea fog events in 2021. Table 8 reports, in 2021, when the sea fog event happened and how long time the sea fog event lasted. Table 9 presents monthly statistics of the number of sea fog events. From these two tables, it can be observed that sea fog is most frequent during the spring season, with no occurrences during the autumn season. Sea fog also occurs during the summer and winter seasons, but the total number of occurrences is lower compared to the spring. In general, the duration of sea fog events lasts around 7 to 8 h, but in July, there were two events with much shorter durations, due to the influence of tropical cyclones during the summer.
Table 8 further compares the performance of the SFD, where the backbone is ViT-Base with the ImageNet and SeaMAE pre-training weights for each sea fog event. Our method outperforms the SFD performance from the ImageNet pre-training weights on any sea fog event. This indicates that the proposed pre-training paradigm is not influenced by specific sea fog events, occurrence dates, or seasons, and consistently presents performance gains for sea fog detection.
During a sea fog event, the position and shape of the sea fog gradually change over time. Although Table 8 calculates the Intersection over Union (IoU) for each sea fog event as a criterion of performance, we aim to further highlight the superiority of our proposed pre-training paradigm for SFD intuitively. To achieve this, we devise high-frequency SFD extraction along the time dimension over the entire duration of one event. This strategy allows for summarizing the sea fog event using a single sea fog mask that captures the climactic sea fog (CSF) locations throughout the entire sea fog event.
For a specific location denoted by pixel ( i , j ) in the study region, if sea fog occurs at that location for more than half of the time steps within the sea fog event duration ( 0 , T ) , the pixel ( i , j ) is defined as a CSF location. Both ground truth sea fog events and predicted sea fog events can be processed to obtain CSF locations in this way, resulting in the constitution of CSF masks.
Figure 9 compares the CSF masks predicted by the ImageNet pre-training model, the SeaMAE pre-training model, and the ground truth sea fog events. The CSF masks of predictions are contrasted with the ground truth CSF masks. The yellow regions represent True Positive predictions, the red regions represent False Negative predictions, and the orange regions represent False Positive predictions. Across all sea fog events, our proposed paradigm consistently generates predictions that most closely resemble the ground truth CSF masks.

4.3. Literature Review and Comparison to Ours

In comparison with related literature, this work’s contributions and significance can be highlighted.

4.3.1. Sea Fog Detection in Bohai Sea and Yellow Sea

We primarily review works that use deep-learning technology to detect sea fog (SFD) [3,4,5,6,7,8,9,10,11,12,13,14]. In recent years, most of them studied the area covering the Bohai Sea and Yellow Sea [3,6,10,11,12,13,14], where the sea fog is a frequent occurrence. Ref. [3] introduces a scSE-LinkNet model that utilizes residual blocks and an attention module to accurately detect daytime sea fog. Ref. [6] presents a correlation context-driven method for SFD using a two-stage superpixel-based fully convolutional network (SFCNet) and an attentive Generative Adversarial Network (GAN). Ref. [10] introduces an unsupervised domain adaptation method that utilizes labeled land fog data to detect sea fog in meteorological satellite imagery. Ref. [11] presents an improved algorithm using one visible (VIS) and one near-infrared (NIR) band of the Himawari-8 satellite, demonstrating successful SFD without cloud mask information. Ref. [12] proposes a two-stage deep-learning strategy for daytime SFD through the creation of a labeled dataset and the Cloud-Aerosol LiDAR with Orthogonal Polarization (CALIOP) vertical feature mask (VFM). Ref. [13] utilizes MODIS images to analyze the temporal and spatial characteristics of sea fog, establishing a threshold algorithm based on specific MODIS bands. Ref. [14] presents a novel approach to weakly supervised semantic segmentation for SFD using point-based annotation and auxiliary information from ICOADS visibility data.
Previous methods oftentimes employ ImageNet pre-training models or learn from scratch on meteorological satellite imagery annotated with sea fog. In contrast, our novel learning paradigm prioritizes large-scale pre-learning with meteorological satellite imagery, capitalizing on the rich availability of multi-spectral geostationary imagery captured by the meteorological satellite. Furthermore, we strengthen this paradigm by harnessing and amending Masked Autoencoders (MAE), for the first time enabling the integration of Vision Transformers into SFD, a pioneering endeavor within this interdisciplinary community. Moreover, our approach extends the scope of pre-training beyond the encoder, embracing the decoder of the SFD network as an essential pre-trained component.

4.3.2. MIM Pre-Training with Satellite Imagery

Recently MIM succeeds in pre-training very large-scale Vision Transformers [15] in many fields [19,20,34,35,36,37], so more and more Vision Transformers pre-trained with satellite imagery have been being proposed [21,22,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56]. Ref. [38] introduces a scalable and adaptive self-supervised Transformer (AST) model for optical satellite imagery interpretation by generating scale-invariant representations through cross-scale collaboration and masking strategies. Ref. [39] presents large vision models specifically designed for VHR satellite imagery interpretation, utilizing rotated varied-size window attention in transformers. Ref. [40] introduces Presto, a pre-trained Remote Sensing Transformer specifically designed for self-supervised learning on remote-sensing pixel-time series data. Ref. [22] introduces RingMo, a remote-sensing foundation model framework that leverages generative self-supervised learning to construct a large-scale dataset—Million-AID—and train models with improved general representation capabilities. Ref. [21] introduces SatMAE, a pre-training framework for satellite imagery based on Masked Autoencoder, leveraging temporal and multi-spectral information. Ref. [41] explores the use of self-supervised MIM and modifications to the transformer architecture for hyper-spectral satellite imagery. Ref. [42] proposes RS-BYOL, a self-supervised learning approach for remote sensing, which leverages multi-spectral and synthetic aperture radar data as an implicit augmentation to learn invariant feature embeddings.
Most of the existing studies in the field of MIM pre-training with satellite imagery use large-scale satellite datasets to pre-train models, tailored to the downstream task of interpreting VHR satellite imagery. These pre-training datasets typically involve various geographical scenarios and various types of satellite imagery, such as multi-high-resolution, multi-spectral, optical RGB, and SAR. However, the marine environment and meteorological satellite imagery have remained unexplored in these works. In comparison to the existing literature, SeaMAE serves as the first contribution in this community to intelligent vision tasks of ocean observation, i.e., SFD, through a novel MAE architecture with architecture characteristics from the SFD network and a large-scale pre-training dataset consisting of meteorological satellite imagery.

4.4. Future Directions

Our future research will consist of a probe into the used fine-tuning dataset, a validation of the deduction proposed at the end of Section 4.1, which will be carried on another challenging interpretation of meteorological imagery, cloud classification in pixel, and an extended application on pre-training SFD deep-learning models in other marine regions.
The SFD dataset used in our study is finely annotated by experts with vast knowledge of sea fog, covering all sea fog events in the Bohai Sea and Yellow Sea over three years, from 2019 to 2021. Considering the scarcity of satisfactory SFD datasets in the community, we will make this dataset publicly available. Apart from benchmarking SFD technology, this dataset can also be utilized for research on cloud classification in meteorological satellite imagery, therefore promoting the intelligent development of marine meteorology. Encyclopedic statistics of the dataset will be analyzed in a separate article, and we will propose a novel SFD method capable of setting state-of-the-art results on this dataset.
In Section 4.1, we draw a deduction that the methodology proposed in this paper is likely to apply to multiple meteorological satellite imagery interpretation tasks, such as pixel-level cloud classification in marine regions. From the perspective of pixel-level classification, i.e., semantic segmentation or scene understanding, SFD is a binary classification, distinguishing pixels as either sea fog or background. On the other hand, cloud classification involves categorizing pixels into ten classes, including alto-cumulus, altostratus, nimbostratus, cumulus, stratocumulus, stratus, cirrus, cirrostratus, and deep convection. We expect to revisit pixel-level cloud classification in meteorological satellite imagery, building upon the pre-training guideline proposed in this paper. Such research will be to explore a pre-training paradigm tailored for cloud classification. Additionally, the deduction drawn from the above results in pre-training for SFD can be examined by the study on cloud classification, one of the most important meteorological imagery interpretation tasks.
In future work, validating the proposed learning paradigm for SFD in other marine regions, or devising specialized pre-training strategies, will both significantly advance the field. We will also proceed in this direction by conducting SFD research in the South China Sea and integrating it with the Bohai Sea and Yellow Sea regions to propose a highly generalized pre-training model tailored for SFD across multiple marine areas.

5. Conclusions

This paper develops a novel learning paradigm for sea fog detection (SFD). We focus on an economically important and sea-fog-vulnerable area in the range of 115.00E to 128.75E and 27.60N to 41.35N. Our learning paradigm improves the SFD performance of canonical learning paradigms by 2–3% sea fog IoU, achieving 64.18 % sea fog IoU. We primarily answer the question that how to pre-train SFD networks. This paper is also the first research utilizing Vision Transformer as the backbone of SFD networks. First, we show that ViT without pre-training will lead to unsatisfactory performance, and pre-training in any manner is of significance. Next, we show that ImageNet pre-training and VHR satellite imagery pre-training can be substituted by the meteorological satellite imagery pre-training. Finally, we show that making the pre-training network architecturally consistent with the sea fog detection model can brings more performance gains, which is mainly attributed to (1) the architectural characteristics of the SFD model are also effective in meteorological sea fog reconstruction and (2) such an architectural consistency enables the decoder part of SFD model to load pre-training knowledge. Even when compared to the latest pre-training methods specialized for satellite imagery, our approach still demonstrates considerable superiority in meteorological satellite imagery pre-training for SFD. We hope this work can inspire creativity in pre-training deep-learning models for interpreting meteorological satellite imagery in the future.
Limitations: This paper investigates the pre-training of deep-learning models for SFD in meteorological satellite imagery. However, due to the availability of large-scale data (more sea fog events in the Bohai Sea and Yellow Sea region), our analysis is limited to SFD in this specific area. Therefore, whether the proposed approach is effective for SFD in other sea regions deserves further investigation.

Author Contributions

Conceptualization, H.Y. and Y.Z.; Methodology, H.Y. and Y.Z.; Software, S.S. and Y.Z.; Validation, S.S. and M.X.; Formal analysis, H.Y.; Investigation, H.Y.; Resources, S.S.; Data curation, S.S. and M.X.; Writing—original draft, H.Y.; Writing—review & editing, M.W. and C.Z.; Visualization, S.S.; Supervision, M.W., C.Z. and B.H.; Project administration, M.W., M.X., C.Z. and B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by National Key R&D Program of China No. 2021YFC3000905.

Data Availability Statement

The data used in this study is not currently available. But we will make it available in a few months’ time as we continue to work on it. Anyone who needs the data can contact us by e-mail for latest information.

Conflicts of Interest

The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Liu, X.; Ruan, Z.; Hu, S.; Wan, Q.; Liu, L.; Luo, Y.; Hu, Z.; Li, H.; Xiao, H.; Lei, W.; et al. The Longmen cloud physics field experiment base, China Meteorological Administration. J. Trop. Meteorol. 2023, 29, 1–15. [Google Scholar]
  2. Finnis, J.; Reid-Musson, E. Managing weather & fishing safety: Marine meteorology and fishing decision-making from a governance and safety perspective. Mar. Policy 2022, 142, 105120. [Google Scholar]
  3. Guo, X.; Wan, J.; Liu, S.; Xu, M.; Sheng, H.; Yasir, M. A scse-linknet deep learning model for daytime sea fog detection. Remote Sens. 2021, 13, 5163. [Google Scholar] [CrossRef]
  4. Zhu, C.; Wang, J.; Liu, S.; Sheng, H.; Xiao, Y. Sea fog detection using U-Net deep learning model based on MODIS data. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  5. Hu, T.; Jin, Z.; Yao, W.; Lv, J.; Jin, W. Cloud Image Retrieval for Sea Fog Recognition (CIR-SFR) Using Double Branch Residual Neural Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 3174–3186. [Google Scholar] [CrossRef]
  6. Huang, Y.; Wu, M.; Guo, J.; Zhang, C.; Xu, M. A correlation context-driven method for sea fog detection in meteorological satellite imagery. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  7. Jeon, H.K.; Kim, S.; Edwin, J.; Yang, C.S. Sea fog identification from GOCI images using CNN transfer learning models. Electronics 2020, 9, 311. [Google Scholar] [CrossRef]
  8. Li, T.; Jin, W.; Fu, R.; He, C. Daytime sea fog monitoring using multimodal self-supervised learning with band attention mechanism. Neural Comput. Appl. 2022, 34, 21205–21222. [Google Scholar] [CrossRef]
  9. Mahdavi, S.; Amani, M.; Bullock, T.; Beale, S. A probability-based daytime algorithm for sea fog detection using GOES-16 imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1363–1373. [Google Scholar] [CrossRef]
  10. Xu, M.; Wu, M.; Guo, J.; Zhang, C.; Wang, Y.; Ma, Z. Sea fog detection based on unsupervised domain adaptation. Chin. J. Aeronaut. 2022, 35, 415–425. [Google Scholar] [CrossRef]
  11. Ryu, H.S.; Hong, S. Sea fog detection based on Normalized Difference Snow Index using advanced Himawari imager observations. Remote Sens. 2020, 12, 1521. [Google Scholar] [CrossRef]
  12. Tang, Y.; Yang, P.; Zhou, Z.; Zhao, X. Daytime Sea Fog Detection Based on a Two-Stage Neural Network. Remote Sens. 2022, 14, 5570. [Google Scholar] [CrossRef]
  13. Wan, J.; Su, J.; Sheng, H.; Liu, S.; Li, J. Spatial and temporal characteristics of sea fog in Yellow Sea and Bohai Sea based on active and passive remote sensing. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 5477–5480. [Google Scholar]
  14. Zhu, X.; Xu, M.; Wu, M.; Zhang, C.; Zhang, B. Annotating Only at Definite Pixels: A Novel Weakly Supervised Semantic Segmentation Method for Sea Fog Recognition. In Proceedings of the 2022 IEEE International Conference on Visual Communications and Image Processing (VCIP), Suzhou, China, 13–16 December 2022; pp. 1–5. [Google Scholar]
  15. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar]
  16. Bao, H.; Dong, L.; Piao, S.; Wei, F. Beit: Bert pre-training of image transformers. arXiv 2021, arXiv:2106.08254. [Google Scholar]
  17. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  18. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  19. Chen, Z.; Agarwal, D.; Aggarwal, K.; Safta, W.; Balan, M.M.; Brown, K. Masked image modeling advances 3d medical image analysis. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 1970–1980. [Google Scholar]
  20. Zhou, L.; Liu, H.; Bae, J.; He, J.; Samaras, D.; Prasanna, P. Self pre-training with masked autoencoders for medical image analysis. arXiv 2022, arXiv:2203.05573. [Google Scholar]
  21. Cong, Y.; Khanna, S.; Meng, C.; Liu, P.; Rozi, E.; He, Y.; Burke, M.; Lobell, D.; Ermon, S. Satmae: Pre-training transformers for temporal and multi-spectral satellite imagery. Adv. Neural Inf. Process. Syst. 2022, 35, 197–211. [Google Scholar]
  22. Sun, X.; Wang, P.; Lu, W.; Zhu, Z.; Lu, X.; He, Q.; Li, J.; Rong, X.; Yang, Z.; Chang, H.; et al. Ringmo: A remote sensing foundation model with masked image modeling. IEEE Trans. Geosci. Remote Sens. 2022, 61, 5612822. [Google Scholar] [CrossRef]
  23. Bessho, K.; Date, K.; Hayashi, M.; Ikeda, A.; Imai, T.; Inoue, H.; Kumagai, Y.; Miyakawa, T.; Murata, H.; Ohno, T.; et al. An introduction to Himawari-8/9—Japan’s new-generation geostationary meteorological satellites. J. Meteorol. Soc. Jpn. Ser. II 2016, 94, 151–183. [Google Scholar] [CrossRef]
  24. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  25. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  26. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  27. Zhou, L.; Zhang, C.; Wu, M. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 182–186. [Google Scholar]
  28. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  29. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support 4th International Workshop, Proceedings of the DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  30. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015. Proceedings, Part III 18. pp. 234–241. [Google Scholar]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Li, Y.; Xie, S.; Chen, X.; Dollar, P.; He, K.; Girshick, R. Benchmarking detection transfer learning with vision transformers. arXiv 2021, arXiv:2111.11429. [Google Scholar]
  34. Feichtenhofer, C.; Li, Y.; He, K. Masked autoencoders as spatiotemporal learners. Adv. Neural Inf. Process. Syst. 2022, 35, 35946–35958. [Google Scholar]
  35. Zhou, J.; Wei, C.; Wang, H.; Shen, W.; Xie, C.; Yuille, A.; Kong, T. ibot: Image bert pre-training with online tokenizer. arXiv 2021, arXiv:2111.07832. [Google Scholar]
  36. Yu, X.; Tang, L.; Rao, Y.; Huang, T.; Zhou, J.; Lu, J. Point-bert: Pre-training 3d point cloud transformers with masked point modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19313–19322. [Google Scholar]
  37. Wang, R.; Chen, D.; Wu, Z.; Chen, Y.; Dai, X.; Liu, M.; Jiang, Y.G.; Zhou, L.; Yuan, L. Bevt: Bert pretraining of video transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14733–14743. [Google Scholar]
  38. He, Q.; Sun, X.; Yan, Z.; Wang, B.; Zhu, Z.; Diao, W.; Yang, M.Y. AST: Adaptive Self-supervised Transformer for optical remote sensing representation. ISPRS J. Photogramm. Remote Sens. 2023, 200, 41–54. [Google Scholar] [CrossRef]
  39. Wang, D.; Zhang, Q.; Xu, Y.; Zhang, J.; Du, B.; Tao, D.; Zhang, L. Advancing plain vision transformer towards remote sensing foundation model. IEEE Trans. Geosci. Remote Sens. 2022, 61, 5607315. [Google Scholar] [CrossRef]
  40. Tseng, G.; Zvonkov, I.; Purohit, M.; Rolnick, D.; Kerner, H. Lightweight, Pre-trained Transformers for Remote Sensing Timeseries. arXiv 2023, arXiv:2304.14065. [Google Scholar]
  41. Scheibenreif, L.; Mommert, M.; Borth, D. Masked Vision Transformers for Hyperspectral Image Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 24–31 January 2023; pp. 2165–2175. [Google Scholar]
  42. Jain, P.; Schoen-Phelan, B.; Ross, R. Self-supervised learning for invariant representations from multi-spectral and sar images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 7797–7808. [Google Scholar] [CrossRef]
  43. Marsocci, V.; Scardapane, S. Continual Barlow Twins: Continual self-supervised learning for remote sensing semantic segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 5049–5060. [Google Scholar] [CrossRef]
  44. Mikriukov, G.; Ravanbakhsh, M.; Demir, B. Deep unsupervised contrastive hashing for large-scale cross-modal text-image retrieval in remote sensing. arXiv 2022, arXiv:2201.08125. [Google Scholar]
  45. Wanyan, X.; Seneviratne, S.; Shen, S.; Kirley, M. DINO-MC: Self-supervised Contrastive Learning for Remote Sensing Imagery with Multi-sized Local Crops. arXiv 2023, arXiv:2303.06670. [Google Scholar]
  46. Li, W.; Chen, K.; Chen, H.; Shi, Z. Geographical knowledge-driven representation learning for remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5405516. [Google Scholar] [CrossRef]
  47. Muhtar, D.; Zhang, X.; Xiao, P. Index your position: A novel self-supervised learning method for remote sensing images semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4411511. [Google Scholar] [CrossRef]
  48. Mall, U.; Hariharan, B.; Bala, K. Change-Aware Sampling and Contrastive Learning for Satellite Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 24–31 January 2023; pp. 5261–5270. [Google Scholar]
  49. Manas, O.; Lacoste, A.; Giró-i Nieto, X.; Vazquez, D.; Rodriguez, P. Seasonal contrast: Unsupervised pre-training from uncurated remote sensing data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9414–9423. [Google Scholar]
  50. Jain, P.; Schoen-Phelan, B.; Ross, R. Multi-modal self-supervised representation learning for earth observation. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 3241–3244. [Google Scholar]
  51. Jain, U.; Wilson, A.; Gulshan, V. Multimodal contrastive learning for remote sensing tasks. arXiv 2022, arXiv:2209.02329. [Google Scholar]
  52. Prexl, J.; Schmitt, M. Multi-Modal Multi-Objective Contrastive Learning for Sentinel-1/2 Imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, BC, Canada, 11–17 October 2023; pp. 2135–2143. [Google Scholar]
  53. Akiva, P.; Purri, M.; Leotta, M. Self-supervised material and texture representation learning for remote sensing tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8203–8215. [Google Scholar]
  54. Li, W.; Chen, H.; Shi, Z. Semantic segmentation of remote sensing images with self-supervised multitask representation learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6438–6450. [Google Scholar] [CrossRef]
  55. Tao, C.; Qi, J.; Zhang, G.; Zhu, Q.; Lu, W.; Li, H. TOV: The original vision model for optical remote sensing image understanding via self-supervised learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 4916–4930. [Google Scholar] [CrossRef]
  56. Scheibenreif, L.; Hanna, J.; Mommert, M.; Borth, D. Self-supervised vision transformers for land-cover segmentation and classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1422–1431. [Google Scholar]
Figure 1. (a) illustrates how sea fog detection (SFD) contributes to marine management. (b) An effective network pipeline on segmenting natural imagery. The upsampling operation is intensive because it interpolates the logits with very large ratios (e.g., 8 or 16) to match the spatial size of the input. (c) A canonical network pipeline for SFD. The most apparent difference lies in the decoder part, where the latent representation is upsampled progressively, and the output layer upsample the logits very slightly, with a ratio of 2.
Figure 1. (a) illustrates how sea fog detection (SFD) contributes to marine management. (b) An effective network pipeline on segmenting natural imagery. The upsampling operation is intensive because it interpolates the logits with very large ratios (e.g., 8 or 16) to match the spatial size of the input. (c) A canonical network pipeline for SFD. The most apparent difference lies in the decoder part, where the latent representation is upsampled progressively, and the output layer upsample the logits very slightly, with a ratio of 2.
Remotesensing 15 04102 g001
Figure 2. (a) Study area. The green curve depicts the coastline. (b) Visualization of 16 AHI bands in grayscale.
Figure 2. (a) Study area. The green curve depicts the coastline. (b) Visualization of 16 AHI bands in grayscale.
Remotesensing 15 04102 g002
Figure 3. (a) Samples from ImageNet dataset. (b) Samples from Million-AID dataset. (c) Samples from Himarwari-8. These samples are synthesized with (B03, B04, B14). The number under every image is the time when it is taken, in a format YYMMDD.
Figure 3. (a) Samples from ImageNet dataset. (b) Samples from Million-AID dataset. (c) Samples from Himarwari-8. These samples are synthesized with (B03, B04, B14). The number under every image is the time when it is taken, in a format YYMMDD.
Remotesensing 15 04102 g003
Figure 4. Illustration of the MIM pre-training and fine-tuning in existing sea fog detection applications.
Figure 4. Illustration of the MIM pre-training and fine-tuning in existing sea fog detection applications.
Remotesensing 15 04102 g004
Figure 5. (a) illustrates our proposed learning paradigm. (b) details our method to output the feature pyramid with only the last feature from ViT. (c) illustrates the hierarchical decoder architecture and components of one decoder block.
Figure 5. (a) illustrates our proposed learning paradigm. (b) details our method to output the feature pyramid with only the last feature from ViT. (c) illustrates the hierarchical decoder architecture and components of one decoder block.
Remotesensing 15 04102 g005
Figure 6. (a) Performance (IoU) change induced by scaling the pre-training data. (b) Performance (IoU) change induced by increasing the pre-training epochs.
Figure 6. (a) Performance (IoU) change induced by scaling the pre-training data. (b) Performance (IoU) change induced by increasing the pre-training epochs.
Remotesensing 15 04102 g006
Figure 7. Comparison of SFD results. Red bounding boxes help to focus the reader’s attention on the key area. (a,e) show the image. (b,f) show the ImageNet pre-training SFD results. (c,g) show the SeaMAE pre-training SFD results. (d,h) show the ground truth of sea fog.
Figure 7. Comparison of SFD results. Red bounding boxes help to focus the reader’s attention on the key area. (a,e) show the image. (b,f) show the ImageNet pre-training SFD results. (c,g) show the SeaMAE pre-training SFD results. (d,h) show the ground truth of sea fog.
Remotesensing 15 04102 g007
Figure 8. Reconstruction Results from SeaMAE based on ViT-B. (a,e) show the masked satellite image. (b,f) show the raw reconstruction output of SeaMAE. (c,g) show the combination of the reconstructed patch with the unmasked patch. (d,h) show the raw satellite image.
Figure 8. Reconstruction Results from SeaMAE based on ViT-B. (a,e) show the masked satellite image. (b,f) show the raw reconstruction output of SeaMAE. (c,g) show the combination of the reconstructed patch with the unmasked patch. (d,h) show the raw satellite image.
Remotesensing 15 04102 g008
Figure 9. CSF masks from ImageNet pre-training prediction, SeaMAE pre-training prediction, and GT in every sea fog event involved within the validation dataset. Yellow pixels represent True Positive predictions. Orange pixels represent False Positive predictions. Red pixels represent False Negative predictions. Zoom in to View.
Figure 9. CSF masks from ImageNet pre-training prediction, SeaMAE pre-training prediction, and GT in every sea fog event involved within the validation dataset. Yellow pixels represent True Positive predictions. Orange pixels represent False Positive predictions. Red pixels represent False Negative predictions. Zoom in to View.
Remotesensing 15 04102 g009
Table 1. Information of AHI observation bands.
Table 1. Information of AHI observation bands.
No.NameTypeDetection Category
B01V1VisibleVegetation, aerosol
B02V2VisibleVegetation, aerosol
B03VSVisibleLow cloud (fog)
B04N1Near InfraredVegetation, aerosol
B05N2Near InfraredCloud phase recognition
B06N3Near InfraredCloud droplet effective radius
B07I4InfraredLow cloud (fog), natural disaster
B08WVInfraredWater vapor density from troposphere to mesosphere
B09W2InfraredWater vapor density in the mesosphere
B10W3InfraredWater vapor density in the mesosphere
B11MIInfraredCloud phase discrimination, sulfur dioxide
B12O3InfraredOzone content
B13IRInfraredCloud image, cloud top
B14L2InfraredCloud image, sea surface temperature
B15I2InfraredCloud image, sea surface temperature
B16COInfraredCloud height
Table 2. Comparison of SOTA methods on our fine-tuning dataset. The bold indicates the highest result.
Table 2. Comparison of SOTA methods on our fine-tuning dataset. The bold indicates the highest result.
BackboneMethodsIoUAcc
Deeplabv3+ [28]56.1068.89
Scselinknet [3]55.2968.18
CNNUNet [31]58.4470.23
UNet++ [29]56.9569.30
Attetion-unet [30]56.3568.94
Table 3. Comparison of meteorological pre-training with ImageNet pre-training and from-scratch initialization on ViT-Ti, ViT-S, and ViT-B. Green numbers are the performance gains derived from changing pre-training data from VHR satellite imagery to meteorological satellite imagery.
Table 3. Comparison of meteorological pre-training with ImageNet pre-training and from-scratch initialization on ViT-Ti, ViT-S, and ViT-B. Green numbers are the performance gains derived from changing pre-training data from VHR satellite imagery to meteorological satellite imagery.
Pre-Train MethodPre-Train DataEncoderFine-Tune DecoderIoUAcc.
SupervisedImageNetCNNUNet58.4470.23
From ScratchViT-TiLinkNet55.0566.48
MAEImageNetViT-TiLinkNet57.7969.78
MAEMillion-AIDViT-TiLinkNet57.5669.63
MAEHimawari-8ViT-TiLinkNet58.63 (+1.07)70.47 (+0.84)
From ScratchViT-SLinkNet56.3068.04
MAEImageNetViT-SLinkNet59.4270.82
MAEMillion-AIDViT-SLinkNet59.6271.35
MAEHimawari-8ViT-SLinkNet60.48 (+0.86)71.87 (+0.52)
From ScratchViT-BLinkNet58.8870.49
MAEImageNetViT-BLinkNet61.6972.53
MAEMillion-AIDViT-BLinkNet61.8772.79
MAEHimawari-8ViT-BLinkNet62.92 (+1.05)73.91 (+1.12)
Table 4. Comparison of meteorological pre-training based on MAE with SeaMAE. Green numbers are the performance gains derived from using SeaMAE instead of MAE. Red numbers are the performance drops derived from removing skip connections. means the skip connection is equipped. means the skip connection is removed.
Table 4. Comparison of meteorological pre-training based on MAE with SeaMAE. Green numbers are the performance gains derived from using SeaMAE instead of MAE. Red numbers are the performance drops derived from removing skip connections. means the skip connection is equipped. means the skip connection is removed.
Pre-Train MethodPre-Train DataEncoderSkip ConnectionIoUAcc.
MAEHimawari-8ViT-Ti58.6370.47
SeaMAEHimawari-8ViT-Ti59.37 (+0.74)70.97 (+0.50)
SeaMAEHimawari-8ViT-Ti59.09 (−0.28)70.71 (−0.28)
MAEHimawari-8ViT-S60.4871.87
SeaMAEHimawari-8ViT-S61.55 (+1.07)72.52 (+0.95)
SeaMAEHimawari-8ViT-S61.12 (−0.43)72.24 (−0.28)
MAEHimawari-8ViT-B62.9273.91
SeaMAEHimawari-8ViT-B63.74 (+0.82)75.12 (+1.21)
SeaMAEHimawari-8ViT-B63.37 (−0.37)74.66 (−0.48)
Table 5. Comparison of loading SeaMAE pre-training weights to the decoder with randomly initializing the decoder. Green numbers are the performance gains derived from initializing the decoder with SeaMAE pre-training weights.
Table 5. Comparison of loading SeaMAE pre-training weights to the decoder with randomly initializing the decoder. Green numbers are the performance gains derived from initializing the decoder with SeaMAE pre-training weights.
Encoder Init.Decoder Init.EncoderIoUAcc.
MAERandomViT-Ti58.6370.47
SeaMAERandomViT-Ti59.1670.94
SeaMAESeaMAEViT-Ti59.37 (+0.21)70.97 (+0.03)
MAERandomViT-S60.4871.87
SeaMAERandomViT-S61.0272.16
SeaMAESeaMAEViT-S61.55 (+0.53)72.52 (+0.36)
MAERandomViT-B62.9273.91
SeaMAERandomViT-B63.4374.25
SeaMAESeaMAEViT-B63.74 (+0.31)74.66 (+0.41)
Table 6. Comparison of different settings of band number in pre-training and fine-tuning in on ViT-B. Red numbers are performance drops compared to using all bands both in pre-training and fine-tuning.
Table 6. Comparison of different settings of band number in pre-training and fine-tuning in on ViT-B. Red numbers are performance drops compared to using all bands both in pre-training and fine-tuning.
Pre-Train Band#Fine-Tune Band#Patch Emb. AdjustmentIoUAcc.
161663.7474.66
3363.08 (−0.66 )73.99 (−0.67)
163Random62.59 (−1.15)73.43 (−1.23)
163Index62.71 (−1.03)73.50 (−1.16)
163Resize63.52 (−0.22)74.48 (−0.18)
Table 7. Comparison of SeaMAE pre-training methods proposed for satellite imagery. Results in bold font are the better ones between SatMAE and RingMo. Green numbers are the performance gains compared to bold results.
Table 7. Comparison of SeaMAE pre-training methods proposed for satellite imagery. Results in bold font are the better ones between SatMAE and RingMo. Green numbers are the performance gains compared to bold results.
Pre-Train MethodPre-Train DataEncoderIoUAcc.
SatMAEHimawari-8ViT-Ti58.5670.28
RingMoHimawari-8ViT-Ti58.2269.90
SeaMAEHimawari-8ViT-Ti59.37 (+0.81)70.97 (+0.69)
SatMAEHimawari-8ViT-S60.2671.55
RingMoHimawari-8ViT-S60.1471.61
SeaMAEHimawari-8ViT-S61.55 (+1.29)72.52 (+0.91)
SatMAEHimawari-8ViT-B62.8374.39
RingMoHimawari-8ViT-B63.0274.42
SeaMAEHimawari-8ViT-B63.74 (+0.72)75.12 (+0.70)
Table 8. Sea fog event records in 2021.: Date, Duration, ImageNet pre-training SFD IoU, and SeaMAE pre-training SFD IoU from top to bottom in every cell. Green numbers are performance gains derived from using the SeaMAE pre-training weights instead of ImageNet pre-training ones.
Table 8. Sea fog event records in 2021.: Date, Duration, ImageNet pre-training SFD IoU, and SeaMAE pre-training SFD IoU from top to bottom in every cell. Green numbers are performance gains derived from using the SeaMAE pre-training weights instead of ImageNet pre-training ones.
Jan. 21stJan. 24thJan. 26thFeb. 11thFeb. 12thFeb. 13th
8 h8 h8 h8 h8 h8 h
49.6940.5261.7076.5161.3246.43
74.28 (+24.59)64.86 (+24.34)72.75 (+11.05)77.63 (+1.12)63.93 (+2.61)59.59 (+13.16)
Feb. 20thMar. 5thMar. 10thMar. 14thMar. 25thMar. 28th
8 h8 h8 h8 h8 h8 h
54.0558.3458.7756.3790.5556.10
63.08 (+9.03)76.61 (+18.27)60.42 (+1.64)58.86 (+2.49)91.12 (+0.57)68.03 (+11.93)
Apr. 26thApr. 27thApr. 28thApr. 29thMay 9thMay 10th
7.5 h8 h6.5 h6.5 h8 h8 h
64.6755.8159.3658.0787.9472.94
65.28 (+0.61)59.64 (+3.83)64.11 (+4.75)62.43 (+4.36)88.08 (+0.14)73.86 (+0.92)
May 30thMay 31stJun. 6thJun. 13thJul. 11thJul. 17th
8 h8 h8 h8 h3 h5.5 h
68.7375.9579.0380.0049.6145.99
70.41 (+1.68)77.81 (+1.86)80.13 (+1.10)80.34 (+0.34)64.96 (+15.35)68.44 (+22.45)
Table 9. Number of sea fog events in every season and every month.
Table 9. Number of sea fog events in every season and every month.
WinterSpringSummer
Jan.Feb.Mar.Apr.MayJun.Jul.
3354422
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, H.; Su, S.; Wu, M.; Xu, M.; Zuo, Y.; Zhang, C.; Huang, B. SeaMAE: Masked Pre-Training with Meteorological Satellite Imagery for Sea Fog Detection. Remote Sens. 2023, 15, 4102. https://doi.org/10.3390/rs15164102

AMA Style

Yan H, Su S, Wu M, Xu M, Zuo Y, Zhang C, Huang B. SeaMAE: Masked Pre-Training with Meteorological Satellite Imagery for Sea Fog Detection. Remote Sensing. 2023; 15(16):4102. https://doi.org/10.3390/rs15164102

Chicago/Turabian Style

Yan, Haotian, Sundingkai Su, Ming Wu, Mengqiu Xu, Yihao Zuo, Chuang Zhang, and Bin Huang. 2023. "SeaMAE: Masked Pre-Training with Meteorological Satellite Imagery for Sea Fog Detection" Remote Sensing 15, no. 16: 4102. https://doi.org/10.3390/rs15164102

APA Style

Yan, H., Su, S., Wu, M., Xu, M., Zuo, Y., Zhang, C., & Huang, B. (2023). SeaMAE: Masked Pre-Training with Meteorological Satellite Imagery for Sea Fog Detection. Remote Sensing, 15(16), 4102. https://doi.org/10.3390/rs15164102

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop