Next Article in Journal
Development of a Reference-Free Indirect Bridge Displacement Sensing System
Next Article in Special Issue
Gated Skip-Connection Network with Adaptive Upsampling for Retinal Vessel Segmentation
Previous Article in Journal
PtOEP–PDMS-Based Optical Oxygen Sensor
Previous Article in Special Issue
Fully Parallel Implementation of Otsu Automatic Image Thresholding Algorithm on FPGA
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Generation of the NIR Spectral Band for Satellite Images with Convolutional Neural Networks

1
Skolkovo Institute of Science and Technology, 143026 Moscow, Russia
2
Institute of Numerical Mathematics of Russian Academy of Sciences, 119333 Moscow, Russia
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5646; https://doi.org/10.3390/s21165646
Submission received: 11 June 2021 / Revised: 17 August 2021 / Accepted: 19 August 2021 / Published: 21 August 2021
(This article belongs to the Special Issue Advances in Image Segmentation: Theory and Applications)

Abstract

:
The near-infrared (NIR) spectral range (from 780 to 2500 nm) of the multispectral remote sensing imagery provides vital information for landcover classification, especially concerning vegetation assessment. Despite the usefulness of NIR, it does not always accomplish common RGB. Modern achievements in image processing via deep neural networks make it possible to generate artificial spectral information, for example, to solve the image colorization problem. In this research, we aim to investigate whether this approach can produce not only visually similar images but also an artificial spectral band that can improve the performance of computer vision algorithms for solving remote sensing tasks. We study the use of a generative adversarial network (GAN) approach in the task of the NIR band generation using only RGB channels of high-resolution satellite imagery. We evaluate the impact of a generated channel on the model performance to solve the forest segmentation task. Our results show an increase in model accuracy when using generated NIR compared to the baseline model, which uses only RGB ( 0.947 and 0.914 F1-scores, respectively). The presented study shows the advantages of generating the extra band such as the opportunity to reduce the required amount of labeled data.

1. Introduction

Machine learning techniques allow researchers to achieve high performance in a wide range of remote sensing tasks by leveraging spectral bands of different wavelengths [1]. One essential spectrum interval for the remote sensing image analysis is represented by the near-infrared (NIR) channel. The classical approaches in landcover classification tasks often use NIR-based spectral indices such as the Normalized Difference Vegetation Index (NDVI) or the Enhanced Vegetation Index (EVI) to assess the vegetation state [2]. This spectral band is widely used in many applications, including forestry [3,4], agriculture [5,6], and general landcover classification [7,8]. However, there are still cases when the NIR band is not presented in the available data [9,10]. Thus, the researchers rely only on RGB. For example, the Maxar Open Data Program [11] provides only RGB images. Many aerial imaging systems are also limited to visible wavelength ranges.
The NIR band cannot be extracted from RGB bands. A simple example is provided in Figure 1. For both the green tree and the green roof, the RGB values are the same. However, the values differ drastically in the NIR spectral range, as the metal roof does not have the vegetation properties that affect the NIR. On the other hand, indirect features can be used to evaluate the NIR value. In general, all roofs have a lower NIR values than any healthy tree during the vegetation period. Therefore, it is possible to make assumptions about the NIR value based on the object’s shape and texture. This study investigates how neural networks can be applied to solve the NIR generation task by learning the statistical distribution of a large unlabeled dataset of satellite images.
In [12], a similar problem of generating the NIR channel from RGB was described. The proposed solution was based on the K-Nearest Neighbor classification algorithm and was focused on the agricultural domain. The researchers show in [12] a high demand for the generated NIR data, which can solve particular problems. However, the neural network approach was beyond the scope of the present study for image generation. In [13], they generated synthetic spectral bands for archive satellite images using Landsat data. Synthetic satellite imagery generation from Sentinel-2 (with the spatial resolution more than 10 meters per pixel) was considered in [14,15]. However, in our work, we were focused on high-resolution satellite images as they provide valuable texture information.
Generative adversarial networks (GANs) have achieved great results in recent years [16]. The basis of this approach consists of two neural network models that are trained to beat each other. The first network (generator) aims to create instances as realistically as possible, and the second network (discriminator) learns to verify whether the instance is fake or real. Conditional GANs (cGAN) have proven to be a promising approach in various fields using additional conditions in the generation process. Conditional GANs were implemented to solve different tasks such as image colorization [17], including infrared input [18] and remote sensing data [19,20,21,22], and style transfer [23,24].
Pix2pix GAN, as described in [24], proposes an image-to-image translation approach. Previous studies have shown a lack of generalization for other problems. Authors of [24] aimed to develop an efficient framework that can be successfully implemented to solve a wide variety of tasks, such as image colorization, synthesizing images from a labeled map, generating land-cover maps from remote sensing images, changing the style, etc. Pix2pix GAN uses a “U-Net”-based architecture as a generator and a convolutional “PatchGAN” as a discriminator. The model was trained to estimate image originality separately for each small region. The authors used the following objective function G * = arg min G max D L c G A N ( G , D ) + λ L L 1 ( G ) to train the model. The Pix2pix approach enhancements were provided by the authors of [25].
One prevalent computer vision task is image colorization, which is required to obtain color images from grayscale ones [26]. One of the earliest works using texture information for this task is [27]. In recent years, GANs (in particular cGANs) have become a popular approach for such a challenge, in particular, in the remote sensing domain [20,28]. In the image colorization task, cGANs take a condition that should be utilized for new image generation. The results for such a task can be evaluated visually. This challenge share similarities with the NIR generation problem. As an input, grayscale images are received, and as an output, an RGB image is created. In contrast, for NIR, we strive to obtain one channel from three channels. Unlike mapping grayscale to RGB, NIR does not include a mixture of RGB; NIR even lies in a distant wavelength region from RGB. It makes the task more challenging. Moreover, in the colorization problem, the choice of color sometimes depends on the statistical distribution in the training set (for example, the color of the car might depend on the number of cars for each color). Such mismatches in colorization might not be treated as a severe mistake, and it does not corrupt the sense of the natural source of objects or phenomena. In contrast, for NIR in vegetation tasks, there is a strong connection between chlorophyll content and the intensity of the channel value [29]. A neural network can extract structure features such as shape and texture characteristics. We attempt to combine them with RGB values to generate the NIR band artificially and save the physical sense of this channel as much as possible.
In the remote sensing domain, the opportunity to work with multiple satellite data simultaneously is essential in various cases [30]. In [31], the authors consider WorldView and Planet imagery. WorldView has a higher spatial resolution, while Planet has a higher temporal resolution. Therefore, by combining these data, researchers can solve particular problems rapidly and with better quality. In [32], Modis and Landsat images fusion was considered in the flood mapping case. In [33], they combine images from WorldView2, Rapid Eye, and PlanetScope platforms to solve the forest degradation problem. When images from several sensors were available, the highest spatial resolution images were always preferred. Therefore, in the remote sensing domain, acquisition dates can vary for different satellites, and for monitoring, it is crucial to work with all available data sources. However, when a computer vision model uses data from different distributions, it can decrease prediction quality. One of the objectives of our study was to examine the importance of the NIR band for cross-domain stability.
In our study, we examine whether the cGAN image generation approach can produce sufficient results for image segmentation purposes. Multiscale contextual features and spatial details are highly important in the remote sensing domain [34]. Therefore, we aim to apply the NIR generation as a feature-engineering method, creating a new feature (NIR reflectance) that is not present in the original feature space (RGB reflectance). We also study original and artificially generated NIR in the cross-domain stability problem, as convolutional neural network (CNN) robustness for various data is vital in the remote sensing domain [35]. We aim to use a vast amount of RGB & NIR data without markup that can be further leveraged in semantic segmentation tasks when NIR is not always available Figure 2.
We propose and validate an efficient approach to produce an artificial NIR band from the RGB satellite image. A state-of-the-art Pix2pix GAN technique is implemented for this task and compared with a common CNN-based approach for the regression task. WorldView-2 high-resolution data are leveraged to conduct image translation from RGB to NIR with further verification on PlanetScope and Spot-5 RGB images. We also investigate how original and artificially generated NIR bands affect both CNN and Random Forest (RF) [36] predictions in forest segmentation tasks compared to only RGB data. The experiments involve two significant practical cases: two data source combinations (PlanetScope and Spot-5) and different amount of labeled training data (the total dataset size for the segmentation task is 500.000 hectares). The contribution of the presented work is as follows:
  • We propose the approach for feature-engineering based on the NIR channel generation via cGANs.
  • We investigate the impact of artificially generated and real NIR data on the model performance in the satellite image segmentation task. We also examine the NIR channel contribution in reducing labeled dataset size with minimum quality loss. The NIR channel for satellite cross-domain stability is considered.

2. Materials and Methods

2.1. Dataset

We leveraged WorldView-2 satellite imagery downloaded from GBDX [37] to train the generative models. For forest segmentation experiments, we used the satellite data provided by the SPOT-5 [38] satellite and the PlanetScope [39] satellite group. The imagery has a high spatial resolution of 2–3 meters per pixel in four spectral channels (red, green, blue, near-infrared). All images were georeferenced and had values equal to the surface reflectance. Overall, two datasets were used in this work:
The first dataset used in this work was for cGAN model training. The dataset consists of RGB and NIR channels from the same satellite (WorldView-2). It covers different regions of Russia and Kazakhstan with approximately the same climate and ecological conditions. The total territory is about 900,000 ha. The datasets consist of varying land cover classes such as crops, forests, non-cultivated fields, and human-made objects. Images with dates from May to September were chosen to represent the high-vegetation period.
The second dataset was used to test the real and artificial NIR channel’s influence compared to the bare RGB image. This dataset includes PlanetScope and Spot-5 imagery. The resolution of images ranges between 2 and 3 meters, depending on the view angle. The markup for the study region consists of the binary masks of the forested areas and other classes in equal proportion, covering 500,000 ha. The labeled markup was used for the binary image segmentation problem. The region was split into test and train parts in the proportion of 0.25 : 0.75 .

2.2. Artificial NIR Channel Generation

To generate the NIR band from RGB, we used cGAN. We chose the Pix2pix approach for this task because it performs quite well for image translation problems [40,41]. For the generator, we used the U-Net [42] architecture with the Resnet-34 [43] encoder. For the discriminator, the PatchGAN as described in [24] with various receptive field sizes was used. The training procedure is shown in Figure 3. There were two models: the generator and the discriminator. The generator was trained to create artificial NIR images, using the RGB image as a conditional input. The discriminator received an RGB image in combination with the alleged NIR image. Then, there were few possible scenarios: (1) the NIR was original, and the discriminator succeeded in ascertaining it; (2) the NIR was fake, but the discriminator failed by treating it as original; (3) the NIR was original, but the discriminator mistook for fake; (4) the NIR was fake, and the discriminator exposed it. Although this model was trained simultaneously, we ultimately strove to receive a high performing generative model, to solve the objective of the study. For further analysis, only the generator was considered. Unlike classical machine learning techniques, which usually work only with one particular point (see [12]), the U-Net generator processes a particular neighborhood and learns how to summarize 3-dimensional information.
We compared the cGAN-based approach with the simple CNN-based one where U-Net with Resnet-34 encoder was trained to solve the regression problem.
We considered the root mean square error (RMSE), mean absolute error (MAE), and mean bias error (MBE) for the model’s performance evaluation as follows:
R M S E = i = 1 n ( y i y i ^ ) 2 n
M A E = i = 1 n | y i y i ^ | n
M B E = i = 1 n ( y i y i ^ ) n
where y ¯ is the mean target value among all pixels, y i ^ is the predicted value of the ith pixel, y i is the target value of the ith pixel, and n is the pixel number.

2.3. Forest Segmentation Task

To empirically evaluate the usefulness of the original and artificially generated NIR channel to solve real image segmentation problems, we considered the forest segmentation task with high-resolution satellite imagery. In this task, a CNN model was trained to ascribe each pixel with the forest content label.
We used the common solution for the image semantic segmentation: U-Net [42] with the ResNet-34 [44] encoder. The chosen architecture is widely implemented in the remote sensing domain [45]. We conducted experiments with different input channels: only RGB, RGB & original NIR, and RGB & generated NIR. The model output was a binary mask of the forest landcover, which was evaluated against the ground truth with an F1-score. We also assessed the original and artificially generated NIR in the same task using classical machine learning approach. We trained a Random Forest (RF) classifier [36]. The RF implementation was from [46] with the default parameters the same as in [36]. Each pixel was considered as an object for the classification.
p r e c i s i o n = T P T P + F P , r e c a l l = T P T P + F N , F 1 = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l
where T P is the True Positive (the number of correctly classified pixels of the given class), F P is the False Positive (the number of pixels classified as the given class while, in fact, being of the other class, and F N is the False Negative (the number of pixels of the given class, that were missed by the method).

2.4. NIR Channel Usage

We conducted an experiment that estimated the dependency of the segmentation quality on the training dataset size in both RGB and RGB & NIR cases. We randomly split and chose 50% and 30% of the initial training dataset (test data were the same for these random splits). The same experiment was repeated both for the SPOT and Planet imagery but separately for each data source.
In the second study, we considered data from different sources (both PlanetScope and SPOT data) simultaneously. Even if we have two images of the same date, region, and resolution but from various providers, sensors systems and image preprocessing can make them radically different from each other. The intensity distribution for images from Spot and Planet are shown in Figure 4. Such differences can be crucial for machine vision algorithms and lead to a reduction in prediction quality. Therefore, it can be treated as a case of a more complex multi-domain satellite segmentation task. To estimate the importance of the original and artificial NIR channels for different satellite data, we conducted the following experiment. The CNN model was trained using the Planet and SPOT data simultaneously. To evaluate the model’s performance, three test sets were considered: only the Planet test images, only the SPOT test images, and both the Planet and SPOT images. The images for Planet and Spot covered the same territory.

2.5. Training Setup

The training of all neural network models was performed on a PC with GTX-1080Ti GPUs, using Keras [47] with a Tensorflow [48] backend. For the simple regression model, the following training parameters were set. An optimizer RMSprop was chosen with a learning rate of 0.001 , which was reduced with patience 5. There were 20 epochs with 100 steps per epoch. The batch size was specified to be 30 with an image size of 256 × 256 pixels [24]. A model based on GAN training parameters was constructed as follows. The loss functions were chosen to be binary cross-entropy and MAE. The optimizer was Adam. The batch size and image size were the same as for the simple model. The models were trained for 600 epochs, 100 steps per epoch, and a batch size of 30. For the Planet data, we also conducted a fine-tuning procedure of the pretrained generative model using a small area without the necessity of markup. For the SPOT data, there was no additional training.

3. Results and Discussion

The results for NIR generation by cGAN are presented in Table 1 for the WorldView, SPOT, and Planet satellite data. All values for real and generated NIR were in the range [ 0 , 1 ] . The simple CNN regression approach showed significantly poor results (the MAE was 0.21 for WorldView). Therefore, we did not select this approach for future study. The principal difference between cGANs and the regression CNN model is the type of loss function. As our experiments show, both MAE and MSE loss in the regression CNN model led to the local optimum, which was far from the global one. The loss function can be affected by the distribution of RGB values. Compared to the regression CNN model, the results of cGAN were significantly closer to the real NIR values.
Another approach to evaluate the generated NIR band involves the forest segmentation task. The segmentation model was trained on the original NIR channels to predict the forest segmentation mask using RGB & generated NIR. The results are presented in Table 2, which shows that the additional NIR channel improved the cross-domain stability of the model. The example of segmentation prediction is shown in Figure 5. The model using the generated NIR provided more accurate results than the model trained only on RGB bands. The original NIR usage obtained an F1-score of 0.953 , the generated NIR obtained an F1-score of 0.947 , and the model using only RGB bands obtained an F1-score of 0.914 . The predicted NIR channel is shown in Figure 6, which confirms a high level of similarity between generated and original bands. Therefore, this approach allows more efficient CNN model usage in practical cases when data from different Basemaps are processed and cross-domain tasks occur.
We also assessed the generated and original NIR bands using classical machine learning approach. Results are presented in Table 2. For RF, the NIR band usage improves the classification quality from 0.841 to 0.877 . The F1-score for the generated NIR is 0.874 . This experiment shows that for the classification approach without spatial information the generated band is also provide significant information.
The results for different dataset sizes are presented in Table 3 and show that leveraging the NIR channel was beneficial in the case of smaller dataset sizes, whereas its effect decreased with the growing amount of the training data.
GANs aim to learn dataset distribution. It is conducted by minimizing the overall distance between the real and the generated distribution. We made an assumption that the dataset size is enough to approximate the distribution. Thus, we train the generator to sample according to the target distribution. The trained generator allowed a high-realistic image-to-image translation (G: { R G B } → NIR) such that the obtained NIR band is similar to those belonging to the target domain.
Example of a green roof is presented in Figure 7. Although, the color of the object is green, NIR value is low. It shows that the model had a sufficient amount of the training samples to learn such cases.
The experiments indicate that the generated NIR provides additional information to the segmentation model. We assume that the generative model incorporates the hidden statistical connections between the spectral channels that can be learned from the significant amount of real RGB and NIR data. As opposed to the segmentation or classification approach, the channel generation does not require the manual ground truth markup to significantly increase the dataset size. Therefore, this approach can be used as a feature-engineering tool to create a new feature similar to the NIR band of multispectral remote sensing imagery.
We set the goal to predict exactly the NIR band instead of vegetation indexes such as NDVI or EVI. These indexes use the NIR band in combination with the Red band. The NIR band generation allows further computation of other indexes without a requirement for extra model training. The future study can be extended by implementing different vegetation indexes. Moreover, in the case of using neural networks with the generated NIR band, it is enough to provide input NIR and Red bands separately (not in the form of the computed indexes) because a neural network can approximate nonlinear functions such as vegetation indexes.
One example of a failure case is a green lake (Figure 8) that might be mistaken for a green lawn. The reason is insufficient representation in the training dataset. Another possible example is an artificial turf such as an open-air stadium. The model can erroneously treat it as a landcover with high NIR value. On the other hand, if we add a significant amount of such samples, it is possible that the model learns such a distribution both.
Pix2pix architecture includes 54M parameters in the generator part and 6M parameters in the discriminator part. The future study can be focused on trainable parameters reduction. Light-weighted neural network models are studied currently and show promising results in the remote sensing domain [34] (just 4M parameters are used).
Training models independently for each data source often leads to better results. However, it is a more expensive approach. In this study, we considered the case when we minimize the cost. In future research, separate models training for each datasource should be studied and analyzed.
Data providers aim to minimize time and other costs while providing imagery to customers. For this purpose, online services for data acquisition are created [49,50] that allows one to analyze data “on the fly”. The most spread and cheap format for such platforms is RGB images, even when original imagery includes more spectral channels. The proposed NIR generation approach can be implemented for such products as “basemaps”. That requires further study.
In the future, we seek to implement this feature-engineering approach to other remote sensing tasks, such as agriculture classification and land-cover semantic segmentation. In addition, the proposed approach holds potential to solve challenges when only drones’ RGB channels are available. Another direction is to combine this feature-engineering approach with different augmentation techniques for remote sensing tasks [51,52].
It is promising to investigate the application of NIR generation methods beyond remote sensing problems in future works. Since NIR provides valuable auxiliary data in plant phenotyping tasks, NIR generation can be extended for greenhouses where high precision is vital [53].

4. Conclusions

The NIR band contains essential properties for landcover tasks. However, in particular cases, this band is not available. This study investigated Pix2pix cGAN implementation for image-to-image translation from RGB space imagery to the NIR band. We proposed an efficient feature-engineering approach based on an artificial NIR band generation. We conducted forest segmentation experiments to assess the importance of the NIR band in cases of small datasets and different satellite data sources. The proposed approach improved the model’s robustness to data source diversity and reduced the requirement to mark the dataset size, which is crucial for machine learning challenges. We assume that this data generation strategy can be implemented in practical tasks that require the NIR channel. This method can be extended to other spectral channels and remote sensing data sources.

Author Contributions

Conceptualization, S.I. and A.T.; methodology, S.I.; software, S.I.; validation, S.I. and A.T.; formal analysis, S.I., A.T., and D.S.; investigation, S.I.; writing—original draft preparation, S.I.; visualization, S.I.; data curation, A.T.; writing—review and editing, S.I., I.O., D.S., and V.I.; supervision, A.T., I.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Science and Higher Education of the Russian Federation (Grant 075-15-2020-801).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maxwell, A.E.; Warner, T.A.; Fang, F. Implementation of machine-learning classification in remote sensing: An applied review. Int. J. Remote Sens. 2018, 39, 2784–2817. [Google Scholar] [CrossRef] [Green Version]
  2. Huete, A.; Justice, C.; Van Leeuwen, W. MODIS vegetation index (MOD13). Algorithm Theor. Basis Doc. 1999, 3, 295–309. [Google Scholar]
  3. Li, W.; Dong, R.; Fu, H.; Yu, L. Large-scale oil palm tree detection from high-resolution satellite images using two-stage convolutional neural networks. Remote Sens. 2019, 11, 11. [Google Scholar] [CrossRef] [Green Version]
  4. Illarionova, S.; Trekin, A.; Ignatiev, V.; Oseledets, I. Neural-Based Hierarchical Approach for Detailed Dominant Forest Species Classification by Multispectral Satellite Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1810–1820. [Google Scholar] [CrossRef]
  5. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  6. Navarro, P.J.; Pérez, F.; Weiss, J.; Egea-Cortines, M. Machine learning and computer vision system for phenotype data acquisition and analysis in plants. Sensors 2016, 16, 641. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Scott, G.J.; England, M.R.; Starms, W.A.; Marcum, R.A.; Davis, C.H. Training deep convolutional neural networks for land–cover classification of high-resolution imagery. IEEE Geosci. Remote Sens. Lett. 2017, 14, 549–553. [Google Scholar] [CrossRef]
  8. Fan, J.; Chen, T.; Lu, S. Unsupervised Feature Learning for Land-Use Scene Recognition. IEEE Trans. Geosci. Remote Sens. 2017, 54, 2250–2261. [Google Scholar] [CrossRef]
  9. Flood, N.; Watson, F.; Collett, L. Using a U-net convolutional neural network to map woody vegetation extent from high resolution satellite imagery across Queensland, Australia. Int. J. Appl. Earth Obs. Geoinf. 2019, 82, 101897. [Google Scholar] [CrossRef]
  10. Alias, B.; Karthika, R.; Parameswaran, L. Classification of High Resolution Remote Sensing Images using Deep Learning Techniques. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; pp. 1196–1202. [Google Scholar]
  11. Satellite Imagery for Natural Disasters|Digital Globe. Available online: https://www.digitalglobe.com/ecosystem/open-data (accessed on 6 February 2021).
  12. De Lima, D.C.; Saqui, D.; Ataky, S.; Jorge, L.A.d.C.; Ferreira, E.J.; Saito, J.H. Estimating Agriculture NIR Images from Aerial RGB Data. In International Conference on Computational Science; Springer: Cham, Switzerland, 2019; pp. 562–574. [Google Scholar]
  13. Gravey, M.; Rasera, L.G.; Mariethoz, G. Analogue-based colorization of remote sensing images using textural information. ISPRS J. Photogramm. Remote Sens. 2019, 147, 242–254. [Google Scholar] [CrossRef]
  14. Abady, L.; Barni, M.; Garzelli, A.; Tondi, B. GAN generation of synthetic multispectral satellite images. In Image and Signal Processing for Remote Sensing XXVI. International Society for Optics and Photonics; SPIE: Bellingham, WA, USA, 2020; Volume 11533, p. 115330L. [Google Scholar]
  15. Mohandoss, T.; Kulkarni, A.; Northrup, D.; Mwebaze, E.; Alemohammad, H. Generating synthetic multispectral satellite imagery from sentinel-2. arXiv 2020, arXiv:2012.03108. [Google Scholar]
  16. Alqahtani, H.; Kavakli-Thorne, M.; Kumar, G. Applications of generative adversarial networks (gans): An updated review. Arch. Comput. Methods Eng. 2019, 28, 525–552. [Google Scholar] [CrossRef]
  17. Nazeri, K.; Ng, E.; Ebrahimi, M. Image colorization using generative adversarial networks. In International Conference on Articulated Motion and Deformable Objects; Springer: Cham, Switzerland, 2018; pp. 85–94. [Google Scholar]
  18. Suárez, P.L.; Sappa, A.D.; Vintimilla, B.X. Infrared image colorization based on a triplet dcgan architecture. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 18–23. [Google Scholar]
  19. Wu, M.; Jin, X.; Jiang, Q.; Lee, S.J.; Guo, L.; Di, Y.; Huang, S.; Huang, J. Remote Sensing Image Colorization Based on Multiscale SEnet GAN. In Proceedings of the 2019 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Huaqiao, China, 19–21 October 2019; pp. 1–6. [Google Scholar]
  20. Li, F.; Ma, L.; Cai, J. Multi-discriminator generative adversarial network for high resolution gray-scale satellite image colorization. In Proceedings of the IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 3489–3492. [Google Scholar]
  21. Tang, R.; Liu, H.; Wei, J. Visualizing Near Infrared Hyperspectral Images with Generative Adversarial Networks. Remote Sens. 2020, 12, 3848. [Google Scholar] [CrossRef]
  22. Singh, P.; Komodakis, N. Cloud-gan: Cloud removal for sentinel-2 imagery using a cyclic consistent generative adversarial networks. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 1772–1775. [Google Scholar]
  23. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  24. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  25. Qu, Y.; Chen, Y.; Huang, J.; Xie, Y. Enhanced pix2pix dehazing network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8160–8168. [Google Scholar]
  26. Wang, H.; Liu, X. Overview of image colorization and its applications. In Proceedings of the 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 March 2021; Volume 5, pp. 1561–1565. [Google Scholar]
  27. Welsh, T.; Ashikhmin, M.; Mueller, K. Transferring Color to Greyscale Images. ACM Trans. Graph. 2002, 21, 277–280. [Google Scholar] [CrossRef]
  28. Wu, M.; Jin, X.; Jiang, Q.; Lee, S.j.; Liang, W.; Lin, G.; Yao, S. Remote sensing image colorization using symmetrical multi-scale DCGAN in YUV color space. Visual Comput. 2020, 1–23. [Google Scholar] [CrossRef]
  29. Yang, P.; van der Tol, C.; Campbell, P.K.; Middleton, E.M. Fluorescence Correction Vegetation Index (FCVI): A physically based reflectance index to separate physiological and non-physiological information in far-red sun-induced chlorophyll fluorescence. Remote Sens. Environ. 2020, 240, 111676. [Google Scholar] [CrossRef]
  30. Vandal, T.J.; McDuff, D.; Wang, W.; Duffy, K.; Michaelis, A.; Nemani, R.R. Spectral Synthesis for Geostationary Satellite-to-Satellite Translation. IEEE Trans. Geosci. Remote. Sens. 2021. [Google Scholar] [CrossRef]
  31. Kwan, C.; Zhu, X.; Gao, F.; Chou, B.; Perez, D.; Li, J.; Shen, Y.; Koperski, K.; Marchisio, G. Assessment of Spatiotemporal Fusion Algorithms for Planet and Worldview Images. Sensors 2018, 18, 1051. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Zhang, F.; Zhu, X.; Liu, D. Blending MODIS and Landsat images for urban flood mapping. Int. J. Remote Sens. 2014, 35, 3237–3253. [Google Scholar] [CrossRef]
  33. Sedano, F.; Lisboa, S.N.; Sahajpal, R.; Duncanson, L.; Ribeiro, N.; Sitoe, A.; Hurtt, G.; Tucker, C.J. The connection between forest degradation and urban energy demand in sub-Saharan Africa: A characterization based on high-resolution remote sensing data. Environ. Res. Lett. 2021, 16, 064020. [Google Scholar] [CrossRef]
  34. He, Q.; Sun, X.; Yan, Z.; Fu, K. DABNet: Deformable Contextual and Boundary-Weighted Network for Cloud Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote. Sens. 2021, 1–16. [Google Scholar] [CrossRef]
  35. Illarionova, S.; Nesteruk, S.; Shadrin, D.; Ignatiev, V.; Pukalchik, M.; Oseledets, I. MixChannel: Advanced Augmentation for Multispectral Satellite Images. Remote Sens. 2021, 13, 2181. [Google Scholar] [CrossRef]
  36. Pal, M. Random forest classifier for remote sensing classification. Int. J. Remote Sens. 2005, 26, 217–222. [Google Scholar] [CrossRef]
  37. GBDX. Available online: https://gbdxdocs.digitalglobe.com/ (accessed on 17 August 2021).
  38. Optical and Radar Data|SPOT. Available online: https://www.intelligence-airbusds.com/optical-and-radar-data/#spot (accessed on 6 February 2021).
  39. Satellite Imagery and Archive|Planet. Available online: https://www.planet.com/products/planet-imagery/ (accessed on 6 February 2021).
  40. Salehi, P.; Chalechale, A. Pix2Pix-based Stain-to-Stain Translation: A Solution for Robust Stain Normalization in Histopathology Images Analysis. In Proceedings of the 2020 International Conference on Machine Vision and Image Processing (MVIP), Tehran, Iran, 18–20 February 2020; pp. 1–7. [Google Scholar]
  41. Ren, H.; Li, J.; Gao, N. Two-stage sketch colorization with color parsing. IEEE Access 2019, 8, 44599–44610. [Google Scholar] [CrossRef]
  42. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  43. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  45. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  46. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  47. Keras. 2019–2020. Available online: https://keras.io/ (accessed on 20 August 2020).
  48. TensorFlow. 2019–2020. Available online: https://github.com/tensorflow/tensorflow (accessed on 20 August 2020).
  49. Securewatch. Available online: https://www.maxar.com/products/securewatch (accessed on 17 August 2021).
  50. OneAtlas. Available online: https://www.intelligence-airbusds.com/imagery/oneatlas/ (accessed on 17 August 2021).
  51. Yu, X.; Wu, X.; Luo, C.; Ren, P. Deep learning in remote sensing scene classification: A data augmentation enhanced convolutional neural network framework. GIScience Remote Sens. 2017, 54, 741–758. [Google Scholar] [CrossRef] [Green Version]
  52. Illarionova, S.; Nesteruk, S.; Shadrin, D.; Ignatiev, V.; Pukalchik, M.; Oseledets, I. Object-Based Augmentation Improves Quality of Remote SensingSemantic Segmentation. arXiv 2021, arXiv:2105.05516. [Google Scholar]
  53. Nesteruk, S.; Shadrin, D.; Pukalchik, M.; Somov, A.; Zeidler, C.; Zabel, P.; Schubert, D. Image Compression and Plants Classification Using Machine Learning in Controlled-Environment Agriculture: Antarctic Station Use Case. IEEE Sens. J. 2021. [Google Scholar] [CrossRef]
Figure 1. Objects with the same spectral values in the RGB range can belong to significantly different classes. For these objects, spectral values beyond the visible range differ. These differences can be illustrated using vegetation indices such as the NDVI in the case of an artificial object and a plant during the vegetation period.
Figure 1. Objects with the same spectral values in the RGB range can belong to significantly different classes. For these objects, spectral values beyond the visible range differ. These differences can be illustrated using vegetation indices such as the NDVI in the case of an artificial object and a plant during the vegetation period.
Sensors 21 05646 g001
Figure 2. A large amount of RGB & NIR data without markup that can be further leveraged in semantic segmentation tasks when NIR is not available in some particular cases.
Figure 2. A large amount of RGB & NIR data without markup that can be further leveraged in semantic segmentation tasks when NIR is not available in some particular cases.
Sensors 21 05646 g002
Figure 3. Training procedure for GAN using the RGB image as an input and the NIR band as a condition.
Figure 3. Training procedure for GAN using the RGB image as an input and the NIR band as a condition.
Sensors 21 05646 g003
Figure 4. Original SPOT and Planet images (without any enhancements) and their RGB spectral values distribution. The histograms were computed within the forest area. Although the presented images are from the summer period, their spectral values differ drastically, as the histogram shows.
Figure 4. Original SPOT and Planet images (without any enhancements) and their RGB spectral values distribution. The histograms were computed within the forest area. Although the presented images are from the summer period, their spectral values differ drastically, as the histogram shows.
Sensors 21 05646 g004
Figure 5. Forest segmentation predictions on the test regions (SPOT). One model was trained just on RGB images; another model used RGB and generated NIR.
Figure 5. Forest segmentation predictions on the test regions (SPOT). One model was trained just on RGB images; another model used RGB and generated NIR.
Sensors 21 05646 g005
Figure 6. Example of generated NIR on the test set. The first row presents the SPOT image; the second row is the WorldView image.
Figure 6. Example of generated NIR on the test set. The first row presents the SPOT image; the second row is the WorldView image.
Sensors 21 05646 g006
Figure 7. Example of a case with a green roof (SPOT image). The green roof has low NIR values both for original and generated NIR bands.
Figure 7. Example of a case with a green roof (SPOT image). The green roof has low NIR values both for original and generated NIR bands.
Sensors 21 05646 g007
Figure 8. Example of a failure case (SPOT image). Green lake is erroneously treated as a surface with high NIR value.
Figure 8. Example of a failure case (SPOT image). Green lake is erroneously treated as a surface with high NIR value.
Sensors 21 05646 g008
Table 1. Error of the artificial NIR band for the test WorldView, SPOT, and Planet imagery.
Table 1. Error of the artificial NIR band for the test WorldView, SPOT, and Planet imagery.
MAERMSEMean Bias
WorldView0.090.310.058
SPOT0.0370.194−0.0029
Planet0.160.410.088
Table 2. The results of the forest segmentation experiments with different data sources. Both the RGB model and the RGB and NIR model were trained on Planet and Spot images simultaneously. The F1-score was computed on the test set individually for Planet and Spot and for the joined Planet and Spot test set.
Table 2. The results of the forest segmentation experiments with different data sources. Both the RGB model and the RGB and NIR model were trained on Planet and Spot images simultaneously. The F1-score was computed on the test set individually for Planet and Spot and for the joined Planet and Spot test set.
U-NetRF
Test imagesRGBRGBRGB andRGBRGBRGB and
and NIRartificial NIR and NIRartificial NIR
SPOT0.9540.9610.960.8740.8920.889
Planet0.8570.9390.9360.8150.8630.861
SPOT + Planet0.9320.960.9450.8360.8760.872
Average0.9140.9530.9470.8410.8770.874
(+0.039)(+0.033) (+0.036)(+0.033)
Table 3. The results for the forest segmentation experiments with different dataset sizes. The F1-score for SPOT and Planet on the test set. The entire data size was 500,000 ha.
Table 3. The results for the forest segmentation experiments with different dataset sizes. The F1-score for SPOT and Planet on the test set. The entire data size was 500,000 ha.
BandsAll Data1/21/3
SPOTRGB0.970.9560.942
RGB and NIR0.970.9630.961
PlanetRGB0.9390.9330.874
RGB and NIR0.950.9420.927
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Illarionova, S.; Shadrin, D.; Trekin, A.; Ignatiev, V.; Oseledets, I. Generation of the NIR Spectral Band for Satellite Images with Convolutional Neural Networks. Sensors 2021, 21, 5646. https://doi.org/10.3390/s21165646

AMA Style

Illarionova S, Shadrin D, Trekin A, Ignatiev V, Oseledets I. Generation of the NIR Spectral Band for Satellite Images with Convolutional Neural Networks. Sensors. 2021; 21(16):5646. https://doi.org/10.3390/s21165646

Chicago/Turabian Style

Illarionova, Svetlana, Dmitrii Shadrin, Alexey Trekin, Vladimir Ignatiev, and Ivan Oseledets. 2021. "Generation of the NIR Spectral Band for Satellite Images with Convolutional Neural Networks" Sensors 21, no. 16: 5646. https://doi.org/10.3390/s21165646

APA Style

Illarionova, S., Shadrin, D., Trekin, A., Ignatiev, V., & Oseledets, I. (2021). Generation of the NIR Spectral Band for Satellite Images with Convolutional Neural Networks. Sensors, 21(16), 5646. https://doi.org/10.3390/s21165646

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop