Next Article in Journal
Recommendation System Algorithms on Location-Based Social Networks: Comparative Study
Previous Article in Journal
A Personalized Compression Method for Steady-State Visual Evoked Potential EEG Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model-Based Underwater Image Simulation and Learning-Based Underwater Image Enhancement Method

Institute of Deep-Sea Science and Engineering, Chinese Academy of Sciences, Sanya 572000, China
*
Author to whom correspondence should be addressed.
Information 2022, 13(4), 187; https://doi.org/10.3390/info13040187
Submission received: 15 February 2022 / Revised: 24 March 2022 / Accepted: 2 April 2022 / Published: 7 April 2022

Abstract

:
Due to the absorption and scattering effects of light in water bodies and the non-uniformity and insufficiency of artificial illumination, underwater images often present various degradation problems, impacting their utility in underwater applications. In this paper, we propose a model-based underwater image simulation and learning-based underwater image enhancement method for coping with various degradation problems in underwater images. We first derive a simplified model for describing various degradation problems in underwater images, then propose a model-based image simulation method that can generate images with a wide range of parameter values. The proposed image simulation method also comes with an image-selection part, which helps to prune the simulation dataset so that it can serve as a training set for learning to enhance the targeted underwater images. Afterwards, we propose a convolutional neural network based on the encoder-decoder backbone to learn to enhance various underwater images from the simulated images. Experiments on simulated and real underwater images with different degradation problems demonstrate the effectiveness of the proposed underwater image simulation and enhancement method, and reveal the advantages of the proposed method in comparison with many state-of-the-art methods.

1. Introduction

Clear underwater images can provide first-hand information about the structure, texture, and color of underwater scenes and objects, and are vital for many underwater operations, such as marine archeology, seafloor mapping, and marine biological research [1,2,3,4,5]. However, due to the complexity of underwater environment, underwater images are prone to several degradation problems [6,7], which impact their utilities in underwater applications. Firstly, underwater images are affected by the wavelength-dependent attenuation of light in the water body, and present different degrees of color deviation depending on light propagation distance and water body property. Secondly, underwater images are degraded by the scattering effects caused by particles in the water body. These particles alter the directions of light, and cause blurriness and contrast reduction in underwater images. In shallow water, the refraction of light rays by the water surface may even cause caustics, which alters the actual appearance of objects of interest [8,9]. Moreover, illumination deficiency caused by light attenuation and non-uniform lighting is also a non-negligible issue, because in regions with insufficient lighting, the visibility of details is often limited. This issue is especially important for images in the deep sea, where non-uniform artificial illumination is the only light source. In Figure 1, we present four samples of underwater images with different degradation conditions for demonstration. Here, Figure 1a–c present degradation problems of bluish color deviation, contrast reduction, and insufficient illumination, respectively, and Figure 1d is an example of a clear underwater image with no degradation. All these images are from the unpaired real-world underwater image subset of the EUVP dataset [10].
Due to the complexity of underwater image degradation problems, the enhancement of underwater images has been a challenging task for a long time. Recently, owing to its great success in many vision tasks, deep learning was introduced in the field of underwater image enhancement. However, unlike many other tasks, the lack of paired training data is still a big obstacle for underwater image enhancement, because ground-truth images, i.e., clear counterparts of degraded underwater images, are almost impossible to get in the real world. There have been several attempts to solve this problem, such as generating simulated underwater images from clear images by using underwater image models [11] or deep-learning networks [10], or producing ground-truth candidates for real underwater images by using existing underwater image enhancement methods [12], but to the best of our knowledge, most of these methods are still limited to common degradation problems such as bluish color deviation and water-caused scattering, and have little control to the quality of generated image sets. As a result, the generated image sets can have narrower or different distributions than real-world underwater images, and methods using these sets as training data tend to have poor performance in solving diverse degradation problems in real-world underwater images.
In this paper, we propose a complete method of underwater image enhancement based on model-based image simulation and deep learning to deal with diverse degradation problems in real underwater scenes. To generate reliable training sets for deep learning, we derive a novel image model to describe the various degradation problems in real water and propose an underwater image simulation and selection method for generating reliable training sets for the targeted real-world images. Then, we propose a convolutional neural network (CNN) based on the encoder-decoder backbone with skip connections [13] and shortcut connections [14] to perform direct translations from degraded images to clear images. The proposed network is trained with the training sets generated by the aforementioned method to deal with various degradation problems in real-water underwater images. Experiments on underwater images with different degradation problems demonstrate the effectiveness and advantages of the proposed method, especially in enhancing images from challenging underwater scenes, such as the deep sea with insufficient lighting. The main contributions of this paper are summarized below:
  • We propose a sophisticated underwater image simulation and selection method for generating training sets for learning-based underwater image enhancement methods. To the best of our knowledge, the proposed method covers a wider range of degradation problems than previous methods and is the first one that can simulate underwater images with uneven lighting and generate training sets specialized for the targeted real-world images.
  • We propose an efficient CNN to learn the translation from degraded underwater images to clear enhanced images. Owing to the smooth transmission of information in the proposed network, textures and structures of the original underwater images can be well-preserved in the enhanced images.
  • The proposed method also outperforms many state-of-the-art methods, especially in enhancing underwater images with challenging degradation problems, such as strong light scattering or insufficient lighting.

2. Related Works

Underwater image degradation is a complex process that is influenced by many factors. However, most of them are hard to know during the underwater image enhancement process, which makes it an underdetermined problem that requires additional constraints to solve. Here, based on the types of these constraints, we classify underwater image enhancement methods into three categories and carry out a brief review of each of them.

2.1. Methods Based on General Image Processing Skills

The first type of underwater image enhancement method uses general image processing skills to improve the quality of underwater images. As mentioned before, the degradation of underwater images can appear in different aspects, including color, texture, contrast, and brightness. Methods of this type are built according to the aspects of image degradation. In 2007, Iqbal et al. [15] proposed an approach to improve the perception of underwater images. They performed contrast stretching on the RGB color space for color contrast equalization, and saturation and intensity stretching on HSI color space to increase the true color and address the lighting problem. Ancuti et al. [16] proposed a fusion-based method to solve the color deviation and contrast reduction problem. Their method combined several inputs generated by white balance, min-max windowing, and their combination, by weight maps of image luminance, contrast, chroma, and saliency. In Ancuti et al.’s later work [17], they proposed a new white balance method, and introduced gamma correction and unsharp masking to further improve the color balance, global contrast, and edge sharpness of input images. Fu et al. [18] used the Retinex theory to decompose the reflectance and illumination of the image, and applied histogram equalization and specification to address the fuzz and under-exposure problems. Similarly, Zhang et al. [19] also used the Retinex theory to obtain the reflectance and illumination component, but they used the simpler gamma correction for illumination adjustment.

2.2. Methods Based on Physical Models

In the second type of method, the complicated underwater image degradation process is described with simplified models and solved as an inverse problem by using statistics and inherent characters of underwater images as constraints. In these methods, the simplified image formation model (IFM) [20] is often used. As shown in Equation (1), IFM models underwater image I k x as the linear combination of clear image J k x and background light B k weighted by transmission map t k x , which is calculated by t k x = e c k d x , where c k is attenuation coefficients and d x is transmission distance.
I k x = J k x t k x + B k 1 t k x ,         k R , G , B ,
Due to the similarity of IFM and the hazy image model in the dark channel prior-based (DCP) haze-removal method in [21], the idea of DCP is also borrowed in underwater image enhancement. In [22], Galdran et al. proposed a red channel prior method to restore the colors associated to short wavelengths in underwater images. Drews Jr. et al. [23] proposed an underwater DCP that utilizes blue and green color channels as the underwater visual information source in the enhancement of underwater images. Carlevaris-Bianco et al. [24] proposed a prior based on the difference of attenuation between different color channels and used it to estimate the scene depth in underwater images. Apart from DCP, other priors and assumptions of the underwater images are also used in this type of method. In [20], Peng and Cosman proposed to exploit image blurriness and light absorption to estimate the background light and transmission map of the IFM, and restore underwater images with these parameters. Liu et al. [25] extended the IFM by assuming locally invariant illumination to incorporate the case with non-uniform light sources, and use the new model to solve the underwater degradation problems caused by the water body and the illumination. Marques et al. [26] utilized DCP to enhance low-lighting underwater images. More specifically, they applied DCP on the inverted underwater image to improve the brightness and contrast of low-lighting regions and proposed a contrast-guided approach for calculating dark channels and transmission maps with patches of varying sizes.

2.3. Methods Based on Deep Learning

Despite the diversity of aforementioned methods, most of them do not perform well in challenging cases. Recently, deep learning was introduced in the enhancement of underwater images due to its success in many other vision tasks. Unlike former two types of methods, learning-based methods use CNNs to capture the latent relationships between degraded and clear images in the training set, and enhance the targeted underwater images with trained networks, so the performance of learning-based methods is highly influenced by the training set in use. Since the clear counterparts of degraded underwater images are usually hard to get in the real world, generating reliable training sets is also a challenging task and is crucial for these methods. In [11], Li et al. proposed a lightweight CNN for underwater image enhancement. For preparing training sets, they used IFM and attenuation coefficients of Jerlov water types to simulate degraded images from 10 different water types. The same training set was also used in [27], where Uplavikar et al. proposed an all-in-one network that adversarially learns the domain agnostic features to generate enhanced underwater images from degraded images of 10 different water types. In [28], Wang et al. proposed a generative adversarial network (GAN) based on an improved IFM to generate underwater images. These images are then used to train a U-Net for underwater image enhancement. Li et al. [29] proposed another GAN for generating underwater images. Their method simulates attenuation and backscattering caused by the water body, as well as vignetting effect of the camera. In [30], Li et al. proposed a network structure with medium transmission-guided multi-color space embedding for coping with the color casts and low contrast problems of underwater image. In [31], Fabbri et al. proposed to use the cycle-consistent adversarial network (CycleGAN) [32] to directly learn to generate degraded underwater images based on two separate groups of clear and degraded real-world images. The generated images are then used to train a U-Net for underwater image enhancement. In [33,34], the same dataset was used to train a multiscale dense GAN and a multilevel feature fusion-based conditional GAN, respectively, for underwater image enhancement. In [10], Islam et al. utilized the same CycleGAN-based method to build a large dataset called EUVP, and proposed a lightweight conditional GAN for fast underwater image enhancement (FUnIE-GAN). Li et al. [12] constructed a real-world underwater image benchmark (UIEB) by using traditional underwater image enhancement methods to generate clear reference images from degraded underwater images. They also proposed a gated fusion network called Water-Net to learn to enhance underwater images. In general, training sets used in learning-based underwater image enhancement methods are mainly built from models, networks, or real-world images. To the best of our knowledge, UIEB is one of the most realistic and diverse training sets for underwater image enhancement, but it only contains 890 image pairs, which limits its utility in training large networks.

3. Methodology of Underwater Image Simulation

To develop a widely applicable learning-based method for underwater image enhancement, we start from the foundation problem of generating reliable training sets, for which we derive a new model to incorporate a wider range of degradation problems and propose a novel method for generating specialized training sets for various real-world underwater images. The generated training sets can then be used to train the CNN that learns to translate degraded underwater images to their clear counterparts. An example of such CNN is given in the next section.

3.1. Proposed Underwater Image Degradation Model

According to McGlamery’s and Jaffe’s works [35,36], monochromatic irradiance received by a camera underwater is composed of the direct component, the forward scattering component, and the backscattering component. Since the forward scattering component has high computational complexity but relatively low influence on pixel values and limited function range, we removed this component from our model for simplicity as done in many previous works [37,38,39].
The direct component mainly describes the energy loss of light underwater caused by attenuation. Its formula in [35,36] is given by
E d x , y | λ = E I x , y | λ M x , y | λ e c λ d x , y cos 4 θ x , y · T l 4 f n 2 · d x , y F l d x , y 2 ,
where E d x , y | λ represents the irradiance of the direct component and E I x , y | λ is the irradiance of incident light underwater. M x , y | λ is scene reflectance, c λ is volume attenuation coefficient and d x , y is transmission distance. θ , T l , f n , and F l are parameters of system geometry and camera lens and are usually hard to know in image simulation. To remove these parameters, we assumed an identical imaging system in air with no water-caused attenuation and derived the following formula.
J x , y | λ = E 0 λ M x , y | λ cos 4 θ x , y · T l 4 f n 2 · d x , y F l d x , y 2 .
Here, J ( x , y | λ ) is the received irradiance with no water, and E 0 λ is the irradiance of incident light in air. By combining Equations (2) and (3), the calculation of the direct component is simplified as
E d x , y | λ = E I x , y | λ J x , y | λ e c λ d x , y E 0 λ .
The backscattering component mainly contributes to the hazy looks of underwater images. In [40], Zhao et al. provided a simple but efficient formula for calculating this component, which is
E b s x , y | λ = E b s , λ 1 e c λ d x , y ,
where E b s , λ is the background light. Its formula under homogeneous illumination is given by
E b s , λ = κ l E s λ c λ Θ β ϕ d ϕ ,
where κ l is a constant of the system’s properties, E s λ is the incident irradiance, β ϕ is the volume-scattering function, and Θ represents all possible scattering angles. To incorporate inhomogeneous illumination under simpler expression, we use E s ( x , y | λ ) to replace E s λ , and use constant κ to replace the product of κ l and Θ β ϕ d ϕ , i.e.,
E b s x , y | λ = κ E s x , y | λ c λ 1 e c λ d x , y .
Finally, to derive a concise image model from the irradiance model, we assume a delta response function of the camera sensor and use it to integrate the total received irradiance (i.e., the combination of Equations (4) and (7)) of all wavelengths. The derived model is given in Equation (8). We use similar symbol expressions as in IFM for better comparison.
I k x , y = L k x , y J k x , y e c k d x , y + κ L s k x , y c k 1 e c k d x , y ,         k R , G , B .
The most significant difference between IFM and our new model is the lighting parameters L k x , y and L s k x , y , which represent the incident light intensities on the photographed scene and the scattering medium, and correspond to E I ( x , y | λ ) / E 0 λ in Equation (4) and E s ( x , y | λ ) in Equation (7), respectively. Apparently, with these parameters, the new model is able to simulate more complicated illumination conditions and analyze their influence on underwater image degradation.

3.2. Generating Reliable Training Sets by Underwater Image Simulation and Selection

With the proposed model, a simulated underwater image can be easily generated by assigning values to corresponding parameters. However, considering the diversity of underwater image degradation problems in the real world, it is very hard to choose the right parameters in the simulation process. Therefore, we propose to adopt another strategy that first expands the value range of model parameters as much as possible, then selects simulated images according to their quality and reliability.

3.2.1. Underwater Image Simulation with Broadened Ranges of Parameter Values

In this subsection, we introduce our strategies for broadening the value ranges of model parameters in the simulation process. Since spatially varying illumination is not included in previous simulation works [2,19,20], we especially proposed a method to effectively simulate variant illumination conditions in underwater scenes.
(1)
Simulation of lighting parameters L k x , y and L s k x , y
In real-world conditions, illumination of underwater images can be extremely complicated. Due to the absorption and scattering effects in the water body, incident light is often degraded before reaching the object, and thus brings initial color deviations to the underwater images. Furthermore, in the deep sea, artificial illumination is the only light source, and its non-uniformity often results in uneven brightness and limited visibility in the underwater images. To simplify the simulation of underwater lighting parameters, we use the combination of the following three basic patterns to approximate real-world underwater illumination: (a) uniform pattern with constant intensities for simulating shallow-water illumination, (b) parallel pattern whose intensities degrade with depth, and (c) artificial pattern that provides spatially uneven illumination. Diagrams and real-world examples of these patterns are presented in Figure 2. For simplification, we also assume L k x , y and L s k x , y are identical and use parameter κ to adjust the relative intensity of backscattering light.
The proposed formula for simulating underwater illumination is shown in Equation (9). In the right half of Equation (9), the first part is the intensity of uniform lighting (i.e., pattern (a)), the second part is the intensity of parallel lighting (i.e., pattern (b)), and the last part is the intensity of artificial lighting (i.e., pattern (c)). Here we use L t k and ω t   t a , b , c to represent the initial intensities and weights of these patterns, and their values are randomly selected from [0.9, 1.0] and [0, 1], respectively, in the simulation process. The sum of all weight parameters is equal to 1.
L k x , y = ω a L a k + ω b L b k e c k Z b + d x , y + i N c ω c , i κ c P x , y | L c , i k , σ i e c k D c , i x , y
The distribution of parallel lighting follows the Beer–Lambert’s Law. Therefore, its simulation requires the initial-light-intensity parameter, light-attenuation parameter, and light-traveling-distance parameter. We randomly picked a light-camera distance Z b from [0, 2 m] and added it to the transmission distance d x , y to approximate the light-traveling-distance parameter. The simulation method of light-attenuation parameter c k will be covered in the following part. The simulation method of transmission distance d x , y is given afterwards.
The simulation of artificial lighting is the most complicated part. In real-world underwater conditions, the distribution of artificial lighting can be influenced by the number, position, orientation, and beam pattern of light sources, which makes it difficult to simulate quickly. For simplification, we fixed the orientation of the light sources to orthographic projection, and used 2D Gaussian distribution to approximate real beam patterns. The difference of arbitrary and fixed light modes is depicted in Figure 3. Clearly, by using these simplifications, the simulation of imbalanced artificial lighting can be done in a much easier way.
In the parameter setting of artificial lighting simulation, the total number of light sources N c is selected from [0, 2] according to the character of the targeted illumination pattern. The beam pattern P x , y | L c , i k , σ i is approximated by a Gaussian distribution with peak value at L c , i k and standard deviation σ i proportional to the width of image by a random rate in [0.2, 1.1]. The distance from scene point to light source is calculated by D c , i x , y = Z l 2 + r l 2 · x x l , i 2 + y y l , i 2 , where Z l is the height of the i -th light source, and is randomly selected from [−1 m, 1 m], x l , i , y l , i is the projection of the i -th light source and is randomly picked from the pixel range, and r l is used to convert pixel distance to meter distance and is proportional to the ratio between the largest transmission distance and the image width, the scaling parameter of which is randomly selected from [0.1, 1]. To ensure the validity of light intensity, we use a scalar κ c to adjust the intensity level, so that the highest value is equal to 1.
In Figure 4, five sets of simulated underwater lighting parameters are presented, which shows the ability of proposed method in simulating diverse illumination conditions, including uneven illumination with artificial light source in use.
(2)
Parameter setting of channel-wise attenuation coefficients c k
To cover a wide range of water types in the simulation, we use attenuation coefficients of Jerlov water types [41] to generate values of channel-wise attenuation coefficients c k , as in [11]. The Jerlov water types include two sets, where types I–III correspond to open ocean water and types 1–9 correspond to coastal water. From types I–III to 1–9, the turbidity of the water body grows gradually, and the color deviation of underwater images varies from blue to green to yellow. To generate c k from attenuation coefficient curves, a straightforward method is to pick one set of values at predefined response wavelengths as done in [40], but considering the diversity of camera sensors, we expand the predefined wavelengths to nearby intervals, and randomly select wavelength values in each simulation. Then, we linearly combine the attenuation coefficients of adjacent Jerlov water types and multiply it with random scalars, so as to introduce additional variation and simulate in-between water types of these Jerlov water types. Mathematically,
c k = r c k · ω · c J λ k | T i + 1 ω · c J λ k | T i + 1
where c J λ k | T i represents the attenuation coefficient of Jerlov water type T i ( i 1 , 7 corresponds to types I–III to 1–7, note that we do not include type IA and IB because they are very similar to type I.) at wavelength λ k (randomly selected within the ± 10   nm intervals of response wavelengths in [40], i.e., λ R = 620 nm, λ G = 540 nm, and λ B = 450 nm), ω is the random weighting parameter within [0, 1], and r c k is the random scalar parameter from [0.9, 1.1]. Apparently, the proposed method can cover a wider range of attenuation coefficients than the Jerlov water type [41] itself.
(3)
Parameter setting of transmission distance d x , y
Instead of using RGB-D data to get paired J k x , y and d x , y as in previous methods [11,28,29], we use megaDepth [42], a CNN model for single-view depth prediction, to estimate the initial transmission distance d 0 x , y from J k x , y , and randomly scale it to get d x , y for the simulation process, mathematically,
d x , y = r d 1 + r d 2 · d 0 x , y min d 0 x , y max d 0 x , y min d 0 x , y r d 3 ,
where r d 1 , r d 2 , and r d 3 are random scalars that are picked from [0, 5 m], [2 m, 10 m], and [0.5, 2], respectively. In this way, the diversity of scenes and distance ranges of the simulated images will not be limited by the finite RGB-D datasets, and thus increases the number and diversity of simulated underwater images.
(4)
Parameter setting of κ
As previously mentioned, κ is a scalar about the camera system and backscattering. In the simulation, it is set as κ = r κ c k / max L k x , y with r κ randomly selected from 0.3 ,   1 , so as to prevent overexposure while controlling the proportion of backscattering.
(5)
Generating underwater image I k x , y
After the simulation of all aforementioned parameters, underwater images can be generated by substituting corresponding parameters in Equation (8). In Figure 5, we present some samples of simulated underwater images generated by the proposed method. Obviously, the proposed method is able to generate various degradation appearances for a given image, including images with deviated color, low contrast, and insufficient lighting.
Note that the diversity shown in Figure 5 is actually lower than the diversity the proposed method can reach, because the proposed method uses continuous values in most parameters. However, higher variance also indicates higher possibility to produce unrealistic simulated images, which should be excluded from the training process for higher efficiency. In the next subsection, we propose a simple but efficient method for selecting images into the training sets.

3.2.2. Generating Reliable Training Sets Based on Image Selection

We use the following two steps to select simulated underwater images into the training sets.
(1)
Selecting simulated underwater images with similar color deviations as the targeted image sets
As shown in Figure 5, the variances of transmission distance, water type, and illumination condition have a big influence on the color deviation of a simulated underwater image. To ensure the similarity of the degradation patterns in the training image sets and the targeted image sets, we only select those simulated images with similar color deviations as the targeted images into the training sets. We use the frequency-based method in [25] to calculate the color tone of an image. In Table 1, we present the calculated color tones of four real-world underwater images and four simulated underwater images, together with their HSV values. Apparently, images with similar color deviations have similar H (hue) and V (brightness) values of their color tones. Therefore, by only selecting simulated images whose H and V values are in the intervals defined by the color tones of targeted images, we can reject simulated images with dissimilar color deviations to the targeted image set. The intervals of H and V values are defined by σ m i n H m i n , σ m a x H m a x and σ m i n V m i n , σ m a x V m a x , respectively, where H m i n ( V m i n ) and H m a x ( V m a x ) are the minimum and maximum values of channel H ( V ) of targeted images, σ m i n and σ m a x are scaling parameters, and are set according to the size of targeted image set. For small set, σ m i n and σ m a x are set as 0.5 and 1.5, respectively. Otherwise, their values are set to 1.
(2)
Selecting simulated underwater images with enough details preserved
The preservation of enough details in the degraded underwater images is very important for providing necessary information in underwater image enhancement. To evaluate the degree of detail preservation in the degraded image, we use the Sobel edge map to quantify the intensity of details and calculate the difference between Sobel edge maps of the degraded image and the clear image. The Sobel edge map of an image is calculated as follows.
I e d g e x , y = k R , G , B 1 0 1 2 0 2 1 0 1 I k x , y 2 + k R , G , B 1 2 1 0 0 0 1 2 1 I k x , y 2
To avoid the influence of original detail intensities of the input image, we use the normalized mean absolute difference (NMAD) to evaluate the degradation degree of a simulated image, i.e., NMAD = x I e d g e x , y J e d g e x , y / x J e d g e x , y , where J e d g e is the edge map of the input clear image. Then, we exclude those simulated images with NMADs higher than a predefined threshold from the training sets, so as to avoid artifacts caused by these images in the network training process. In practice, the threshold is set to 0.75. In Table 2, we present a clear input image and several simulated images with their Sobel edge maps and NMAD scores. Clearly, the proposed method mainly rejects simulated images with extremely low contrast or very limited visibility.
Apparently, the proposed selection method provides a simple way to generate specialized training sets according to the characters of targeted real-world underwater images. For underwater tasks within fixed regions, it helps to remove irrelevant degradation problems from the training sets, thus improves the efficiency of network training. In Table 3, we present some samples of three different training sets generated by the proposed method. In the simulation, the input clear images are from the non-degraded real-world images in the EUVP dataset [10], and the targeted real-world images are from the degraded real-world images in the EUVP dataset [10] and the UIEB dataset [12]. Here, the targeted images are specially selected from very different underwater scenes and grouped into three sets based on their degradation types, i.e., bluish color deviation in Set-A, strong scattering in Set-B, and insufficient lighting in Set-C. Obviously, the proposed method manages to simulate all the targeted degradation problems in the corresponding training set images, which verifies its effectiveness in generating specialized training sets for the targeted real-world images. To the best of our knowledge, it is the first method that can simulate underwater images with insufficient lighting and produce customized training sets for targeted underwater images.

4. A Modular CNN for Learning to Enhance Degraded Underwater Images

In this section, we provide a detailed introduction of the architecture and object function of a modular CNN, which is to be trained with the generated training sets for enhancing targeted degraded images.

4.1. Network Architecture

The enhancement of a degraded underwater image requires to improve the overall quality of the image without losing useful information or introducing artifacts. To deal with the degradation problems presented in different image scales, we use the backbone of encoder-decoder structure to build a modular CNN to extract image features in different resolutions and construct the enhanced images. The architecture of the proposed network is shown in Figure 6.
The proposed network consists of 8 encoder blocks and 8 decoder blocks, and each pair but the last one has a skip connection that sends the output of encoder block to decoder block. All encoder blocks have same structure, and so do the first 7 blocks in the decoder part. The last decoder block produces enhancement results and has a special structure. Details of the encoder block, common decoder block, and output decoder block are shown in Figure 7.
As shown in Figure 7, the encoder block contains two convolution layers together with two sets of instance normalization [43] and leaky ReLU [44] operations for feature map normalization and nonlinear activation. The negative slope of leaky ReLU is set as 0.2. The first convolution layer in the encoder block is 4 × 4 stride 2 convolution layer, which extracts feature maps with halved resolution from the input. The number of output feature maps of the first convolution layer starts from 64 and doubles gradually until reaching 512. After the first set of operations, we use a 3 × 3 stride 1 convolution layer for producing deeper features with same resolution and number. To ensure that useful information is well preserved in the block output, we also build an open path by shortcut connection [14] in this set of operations.
The decoder block is also composed of two sets of convolution, instance normalization, and leaky ReLU operations, but at the beginning of the first set, we add a bilinear interpolation operation to up-sample the input feature maps, then use a 3 × 3 stride 1 convolution to get deeper feature maps from the upscaled input. The number of output feature maps is changed inversely to that in the encoder part. Considering the possible information loss in the long path of feature map transmission in the backbone of the network, we also add skip connections to directly pass the output feature maps of the encoder blocks to the corresponding decoder blocks, as done in the U-Net [13]. More specifically, the skip connection creates a bypath for the feature maps produced by the encoder block so that they can be directly concatenated after the decoded feature maps of the same resolution as them and fed into the following decoder block of the next resolution level.
The output block is similar to the decoder block but uses tanh operation for nonlinear activation after the second convolution operation and directly generates output images from it. The default value range of output images in the network is [−1, 1], so an intensity scaling operation is added to change the value range to [0, 1].

4.2. Objective Function

To improve the accuracy of the output images produced by the proposed network, we use a combination of a local loss function and a perceptually motivated loss function to constrain the network training process, which has been proven to be effective by experiments on similar vision tasks such as image denoising and demosaicking in [45].
The local loss function evaluates the pixel-wise difference between the network outputs and the ground-truth images, and mainly penalizes errors in color and brightness. In this work, we use L1 loss function to calculate the local loss according to [45]. The formula of L1 loss function is
L 1 = 1 N x I o u t x , y J x , y ,
where I o u t is the output image of the proposed network, J is the ground truth image (i.e., the clear image in the simulation process), and N is the total pixel number.
The perceptually motivated loss function is complementary to the local loss function and helps to improve the quality of output images in terms of characters sensitive to human visual systems. In this work, we use the structure similarity metric (SSIM) to evaluate the perceptual error of the image in the network training process. The SSIM of the output image with respect to the ground truth image is calculated as follows.
S S I M x , y = 2 μ I o u t x , y μ J x , y + C 1 μ I o u t 2 x , y + μ J 2 x , y + C 1 · 2 σ I o u t J x , y + C 2 σ I o u t 2 x , y + σ J 2 x , y + C 2
Here, μ I o u t x , y and μ J x , y are the average values of pixels intensities in a local region of x , y in the output image and ground truth image, respectively. Similarly, σ I o u t 2 x , y and σ J 2 x , y are the variances of these local region pixel intensities in the output image and ground truth image. σ I o u t J x , y is the covariance of pixel intensities in these local regions of the output image and the ground truth image. The size of local regions is set as 11 × 11 . C 1 and C 2 are constants and are set as C 1 = 0.01 L 2 and C 2 = 0.03 L 2 , where L represents the intensity range of images. Since the SSIM value grows with the improvement of image similarity, the loss function of SSIM is simply built as
L S S I M = 1 1 N   x S S I M x .
The final objective function is the linear combination of the L1 loss and the SSIM loss, i.e.,
L = L 1 + L S S I M .

5. Experimental Results

This section conducts a thorough evaluation of the proposed method, including the performance of proposed deep-learning network, the utility of proposed simulation method, and the performance of the whole enhancement method with both network and simulated data in use.
In the preparation of simulation dataset for network training, 10,350 clear images selected from EUVP dataset [10] and Microsoft COCO dataset [46] were used as inputs, and produced a simulation dataset with over 10,000 image pairs after selection. All images were regulated to the size of 256 × 256 . The simulation dataset was split into training set and validation set by the ratio of 9:1. In the network-training process, the proposed network was trained over 200 epochs with the Adam optimizer ( β 1 = 0.9 ,   β 2 = 0.999 ) . The batch size of training data was set as 24, and the learning rate was set as 0.0002. The proposed network was implemented on the PyTorch framework. All experiments were conducted on a workstation with an Intel Xeon E5-1650 CPU, 48 GB RAM, and two NVIDIA TITAN X GPUs.
The performance of the proposed method was compared against three state-of-the-art learning-based methods and two non-learning methods, which are All-In-One [27], FUnIE-GAN [10], Water-Net [12], Liu’s method [25], and Zhang’s method [19]. The former three are learning-based methods, and are trained with Li’s dataset [11], the EUVP dataset [10], and the UIEB dataset [12], respectively. The latter two are non-learning methods. More specifically, the second to last method is based on physical model, and the last method is built with general image processing skills. To control variables in the evaluation of proposed network and simulation method, networks from the three learning-based methods were retrained on the same simulation dataset produced by the proposed simulation method and generated three new network entities, named All-In-One*, FUnIE-GAN*, and Water-Net*, respectively.

5.1. Evaluation on Simulated Underwater Images with Various Degradation Problems

The first experiment was conducted on simulated underwater images produced by the proposed simulation method. To evaluate the performance of proposed network and simulation method thoroughly, all aforementioned learning-based methods and their retrained entities were included in this experiment, where the performance of proposed network was evaluated by comparing it with All-In-One*, FUnIE-GAN*, and Water-Net*, and the difference of the dataset produced by proposed simulation method and the three datasets used in training All-In-One, FUnIE-GAN, and Water-Net (i.e., Li’s dataset, EUVP dataset, and UIEB dataset) was evaluated by comparing All-In-One, FUnIE-GAN and Water-Net with All-In-One*, FUnIE-GAN*, and Water-Net*, respectively. Underwater images used in this experiment cover a wide range of degradation problems, including color deviation, scattering, and insufficient lighting. In Figure 8, samples of input underwater images and the enhancement results are presented. For objective evaluation, the quality of enhance images were also evaluated by quantitative metrics, which were MSE (mean square error), PSNR (peak signal to noise ratio), and SSIM. The quantitative evaluation results are shown in Table 4.
As shown in Figure 8 and Table 4, the proposed network produced images very similar to the ground truth and had much better scores than All-In-One*, FUnIE-GAN* and Water-Net*, which shared a same training set with the proposed network. Since the tested images were also produced by the proposed simulation method, this experimental result proves that the proposed network is able to learn the underlying relationship between the image pairs in the training set and has a better performance than the other three networks. The average runtime of each network is also presented in Table 4, which shows that the proposed network has a medium speed among all tested networks.
The difference between the simulation dataset produced by the proposed method and the three other datasets used in All-In-One, FUnIE-GAN and Water-Net is also very clear. As can been seen in Figure 8, images produced by entities with same network structure and different training data (i.e., All-In-One vs. All-In-One*, FUnIE-GAN vs. FUnIE-GAN*, Water-Net vs. Water-Net vs. Water-Net) are very different. In Table 4, entities trained with the simulation dataset produced by proposed method have better scores. This experimental result does not necessarily prove the superiority of the proposed simulation method, but reveals that the dataset produced by the proposed method has a different distribution with the other three dataset (i.e., Li’s dataset [11], EUVP dataset [10], and UIEB dataset [12]).

5.2. Evaluation on Real-World Underwater Images with Various Degradation Problems

The second experiment evaluated the performance of proposed method on real underwater images. More specifically, the performance of the whole proposed method was evaluated by comparing it with all aforementioned methods in this section, including six network entities and two non-learning methods, and the applicability of different training datasets was assessed by comparing the performance of all learning-based methods. For reproducibility, all tested images in this experiment were collected from open sources such as published datasets (e.g., EUVP dataset [10] and the UIEB dataset [12]) and online sharing websites (e.g., Flickr https://www.flickr.com (accessed on 20 August 2020)). Samples of experimental results are shown in Figure 9. To evaluate the performance of each method in challenging cases, we especially select some severely degraded images in this experiment.
Since ground-truth images are not available, quantitative evaluation is conducted by using metrics that describe the color deviation, edge intensity, and information content of images, so as to quantify the improvement of image quality for images with different degradation problems after enhancement. The evaluation results of all tested methods are presented in Table 5. We did not adopt existing non-reference quality measurements of underwater, such as UCIQE [47] or UIQM [48], because their utilities are still in doubt, especially in evaluating underwater images with insufficient lighting [2,3]. The color deviation metric used in this experiment is from Li’s work [49], where the average values and standart deviations of channel a and b of the Lab color space were used to evaluate the degree of color cast of an image. The edge intensity metric evaluates the richness of details in an image, and was obtained by calculating the average intensity of Sobel edge map of the image. The information content metric was evaluated by entropy, which evaluates the richness of grayscales in the image. The latter two metrics are complementary in evaluating the visibility of details in the image.
In Figure 10 and Table 6, we also present a simple comparison of the UIQM metrics (including UICM for colorfulness, UISM for sharpness, UIConM for contrast, and UIQM for the whole image) and the three metrics used in this study. Sample images shown in Figure 10 are taken from Figure 5, including a high-quality image and three simulated degraded images. Clearly, the three metrics used in this study correctly assigned high scores to high-quality images, while the UISM metric seemed to prefer images with uneven lighting even though it does not have richer details.
As shown by Figure 9 and Table 5, the proposed method significantly improved the quality of input underwater images. Qualitatively speaking, it restored the deviated color tones in tested images, removed hazy looks, and improved the degrees of brightness and visibility in low-light regions of the tested images. The quantitative result in Table 5 is consistent with the qualitative result, where output images of the proposed method get much better scores than the original images.
The two non-learning methods, i.e., Liu’s method [25] and Zhang’s method [19], also improved the quality of tested images, but they tended to produce pseudo color in the enhancement results of images with strong color deviation, such as those in the 3rd and 4th rows, and could not fully restore the illumination balance in low-light images as shown in the 4th and last row of Figure 9. Accordingly, their scores of edge intensity and information content were lower than the proposed method. The color deviation score of Liu’s method [25] was also worse than the proposed method, and Zhang’s method [19] obtained a better score in color deviation than the proposed method mainly due to the high color-richness of its result images.
The three original learning-based methods, i.e., All-In-One [27], FUnIE-GAN [10] and Water-Net [12], did not perform well in this experiment. All-In-One and FUnIE-GAN failed in almost all tests and produced false color and textures in the enhanced images. Water-Net could not improve the visibility of low-light regions in the tested images. Their scores were also worse than those of the proposed method and the two non-learning methods.
The three retrained networks, i.e., All-In-One*, FUnIE-GAN*, and Water-Net*, performed better than the corresponding original networks but worse than the proposed method. In Figure 9, all result images of the retrained networks have more balanced color than those of their original counterparts, and All-In-One* and Water-Net* also managed to improve the visibility of low light regions in the tested images. The quantitative results of All-In-One* and FUnIE-GAN* are also better than their corresponding original network as shown in Table 5. Water-Net* sees better result in terms of edge intensity but worse scores in color deviation and information content, which might be due to its low color richness and small intensity range.
Overall, the proposed method shows its advantages against the competing methods in this experiment and is especially good at improving the visual quality of low-light underwater images. Networks trained with a simulation dataset generated by the proposed method also performed better, which implies the superiority of the proposed image simulation method in generating realistic images for training underwater image enhancement networks.

6. Conclusions

In this paper, we proposed a complete method for enhancing underwater images with different degradation problems. To utilize the superpower of deep learning, we first proposed a training set generating method to produce specified and reliable training data for the targeted underwater images, then proposed a convolutional neural network for translating degraded underwater images to their clear counterparts by learning from the image pairs in the training set. Experiments on underwater images with different degradation problems demonstrated the effectiveness of the proposed method and its advantages against many state-of-the-art methods, especially in enhancing images from challenging underwater scenes, such as the deep sea with insufficient lighting. In the future, we plan to work on the improvement of the network structure and increase the diversity of image pairs in the training set, so as to further improve the efficiency and performance of the proposed method.

Author Contributions

Conceptualization, Y.L., H.X. and K.S.; methodology, Y.L.; software, Y.L. and C.L.; validation, B.Z., J.Y. and B.L.; formal analysis, X.Q.; investigation, B.Z. and B.L.; resources, H.X. and K.S.; data curation, Y.L. and C.L.; writing—original draft preparation, Y.L.; writing—review and editing, H.X. and K.S.; visualization, B.Z.; supervision, H.X.; project administration, C.L.; funding acquisition, B.Z. and K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China, grant number 2018YFC0307905 and the National Natural Science Foundation of China, grant number 61972240.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank the associate editor and the anonymous reviewers for their constructive comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, X.; Yang, Y.-H. A Closed-Form Solution to Single Underwater Camera Calibration Using Triple Wavelength Dispersion and Its Application to Single Camera 3D Reconstruction. IEEE Trans. Image Process. 2017, 26, 4553–4561. [Google Scholar] [CrossRef] [PubMed]
  2. Lu, H.; Uemura, T.; Wang, D.; Zhu, J.; Huang, Z.; Kim, H. Deep-Sea Organisms Tracking Using Dehazing and Deep Learning. Mob. Netw. Appl. 2018, 25, 1008–1015. [Google Scholar] [CrossRef]
  3. Agrafiotis, P.; Drakonakis, G.I.; Georgopoulos, A.; Skarlatos, D. The Effect of Underwater Imagery Radiometry on 3D Reconstruction and Orthoimagery. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, XLII-2/W3, 25–31. [Google Scholar] [CrossRef] [Green Version]
  4. Chuang, M.-C.; Hwang, J.-N.; Williams, K. A Feature Learning and Object Recognition Framework for Underwater Fish Images. IEEE Trans. Image Process. 2016, 25, 1862–1872. [Google Scholar] [CrossRef]
  5. Lee, D.; Kim, G.; Kim, D.; Myung, H.; Choi, H.-T. Vision-based object detection and tracking for autonomous navigation of underwater robots. Ocean Eng. 2012, 48, 59–68. [Google Scholar] [CrossRef]
  6. Wang, Y.; Song, W.; Fortino, G.; Qi, L.-Z.; Zhang, W.; Liotta, A. An Experimental-Based Review of Image Enhancement and Image Restoration Methods for Underwater Imaging. IEEE Access 2019, 7, 140233–140251. [Google Scholar] [CrossRef]
  7. Yang, M.; Hu, J.; Li, C.; Rohde, G.; Du, Y.; Hu, K. An In-Depth Survey of Underwater Image Enhancement and Restoration. IEEE Access 2019, 7, 123638–123657. [Google Scholar] [CrossRef]
  8. Forbes, T.; Goldsmith, M.; Mudur, S.; Poullis, C. DeepCaustics: Classification and Removal of Caustics from Underwater Imagery. IEEE J. Ocean. Eng. 2018, 44, 728–738. [Google Scholar] [CrossRef]
  9. Reggiannini, M.; Moroni, D. The Use of Saliency in Underwater Computer Vision: A Review. Remote Sens. 2020, 13, 22. [Google Scholar] [CrossRef]
  10. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef] [Green Version]
  11. Li, C.; Anwar, S.; Porikli, F. Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recognit. 2020, 98, 107038. [Google Scholar] [CrossRef]
  12. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2020, 29, 4376–4389. [Google Scholar] [CrossRef] [Green Version]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  15. Iqbal, K.; Salam, R.A.; Osman, A.; Talib, A.Z. Underwater Image Enhancement Using an Integrated Colour Model. IAENG Int. J. Comput. Sci. 2007, 32, 239–244. [Google Scholar]
  16. Ancuti, C.O.; Ancuti, C.; Haber, T.; Bekaert, P. Fusion-based restoration of the underwater images. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 1557–1560. [Google Scholar] [CrossRef]
  17. Ancuti, C.O.; Ancuti, C.; De Vleeschouwer, C.; Bekaert, P. Color Balance and Fusion for Underwater Image Enhancement. IEEE Trans. Image Process. 2018, 27, 379–393. [Google Scholar] [CrossRef] [Green Version]
  18. Fu, X.; Zhuang, P.; Huang, Y.; Liao, Y.; Zhang, X.-P.; Ding, X. A retinex-based enhancing approach for single underwater image. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 4572–4576. [Google Scholar] [CrossRef]
  19. Zhang, W.; Li, G.; Ying, Z. A new underwater image enhancing method via color correction and illumination adjustment. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar] [CrossRef]
  20. Peng, Y.-T.; Cosman, P. Underwater Image Restoration Based on Image Blurriness and Light Absorption. IEEE Trans. Image Process. 2017, 26, 1579–1594. [Google Scholar] [CrossRef]
  21. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [CrossRef]
  22. Galdran, A.; Pardo, D.; Picon, A.; Alvarez-Gila, A. Automatic Red-Channel underwater image restoration. J. Vis. Commun. Image Represent. 2015, 26, 132–145. [Google Scholar] [CrossRef] [Green Version]
  23. Drews, P., Jr.; do Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission Estimation in Underwater Single Images. In Proceedings of the 2013 IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 2–8 December 2013; pp. 825–830. [Google Scholar] [CrossRef]
  24. Carlevaris-Bianco, N.; Mohan, A.; Eustice, R.M. Initial results in underwater single image dehazing. In Proceedings of the OCEANS 2010 MTS/IEEE SEATTLE, Seattle, WA, USA, 20–23 September 2010; pp. 1–8. [Google Scholar]
  25. Liu, Y.; Xu, H.; Shang, D.; Li, C.; Quan, X. An Underwater Image Enhancement Method for Different Illumination Conditions Based on Color Tone Correction and Fusion-Based Descattering. Sensors 2019, 19, 5567. [Google Scholar] [CrossRef] [Green Version]
  26. Marques, T.P.; Albu, A.B.; Hoeberechts, M. A Contrast-Guided Approach for the Enhancement of Low-Lighting Underwater Images. J. Imaging 2019, 5, 79. [Google Scholar] [CrossRef] [Green Version]
  27. Uplavikar, P.; Wu, Z.; Wang, Z. All-In-One Underwater Image Enhancement using Domain-Adversarial Learning. arXiv 2019, arXiv:1905.13342. [Google Scholar]
  28. Wang, N.; Zhou, Y.; Han, F.; Zhu, H.; Zheng, Y. UWGAN: Underwater GAN for Real-world Underwater Color Restoration and Dehazing. arXiv 2019, arXiv:1912.10269. [Google Scholar]
  29. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef] [Green Version]
  30. Li, C.; Anwar, S.; Hou, J.; Cong, R.; Guo, C.; Ren, W. Underwater Image Enhancement via Medium Transmission-Guided Multi-Color Space Embedding. IEEE Trans. Image Process. 2021, 30, 4985–5000. [Google Scholar] [CrossRef]
  31. Fabbri, C.; Islam, J.; Sattar, J. Enhancing Underwater Imagery Using Generative Adversarial Networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 7159–7165. [Google Scholar] [CrossRef] [Green Version]
  32. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  33. Guo, Y.; Li, H.; Zhuang, P. Underwater Image Enhancement Using a Multiscale Dense Generative Adversarial Network. IEEE J. Ocean. Eng. 2020, 45, 862–870. [Google Scholar] [CrossRef]
  34. Liu, X.; Gao, Z.; Chen, B.M. MLFcGAN: Multilevel Feature Fusion-Based Conditional GAN for Underwater Image Color Correction. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1488–1492. [Google Scholar] [CrossRef] [Green Version]
  35. Jaffe, J. Computer modeling and the design of optimal underwater imaging systems. IEEE J. Ocean. Eng. 1990, 15, 101–111. [Google Scholar] [CrossRef]
  36. McGlamery, B.L. A Computer Model for Underwater Camera Systems. SPIE Proc. 1980, 208, 221–231. [Google Scholar] [CrossRef]
  37. Schechner, Y.; Karpel, N. Clear underwater vision. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; Volume 1, pp. 536–543. [Google Scholar] [CrossRef] [Green Version]
  38. Schechner, Y.Y.; Karpel, N. Recovery of Underwater Visibility and Structure by Polarization Analysis. IEEE J. Ocean. Eng. 2005, 30, 570–587. [Google Scholar] [CrossRef] [Green Version]
  39. Akkaynak, D.; Treibitz, T.; Shlesinger, T.; Loya, Y.; Tamir, R.; Iluz, D. What is the Space of Attenuation Coefficients in Underwater Computer Vision? In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 568–577. [Google Scholar]
  40. Zhao, X.; Jin, T.; Qu, S. Deriving inherent optical properties from background color and underwater image enhancement. Ocean Eng. 2015, 94, 163–172. [Google Scholar] [CrossRef]
  41. Mobley, C.D. Light and Water: Radiattive Transfer in Natural Waters; Academic Press: New York, NY, USA, 1994. [Google Scholar]
  42. Li, Z.; Snavely, N. MegaDepth: Learning Single-View Depth Prediction from Internet Photos. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2041–2050. [Google Scholar]
  43. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance Normalization: The Missing Ingredient for Fast Stylization. arXiv 2016, arXiv:1607.08022. [Google Scholar]
  44. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th ICML Workshop on Deep Learning for Audio, Speech and Language Processing, Atlanta, GA, USA, 16–21 June 2013; Volume 28. [Google Scholar]
  45. Zhao, H.; Gallo, O.; Frosio, I.; Kautz, J. Loss Functions for Image Restoration with Neural Networks. IEEE Trans. Comput. Imaging 2017, 3, 47–57. [Google Scholar] [CrossRef]
  46. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In European Conference on Computer Vision, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; Volume 8693, pp. 740–755. [Google Scholar]
  47. Yang, M.; Sowmya, A. An Underwater Color Image Quality Evaluation Metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
  48. Panetta, K.; Gao, C.; Agaian, S. Human-Visual-System-Inspired Underwater Image Quality Measures. IEEE J. Ocean. Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  49. Li, F.; Wu, J.; Wang, Y.; Zhao, Y.; Zhang, X. A color cast detection algorithm of robust performance. In Proceedings of the 2012 IEEE Fifth International Conference on Advanced Computational Intelligence (ICACI), Nanjing, China, 18–20 October 2012; pp. 662–664. [Google Scholar]
Figure 1. Underwater images with different degradation conditions. (a) An underwater image with bluish color deviation. (b) An underwater image with low contrast. (c) An underwater image with insufficient illumination. (d) An underwater image with no visible degradation.
Figure 1. Underwater images with different degradation conditions. (a) An underwater image with bluish color deviation. (b) An underwater image with low contrast. (c) An underwater image with insufficient illumination. (d) An underwater image with no visible degradation.
Information 13 00187 g001
Figure 2. Approximating complicated real-world underwater illumination with the combination of uniform lighting, parallel lighting, and artificial lighting.
Figure 2. Approximating complicated real-world underwater illumination with the combination of uniform lighting, parallel lighting, and artificial lighting.
Information 13 00187 g002
Figure 3. The difference of arbitrary and fixed light modes. (a) Diagram of artificial lighting with arbitrary orientation. (b) Diagram of artificial lighting with orthographic orientation. In the diagrams, X l , Y l , Z l represents the location of light source, x l , y l represents the project of light center, X c , Y c , Z c is the location of camera, x , y is the location of a photographed scene point, d represents the distance from scene point to camera, D is the distance from scene point to light source.
Figure 3. The difference of arbitrary and fixed light modes. (a) Diagram of artificial lighting with arbitrary orientation. (b) Diagram of artificial lighting with orthographic orientation. In the diagrams, X l , Y l , Z l represents the location of light source, x l , y l represents the project of light center, X c , Y c , Z c is the location of camera, x , y is the location of a photographed scene point, d represents the distance from scene point to camera, D is the distance from scene point to light source.
Information 13 00187 g003
Figure 4. Examples of simulated underwater lighting parameters. (a) A clear image used to calculate scene structure (taken from the EUVP dataset). (b) Simulated transmission distance d x , y of the scene (method of simulating d x , y is given in the following parts). (c) Simulation results of underwater lighting parameters generated by using the transmission distance d x , y in (b).
Figure 4. Examples of simulated underwater lighting parameters. (a) A clear image used to calculate scene structure (taken from the EUVP dataset). (b) Simulated transmission distance d x , y of the scene (method of simulating d x , y is given in the following parts). (c) Simulation results of underwater lighting parameters generated by using the transmission distance d x , y in (b).
Information 13 00187 g004
Figure 5. Samples of simulated underwater images generated by the proposed method. The first column presents clear images taken from the EUVP datasets, and the second to last columns are underwater images simulated based on the clear images in the first column.
Figure 5. Samples of simulated underwater images generated by the proposed method. The first column presents clear images taken from the EUVP datasets, and the second to last columns are underwater images simulated based on the clear images in the first column.
Information 13 00187 g005
Figure 6. Architecture of the proposed underwater image enhancement network. Dotted-line frame (a) is an encoder block, (b) is a decoder block, (c) is the output block. All encoder blocks have same structure, and so are the decoder blocks but the last output block. Details of the encoder block, decoder block, and output block are given in Figure 7.
Figure 6. Architecture of the proposed underwater image enhancement network. Dotted-line frame (a) is an encoder block, (b) is a decoder block, (c) is the output block. All encoder blocks have same structure, and so are the decoder blocks but the last output block. Details of the encoder block, decoder block, and output block are given in Figure 7.
Information 13 00187 g006
Figure 7. Detailed structure diagrams of basic blocks in the proposed network. (a) The structure of encoder block (the dotted-line frame (a) in Figure 6). (b) The structure of decoder block (the dotted-line frame (b) in Figure 6). (c) The structure of output block (the dotted-line frame (c) in Figure 6).
Figure 7. Detailed structure diagrams of basic blocks in the proposed network. (a) The structure of encoder block (the dotted-line frame (a) in Figure 6). (b) The structure of decoder block (the dotted-line frame (b) in Figure 6). (c) The structure of output block (the dotted-line frame (c) in Figure 6).
Information 13 00187 g007
Figure 8. Qualitative evaluation on simulated underwater images. From left to right are original underwater images, enhanced images produced by each network entity, and the corresponding ground-truth images.
Figure 8. Qualitative evaluation on simulated underwater images. From left to right are original underwater images, enhanced images produced by each network entity, and the corresponding ground-truth images.
Information 13 00187 g008
Figure 9. Qualitative comparisons on real-world underwater images. From left to right are original underwater images and enhanced images produced by each tested enhancement method.
Figure 9. Qualitative comparisons on real-world underwater images. From left to right are original underwater images and enhanced images produced by each tested enhancement method.
Information 13 00187 g009
Figure 10. Sample images used for comparing the UIQM metrics and the three metrics used in this study. These images are taken from Figure 5, where (a) is the high-quality image and (bd) are simulated underwater images with various degradation problems. Corresponding evaluation results are presented in Table 6.
Figure 10. Sample images used for comparing the UIQM metrics and the three metrics used in this study. These images are taken from Figure 5, where (a) is the high-quality image and (bd) are simulated underwater images with various degradation problems. Corresponding evaluation results are presented in Table 6.
Information 13 00187 g010
Table 1. Color tones and HSVs of real-world underwater images and simulated underwater images.
Table 1. Color tones and HSVs of real-world underwater images and simulated underwater images.
Real-World Underwater ImagesSimulated Underwater Images
Images Information 13 00187 i001 Information 13 00187 i002 Information 13 00187 i003 Information 13 00187 i004 Information 13 00187 i005 Information 13 00187 i006 Information 13 00187 i007 Information 13 00187 i008
Color tones Information 13 00187 i009 Information 13 00187 i010 Information 13 00187 i011 Information 13 00187 i012 Information 13 00187 i013 Information 13 00187 i014 Information 13 00187 i015 Information 13 00187 i016
HSVs[185, 0.54, 0.56][151, 0.45, 0.33][61, 0.37, 0.50][75, 0.85, 0.28][172, 0.40, 0.50][150, 0.57, 0.48][48, 0.29, 0.62][48, 0.57, 0.35]
Table 2. Sobel edge maps of a clear image and simulated degraded images and their NMAD scores.
Table 2. Sobel edge maps of a clear image and simulated degraded images and their NMAD scores.
Clear ImageSimulated Underwater Images
Images Information 13 00187 i017 Information 13 00187 i018 Information 13 00187 i019 Information 13 00187 i020 Information 13 00187 i021 Information 13 00187 i022
Sobel-edge maps Information 13 00187 i023 Information 13 00187 i024 Information 13 00187 i025 Information 13 00187 i026 Information 13 00187 i027 Information 13 00187 i028
NMADs-0.340.590.930.620.78
Table 3. Sample images of three training sets generated by the proposed method.
Table 3. Sample images of three training sets generated by the proposed method.
Group Targeted Image SamplesReal-World Clear ImagesSimulated Images after Selection
Set-A Information 13 00187 i029 Information 13 00187 i030 Information 13 00187 i031
Set-B Information 13 00187 i032 Information 13 00187 i033 Information 13 00187 i034
Set-C Information 13 00187 i035 Information 13 00187 i036 Information 13 00187 i037
Table 4. Quantitative evaluation on simulated underwater images. means smaller scores are better. means larger scores are better.
Table 4. Quantitative evaluation on simulated underwater images. means smaller scores are better. means larger scores are better.
MSE↓PSNR↑SSIMTime (ms)
Original0.0217 16.6390 0.5343 -
All-In-One0.0170 17.7063 0.6543 13.9641
All-In-One*0.0053 22.7201 0.8050 -
FUnIE-GAN0.0266 15.7483 0.5671 7.5361
FUnIE-GAN*0.0211 16.7601 0.6969 -
Water-Net0.0180 17.4526 0.6913 80.4657
Water-Net*0.0176 17.5337 0.6894 -
Proposed0.002925.42090.861627.7387
Table 5. Quantitative evaluation on real-world underwater images. means smaller scores are better. means larger scores are better.
Table 5. Quantitative evaluation on real-world underwater images. means smaller scores are better. means larger scores are better.
Color Deviation↓Edge Intensity↑Information Content↑
Original2.90420.12866.8242
All-In-One1.59170.19416.9834
All-In-One*1.22180.19657.2118
FUnIE-GAN2.42290.15027.1567
FUnIE-GAN*1.32530.23807.3171
Water-Net0.85200.16037.2225
Water-Net*1.04760.17017.0262
Liu’s method0.43120.25937.3229
Zhang’s method0.11320.24567.4529
Proposed0.29880.26647.5247
Table 6. Comparison of UIQM metrics and the three metrics used in this study on images from Figure 10. means smaller scores are better. means larger scores are better.
Table 6. Comparison of UIQM metrics and the three metrics used in this study on images from Figure 10. means smaller scores are better. means larger scores are better.
ImageUICM↑UISM↑UIConM↑UIQM↑Edge Intensity↑Color Deviation↓Information Content↑
Figure 10a6.125213.82330.28335.26770.17870.86227.2844
Figure 10b3.654314.18870.27545.27750.15392.33817.1077
Figure 10c0.800713.00620.20134.58320.06936.00976.2114
Figure 10d3.608114.81730.17765.11240.11501.22316.4453
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, Y.; Xu, H.; Zhang, B.; Sun, K.; Yang, J.; Li, B.; Li, C.; Quan, X. Model-Based Underwater Image Simulation and Learning-Based Underwater Image Enhancement Method. Information 2022, 13, 187. https://doi.org/10.3390/info13040187

AMA Style

Liu Y, Xu H, Zhang B, Sun K, Yang J, Li B, Li C, Quan X. Model-Based Underwater Image Simulation and Learning-Based Underwater Image Enhancement Method. Information. 2022; 13(4):187. https://doi.org/10.3390/info13040187

Chicago/Turabian Style

Liu, Yidan, Huiping Xu, Bing Zhang, Kelin Sun, Jingchuan Yang, Bo Li, Chen Li, and Xiangqian Quan. 2022. "Model-Based Underwater Image Simulation and Learning-Based Underwater Image Enhancement Method" Information 13, no. 4: 187. https://doi.org/10.3390/info13040187

APA Style

Liu, Y., Xu, H., Zhang, B., Sun, K., Yang, J., Li, B., Li, C., & Quan, X. (2022). Model-Based Underwater Image Simulation and Learning-Based Underwater Image Enhancement Method. Information, 13(4), 187. https://doi.org/10.3390/info13040187

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop