Next Article in Journal
New Explicit Propagating Solitary Waves Formation and Sensitive Visualization of the Dynamical System
Next Article in Special Issue
Application of Empirical Bode Analysis for Delay-Margin Evaluation of Fractional-Order PI Controller in a Renewable Distributed Hybrid System
Previous Article in Journal
LFM Signal Parameter Estimation via FTD-FRFT in Impulse Noise
Previous Article in Special Issue
Analysis of the Charge Density Variation Caused by the Physical Properties of the Electrodes of Lithium-Ion Batteries
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder

School of Marine Science and Technology, Northwestern Polytechnical University, Xi’an 710068, China
*
Author to whom correspondence should be addressed.
Fractal Fract. 2023, 7(1), 70; https://doi.org/10.3390/fractalfract7010070
Submission received: 12 November 2022 / Revised: 30 December 2022 / Accepted: 2 January 2023 / Published: 9 January 2023
(This article belongs to the Special Issue Fractional-Order Circuits, Systems, and Signal Processing)

Abstract

:
As an essential low-level computer vision task for remotely operated underwater robots and unmanned underwater vehicles to detect and understand the underwater environment, underwater image enhancement is facing challenges of light scattering, absorption, and distortion. Instead of using a specific underwater imaging model to mitigate the degradation of underwater images, we propose an end-to-end underwater-image-enhancement framework that combines fractional integral-based Retinex and an encoder–decoder network. The proposed variant of Retinex aims to alleviate haze and color distortion in the input image while preserving edges to a large extent by utilizing a modified fractional integral filter. The encoder–decoder network with channel-wise attention modules trained in an unsupervised manner to overcome the lack of paired underwater image datasets is designed to refine the output of the Retinex. Our framework was evaluated under qualitative and quantitative metrics on several public underwater image datasets and yielded satisfactory enhancement results on the evaluation set.

1. Introduction

About 70% of the earth’s surface is covered by water, so exploring and utilizing marine resources benefits humankind greatly. However, vision-based underwater detection tasks face many challenges caused by the poor quality of underwater images. On the one hand, underwater images always present awful visibility due to the scattering and absorption of light while it propagates in the water. On the other hand, the more rapid attenuation of red light than the green and blue light in the water results in severe distortion of color rendition in digital images.
Similar to the physical dehazing model [1,2] on land, there are some classical underwater image restoration algorithms based on physical models, for example, methods based on the Jaffe–McGlamery underwater imaging model [3,4]. Sea-thru [5] is a representative algorithm that combines the physical model and the principle of digital image processing. Researchers from the Univ. of Haifa developed a model that takes an RGB-D image as input, estimates backscatter in a way inspired by the dark channel prior, and uses an optimization framework to obtain the range-dependent attenuation coefficient. However, some models based on the physical assumption of underwater imaging rely on extra devices, such as the Laser Underwater Camera Imaging Enhancer [6] and Coulter Counter [7], to acquire specific parameters, and even the Sea-thru requires an additional depth sensor. Thus, computer-vision algorithms that are independent of expensive measurements are preferred for general underwater image enhancement tasks.
First introduced by Land and McCann [8], the Retinex theory has been extensively applied in multiple research areas such as dehazing, enhancement of remote sensing images, and underwater images. One of the most noted variants of Retinex, called Multi-Scale Reinex [9], captured light changes under different scales and removed the light from the input image to simultaneously achieve dynamic range compression, color consistency, and lightness rendition. A color restoration function based on empirical parameters was applied to the output after Retinex processing to recover authentic color from the degraded image. As most Retinex algorithms employ Gaussian operators to estimate the light image, note that underwater images are generally used for close-range detection of the reef, shellfish, ancient ruins, and underwater vehicles, and sharp edges naturally exist in them. Hence, the Gaussian filtering operator used by classical MSR algorithms may cause over-smoothing and a lack of details. According to Qi Wang et al. [10], fractional calculus operators excel at processing information with weak derivatives that the texture structure contains. More specifically, the fractional differential operator has the characteristics of memorizing and a higher signal-to-noise ratio than the integer differential operator, which enables the fractional operator to capture more detailed information. For the fractional integral operator, it attenuates the high-frequency portion of the signal dramatically while enhancing the low part to a large extent. The denoising method based on fractional integral was first proposed by Huang et al. [11], who constructed an eight-direction fractional integral operator to filter input noise from the original image directionally. We noticed that most of the fractional integral-based filters were limited to a 3 × 3 or 5 × 5 size, which was not compatible with the Retinex theory that the light image should be estimated by a much larger filter to capture a trend of slow change of the light. In the Multi-Scale Retinex, a three-scale parallel Retinex network always uses 15, 80, and 250 as sigma values for the Gaussian filter to predict the light, leading to filters of size 45 × 45 , 240 × 240 , 750 × 750 , respectively. For such a considerable filter size, the fractional method could either expand the filter kernel size or carry out recurrent filtering. However, the former will not work because the filter became sparse and resulted in a “fringe phenomenon” on the image, while the latter required significant time cost.
With the development of the use of deep learning theory in image processing, approaches based on convolutional neural networks have shown potential for image enhancement tasks. For underwater image dehazing, J. Perez et al. introduced a model based on classical CNN architecture [12]. This work utilized a dataset of pairs of raw and restored images for training a network so that clear images could be restored from degraded inputs. Based on the multi-branch design, the UIE-Net proposed by Yang Wang et al. [13] firstly extracts features from the input image using a sharing network and then uses two subnets—the color correction network (CC-Net) and the haze removal network (HR-Net)—to simultaneously achieving color correction and haze removal. Taking into account that pretraining is necessary for these supervised deep learning-based approaches, a problem occurred, namely that there were limited underwater image datasets available for training the networks. To overcome the lack of paired underwater images, an amount of paired datasets for image enhancement tasks were proposed, for example, the UIEB dataset [14] proposed by Li et al. Containing 950 authentic underwater images under various light conditions with 890 paired images in them, the UIEB dataset uses classical enhancement methods and artificial selection to generate reference underwater images. Although paired image-based datasets boost supervised methods, we should be aware that acquiring absolute ground truth for underwater image enhancement tasks is practically impossible. In fact, since a significant portion of the paired datasets were synthetic, models trained on these datasets yielded undesirable generalization performance and sometimes failed on authentic underwater images.
To surmount the challenges, methods based on unsupervised learning could be a possible solution. One direction for unsupervised underwater image enhancement is to utilize depth-guided networks, represented by the WaterGAN [15]. While the first part of the WaterGAN aims to generate underwater images from the in-air ones, the second part restores underwater images by successively passing the input through a depth estimation network and a color correction network. Taking an underwater image and corresponding output relative depth map of the depth estimation network as input, the color-correction network restores the color of the input image. The other direction is based on pure computer vision, independent of any specific underwater imaging model. Inspired by unsupervised methods on low-light image enhancement, image translation, and style transfer, Lu et al. proposed an adaptive algorithm with multi-scale cycle GAN and dark channel prior (MCycleGAN) [16]. By using the dark channel prior to obtain the transmission map of an image and designing an adaptive loss function to improve underwater image quality, the raw images being multi-scale calculated were able to convert to clear results. The TACL [17] proposed by Risheng Liu et al. utilized a bilateral constrained closed-loop adversarial enhancement module to preserve more informative features and embedded a task-aware feedback module in the enhancement process to narrow the gap between visually-oriented and detection-favorable target images.
Rather than the above-mentioned unsupervised methods, which mostly used GAN to generate enhancement results or classical underwater image restoration algorithms based on specific physical models, in this paper, by combining the Retinex algorithm and unsupervised image enhancement approaches, we propose an end-to-end underwater-image-enhancement framework which generates the enhanced image primarily using the Retinex. The underwater image pre-enhanced by the Retinex will be further improved by an encoder–decoder network trained in an unsupervised style to yield better contrast and luminance enhancement, as well as more satisfactory perceptual performance.
The main contributions of this paper are as follows:
  • A fractional integral-based Retinex and an improved fractional-order integral operator, which eliminated the drawback of classical fractional integral operator for large-kernel filtering and resulted in more accurate estimation for light images, was proposed in this paper.
  • An effective unsupervised encoder–decoder network requiring no adversarial training and yielding perceptually pleasing results was employed to refine the output of the previous Retinex model.
  • Combining the fractional integral-based Retinex and unsupervised autoencoder mentioned above, the proposed end-to-end framework for underwater image enhancement was evaluated on several public datasets and produced impressive results.
The rest of this paper is organized as follows: In Section 2, the mathematical background of fractional double integral, on which the proposed variant of Retinex is based, is briefly introduced. In Section 3, we first propose the FDIF-MSR algorithm and then illustrate an end-to-end framework based on FDIF-MSR and the unsupervised encoder–decoder network. In Section 4, the proposed image enhancement model is evaluated on three public datasets, and some of the results are shown in this paper. In Section 5, the conclusion is given.

2. Mathematical Background

2.1. The Definition of Fractional Derivatives

There are many definitions of fractional-order derivatives. However, Riemann–Liouville (R-L), Grünwald–Letnikov (G-L), and Caputo gave the three most commonly used definitions of fractional derivatives. The Grünwald–Letnikov definition is deduced from the expression of integer-order differential, whereas the other two are derived from the integer-order integral Cauchy formula.
  • Let α be a positive real number. When n 1 α < n , where n is a positive integer, the left-hand Riemann–Liouville fractional derivatives can be written as:
    a D t α f ( t ) = d n d t n ( 1 Γ ( n α ) a t f ( τ ) ( t τ ) α n + 1 d x )
    where α is called the order of the R-L derivative.
  • The Grünwald–Letnikov definition of fractional derivatives is defined as
    a D t α f ( t ) = lim h 0 h α j = 0 [ ( t a ) / h ] ( 1 ) j α j f ( t j h )
    where α j are the binomial coefficients, [ · ] denotes the integer part.
  • We have Caputo’s definition of fractional derivatives, which is defined as:
    a C D t α f ( t ) = 1 Γ ( n α ) a t f ( n ) ( τ ) ( t τ ) α n + 1 d x
There is an equivalent relation between Caputo and R-L derivatives such that for a positive real number α , satisfied 0 n 1 < α < n , if function f ( t ) defined on the interval [ a , b ] has continuous derivatives of n 1 order and integrable, we have
a D t α f ( t ) = a C D t α f ( t ) + j = 0 n 1 f ( j ) ( a ) ( t a ) j α Γ ( 1 + j α )
Similarly, R-L and G-L derivatives are also equivalent when f ( t ) satisfies the above conditions. The G-L definition is the most commonly used for the numerical calculation of fractional derivatives. By extending the binomial coefficients to the field of real numbers, a more general form of the G-L derivatives can be written:
a D t α = lim h 0 h α j = 0 [ ( t a ) h ] ( 1 ) j Γ ( 1 + α ) Γ ( α j + 1 ) Γ ( j + 1 ) f ( t j h )
Note that all definitions of fractional derivatives above are left-handed; the corresponding right-hand derivative of the G-L definition is defined similarly by the expression
a D t α = lim h 0 h α j = 0 [ ( t a ) h ] ( 1 ) j Γ ( 1 + α ) Γ ( α j + 1 ) Γ ( j + 1 ) f ( t + j h )

2.2. Fractional Integral

To extend fractional differential to integral, according to fractional operator theory, we use integral operator I 1 instead of the differential operator D and fractional order ν instead of α . For the G-L definition, the fractional integral of order ν can be written as:
a I t ν = lim h 0 h ν j = 0 [ ( t a ) h ] ( 1 ) j Γ ( 1 ν ) Γ ( 1 ν j ) Γ ( j + 1 ) f ( t j h ) = lim h 0 h ν ( f ( t ) + ν f ( t h ) + ν ( ν + 1 ) 0 f ( t 2 h ) + )
Replace the coefficients of f ( t j h ) with ω j ( ν ) ; the equation becomes
a I t ν = lim h 0 h ν j = 0 [ ( t a ) h ] ω j ( ν ) f ( t j h )
while ω j ( ν ) can be recurrently calculated by
ω 0 ( ν ) = 1 ω j ( ν ) = ( 1 1 ν j ) ω j 1 ( ν )

2.3. Fractional Double Integral

The fractional integral can be enlarged for a double integral. Assuming that x ( a , c ) and y ( b , d ) , for rectangular domain ( a , c ) × ( b , d ) in R 2 , consider α -order fractional integral derived from the left-hand G-L definition:
b I y α ( a I x α f ( x , y ) X L ) Y L = lim h y 0 h α k = 0 [ ( y b ) / h y ] { ω k ( α ) [ lim h x 0 h α j = 0 [ ( x a ) / h x ] ω j ( α ) f ( x j h , y k h ) ] }
For casual signal f ( x , y ) , divide the interval ( a , c ) and ( b , d ) into equal parts using step h x = h y = 1 , which also conforms to the fact that the sampling step on a 2-D image matrix is one unit; then, we can eliminate the limit sign and obtain
b I y α ( a I x α f ( x , y ) X L ) Y L = k = 0 [ y b ] { ω k ( α ) [ j = 0 [ x a ] ω j ( α ) f ( x j h , y k h ) ] }
where ω j ( α ) and ω k ( α ) can be calculated by Equation (9), respectively.

3. Proposed Method

In this section, we first put forward an improved fractional integral filter and then combine it with the MSRCR algorithm. Furthermore, an unsupervised encoder–decoder network is utilized to improve the quality of images processed by the Retinex. The pipeline of the proposed method is shown in Figure 1.

3.1. Fractional Double Integral Filter (FDIF)

A classical eight-direction fractional integral filter is constructed as shown in Figure 2. Recalling Equation (8), we can note that a more simplified version of left fractional single integral by replacing variable h with unit one can be written as:
a I t α = j = 0 t a ω j ( α ) f ( t j h )
Generalizing this simplified equation to eight directions, involving directions along the positive and negative x-axis, positive and negativey -axis, and the 45°, 135°, 225°, and 315° directions along the positive x-axis in the counter-clockwise direction, we were able to construct the eight-direction fractional integral operator.
The coefficients of the classical eight-direction filter are as follows:
W f 0 = 1 W f 1 = ν ( ν + 1 ) 2 W f 2 = ν ( ν + 1 ) ( ν + 2 ) 6 W f m = ν ( ν + 1 ) ( ν + 2 ) ( ν + m 1 ) m !
Specifically when ν = 1 , W f i = 1 where i = 0 , 1 , , m .
As we mentioned above and show in Figure 3, such a filter presents a “fringe phenomenon” when the filter matrix size grows and becomes sparse.
To address the issue, we replace the fractional integral filter by introducing a fractional double integral. According to Equation (11), the coefficients in fractional integral filter kernel are rearranged as in Figure 4. Considering the upper left part of the fractional filter deduced from the lefthand G-L definition of the fractional integral, it is obvious that along the x-axis and y-axis, coefficients are consistent with those used in the eight-direction filter, while they decay exponentially along the diagonal. Mathematically, the upper right part can be obtained by using the right-hand and left-hand fractional integral formula for x and y, and the Equation (11) becomes
b I y α a I x α f ( x , y ) X R , Y L = k = 0 y b ω k ( α ) j = 0 x a ω j ( α ) f ( x + j h , y k h )
The other two parts can be obtained similarly. By using the additivity property of the convolution operation and assuming the four parts of the filter be F i , i = 1 , 2 , 3 , 4 , four quadrants of integral can be calculated by one single kernel
F F D I F = i = 1 4 F i
A normalization operation is performed after the FDIF kernel is calculated to avoid introducing extra energy into the image matrix. A visualized version of the FDIF kernel is shown in Figure 4.

3.2. Multi-Scale Retinex with FDIF

In this section, let us briefly recall the basic structure of the MSRCR algorithm and then optimize it with our FDIF filter.

3.2.1. Retinex Theory

Briefly, the Retinex models the imaging process to show that objects in the image have particular reflection properties, and therefore, it can be found that incident light led to various color performances. The essential purpose of the Retinex model is to remove the influence of the light so that objects’ characteristics can be recovered from the noised image. The primitive Retinex modeled the light as multiplicative noise such that
R ( x , y ) = I ( x , y ) × L ( x , y )
The single-scale Retinex process is given by
log R ( x , y ) = log I ( x , y ) log L ( x , y ) = log I ( x , y ) log [ F ( x , y ) I ( x , y ) ]
where R ( x , y ) represents the enhanced image, I ( x , y ) represents the original input image, and L ( x , y ) denotes the light image, which is also considered as noised image, and the light image can be estimated by F ( x , y ) I ( x , y ) . Most of the time, assuming the light changes slowly, the Gaussian kernel consequently becomes an appropriate implementation of the light estimation operator. The Gaussian surround function is given by Jobson [9] as
F G a u s s i a n ( x , y ) = K e r 2 c 2
where c denotes Gaussian surround space constant, r denotes the distance from current pixel ( x , y ) to the center ( x 0 , y 0 ) , and K is chosen so that the kernel brings in no extra gain. The MSR integrates Gaussian surround functions or, more generally, light estimation operators under various scales. The multi-scale calculation can be expressed by
R = i = 1 n R i
where n denotes the number of scales, and the overall R can be obtained by summation of images enhanced by different scales of a basic Retinex process.
In Retinex theory, a small-scale Gaussian kernel can achieve dynamic range compression, while a sizeable Gaussian kernel specializes in tonal and color rendition. A third intermediate scale combines both advantages of the small kernel and the large kernel; meanwhile, it eliminates the “halo” artifacts near sharp edges. The principle for choosing the Gaussian surround scale, proposed by Jobson [9], is that the sigma values of the Gaussian filter are 15, 80, and 250, corresponding to the three scales. According to the three-sigma rule, the filters’ kernel size should be 45 × 45 , 240 × 240 , and 750 × 750 .

3.2.2. Combination of FDIF and MSR

In this section, a modified MSRCR algorithm based on the fractional double integral filter is proposed. By directly superseding the Gaussian filter in the single-scale Retinex process, the FDIF operator estimates the light image while preserving edges in the original noised image. As shown in Figure 5, when using the same scale of the filtering kernel, the FDIF implements similar light estimation results to the Gaussian filter, which means our FDIF is effective in the light-estimation aspect. Moreover, the FDIF version of light estimation preserves edges better, while the Gaussian filter blurs them and therefore causes “over-smoothing” on the estimation image.

3.2.3. Color Restoration

Since underwater images often present disastrous color distortion, a color restoration method is decisive to the algorithm’s overall performance. One of the classical color restoration approaches, gray-world white balancing, shows stability and validity in underwater color restoration tasks [20]. However, the Retinex processing of images sometimes brings about local or global violations for the gray-world theory, which assumes that the average intensities for R, G, and B channels tend to be constant. Physically, the assumption supposes the average reflection of light from natural objects is a constant. A logarithmic color restoration function (CRF) is therefore proposed by Jobson [9]:
C i ( x , y ) = β log [ α I i ( x , y ) ] = β { log [ α I i ( x , y ) ] log [ i = 1 S I i ( x , y ) ] }
where C i ( x , y ) denotes the CRF for channel i, I i ( x , y ) denotes the intensity on channel i, S = 3 representsan RGB image, and α = 125 and β = 46 are empirical parameters for the underwater scene. In fact, I i ( x , y ) represents a gray-world white balancing process. Different values of empirical parameters α and β were also tested in Figure 6, and the chosen values are believed to be optimal.
As an image enhancement method based on the domain transform, the Retinex enhanced the image by converting the pixel matrix to the logarithmic domain and reducing noise. After the Retinex process, an inverse transformation called quantization, which converts the continuous logarithmic pixels back to RGB space, should be imposed, and the process determines the performance to a large extent. Due to the wide dynamic range of the logarithmic images, most Retinex-based frameworks apply a canonical gain method for inverse transform instead of linear quantization:
R o u t ( x , y ) = G × C i ( x , y ) × log R ( x , y ) + b
Here, G and b are empirical parameters, which are chosen to be 192 and 30 , and R o u t ( x , y ) presents the final output of the MSRCR. However, for underwater images, this canonical gain measure results in gray-out, which means the output image appears to turn gray and lacks fresh colors. To refine the algorithm, Parthasarathy’s method [21] has been proven to be effective for underwater images. By clipping the largest and smallest pixels of the logarithmic image to fixed values, the highly deviated maximum and minimum values are excluded from the quantization process. Then, the linear quantization method is used to transform the other values to RGB space, and the excluded largest and smallest parts of the pixels are set to 255 and 0, respectively. Furthermore, we utilize a gamma correction to the transformed pixel matrix, and compared to Parthasarathy’s method, ours accomplished better color rendition.

3.3. Unsupervised Encoder–Decoder Network

After the Retinex with FDIF preliminarily enhanced the input underwater image, an unsupervised autoencoder network was employed to improve the image quality. For the impossibility of obtaining both raw data of an underwater image and its ground truth simultaneously in the real-world environment, the network is designed to be trained in an unsupervised manner.

3.3.1. Network Architecture

Based on the prevailing U-Net [22] architecture, our encoder–decoder network takes a pre-enhanced underwater image as its input and generates a further improved output image. As depicted in Figure 7, the model is composed of the basic U-Net structure and attention modules. Four loss functions were taken into account so the training could converge.
First, the proposed framework takes an underwater image as input and applies the FDIF-MSR to the image. Then, the enhanced image will be refined by the unsupervised encoder–decoder network to alleviate noise brought in by the Retinex algorithm and improve detailed performance. The overall pipeline is shown in Figure 7.

3.3.2. Attention Module

In this paper, a squeeze-and-excitation module [23] first proposed by Jie Hu et al. was integrated into our network. By squeezing each input feature channel into a descriptor, the SE-Block concentrates on exploiting channel-wise dependency and therefore expresses the whole image. The excitation operator maps the descriptors into channel weights to extend conventional local receptive fields to a representation for a cross-channel fusion of features, which can be regarded as a self-attention function.

3.3.3. Loss Function

As for unsupervised learning, the loss function largely determines whether the neural network can effectively fit the data features. Based on the proposed encoder–decoder network, we introduce four losses to guide the training process.
  • Color Loss. To mitigate the color difference between pre-enhanced and post-enhanced underwater images, color loss was introduced into our model. The color loss function is defined by the angle between the input and output pixel vectors:
    L c o l o r = i , j ( I i , j , F ( I i , j ) )
    where F denotes a pixel matrix transform, and I i , j represents a single pixel in the matrix. As L 2 distance is widely used in image-processing tasks, a disadvantage has shown that the L 2 norm only calculates the numerical difference between the pixels, but the directional difference of the pixel vectors cannot be measured. In the proposed model, the color loss aims to narrow the gap of the angle between pixel vectors but not introduce too much computational cost.
  • Mix- L 1 -SSIM Loss. Since the network is designed to learn to produce visually pleasing images, it is natural that a perceptually motivated loss function should be adopted in the training pipeline. Structural Similarity, also known as SSIM, is defined as:
    S S I M ( I x , I y ) = 2 μ x μ y + C 1 μ x 2 + μ y 2 + C 1 · 2 σ x y + C 2 σ x 2 + σ y 2 + C 2
    where μ x and σ x denote mean and variance of pixel matrix x, respectively. Constants C 1 and C 2 are determined by the dynamic range of the pixel. Consequently, the loss function of SSIM can be written as:
    L S S I M = 1 S S I M ( I i n p u t , I o u t p u t )
    By fine-tuning constants in Equation (23), multi-scale SSIM can be expressed by
    L M S _ S S I M = 1 M S _ S S I M ( I i n p u t , I o u t p u t )
    where the multi-scale SSIM is defined as:
    M S _ S S I M ( I x , I y ) = m = 1 M ( 2 μ x μ y + C 1 μ x 2 + μ y 2 + C 1 ) β m · ( 2 σ x y + C 2 σ x 2 + σ y 2 + C 2 ) γ m
    According to Zhou Wang et al. [24], the MS-SSIM excels at preserving the contrast in high-frequency areas and supplies more flexibility than SSIM in incorporating the variations of viewing conditions. However, shifting of colors can be introduced by MS-SSIM loss, which may result in monotonous color rendition. To achieve better color and luminance performance, the L 1 loss, which aims to maintain color and luminance stability, is combined with the MS-SSIM loss, and the Mix- L 1 -SSIM Loss can be written as
    L M i x _ L 1 _ S S I M = α L M S _ S S I M + ( 1 α ) L 1
    where empirical parameter α is chosen to be 0.7 in our model.
  • Perceptual Loss. First proposed by Justin Johnson et al., perceptual loss has been proved to be valuable by numerous unsupervised models on image super-resolution and style transfer tasks. Instead of L 1 or L 2 loss, which exactly matches pixels of target image y ^ with input y, the perceptual loss encourages y ^ to have a similar feature representation to y, which can be regarded as constraining semantic changes during the image-enhancement process. The feature reconstruction loss can be defined as:
    L p e r c e p t u a l ( y ^ , y ) = 1 C j × H j × W j ϕ j ( y ^ ) ϕ j ( y ) 2 2
    where C j , H j , and W j represent the channel, height, and width of the feature map, respectively, and ϕ j denotes a feature extraction operator. We utilize a VGG-19 pretrained model to extract features of multiple layers from the image y ^ and y and then calculate the Euclidean distance between them to measure the difference. By minimizing feature-reconstruction perceptual loss, the model is able to produce visually indistinguishable output image y ^ from y.
  • Total Variation Loss. To prevent over-fitting and encourage the model to have better generalization capability, we use total variation loss in addition. In a two-dimensional continuous framework, the total variation loss is defined by:
    L T V = Ω u x 2 + u y 2 d x d y
    where u x = u x , u y = u y and u denotes the image, x , y Ω . As u x represents derivative along the x-axis, by minimizing u x , we can constrain the luminance difference between two adjacent pixels, and therefore, the overall noise of the output image can be suppressed.

4. Experiments

Elaborate experiments were conducted on public underwater image datasets, and both qualitative methods and quantitative methods were used for evaluating our algorithm. The datasets we used in this paper and some implementation details should be explained before we show the performance of our framework.

4.1. Datasets and Implementation Details

Three public underwater image datasets were used in our paper: Underwater Image Enhancement Benchmark (UIEB) [14], which includes 890 raw underwater images, OceanDark [25] for low-light underwater image enhancement, including 183 images of low-light or unbalanced-light condition, and Stereo Underwater Image Dataset [26] by Katherine A. Skinner et al.
Firstly, to train our unsupervised model, we inspected the entire UIEB and subjectively picked 782 high-quality samples from the 890 images as our training set, which contains about 88% data of the original UIEB. The training set was expanded to four times that of the original selected images by simply rotating the images 90 , 180 , and 270 .
Secondly, the FDIF-MSR was applied to the pre-processed dataset to attain preliminary enhanced images. Since most parameters of the FDIF-MSR algorithm comprising fractional orders are determined empirically and independent of the subsequent encoder–decoder network, we saved the preliminary enhanced images from the training set after the FDIF-MSR parameters were decided and the parameters were believed to be optimal.
Next, for one particular unsupervised training process, we utilized the saved enhanced images as input and fine-tuned the network hyper-parameters. All alterable hyper-parameters in our framework are shown in Table 1.
The unsupervised encoder–decoder network was trained on a single RTX-3090 GPU, with two Xeon Silver 4210 CPUs and 128 GB memory. We implemented the algorithm in Python 3.8.12 and used PyTorch framework version 1.11.0.

4.2. Evaluation

The evaluation of image enhancement has always been a significant challenge in that the human visual system is quite different from that of machines, and it is of great difficulty to measure what a visually pleasing image is by discrete calculation. To tackle this issue, we tried to evaluate our model in both qualitative and quantitative ways, and we obtained the conclusion that it achieves excellent performance on the validation set.

Qualitative Evaluation

As mentioned above, underwater images face two major challenges, one that light scattering and absorption cause poor visibility, and another that various attenuation characteristics of different frequencies of light result in terrible color distortion. In Figure 8, and Figure 9, our algorithm is tested on both hazy and color distortion examples.
Such hazy images in Figure 8 are common in real underwater images taken by unmanned undersea vehicles (UUV), and our framework achieved remarkable results on these images in that heavy haze brought in by light scattering and absorption was removed while unnatural color deviation was mitigated as well.
In the first and second columns of Figure 9, the entire images suffered from color casts of blue and green, respectively, but our framework has effectively corrected the distortion. As for the third and fourth columns, the framework mainly focuses on eliminating color distortion of the foreground object in the images; thereby, the divers’ and their equipment’s natural color has been recovered observably.
Furthermore, as most underwater images were photographed under extremely low-light conditions, we tested our model for that as well, and some of the results are shown in Figure 10.
In the case of the low-light condition, our framework achieves terrific results in that details in dark areas were enhanced vastly, while objects in bright areas were not over-enhanced and retained normal brightness.
As shown in Figure 11, we selected some images from the three above-mentioned public datasets for conducting the visual comparison. The chosen images were taken from diverse and challenging underwater scenes that include underwater haze, distortion of color, and low-light conditions. One can observe that our algorithm presented high robustness and satisfactory results under various underwater environments. Compared to the state-of-the-art method, TACL, our model shows equal effectiveness in color correction and low-light enhancement and performs with better haze-removal results (the sixth row) with clear details and sharp structures. Specifically, among the third row of the images, Chen’s method did not enhance the input image much; the UWCNN blurred objects in the shadows; the FUnIE-GAN brought undesirable color distortion; and the TACL failed to balance objects under different lighting conditions. However, our model has successfully overcome these problems.

4.3. Quantitative Evaluation

To make our analysis more convincing, some of the widely used full-reference image quality evaluation metrics were applied to our algorithm. Firstly, we considered the PSNR metric. The PSNR, also known as peak signal-to-noise ratio, is defined as
P S N R = 10 × l o g 10 ( 255 2 M S E )
where the M S E represents the pixel-wise difference between the input and output images by the expression
M S E = 1 M N i = 1 M j = 1 N ( I i j I i j ) 2
where M and N represent the height and width of the image, respectively, and I i j indicates a specific pixel at ( i , j ) .
Secondly, we utilize the SSIM metric, which was defined above by Equation (23). The SSIM is based on the hypothesis that the human’s visual system excels at perceiving structural information from real scenes, and hence the structural similarity can be an appropriate approximation to the image quality. Furthermore, two significant image properties, contrast and luminance, were also considered in our evaluation process.
Since not all of the images in the datasets we used are of high quality in terms of both color and texture, and quite a few images from the datasets were monotonous in the scene and tonality, we only show the evaluation results on a subset of the typical and high-quality samples instead of evaluating on the entire datasets. Each image we subjectively selected for the evaluation is based on the fact that most of the underwater images that required an enhancement were taken by underwater vehicles or divers in areas of poor visibility, so we specifically excluded those photographed near the surface of the water with strong natural uniform light.
The evaluation results are shown in Figure 12 and Table 2.
C E R = 1 C c = 1 C V a r ( I c )
where C denotes the number of channels and V a r ( I c ) denotes the variance of the image luminance on channel c.
L E R = 1 M N i = 1 M j = 1 N I o u t ( i , j ) I i n ( i , j ) I i n ( i , j )
As shown in Figure 11, the proposed model performed better than many other popular methods on underwater images. In Table 3, we provide the average PSNR, LER, CER, and SSIM values over the evaluation set used in Figure 11 for comparing the results of our model and the state-of-the-art methods. Since the four metrics are independent and lack a uniform representation of the enhancement performance, to quantitatively evaluate the image enhancement performances better, here, we introduced a uniform overall score that takes all four impact factors into account. The score is calculated by:
S c o r e = α × P S N R + β × | L E R C E R | 2 + γ × S S I M
where weight coefficients α = 0.6 , β = 0.5 , and γ = 2 were selected so the overall score could better represent the image enhancement performance. In Equation (34), we designed | L E R C E R | 2 to be a penalty term because unmatched luminance and contrast enhancement harm the visual quality of the image such that the images may have good results on quantitative values but fail on the human perceptual system. The results indicate that our model has similar LER and CER performance to the TACL but higher PSNR and SSIM and that our model presented even better results than TACL, the SOTA. Although the UWCNN and FUnIE-GAN yield better luminance enhancement rates, the two methods have bad results in terms of contrast enhancement, so they result in less pleasing enhanced images, as shown in Figure 11. Among all five methods, our model yielded the best overall score and presented visually pleasing results.
To make our paper more convincing, an ablation study was imposed on the proposed model. Models were evaluated on the dataset we used in Figure 12. Experiments were organized as follows:
(1)
Model No. 1: No encoder–decoder network is used for refining the result of the proposed FDIF-Retinex;
(2)
Model No. 2 to 5: The network is trained by the specific combination of loss function;
(3)
Model No. 6: The SE-Block is replaced with direct residual connections.
(4)
Model No. 7: The proposed full model.
As shown in Table 4, the FDIF-Retinex without post-enhancement (No. 1) has higher PSNR, LER, and CER but lower SSIM and overall score compared to the full model. This indicates that a post-enhancement network benefits the enhancement process by suppressing over-enhancement, which brings about more visually pleasing results. Through the results of the ablation study, we can see that if the model is trained by TV loss only or color loss with perceptual loss, the models perform poorly on the evaluation set and have a negative impact on the contrast enhancement rate. The models trained by perceptual loss only or perceptual loss and mixed loss have better performance than the above, but we want more natural results to be achieved=. To train the network, by combining the four loss functions, we can achieve a balance between luminance and contrast enhancement performance so that the output images will not be too bright or unnatural. Considering replacing residual connections with the SE-Block, the proposed model yielded better results on contrast enhancement and remained equal luminance enhancement capability, SSIM, and a higher PSNR. According to the experiments’ results, our method adopted an appropriate loss function group for training and an influential network architecture to improve underwater image quality.

5. Conclusions

In this paper, we proposed an end-to-end underwater-image-enhancement framework that excels at color restoration and haze removal for underwater scenes. Based on the fractional double integral filter, the proposed FDIF algorithm yielded better results on edge preservation than the widely used Gaussian version of multi-scale Retinex. An unsupervised encoder–decoder network that integrates an advanced attention mechanism and well-designed loss functions was utilized to further improve the quality of the enhanced images. Both qualitative evaluation and quantitative evaluation showed the effectiveness of the proposed framework, which achieved superb performance across multiple datasets and various underwater environments. In the future, we are planning to deploy the proposed underwater-image-enhancement model on embedded devices and test it in a real-world environment. We hope the model will provide a brand new view of underwater-images-enhancement methods, and we believe that the proposed model could benefit other downstream tasks such as object detection and 3D reconstruction.

Author Contributions

Conceptualization, Y.Y.; methodology, Y.Y. and C.Q.; software, C.Q.; validation, C.Q.; resources, Y.Y.; data curation, C.Q.; writing—original draft preparation, C.Q.; writing—review and editing, Y.Y.; visualization, C.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Key Research and Development Program [grant numbers 2021YFC2803000, 2020FYB1313200]; the National Natural Science Foundation of China [grant numbers 52001260], and the National Basic Scientific Research Program [grant numbers JCKY2019207A019].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in https://li-chongyi.github.io/proj_benchmark.html for the UIEB (accessed on 14 April 2022), https://sites.google.com/view/oceandark/home for the OceanDark (accessed on 12 September 2022), https://github.com/dlut-dimt/Realworld-Underwater-Image-Enhancement-RUIE-Benchmark for the RUIE (accessed on 13 December 2021), and https://github.com/kskin/data for the Stereo Underwater Image Dataset (accessed on 28 June 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MSRMulti-Scale Retinex
MSRCRMulti-Scale Retinex with Color Restoration
CNNConvolutional Neural Network
GANGenerative Adversarial Network
FDIFFractional Double Integral Filter
SSIMStructural Similarity
TVTotal Variation
UUVUnmanned Underwater Vehicle
PSNRPeak Signal-to-Noise Ratio
MSEMean Square Error
LERLuminance Enhancement Rate
CERContrast Enhancement Rate

References

  1. Narasimhan, S.G.; Nayar, S.K. Chromatic framework for vision in bad weather. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2000 (Cat. No. PR00662), Hilton Head, SC, USA, 15 June 2000; Volume 1, pp. 598–605. [Google Scholar]
  2. Narasimhan, S.G.; Nayar, S.K. Vision and the atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
  3. McGlamery, B. Computer analysis and simulation of underwater camera system performance. SIO Ref 1975, 75, 1–55. [Google Scholar]
  4. Jaffe, J.S. Computer modeling and the design of optimal underwater imaging systems. IEEE J. Ocean. Eng. 1990, 15, 101–111. [Google Scholar] [CrossRef]
  5. Akkaynak, D.; Treibitz, T. Sea-thru: A method for removing water from underwater images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1682–1691. [Google Scholar]
  6. Hou, W.; Gray, D.J.; Weidemann, A.D.; Fournier, G.R.; Forand, J. Automated underwater image restoration and retrieval of related optical properties. In Proceedings of the 2007 IEEE International Geoscience and Remote Sensing Symposium, Barcelona, Spain, 23–27 July 2007; pp. 1889–1892. [Google Scholar]
  7. Del Grosso, V. Modulation transfer function of water. In Proceedings of the OCEAN 75 Conference, Brighton, UK, 16–21 March 1975; pp. 331–347. [Google Scholar] [CrossRef]
  8. Land, E.H.; McCann, J.J. Lightness and retinex theory. Josa 1971, 61, 1–11. [Google Scholar] [CrossRef] [PubMed]
  9. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Wang, Q.; Ma, J.; Yu, S.; Tan, L. Noise detection and image denoising based on fractional calculus. Chaos Solitons Fractals 2020, 131, 109463. [Google Scholar] [CrossRef]
  11. Guo, H.; Li, X.; Chen, Q.-L.; Wang, M.-R. Image denoising using fractional integral. In Proceedings of the 2012 IEEE International Conference on Computer Science and Automation Engineering (CSAE), Zhangjiajie, China, 25–27 May 2012; Volume 2, pp. 107–112. [Google Scholar]
  12. Perez, J.; Attanasio, A.C.; Nechyporenko, N.; Sanz, P.J. A Deep Learning Approach for Underwater Image Enhancement. In Proceedings of the Biomedical Applications Based on Natural and Artificial Computing, Corunna, Spain, 19–23 June 2017; Ferrández Vicente, J.M., Álvarez-Sánchez, J.R., de la Paz López, F., Toledo Moreo, J., Adeli, H., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 183–192. [Google Scholar]
  13. Wang, Y.; Zhang, J.; Cao, Y.; Wang, Z. A deep CNN method for underwater image enhancement. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1382–1386. [Google Scholar]
  14. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef] [Green Version]
  16. Lu, J.; Li, N.; Zhang, S.; Yu, Z.; Zheng, H.; Zheng, B. Multi-scale adversarial network for underwater image restoration. Optics and Laser Technology 2019, 110, 105–113. [Google Scholar] [CrossRef]
  17. Liu, R.; Jiang, Z.; Yang, S.; Fan, X. Twin adversarial contrastive learning for underwater image enhancement and beyond. IEEE Trans. Image Process. 2022, 31, 4922–4936. [Google Scholar] [CrossRef] [PubMed]
  18. Liu, R.; Fan, X.; Zhu, M.; Hou, M.; Luo, Z. Real-world underwater enhancement: Challenges, benchmarks, and solutions under natural light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  19. Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
  20. Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 81–88. [Google Scholar]
  21. Parthasarathy, S.; Sankaran, P. An automated multi scale retinex with color restoration for image enhancement. In Proceedings of the 2012 National Conference on Communications (NCC), Kharagpur, India, 3–5 February 2012; pp. 1–5. [Google Scholar]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
  23. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  24. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
  25. Marques, T.P.; Albu, A.B. L2UWE: A Framework for the Efficient Enhancement of Low-Light Underwater Images Using Local Contrast and Multi-Scale Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  26. Skinner, K.A.; Zhang, J.; Olson, E.A.; Johnson-Roberson, M. Uwstereonet: Unsupervised learning for depth estimation and color correction of underwater stereo imagery. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 7947–7954. [Google Scholar]
  27. Chen, X.; Zhang, P.; Quan, L.; Yi, C.; Lu, C. Underwater image enhancement based on deep learning and image formation model. arXiv 2021, arXiv:2101.00991. [Google Scholar]
  28. Li, C.; Anwar, S.; Porikli, F. Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recognit. 2020, 98, 107038. [Google Scholar] [CrossRef]
  29. Islam, M.J.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
Figure 1. Diagram of the proposed end-to-end underwater-image-enhancement framework.
Figure 1. Diagram of the proposed end-to-end underwater-image-enhancement framework.
Fractalfract 07 00070 g001
Figure 2. The classical 8-direction fractional integral filter on a 2-D grid.
Figure 2. The classical 8-direction fractional integral filter on a 2-D grid.
Fractalfract 07 00070 g002
Figure 3. The fringe phenomenon caused by a sparse filter matrix. A sparse filter could only capture changes of pixels in the 8 directions; therefore starlike artifacts were found in the denoised image. The rightmost image patch shows the result of the FDI filter, and the artifacts were perfectly removed. We also made a quantitative comparison between the input image and filtered images: S S I M ( F I , i n p u t ) = 0.939 and S S I M ( F D I , i n p u t ) = 0.947 . This image is from the RUIE dataset [18], and the SSIM can be calculated by Equation (23).
Figure 3. The fringe phenomenon caused by a sparse filter matrix. A sparse filter could only capture changes of pixels in the 8 directions; therefore starlike artifacts were found in the denoised image. The rightmost image patch shows the result of the FDI filter, and the artifacts were perfectly removed. We also made a quantitative comparison between the input image and filtered images: S S I M ( F I , i n p u t ) = 0.939 and S S I M ( F D I , i n p u t ) = 0.947 . This image is from the RUIE dataset [18], and the SSIM can be calculated by Equation (23).
Fractalfract 07 00070 g003
Figure 4. FDIF kernel based on the fractional double integral. (a): Coefficients in the kernel; (b): Visualized version of the kernel by Matplotlib [19]. Unlike in the 8-direction operator, all of the coefficients in the kernel are unequal to zero as long as 1 α j . Since we assume α > 0 and ω 0 ( α ) = 1 , the coefficients are all nonzero. In particular, the coefficients all become one when the fractional order α equals 1, and the filter becomes an average filter. Coincidentally, some of the Retinex implementations use average filtering in practice to estimate the light to lower computational costs.
Figure 4. FDIF kernel based on the fractional double integral. (a): Coefficients in the kernel; (b): Visualized version of the kernel by Matplotlib [19]. Unlike in the 8-direction operator, all of the coefficients in the kernel are unequal to zero as long as 1 α j . Since we assume α > 0 and ω 0 ( α ) = 1 , the coefficients are all nonzero. In particular, the coefficients all become one when the fractional order α equals 1, and the filter becomes an average filter. Coincidentally, some of the Retinex implementations use average filtering in practice to estimate the light to lower computational costs.
Fractalfract 07 00070 g004
Figure 5. Comparison between the Gaussian and FDIF-based light estimation, k e r n e l _ s i z e = 45 , 243 , 753 from left to the right. The kernel size is chosen to be as close to the original MSR framework as possible to ensure accuracy, while the size must be odd for FDIF. In the first column, we can see that the FDIF version has kept the texture of the metal connector at the end of the breathing hoses, while the edges could barely be seen in the Gaussian version.
Figure 5. Comparison between the Gaussian and FDIF-based light estimation, k e r n e l _ s i z e = 45 , 243 , 753 from left to the right. The kernel size is chosen to be as close to the original MSR framework as possible to ensure accuracy, while the size must be odd for FDIF. In the first column, we can see that the FDIF version has kept the texture of the metal connector at the end of the breathing hoses, while the edges could barely be seen in the Gaussian version.
Fractalfract 07 00070 g005
Figure 6. Empirical parameters for (a): α = 20 , β = 20 ; (b): α = 125 , β = 46 ; (c): α = 219 , β = 69 . Compared to the optimal parameters in (b), small values for (a) brought about unsatisfactory correction for color distortion, while large values for (c) resulted in gray-out.
Figure 6. Empirical parameters for (a): α = 20 , β = 20 ; (b): α = 125 , β = 46 ; (c): α = 219 , β = 69 . Compared to the optimal parameters in (b), small values for (a) brought about unsatisfactory correction for color distortion, while large values for (c) resulted in gray-out.
Fractalfract 07 00070 g006
Figure 7. The structure of the proposed end-to-end underwater-image-enhancement framework.
Figure 7. The structure of the proposed end-to-end underwater-image-enhancement framework.
Fractalfract 07 00070 g007
Figure 8. The hazy examples were chosen from the Stereo Underwater Image Dataset [26]. Although the dataset provided paired images taken by stereo cameras, we treat the images as if they were taken by monocular cameras in our work. Top row: hazy images; Bottom row: corresponding enhanced images.
Figure 8. The hazy examples were chosen from the Stereo Underwater Image Dataset [26]. Although the dataset provided paired images taken by stereo cameras, we treat the images as if they were taken by monocular cameras in our work. Top row: hazy images; Bottom row: corresponding enhanced images.
Fractalfract 07 00070 g008
Figure 9. Some color distortion examples from the UIEB dataset. Top row: images with color casts; Bottom row: corresponding enhanced images.
Figure 9. Some color distortion examples from the UIEB dataset. Top row: images with color casts; Bottom row: corresponding enhanced images.
Fractalfract 07 00070 g009
Figure 10. Three low-light examples on the left were chosen from the OceanDark dataset, and the other one was from the UIEB. Top row: images with unbalanced light; Bottom row: corresponding enhanced images.
Figure 10. Three low-light examples on the left were chosen from the OceanDark dataset, and the other one was from the UIEB. Top row: images with unbalanced light; Bottom row: corresponding enhanced images.
Fractalfract 07 00070 g010
Figure 11. Images for qualitative comparison between our method and others, including a method based on deep learning and image formation model by Chen et al., UWCNN, FUnIE-GAN and TACL. (a) Input image. (b) Chen et al. [27]. (c) UWCNN [28]. (d) FUnIE-GAN [29]. (e) TACL [17]. (f) Ours.
Figure 11. Images for qualitative comparison between our method and others, including a method based on deep learning and image formation model by Chen et al., UWCNN, FUnIE-GAN and TACL. (a) Input image. (b) Chen et al. [27]. (c) UWCNN [28]. (d) FUnIE-GAN [29]. (e) TACL [17]. (f) Ours.
Fractalfract 07 00070 g011
Figure 12. Images for quantitative evaluation. Top two rows: UIEB-65, UIEB-66, UIEB-99, UIEB-416, UIEB-715 from the UIEB dataset and corresponding post-enhancement images; Middle two rows: images numbered HIMB-1 to HIMB-5 from the Stereo Underwater Image Dataset and their corresponding post-enhancement images; Bottom two rows: OceanDark-2, OceanDark-5, OceanDark-145, OceanDark-155, OceanDark-164 from the OceanDark dataset and corresponding post-enhancement images.
Figure 12. Images for quantitative evaluation. Top two rows: UIEB-65, UIEB-66, UIEB-99, UIEB-416, UIEB-715 from the UIEB dataset and corresponding post-enhancement images; Middle two rows: images numbered HIMB-1 to HIMB-5 from the Stereo Underwater Image Dataset and their corresponding post-enhancement images; Bottom two rows: OceanDark-2, OceanDark-5, OceanDark-145, OceanDark-155, OceanDark-164 from the OceanDark dataset and corresponding post-enhancement images.
Fractalfract 07 00070 g012
Table 1. Alterable hyper-parameters in the proposed model.
Table 1. Alterable hyper-parameters in the proposed model.
Hyper-ParameterValue
training epoch200
initial learning rate0.01
λ T V 0.5
λ c o l o r 0.2
λ p e r c e p t u a l 1.0
λ m i x 0.4
α f r a c 0.75
σ 45,81,251
α 125
β 46
high clip0.01
low clip0.01
γ 0.7
Table 2. Quantitative evaluation results on the datasets.
Table 2. Quantitative evaluation results on the datasets.
Image NumberPSNRCER 1LERSSIM
UIEB-6527.6240.8275.4710.692
UIEB-6627.6270.9243.6530.772
UIEB-9928.1252.1921.4510.692
UIEB-41627.7561.6261.7320.568
UIEB-71527.809−0.6071.1950.803
HIMB-127.8192.0041.1920.632
HIMB-227.8333.0811.3090.427
HIMB-327.7511.8771.2550.435
HIMB-427.9460.7011.1500.890
HIMB-527.7972.7281.0920.625
OceanDark-227.768−0.3580.8390.899
OceanDark-527.929−0.6631.2440.910
OceanDark-14527.969−0.2670.7760.905
OceanDark-15527.666−0.2871.7780.797
OceanDark-16427.703−0.2321.7710.803
1 CER and LER in Equations (32) and (33) refer to contrast enhancement rate and luminance enhancement rate, respectively. They can be calculated by the contrast and luminance difference between input and output images, respectively. The values are normalized by the input.
Table 3. Quantitative comparison between the proposed model and other methods in Figure 11.
Table 3. Quantitative comparison between the proposed model and other methods in Figure 11.
ModelPSNRLERCERSSIMScore
Chen et al.27.2811.6991.9650.82517.983
UWCNN27.0262.026−0.1050.78415.513
FUnIE-GAN27.3132.0290.6980.77717.056
TACL27.2461.5901.9860.77017.809
Ours27.8021.5831.9550.84918.310
Best27.802 (Ours)2.029 (FUnIE-GAN)1.986 (TACL)0.849 (Ours)18.310 (Ours)
Table 4. Ablation study results on network architecture and loss functions 1.
Table 4. Ablation study results on network architecture and loss functions 1.
Model No.TV LossColor LossPercep. LossMix LossSE-BlockPSNRLERCERSSIMScore
1-----27.8841.8652.1020.60517.912
2---27.9792.239−1.0000.66412.870
3--27.8243.385−0.3900.57410.717
4---27.8591.6381.1260.63417.852
5--27.7481.7301.7690.62517.898
6-27.7771.7051.1070.66217.811
7 227.8111.6571.3950.64717.946
1 The results are reported for the dataset we used in Figure 12. 2 The No. 7 model refers to enhancing the image by FDIF-Retinex only without a post-enhancement network.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, Y.; Qin, C. An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder. Fractal Fract. 2023, 7, 70. https://doi.org/10.3390/fractalfract7010070

AMA Style

Yu Y, Qin C. An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder. Fractal and Fractional. 2023; 7(1):70. https://doi.org/10.3390/fractalfract7010070

Chicago/Turabian Style

Yu, Yang, and Chenfeng Qin. 2023. "An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder" Fractal and Fractional 7, no. 1: 70. https://doi.org/10.3390/fractalfract7010070

APA Style

Yu, Y., & Qin, C. (2023). An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder. Fractal and Fractional, 7(1), 70. https://doi.org/10.3390/fractalfract7010070

Article Metrics

Back to TopTop