Author Contributions
Conceptualization, methodology, software, validation, writing—original draft J.Y.; investigation, methodology, software, writing—original draft, writing—review and editing Y.W. and C.W.; resources, data curation, project administration, H.F.; writing—original draft preparation, writing—review and editing, J.H.; supervision, project administration, A.G. and C.W. All authors have read and agreed to the published version of the manuscript.
Figure 1.
Comparison of enhanced results obtained by different UIE methods. The proposed LEPF-Net improves brightness, color, contrast, and detail recovery compared to other methods.
Figure 1.
Comparison of enhanced results obtained by different UIE methods. The proposed LEPF-Net improves brightness, color, contrast, and detail recovery compared to other methods.
Figure 2.
The network structure of LEPF-Net.
Figure 2.
The network structure of LEPF-Net.
Figure 3.
Visual result of a pixel-level curve parameter map . (a) is a raw underwater image, (b) is the corresponding curve parameter map , and (c) is the enhanced image by the proposed LEPF-Net.
Figure 3.
Visual result of a pixel-level curve parameter map . (a) is a raw underwater image, (b) is the corresponding curve parameter map , and (c) is the enhanced image by the proposed LEPF-Net.
Figure 4.
Comparisons of loss function curves, PSNR curves, and SSIM curves with/without using skip connection.
Figure 4.
Comparisons of loss function curves, PSNR curves, and SSIM curves with/without using skip connection.
Figure 5.
Visual examples of pixel attention map and channel attention map of a bluish underwater image. (a) is an underwater image with a large and dense background of bluish pixels, (b) is its corresponding pixel attention map, and (c) is the channel attention weight map for each pixel, the larger the pixel value, the greater the attention it receives.
Figure 5.
Visual examples of pixel attention map and channel attention map of a bluish underwater image. (a) is an underwater image with a large and dense background of bluish pixels, (b) is its corresponding pixel attention map, and (c) is the channel attention weight map for each pixel, the larger the pixel value, the greater the attention it receives.
Figure 6.
Qualitative comparisons on synthetic underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 6.
Qualitative comparisons on synthetic underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 7.
Qualitative comparisons on bluish underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 7.
Qualitative comparisons on bluish underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 8.
Qualitative comparisons on greenish underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 8.
Qualitative comparisons on greenish underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 9.
Qualitative comparisons on low-illuminated underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 9.
Qualitative comparisons on low-illuminated underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 10.
Qualitative comparisons on yellowish underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 10.
Qualitative comparisons on yellowish underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 11.
Qualitative comparisons on shallow water underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 11.
Qualitative comparisons on shallow water underwater images. The last row is the enlarged pictures of the red boxes in the first row.
Figure 12.
In the form of radar map, the no-reference quality assessment results of 10 images in
Figure 7,
Figure 8,
Figure 9,
Figure 10 and
Figure 11 are represented. The indicators are Entropy, BIQI, NIQE, AG, and JPEG. The larger the area of the yellow area, the better. The score ranges from 0 (worst) to 5 (best).
Figure 12.
In the form of radar map, the no-reference quality assessment results of 10 images in
Figure 7,
Figure 8,
Figure 9,
Figure 10 and
Figure 11 are represented. The indicators are Entropy, BIQI, NIQE, AG, and JPEG. The larger the area of the yellow area, the better. The score ranges from 0 (worst) to 5 (best).
Figure 13.
The framework of LE-ICurve, which is used to iteratively enhance the input image . Different curve degree mapping and iteration number n correspond to different output results. The horizontal axis represents the input pixel value, and the vertical axis represents the output pixel value.
Figure 13.
The framework of LE-ICurve, which is used to iteratively enhance the input image . Different curve degree mapping and iteration number n correspond to different output results. The horizontal axis represents the input pixel value, and the vertical axis represents the output pixel value.
Figure 14.
The comparison results of different loss function configurations. Because the raw image (a) is a yellowish underwater image, its histogram is the least close to that of the reference image (f) on the blue-green channel. (b) shows an overexposure effect, and its red-green channel has a certain gap compared with (f). (c,d) have artifacts, which lead to the loss of image detail information. In the histogram of the former, its blue-green channel is much different from that of (f), while in the histogram of the latter, the three channels R, G, B are very different from those of (f). Only (e) in various configurations shows relatively good quality, and its histogram is closest to that of (f).
Figure 14.
The comparison results of different loss function configurations. Because the raw image (a) is a yellowish underwater image, its histogram is the least close to that of the reference image (f) on the blue-green channel. (b) shows an overexposure effect, and its red-green channel has a certain gap compared with (f). (c,d) have artifacts, which lead to the loss of image detail information. In the histogram of the former, its blue-green channel is much different from that of (f), while in the histogram of the latter, the three channels R, G, B are very different from those of (f). Only (e) in various configurations shows relatively good quality, and its histogram is closest to that of (f).
Figure 15.
The ablation study. Since the raw image (a) is a bluish underwater image, its histogram is the least close to that of the reference image (f) on the blue channel. Both (b,c) show some grid artifacts, but the former has the most serious grid artifacts. (d) has serious color deviation and obvious gridding. In all configurations, the result image of (e) performs best, and its histogram is relatively close to that of (f). (e) does not show multicolor bias and artifacts.
Figure 15.
The ablation study. Since the raw image (a) is a bluish underwater image, its histogram is the least close to that of the reference image (f) on the blue channel. Both (b,c) show some grid artifacts, but the former has the most serious grid artifacts. (d) has serious color deviation and obvious gridding. In all configurations, the result image of (e) performs best, and its histogram is relatively close to that of (f). (e) does not show multicolor bias and artifacts.
Figure 16.
Object detection results by applying the YOLO-v4.
Figure 16.
Object detection results by applying the YOLO-v4.
Figure 17.
Local keypoints matching results by applying the SIFT operator.
Figure 17.
Local keypoints matching results by applying the SIFT operator.
Figure 18.
Salient detection results by applying the PoolNet+.
Figure 18.
Salient detection results by applying the PoolNet+.
Figure 19.
Edge detection results by applying Canny operator.
Figure 19.
Edge detection results by applying Canny operator.
Table 1.
The proposed LEPF-Net architecture.
Table 1.
The proposed LEPF-Net architecture.
Number | Layer Description | Output Size |
---|
Encoder |
#1 | Conv(3,64,3,1)+IN+ReLU | 350 × 350 × 64 |
#2 | Conv(64,64,3,1)+IN+ReLU | 350 × 350 × 64 |
#3 | Conv(64,64,3,2)+IN+ReLU | 350 × 350 × 64 |
LEB |
#1 | Conv(64,64,3,1)+IN+ReLU | 350 × 350 × 64 |
#2 | Conv(64,64,3,1)+IN+ReLU | 350 × 350 × 64 |
#3 | Conv(3,3,3,1) | 350 × 350 × 3 |
#4 | Conv(3,32,1,1)+ReLU | 350 × 350 × 32 |
PF-SubNet |
#1 | AAP+Conv(192,4,1,1)+ReLU | 350 × 350 × 4 |
#2 | Conv(4,192,1,1)+ReLU | 350 × 350 × 192 |
#3 | Conv(192,8,1,1)+ReLU | 350 × 350 × 8 |
#4 | Conv(8,192,1,1)+ReLU | 350 × 350 × 192 |
Decoder |
#1 | Conv(64,64,4,2)+IN+ReLU | 350 × 350 × 64 |
#2 | Conv(64,64,3,1)+IN+ReLU | 350 × 350 × 64 |
#3 | Conv(64,3,1,1) | 350 × 350 × 3 |
Table 2.
Training details of different compared algorithms, including the framework, training epoch, learning rate, batch size, and image size.
Table 2.
Training details of different compared algorithms, including the framework, training epoch, learning rate, batch size, and image size.
Methods | UDCP | ULAP | UWGAN | UGAN | UWCNN | DUIENet | UIEIFM | UIE-WD | UIEC2 -Net | Ours |
---|
Frame | MatLab | MatLab | TensorFlow | TensorFlow | TensorFlow | TensorFlow | PyTorch | PyTorch | PyTorch | PyTorch |
---|
Epochs | - | - | 50 | 50 | 50 | 50 | 50 | 50 | 50 | 50 |
Learning_rate | - | - | | | 10 | 10 | 10 | 10 | 10 | 10 |
Batch_size | - | - | 8 | 8 | 8 | 8 | 4 | 4 | 4 | 4 |
Image_size | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 | 350 × 350 |
Table 3.
Full reference image quality assessment of the test900. The best and second-best results are marked in red and blue.
Table 3.
Full reference image quality assessment of the test900. The best and second-best results are marked in red and blue.
Method | PSNR | SSIM |
---|
raw | 14.4978 | 0.6948 |
UDCP | 13.6778 | 0.6373 |
ULAP | 14.3739 | 0.6989 |
UWGAN | 13.3007 | 0.7083 |
UGAN | 16.2636 | 0.6625 |
UWCNN | 14.8638 | 0.7419 |
DUIENet | 16.1073 | 0.7734 |
UIEIFM | 17.5494 | 0.7123 |
UIE-WD | 20.4510 | 0.8661 |
UIEC2 -Net | 20.5442 | 0.8486 |
Ours | 24.1325 | 0.8733 |
Table 4.
Full-reference image quality assessment of test90, the best and second-best results are marked in red and blue.
Table 4.
Full-reference image quality assessment of test90, the best and second-best results are marked in red and blue.
Method | PSNR | SSIM |
---|
raw | 18.2701 | 0.8151 |
UDCP | 11.1646 | 0.5405 |
ULAP | 18.6789 | 0.8194 |
UWGAN | 18.6209 | 0.8454 |
UGAN | 21.3031 | 0.7691 |
UWCNN | 18.2851 | 0.8150 |
DUIENet | 16.2906 | 0.7884 |
UIEIFM | 17.4574 | 0.6583 |
UIE-WD | 19.0074 | 0.7872 |
UIEC2-Net | 24.5663 | 0.9346 |
Ours | 26.2857 | 0.9000 |
Method | Entropy | NIQE | BIQI | AG | JPEG |
---|
Raws | 6.5081 | 39.8008 | 49.0445 | 4.7933 | 11.2021 |
UDCP | 5.8717 | 39.8005 | 39.3064 | 5.6096 | 10.6871 |
ULAP | 6.4558 | 39.8006 | 35.9631 | 7.5027 | 10.2069 |
UWGAN | 6.549 | 39.8009 | 36.9106 | 6.4002 | 10.6459 |
UGAN | 7.3468 | 39.8005 | 37.5143 | 8.6566 | 10.1087 |
UWCNN | 6.5177 | 39.8008 | 44.8743 | 4.9923 | 11.0149 |
DUIENet | 6.6314 | 39.8011 | 43.2616 | 5.9285 | 10.5725 |
UIEIFM-Net | 5.9095 | 39.8009 | 45.7180 | 4.1659 | 11.2879 |
UIE-WD | 6.7663 | 39.8026 | 31.8501 | 10.2489 | 9.7099 |
UIEC2-Net | 7.4661 | 39.8016 | 33.5492 | 9.1011 | 9.7671 |
Ours | 7.4682 | 39.8022 | 29.9085 | 11.0408 | 9.2780 |
References | 7.4292 | 39.8014 | 35.9838 | 8.7786 | 9.7458 |
Table 6.
The quantitative results of different loss function configurations; the best and second-best results are marked in red and blue.
Table 6.
The quantitative results of different loss function configurations; the best and second-best results are marked in red and blue.
| Only MSE | Only L1 | MSE + TV | L1 + TV |
ine MSE | ✓ | | ✓ | |
L1 | | ✓ | | ✓ |
TV | | | ✓ | ✓ |
PSNR | 25.3233 | 25.9720 | 25.9342 | 26.2943 |
SSIM | 0.8852 | 0.8937 | 0.8916 | 0.9025 |
Table 7.
The quantitative results of ablation study; the best and second-best results are marked in red and blue.
Table 7.
The quantitative results of ablation study; the best and second-best results are marked in red and blue.
| No configuration | Only LEBG | Only PF-SubNet | Full configuration |
LEBG | | ✓ | | ✓ |
PF-SubNet | | | ✓ | ✓ |
PSNR | 25.4886 | 25.6389 | 26.1559 | 26.2857 |
SSIM | 0.8752 | 0.8850 | 0.8902 | 0.9000 |