An Underwater Image Enhancement Algorithm Based on Generative Adversarial Network and Natural Image Quality Evaluation Index
Abstract
:1. Introduction
- (1)
- There are some low contrast underwater images in the truth samples of the training set, so the training effect of the model is not good.
- (2)
- The function that can improve the image contrast is not added in the FUnIE-GAN algorithm, so the contrast of the generated image is not high.
- (1)
- To solve the problem of low contrast images in truth datasets, this paper filters the images into truth images based on EUVP datasets to screen out truth images that meet the requirements.
- (2)
- To solve the problem of the low contrast of the generated image, this paper takes the NIQE as a part of the loss function of the generator in FUnIE-GAN and becomes its enhancement index.
- (3)
- To make the discriminant factors more diversified, this paper adds NIQE as FUnIE-GAN to the structure of the discriminator as part of the discriminator, which makes the resulting image more uniform in the color histogram distribution and more consistent with the perception of the human eye; this makes the generated image exceed the effect of the truth image set in the existing dataset.
- (4)
- In FUnIE-GAN-NIQE, there are four loss items in the loss function of the generator, including the adversarial loss function of the standard conditional GAN, L1 loss, content loss, and image quality loss. The weight loss of each part will affect the training result of the generator in the whole network; thus, this paper proposes to train 10 generators and 10 discriminators, traverse the weights of three parts (L1 loss, content loss, and image quality loss), and select the best generator among the 10 generators to generate the image. This method not only enhances the underwater image, but it can be applied to the enhancement of non-underwater images.
2. Fast Underwater Image Enhancement Algorithm
2.1. Network Architecture of Fast Underwater Image Enhancement Algorithm
2.1.1. Generator Network Architecture
2.1.2. Discriminator Network Architecture
2.2. Fast Underwater Image Enhancement Loss Function
- (1)
- In the FUnIE-GAN algorithm, the function is the adversarial loss function of the standard conditional GAN:In Equation (1), X and Y represent lossy image and truth image, respectively, Z is random noise, G is generator, D is discriminator and E is mathematical expectation. The lossy image X is input into the generator as a condition; together with random noise Z, the input of the generator is and the output of the generator is . The discriminator needs to distinguish a pair of , images so that it can not only generate real images, but also ensure that the generated images match the input images.
- (2)
- The function is a global similarity loss that aims to enable the generator to sample from the global similarity space. The generated image is closer to the true image in the global appearance. The calculation formula is shown in Equation (2):
- (3)
- The function is a loss of content to encourage the generator to generate images with advanced features similar to true images. In the FUnIE-GAN algorithm, the image content function is defined as the extraction of high-level feature functions. The advanced features extracted in this paper are extracted from the block5-conv2 layer of the pre-training network VGG-19. The formula for calculating content loss is shown in Equation (3):
3. Algorithm of Fast Underwater Image Enhancement Algorithm Based on NIQE Index
Algorithm 1 FUnIE-GAN-NIQE algorithm flow. |
|
3.1. Screening of Datasets
- (1)
- Four mainstream underwater image enhancement methods (CycleGAN [23], FUnIE-GAN [22], FUnIE-GAN-Up [22], UGAN [21]) were used to enhance the lossy images in the “Underwater Dark” dataset, and four enhanced image sets were obtained. The four image sets were named , , and , respectively. Then, each image in four datasets was evaluated by the NIQE index. Because the existing four mainstream methods could not effectively restore all images in specific applications to remove random errors and improve the accuracy of datasets, this paper used the method of error analysis to choose the four values of each image. The main implementation step was to remove the maximum value and the minimum value, and then take the average of the two values to obtain the NIQE value of the damaged image after effective enhancement.
- (2)
- The average value of NIQE obtained by the mainstream method in 4 was compared with the NIQE value of the corresponding truth image, and the truth image was screened by the error analysis method. The specific calculation formula is as follows:Among them, is the average value of NIQE obtained by four mainstream methods and is the NIQE value of the corresponding truth image. The truth images whose error fluctuation range was greater than were eliminated. Through the screening of truth images, a total of 153 truth images, which did not meet the requirements of the NIQE index, were screened out. Figure 4 shows the truth images selected in “Underwater Dark” that did not meet the requirements.
3.2. Loss Function of the Generator
3.3. Structure of Discriminator
3.4. Network Structure
4. Analysis of Experimental Results
4.1. The Experimental Configuration
4.2. Dataset Setting
4.3. Experiment
4.3.1. Subjective Assessment
4.3.2. Objective Assessment
4.4. Engineering Analysis
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Li, Y.; Lu, H.; Zhang, L.; Li, J.; Serikawa, S. Real-time visualization system for deep-sea surveying. Math. Probl. Eng. 2014, 2014, 437071. [Google Scholar] [CrossRef] [Green Version]
- Ahn, J.; Yasukawa, S.; Sonoda, T.; Nishida, Y.; Ishii, K.; Ura, T. An optical image transmission system for deep sea creature sampling missions using autonomous underwater vehicle. IEEE J. Ocean. Eng. 2018, 45, 350–361. [Google Scholar] [CrossRef]
- Sai, S.; Sai, I. Artificial Object Images Synthesis in Underwater Robot Vision System. In Proceedings of the 2020 International Conference on Industrial Engineering, Applications and Manufacturing, Sochi, Russia, 18–22 May 2020; pp. 1–6. [Google Scholar]
- Song, W.; Wang, Y.; Huang, D.; Tjondronegoro, D. A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration. In Proceedings of the Pacific Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; pp. 678–688. [Google Scholar]
- Iwamoto, Y.; Hashimoto, N.; Chen, Y. Fast Dark Channel Prior Based Haze Removal from a Single Image. In Proceedings of the 2018 14th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, Huangshan, China, 28–30 July 2018; pp. 458–461. [Google Scholar]
- Jobson, D.; Rahman, Z.; Woodell, G. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef] [PubMed]
- Singhai, J.; Rawat, P. Image enhancement method for underwater, ground and satellite images using brightness preserving histogram equalization with maximum entropy. In Proceedings of the International Conference on Computational Intelligence and Multimedia Applications, Sivakasi, India, 13–15 December 2007; pp. 507–512. [Google Scholar]
- Isola, P.; Zhu, J.; Zhou, T.; Efros, A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
- Choi, Y.; Choi, M.; Kim, M.; Ha, J.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
- He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [PubMed]
- Rahman, Z.; Jobson, D.; Woodell, G. Multi-Scale Rretinex for Color Image Enhancement. In Proceedings of the 3rd IEEE International Conference on Image Processing, Lausanne, Switzerland, 16–19 September 1996; pp. 1003–1006. [Google Scholar]
- Jobson, D.; Rahman, Z.; Woodell, G. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Bing, X.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8 December 2014; pp. 2672–2680. [Google Scholar]
- Chen, B.; Xia, M.; Huang, J. MFANet: A Multi-Level Feature Aggregation Network for Semantic Segmentation of Land Cover. Remote Sens. 2021, 13, 731. [Google Scholar] [CrossRef]
- Xia, M.; Wang, T.; Zhang, Y.; Liu, J.; Xu, Y. Cloud/shadow segmentation based on global attention feature fusion residual network for remote sensing imagery. Int. J. Remote Sens. 2021, 42, 2022–2045. [Google Scholar] [CrossRef]
- Xia, M.; Cui, Y.; Zhang, Y.; Xu, Y.; Xu, Y. DAU-Net: A novel water areas segmentation structure for remote sensing image. Int. J. Remote Sens. 2021, 42, 2594–2621. [Google Scholar] [CrossRef]
- Xia, M.; Liu, W. Non-intrusive load disaggregation based on composite deep long short-term memory network. Expert Syst. Appl. 2020, 160, 113669. [Google Scholar] [CrossRef]
- Xia, M.; Zhang, X.; Liu, W.; Weng, L.; Xu, Y. Multi-Stage Feature Constraints Learning for Age Estimation. IEEE Trans. Inf. Forensics Secur. 2020, 15, 2417–2428. [Google Scholar] [CrossRef]
- Arjovsky, M.; Bottou, L. Towards principled methods for training generative adversarial networks. arXiv 2017, arXiv:1701.04862. [Google Scholar]
- Li, J.; Skinner, K.; Eustice, R.; Johnson-Roberson, M. WaterGAN: Unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef] [Green Version]
- Fabbri, C.; Islam, M.; Sattar, J. Enhancing underwater imagery using generative adversarial networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation, Brisbane, QLD, Australia, 21–25 May 2018; pp. 7159–7165. [Google Scholar]
- Islam, M.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef] [Green Version]
- Zhu, J.; Park, T.; Isola, P.; Efros, A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 25 December 2017; pp. 2223–2232. [Google Scholar]
- Hu, K.; Zhang, Y. An Underwater Image Enhancement Algorithm Based on MSR Parameter Optimization. J. Mar. Sci. Eng. 2020, 8, 741. [Google Scholar] [CrossRef]
- Mittal, A.; Soundararajan, R.; Bovik, A. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
- Yang, M.; Sowmya, A. An underwater color image quality evaluation metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
Generator Number | |||
---|---|---|---|
Generator 1 | 6.3 | 2.7 | 1 |
Generator 2 | 5.6 | 2.4 | 2 |
Generator 3 | 4.9 | 2.1 | 3 |
Generator 4 | 4.2 | 1.8 | 4 |
Generator 5 | 3.5 | 1.5 | 5 |
Generator 6 | 2.8 | 1.2 | 6 |
Generator 7 | 2.1 | 0.9 | 7 |
Generator 8 | 1.4 | 0.6 | 8 |
Generator 9 | 0.7 | 0.3 | 9 |
Generator 10 | 0.07 | 0.03 | 9.9 |
Dataset Name | Paired Datasets | Validation | Total Number of Image Samples |
---|---|---|---|
Underwater Dark | 5550 | 570 | 11,670 |
Underwater ImageNet | 3700 | 1270 | 8670 |
Underwater Scenes | 2185 | 130 | 4500 |
Damaged Images | Truth Images | Validation | Total Number of Image Samples |
---|---|---|---|
3195 | 3140 | 330 | 6665 |
Scence | Scence 1 | Scence 2 | Scence 3 | Scence 4 | Scence 5 | |
---|---|---|---|---|---|---|
Image | ||||||
The original image | 3.832 | 4.432 | 5.856 | 6.023 | 4.579 | |
DCP | 3.683 | 4.367 | 5.856 | 6.114 | 4.681 | |
StarGAN | 3.465 | 3.700 | 5.011 | 5.705 | 4.196 | |
MSRCR | 3.740 | 4.383 | 5.490 | 5.728 | 4.452 | |
FUnIE-GAN | 3.306 | 4.929 | 5.228 | 6.208 | 4.655 | |
UGAN | 3.727 | 4.054 | 5.626 | 6.963 | 5.314 | |
CycleGAN | 3.403 | 4.323 | 5.704 | 5.842 | 5.124 | |
The proposed algorithm | 3.262 | 4.400 | 4.990 | 5.611 | 4.146 |
Scence | Scence 1 | Scence 2 | Scence 3 | Scence 4 | Scence 5 | |
---|---|---|---|---|---|---|
Image | ||||||
The original image | 0.603 | 0.524 | 0.517 | 0.535 | 0.561 | |
DCP | 0.603 | 0.498 | 0.522 | 0.534 | 0.601 | |
StarGAN | 0.594 | 0.623 | 0.568 | 0.567 | 0.591 | |
MSRCR | 0.523 | 0.560 | 0.411 | 0.426 | 0.453 | |
FUnIE-GAN | 0.627 | 0.627 | 0.541 | 0.565 | 0.616 | |
UGAN | 0.611 | 0.593 | 0.565 | 0.582 | 0.600 | |
CycleGAN | 0.576 | 0.545 | 0.497 | 0.531 | 0.594 | |
The proposed algorithm | 0.631 | 0.593 | 0.570 | 0.583 | 0.624 |
Algorithm | DCP | StarGAN | MSRCR | FUnIE-GAN | UGAN | CycleGAN | The Proposed Algorithm | |
---|---|---|---|---|---|---|---|---|
Index | ||||||||
NIQE | 4.95 ± 1.80 | 4.80 ± 1.83 | 5.12 ± 1.78 | 4.70 ± 1.92 | 4.68 ± 1.60 | 4.70 ± 1.72 | 4.65 ± 1.68 | |
UCIQE | 0.49 ± 0.06 | 0.54 ± 0.07 | 0.50 ± 0.08 | 0.56 ± 0.05 | 0.56 ± 0.07 | 0.54 ± 0.06 | 0.58 ± 0.06 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hu, K.; Zhang, Y.; Weng, C.; Wang, P.; Deng, Z.; Liu, Y. An Underwater Image Enhancement Algorithm Based on Generative Adversarial Network and Natural Image Quality Evaluation Index. J. Mar. Sci. Eng. 2021, 9, 691. https://doi.org/10.3390/jmse9070691
Hu K, Zhang Y, Weng C, Wang P, Deng Z, Liu Y. An Underwater Image Enhancement Algorithm Based on Generative Adversarial Network and Natural Image Quality Evaluation Index. Journal of Marine Science and Engineering. 2021; 9(7):691. https://doi.org/10.3390/jmse9070691
Chicago/Turabian StyleHu, Kai, Yanwen Zhang, Chenghang Weng, Pengsheng Wang, Zhiliang Deng, and Yunping Liu. 2021. "An Underwater Image Enhancement Algorithm Based on Generative Adversarial Network and Natural Image Quality Evaluation Index" Journal of Marine Science and Engineering 9, no. 7: 691. https://doi.org/10.3390/jmse9070691
APA StyleHu, K., Zhang, Y., Weng, C., Wang, P., Deng, Z., & Liu, Y. (2021). An Underwater Image Enhancement Algorithm Based on Generative Adversarial Network and Natural Image Quality Evaluation Index. Journal of Marine Science and Engineering, 9(7), 691. https://doi.org/10.3390/jmse9070691