An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder
Abstract
:1. Introduction
- A fractional integral-based Retinex and an improved fractional-order integral operator, which eliminated the drawback of classical fractional integral operator for large-kernel filtering and resulted in more accurate estimation for light images, was proposed in this paper.
- An effective unsupervised encoder–decoder network requiring no adversarial training and yielding perceptually pleasing results was employed to refine the output of the previous Retinex model.
- Combining the fractional integral-based Retinex and unsupervised autoencoder mentioned above, the proposed end-to-end framework for underwater image enhancement was evaluated on several public datasets and produced impressive results.
2. Mathematical Background
2.1. The Definition of Fractional Derivatives
- Let be a positive real number. When , where n is a positive integer, the left-hand Riemann–Liouville fractional derivatives can be written as:
- The Grünwald–Letnikov definition of fractional derivatives is defined as
- We have Caputo’s definition of fractional derivatives, which is defined as:
2.2. Fractional Integral
2.3. Fractional Double Integral
3. Proposed Method
3.1. Fractional Double Integral Filter (FDIF)
3.2. Multi-Scale Retinex with FDIF
3.2.1. Retinex Theory
3.2.2. Combination of FDIF and MSR
3.2.3. Color Restoration
3.3. Unsupervised Encoder–Decoder Network
3.3.1. Network Architecture
3.3.2. Attention Module
3.3.3. Loss Function
- Color Loss. To mitigate the color difference between pre-enhanced and post-enhanced underwater images, color loss was introduced into our model. The color loss function is defined by the angle between the input and output pixel vectors:
- Mix--SSIM Loss. Since the network is designed to learn to produce visually pleasing images, it is natural that a perceptually motivated loss function should be adopted in the training pipeline. Structural Similarity, also known as SSIM, is defined as:
- Perceptual Loss. First proposed by Justin Johnson et al., perceptual loss has been proved to be valuable by numerous unsupervised models on image super-resolution and style transfer tasks. Instead of or loss, which exactly matches pixels of target image with input y, the perceptual loss encourages to have a similar feature representation to y, which can be regarded as constraining semantic changes during the image-enhancement process. The feature reconstruction loss can be defined as:
- Total Variation Loss. To prevent over-fitting and encourage the model to have better generalization capability, we use total variation loss in addition. In a two-dimensional continuous framework, the total variation loss is defined by:
4. Experiments
4.1. Datasets and Implementation Details
4.2. Evaluation
Qualitative Evaluation
4.3. Quantitative Evaluation
- (1)
- Model No. 1: No encoder–decoder network is used for refining the result of the proposed FDIF-Retinex;
- (2)
- Model No. 2 to 5: The network is trained by the specific combination of loss function;
- (3)
- Model No. 6: The SE-Block is replaced with direct residual connections.
- (4)
- Model No. 7: The proposed full model.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
MSR | Multi-Scale Retinex |
MSRCR | Multi-Scale Retinex with Color Restoration |
CNN | Convolutional Neural Network |
GAN | Generative Adversarial Network |
FDIF | Fractional Double Integral Filter |
SSIM | Structural Similarity |
TV | Total Variation |
UUV | Unmanned Underwater Vehicle |
PSNR | Peak Signal-to-Noise Ratio |
MSE | Mean Square Error |
LER | Luminance Enhancement Rate |
CER | Contrast Enhancement Rate |
References
- Narasimhan, S.G.; Nayar, S.K. Chromatic framework for vision in bad weather. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2000 (Cat. No. PR00662), Hilton Head, SC, USA, 15 June 2000; Volume 1, pp. 598–605. [Google Scholar]
- Narasimhan, S.G.; Nayar, S.K. Vision and the atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
- McGlamery, B. Computer analysis and simulation of underwater camera system performance. SIO Ref 1975, 75, 1–55. [Google Scholar]
- Jaffe, J.S. Computer modeling and the design of optimal underwater imaging systems. IEEE J. Ocean. Eng. 1990, 15, 101–111. [Google Scholar] [CrossRef]
- Akkaynak, D.; Treibitz, T. Sea-thru: A method for removing water from underwater images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1682–1691. [Google Scholar]
- Hou, W.; Gray, D.J.; Weidemann, A.D.; Fournier, G.R.; Forand, J. Automated underwater image restoration and retrieval of related optical properties. In Proceedings of the 2007 IEEE International Geoscience and Remote Sensing Symposium, Barcelona, Spain, 23–27 July 2007; pp. 1889–1892. [Google Scholar]
- Del Grosso, V. Modulation transfer function of water. In Proceedings of the OCEAN 75 Conference, Brighton, UK, 16–21 March 1975; pp. 331–347. [Google Scholar] [CrossRef]
- Land, E.H.; McCann, J.J. Lightness and retinex theory. Josa 1971, 61, 1–11. [Google Scholar] [CrossRef] [PubMed]
- Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Wang, Q.; Ma, J.; Yu, S.; Tan, L. Noise detection and image denoising based on fractional calculus. Chaos Solitons Fractals 2020, 131, 109463. [Google Scholar] [CrossRef]
- Guo, H.; Li, X.; Chen, Q.-L.; Wang, M.-R. Image denoising using fractional integral. In Proceedings of the 2012 IEEE International Conference on Computer Science and Automation Engineering (CSAE), Zhangjiajie, China, 25–27 May 2012; Volume 2, pp. 107–112. [Google Scholar]
- Perez, J.; Attanasio, A.C.; Nechyporenko, N.; Sanz, P.J. A Deep Learning Approach for Underwater Image Enhancement. In Proceedings of the Biomedical Applications Based on Natural and Artificial Computing, Corunna, Spain, 19–23 June 2017; Ferrández Vicente, J.M., Álvarez-Sánchez, J.R., de la Paz López, F., Toledo Moreo, J., Adeli, H., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 183–192. [Google Scholar]
- Wang, Y.; Zhang, J.; Cao, Y.; Wang, Z. A deep CNN method for underwater image enhancement. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1382–1386. [Google Scholar]
- Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef] [Green Version]
- Lu, J.; Li, N.; Zhang, S.; Yu, Z.; Zheng, H.; Zheng, B. Multi-scale adversarial network for underwater image restoration. Optics and Laser Technology 2019, 110, 105–113. [Google Scholar] [CrossRef]
- Liu, R.; Jiang, Z.; Yang, S.; Fan, X. Twin adversarial contrastive learning for underwater image enhancement and beyond. IEEE Trans. Image Process. 2022, 31, 4922–4936. [Google Scholar] [CrossRef] [PubMed]
- Liu, R.; Fan, X.; Zhu, M.; Hou, M.; Luo, Z. Real-world underwater enhancement: Challenges, benchmarks, and solutions under natural light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
- Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
- Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 81–88. [Google Scholar]
- Parthasarathy, S.; Sankaran, P. An automated multi scale retinex with color restoration for image enhancement. In Proceedings of the 2012 National Conference on Communications (NCC), Kharagpur, India, 3–5 February 2012; pp. 1–5. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
- Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
- Marques, T.P.; Albu, A.B. L2UWE: A Framework for the Efficient Enhancement of Low-Light Underwater Images Using Local Contrast and Multi-Scale Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
- Skinner, K.A.; Zhang, J.; Olson, E.A.; Johnson-Roberson, M. Uwstereonet: Unsupervised learning for depth estimation and color correction of underwater stereo imagery. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 7947–7954. [Google Scholar]
- Chen, X.; Zhang, P.; Quan, L.; Yi, C.; Lu, C. Underwater image enhancement based on deep learning and image formation model. arXiv 2021, arXiv:2101.00991. [Google Scholar]
- Li, C.; Anwar, S.; Porikli, F. Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recognit. 2020, 98, 107038. [Google Scholar] [CrossRef]
- Islam, M.J.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
Hyper-Parameter | Value |
---|---|
training epoch | 200 |
initial learning rate | 0.01 |
0.5 | |
0.2 | |
1.0 | |
0.4 | |
0.75 | |
45,81,251 | |
125 | |
46 | |
high clip | 0.01 |
low clip | 0.01 |
0.7 |
Image Number | PSNR | CER 1 | LER | SSIM |
---|---|---|---|---|
UIEB-65 | 27.624 | 0.827 | 5.471 | 0.692 |
UIEB-66 | 27.627 | 0.924 | 3.653 | 0.772 |
UIEB-99 | 28.125 | 2.192 | 1.451 | 0.692 |
UIEB-416 | 27.756 | 1.626 | 1.732 | 0.568 |
UIEB-715 | 27.809 | −0.607 | 1.195 | 0.803 |
HIMB-1 | 27.819 | 2.004 | 1.192 | 0.632 |
HIMB-2 | 27.833 | 3.081 | 1.309 | 0.427 |
HIMB-3 | 27.751 | 1.877 | 1.255 | 0.435 |
HIMB-4 | 27.946 | 0.701 | 1.150 | 0.890 |
HIMB-5 | 27.797 | 2.728 | 1.092 | 0.625 |
OceanDark-2 | 27.768 | −0.358 | 0.839 | 0.899 |
OceanDark-5 | 27.929 | −0.663 | 1.244 | 0.910 |
OceanDark-145 | 27.969 | −0.267 | 0.776 | 0.905 |
OceanDark-155 | 27.666 | −0.287 | 1.778 | 0.797 |
OceanDark-164 | 27.703 | −0.232 | 1.771 | 0.803 |
Model | PSNR | LER | CER | SSIM | Score |
---|---|---|---|---|---|
Chen et al. | 27.281 | 1.699 | 1.965 | 0.825 | 17.983 |
UWCNN | 27.026 | 2.026 | −0.105 | 0.784 | 15.513 |
FUnIE-GAN | 27.313 | 2.029 | 0.698 | 0.777 | 17.056 |
TACL | 27.246 | 1.590 | 1.986 | 0.770 | 17.809 |
Ours | 27.802 | 1.583 | 1.955 | 0.849 | 18.310 |
Best | 27.802 (Ours) | 2.029 (FUnIE-GAN) | 1.986 (TACL) | 0.849 (Ours) | 18.310 (Ours) |
Model No. | TV Loss | Color Loss | Percep. Loss | Mix Loss | SE-Block | PSNR | LER | CER | SSIM | Score |
---|---|---|---|---|---|---|---|---|---|---|
1 | - | - | - | - | - | 27.884 | 1.865 | 2.102 | 0.605 | 17.912 |
2 | √ | - | - | - | √ | 27.979 | 2.239 | −1.000 | 0.664 | 12.870 |
3 | - | √ | √ | - | √ | 27.824 | 3.385 | −0.390 | 0.574 | 10.717 |
4 | - | - | √ | - | √ | 27.859 | 1.638 | 1.126 | 0.634 | 17.852 |
5 | - | - | √ | √ | √ | 27.748 | 1.730 | 1.769 | 0.625 | 17.898 |
6 | √ | √ | √ | √ | - | 27.777 | 1.705 | 1.107 | 0.662 | 17.811 |
7 2 | √ | √ | √ | √ | √ | 27.811 | 1.657 | 1.395 | 0.647 | 17.946 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yu, Y.; Qin, C. An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder. Fractal Fract. 2023, 7, 70. https://doi.org/10.3390/fractalfract7010070
Yu Y, Qin C. An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder. Fractal and Fractional. 2023; 7(1):70. https://doi.org/10.3390/fractalfract7010070
Chicago/Turabian StyleYu, Yang, and Chenfeng Qin. 2023. "An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder" Fractal and Fractional 7, no. 1: 70. https://doi.org/10.3390/fractalfract7010070
APA StyleYu, Y., & Qin, C. (2023). An End-to-End Underwater-Image-Enhancement Framework Based on Fractional Integral Retinex and Unsupervised Autoencoder. Fractal and Fractional, 7(1), 70. https://doi.org/10.3390/fractalfract7010070