A 5K Efficient Low-Light Enhancement Model by Estimating Increment between Dark Image and Transmission Map Based on Local Maximum Color Value Prior
Abstract
:1. Introduction
- A novel method for estimating a three-channel transmission map by estimating the incremental values from the dark image to the transmission map.
- Introducing a lightweight U-Net structure transmission estimation module.
- Creation of a lightweight correction module for color correction and denoise.
- Proposing a joint loss function specifically designed for optimizing the training process of the low-light enhancement model.
2. Related Work
2.1. Atmospheric Scattering-Based Model
2.2. Retinex-Based Model
2.3. Transformer-Based Model
2.4. AI-Generated Content-Based Model
2.5. Challenges
3. Methodology
3.1. Motivation
3.2. Dark Image Increment Estimation (DIIE) for Transmission Map
Algorithm 1 Overview of our proposed model | |
1: Input: I: Dark image: [C × H × W] | |
2: Output: E Enhanced image: [C × H × W] | |
3: procedure Model(I) | |
4: | ▹ Estimates the increment f |
5: | ▹ Gets 3-channel transmission map t |
6: | ▹ Gets scene radiance image J |
7: | ▹ Generates the final enahnced image E |
8: return E | |
9: end procedure |
3.3. Model Architecture: TEM and CM Module
3.3.1. Transmission Estimation Module
- Encoder: does not contain any trainable parameters but includes three AvgPooling layers. These layers generate pooling images at 1/2, 1/4, and 1/8 scales of the original image. These pooling images will be fed into the Decoder.
- Skip Block: has a structure of conv3×3 + ResBlock + ResBlock + ResBlock + conv3×3 with inputs and outputs being 3-channel features. It is primarily responsible for extracting global features from the smallest 1/8 scale pooling image, which will be fed into the first layer of the Decoder.
- Decoder: contains most of our trainable parameters and consists of four T_UP layers. The structure of the T_UP layer is similar to the Skip Block, but we have added a Deconvolution layer for upscaling the pooling images from the Encoder, as shown in the T_UP diagram in Figure 5.
3.3.2. Correction Module
- Color Correction Block: To adjust color biases, we concatenate the dark source image I and the scene radiance image J, forming a 6-channel feature metric . This combined feature metric is then fed to the Color Correction block. The output of the Color Correction block is a 3-channel matrix c in the range of 0–1, and the result of represents the color-corrected result.
- Denoise Block: Inspired by the Zero-Shot Noise2Noise model [36], we utilize two convolutional layers to estimate and subtract noise from the image, thereby producing a denoised image or, in other words, the final enhanced image E.
3.4. Joint Loss
- (1)
- Full-Reference Loss Functions
- Mean Absolute (L1) Loss. L1 Loss is a common loss function for various image-to-image tasks such as image reconstruction, image super-resolution and image denoise. It is an effective loss function as it directly computes the mean distance between the enhanced result and the ground truth. We propose L1 loss to compare the final enhanced image and the ground truth data.
- Root Mean Squared Log Error (RMSLE) loss. RMSLE loss utilizes the logarithm function based on the root mean squared error, which can reduce the impact of large differences between a few values and the ground truth in the overall error calculation. Thus, RMSLE loss allows for localized small errors. We use RMSLE loss to measure the difference between the scene radiance and the ground truth because scene radiance is not the final enhanced image processed by the correction module. Therefore, we tolerate some errors in this comparison to overcome the overfitting.
- Structural Similarity (SSIM) loss. SSIM loss compares two images’ brightness, contrast, and structure, providing a better metric of human visual perception of image differences. We propose the SSIM loss to measure the loss between the enhanced image and the ground truth data.
- (2)
- No-Reference Loss Functions
- Illumination Smoothness Loss. The Illumination Smoothness Loss calculates the noise level of the enhanced image by identifying prominent points on smooth surfaces using gradient operations in both horizontal and vertical directions. Through statistical analysis of these prominent points, the loss function quantifies the noise level present in the enhanced image.loss aims to minimize noise and irregularities in the output image, resulting in smoother images with fewer prominent noise points.
- Color Constancy Loss, based on the Gray-World color constancy hypothesis [37], states that the average intensity of any channel among the RGB channels should be gray. Therefore, we propose Color Constancy Loss as a measure of whether the overall color of the enhanced image is correct.loss focuses on maintaining color balance across RGB channels. It ensures overall enhancement without color shifts, thereby improving the overall visual quality of the enhanced images.
4. Experiments
4.1. Implementation Details
- Experimentation Platform: We implement our model on a Ubuntu 22.04 system equipped with an Intel Core i5-4690 3.50 GHz CPU from Intel corporation, Santa Clara, California, United States, an Nvidia GeForce RTX 3090 graphics card from Nvidia corporation, Santa Clara, California, United States, and 16 GB of memory from Samsung, Suwon-si, South Korea. The deep learning training framework used for the model was PyTorch 1.13.1.
- Training Hyperparameters: We employed the Adam optimizer with exponential decay rate parameters and , setting the initial learning rate to . Additionally, we employed the OneCycleLR learning rate scheduler to dynamically adjust the learning rate during training, and the training duration for each experiment consisted of 20 epochs.
- Datasets: Three different datasets were used to demonstrate the performance of our model in various low-light enhancement (LLE) tasks.
4.2. Performance Evaluation
4.3. Ablation Study
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
ASM | atmospheric scattering model |
CM | Correction Module |
DCP | dark channel prior |
DIIE | Dark Image Increment Estimation |
FLOPs | floating point operations |
GT | ground truth |
LLE | low-light enhancement |
LMCV | Local Maximum Color Value |
LOL | low-light dataset |
PSNR | peak signal to noise ratio |
SOTA | state of the art |
SSIM | structural similarity |
TEM | Transmission Estimation Module |
References
- Wang, Y.F.; Liu, H.M.; Fu, Z.W. Low-light image enhancement via the absorption light scattering model. IEEE Trans. Image Process. 2019, 28, 5679–5690. [Google Scholar] [CrossRef] [PubMed]
- Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
- Dong, X.; Wen, J. Low Lighting Image Enhancement Using Local Maximum Color Value Prior; Frontiers of Computer Science; Springer: Berlin/Heidelberg, Germany, 2016; Volume 10, pp. 147–156. [Google Scholar]
- Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
- Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
- Cui, Z.; Li, K.; Gu, L.; Su, S.; Gao, P.; Jiang, Z.; Qiao, Y.; Harada, T. You Only Need 90K Parameters to Adapt Light: A Light Weight Transformer for Image Enhancement and Exposure Correction. In Proceedings of the 33rd British Machine Vision Conference 2022, London, UK, 21–24 November 2022. [Google Scholar]
- Panagiotou, S.; Bosman, A.S. Denoising Diffusion Post-Processing for Low-Light Image Enhancement. arXiv 2023, arXiv:2303.09627. [Google Scholar]
- Zhou, D.; Yang, Z.; Yang, Y. Pyramid Diffusion Models for Low-light Image Enhancement. arXiv 2023, arXiv:2305.10028. [Google Scholar]
- Wu, G.; Jin, C. DiffLIE: Low-Light Image Enhancment based on Deep Diffusion Model. In Proceedings of the 2023 3rd International Symposium on Computer Technology and Information Science (ISCTIS), Chengdu, China, 16–18 June 2023; pp. 522–526. [Google Scholar]
- Lv, X.; Dong, X.; Jin, Z.; Zhang, H.; Song, S.; Li, X. L 2 DM: A Diffusion Model for Low-Light Image Enhancement. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Shenzhen, China, 14–17 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 130–145. [Google Scholar]
- Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Diff-retinex: Rethinking low-light image enhancement with a generative diffusion model. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 12302–12311. [Google Scholar]
- McCartney, E.J. Optics of the Atmosphere: Scattering by Molecules and Particles; John Wiley and Sons, Inc.: New York, NY, USA, 1976. [Google Scholar]
- Land, E.H. The Retinex; The Scientific Research Honor Society: Triangle Park, NC, USA, 1965; pp. 217–227. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
- Cao, Y.; Li, S.; Liu, Y.; Yan, Z.; Dai, Y.; Yu, P.S.; Sun, L. A comprehensive survey of ai-generated content (aigc): A history of generative ai from gan to chatgpt. arXiv 2023, arXiv:2303.04226. [Google Scholar]
- Liu, W.; Zhao, P.; Zhang, B.; Xuan, W. A low-light image enhancement method based on atmospheric scattering model. In Proceedings of the 2022 2nd International Conference on Computer Graphics, Image and Virtualization (ICCGIV), Chongqing, China, 23–25 September 2022; pp. 145–150. [Google Scholar]
- Jeon, J.J.; Park, J.Y.; Eom, I.K. Low-light image enhancement using gamma correction prior in mixed color spaces. Pattern Recognit. 2024, 146, 110001. [Google Scholar] [CrossRef]
- Makwana, D.; Deshmukh, G.; Susladkar, O.; Mittal, S. LIVENet: A novel network for real-world low-light image denoising and enhancement. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 1–10 January 2024; pp. 5856–5865. [Google Scholar]
- Li, C.; Guo, C.; Loy, C.C. Learning to enhance low-light image via zero-reference deep curve estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4225–4238. [Google Scholar] [CrossRef] [PubMed]
- Ma, L.; Liu, R.; Zhang, J.; Fan, X.; Luo, Z. Learning deep context-sensitive decomposition for low-light image enhancement. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 5666–5680. [Google Scholar] [CrossRef] [PubMed]
- Wang, R.; Zhang, Q.; Fu, C.W.; Shen, X.; Zheng, W.S.; Jia, J. Underexposed photo enhancement using deep illumination estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 6849–6857. [Google Scholar]
- Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-Light Image/Video Enhancement Using CNNs. Proc. BMVC 2018, 220, 4. [Google Scholar]
- Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
- Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
- Lu, K.; Zhang, L. TBEFN: A two-branch exposure-fusion network for low-light image enhancement. IEEE Trans. Multimed. 2020, 23, 4093–4105. [Google Scholar] [CrossRef]
- Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse gradient regularized deep retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef] [PubMed]
- Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. From fidelity to perceptual quality: A semi-supervised approach for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3063–3072. [Google Scholar]
- Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond brightening low-light images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
- Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage Retinex-based Transformer for Low-light Image Enhancement. arXiv 2023, arXiv:2303.06705. [Google Scholar]
- Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A General u-Shaped Transformer for Image Restoration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 17683–17693. [Google Scholar]
- Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient Transformer for High-Resolution Image Restoration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
- Jie, H.; Zuo, X.; Gao, J.; Liu, W.; Hu, J.; Cheng, S. Llformer: An efficient and real-time lidar lane detection method based on transformer. In Proceedings of the 2023 5th International Conference on Pattern Recognition and Intelligent Systems, Shenyang, China, 28–30 July 2023; pp. 18–23. [Google Scholar]
- Brateanu, A.; Balmez, R.; Avram, A.; Orhei, C. LYT-Net: Lightweight YUV Transformer-based Network for Low-Light Image Enhancement. arXiv 2024, arXiv:2401.15204. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 18th International Conference—Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
- Targ, S.; Almeida, D.; Lyman, K. Resnet in resnet: Generalizing residual architectures. arXiv 2016, arXiv:1603.08029. [Google Scholar]
- Mansour, Y.; Heckel, R. Zero-Shot Noise2Noise: Efficient Image Denoising without any Data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14018–14027. [Google Scholar]
- Cepeda-Negrete, J.; Sanchez-Yanez, R.E. Gray-world assumption on perceptual color spaces. In Proceedings of the 6th Pacific-Rim Symposium, PSIVT 2013—Image and Video Technology, Guanajuato, Mexico, 28 October–1 November 2013; Springer: Berlin/Heidelberg, Germany, 2014; pp. 493–504. [Google Scholar]
- Bychkovsky, V.; Paris, S.; Chan, E.; Durand, F. Learning Photographic Global Tonal Adjustment with a Database of Input/Output Image Pairs. In Proceedings of the Twenty-Fourth IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011. [Google Scholar]
- Wang, H.; Xu, K.; Lau, R.W. Local color distributions prior for image enhancement. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 343–359. [Google Scholar]
- Zeng, H.; Cai, J.; Li, L.; Cao, Z.; Zhang, L. Learning image-adaptive 3d lookup tables for high performance photo enhancement in real-time. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2058–2073. [Google Scholar] [CrossRef] [PubMed]
- Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12299–12310. [Google Scholar]
- Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. Maxim: Multi-axis mlp for image processing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LO, USA, 18–24 June 2022; pp. 5769–5780. [Google Scholar]
- Hu, Y.; He, H.; Xu, C.; Wang, B.; Lin, S. Exposure: A white-box photo post-processing framework. ACM Trans. Graph. (TOG) 2018, 37, 1–17. [Google Scholar] [CrossRef]
- Ignatov, A.; Kobyshev, N.; Timofte, R.; Vanhoey, K.; Van Gool, L. Dslr-quality photos on mobile devices with deep convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3277–3285. [Google Scholar]
Methods | LOL-V1 | LOL-V2-Real | Efficiency | |||
---|---|---|---|---|---|---|
PSNR ↑ | SSIM ↑ | PSNR ↑ | SSIM ↑ | FLOPs (G) ↓ | #Params (M) ↓ | |
LIME [2] | 16.67 | 0.560 | 15.24 | 0.470 | - | - |
Zero-DCE [5] | 14.83 | 0.531 | 14.32 | 0.511 | 2.53 | 0.08 |
RetiNexNet [23] | 16.77 | 0.562 | 18.37 | 0.723 | 587.47 | 0.84 |
MBLLEN [22] | 17.90 | 0.702 | 18.00 | 0.715 | 19.95 | 20.47 |
DRBN [27] | 19.55 | 0.746 | 20.13 | 0.820 | 37.79 | 0.58 |
3D-LUT [40] | 16.35 | 0.585 | 17.59 | 0.721 | 7.67 | 0.6 |
KIND [4] | 20.86 | 0.790 | 19.74 | 0.761 | 356.72 | 8.16 |
UFormer [30] | 16.36 | 0.771 | 18.82 | 0.771 | 12.00 | 5.29 |
IPT [41] | 16.27 | 0.504 | 19.80 | 0.813 | 2087.35 | 115.63 |
MAXIM [42] | 23.43 | 0.863 | 22.86 | 0.818 | 216.00 | 14.14 |
IAT [6] | 23.38 | 0.809 | 23.50 | 0.824 | 1.44 | 0.09 |
IAT (local) [6] | 20.20 | 0.782 | 20.30 | 0.789 | 1.31 | 0.02 |
Ours | 19.4635 | 0.6144 | 20.6379 | 0.6878 | 0.512 | 0.0047 |
Metric | White-Box [43] | DPED [44] | D-UPE [21] | 3D-LUT [40] | IAT [6] | Ours |
---|---|---|---|---|---|---|
PSNR ↑ | 18.57 | 21.76 | 23.04 | 25.21 | 25.32 | 20.68 |
SSIM ↑ | 0.701 | 0.871 | 0.893 | 0.922 | 0.920 | 0.832 |
#Params (M) ↓ | - | - | 1.0 | 0.6 | 0.09 | 0.0047 |
Metric | LIME [2] | RetinexNet [23] | Zero-DCE [5] | D-UPE [21] | LCDP [39] | Ours |
---|---|---|---|---|---|---|
PSNR ↑ | 17.335 | 19.250 | 12.587 | 20.970 | 23.239 | 20.016 |
SSIM ↑ | 0.686 | 0.704 | 0.653 | 0.818 | 0.842 | 0.793 |
#Params (M) ↓ | - | 0.84 | 0.08 | 1.0 | 0.28 | 0.0047 |
Methods | LOL-V1 | LOL-V2-Real | ||
---|---|---|---|---|
PSNR ↑ | SSIM ↑ | PSNR ↑ | SSIM ↑ | |
w/o CM | 17.8738 | 0.5816 | 18.6065 | 0.5916 |
w/o | 18.8969 | 0.6440 | 19.5882 | 0.6877 |
w/o | 18.3100 | 0.5754 | 19.8165 | 0.6646 |
w/o | 18.8289 | 0.6354 | 19.5118 | 0.6850 |
w/o | 19.2315 | 0.6125 | 18.7706 | 0.6238 |
w/o | 18.3861 | 0.6227 | 20.0880 | 0.7022 |
Ours | 19.4635 | 0.6144 | 20.6379 | 0.6878 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Deng, Q.; Choo, D.; Ji, H.; Lee, D. A 5K Efficient Low-Light Enhancement Model by Estimating Increment between Dark Image and Transmission Map Based on Local Maximum Color Value Prior. Electronics 2024, 13, 1814. https://doi.org/10.3390/electronics13101814
Deng Q, Choo D, Ji H, Lee D. A 5K Efficient Low-Light Enhancement Model by Estimating Increment between Dark Image and Transmission Map Based on Local Maximum Color Value Prior. Electronics. 2024; 13(10):1814. https://doi.org/10.3390/electronics13101814
Chicago/Turabian StyleDeng, Qikang, Dongwon Choo, Hyochul Ji, and Dohoon Lee. 2024. "A 5K Efficient Low-Light Enhancement Model by Estimating Increment between Dark Image and Transmission Map Based on Local Maximum Color Value Prior" Electronics 13, no. 10: 1814. https://doi.org/10.3390/electronics13101814
APA StyleDeng, Q., Choo, D., Ji, H., & Lee, D. (2024). A 5K Efficient Low-Light Enhancement Model by Estimating Increment between Dark Image and Transmission Map Based on Local Maximum Color Value Prior. Electronics, 13(10), 1814. https://doi.org/10.3390/electronics13101814