Next Article in Journal
Attosecond-Level Delay Sensing via Temporal Quantum Erasing
Next Article in Special Issue
Atmospheric Turbulence Degraded Video Restoration with Recurrent GAN (ATVR-GAN)
Previous Article in Journal
Force-Induced Visualization of Nucleic Acid Functions with Single-Nucleotide Resolution
Previous Article in Special Issue
Text Recognition Model Based on Multi-Scale Fusion CRNN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Survey of Deep Learning-Based Low-Light Image Enhancement

1
School of Information Engineering, Henan Institute of Science and Technology, Xinxiang 453003, China
2
Institute of Computer Applications, Henan Institute of Science and Technology, Xinxiang 453003, China
3
School of Internet, Anhui University, Hefei 230039, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(18), 7763; https://doi.org/10.3390/s23187763
Submission received: 10 August 2023 / Revised: 29 August 2023 / Accepted: 2 September 2023 / Published: 8 September 2023

Abstract

:
Images captured under poor lighting conditions often suffer from low brightness, low contrast, color distortion, and noise. The function of low-light image enhancement is to improve the visual effect of such images for subsequent processing. Recently, deep learning has been used more and more widely in image processing with the development of artificial intelligence technology, and we provide a comprehensive review of the field of low-light image enhancement in terms of network structure, training data, and evaluation metrics. In this paper, we systematically introduce low-light image enhancement based on deep learning in four aspects. First, we introduce the related methods of low-light image enhancement based on deep learning. We then describe the low-light image quality evaluation methods, organize the low-light image dataset, and finally compare and analyze the advantages and disadvantages of the related methods and give an outlook on the future development direction.

1. Introduction

Due to the development of technology and the continuous improvement of photographic equipment, we have higher and higher requirements for the quality of the images we capture, but we often have difficulty obtaining suitable images because of the interference of environmental factors. Uneven lighting, low lighting, and other factors like backlighting can result in imperfect image information, diminishing the overall quality of captured images. Figure 1 shows an example of an image under suboptimal lighting conditions. Consequently, these issues can have a cascading effect on advanced tasks such as object recognition, detection, and classification. As artificial intelligence technologies continue to evolve, the associated industries are also changing, and thus the requirements for related downstream tasks are increasing. The quality of tasks completed in the image processing area [1,2,3,4,5] can greatly affect the efficiency of upstream tasks.
In daily life, we often encounter uncontrollable environmental or equipment factors that cause uneven lighting, darkness, backlighting, and blurring of captured images [6,7,8,9]. However, we have demand for high-quality images. Superior image quality is crucial for everyday scenarios and holds significant importance across various sectors [10,11,12,13,14], including intelligent transportation and vision monitoring. Therefore, quality enhancement of images has become a subject worthy of further exploration.
The enhancement of low-light images holds a significant role in image processing. This involves enhancing the visual quality of images captured in low-light conditions by adjusting the contrast and brightness levels, thus improving visibility [15,16,17]. Conventional techniques for enhancing low-light images often revolve around statistical learning, including approaches like local exposure compensation algorithms. While these methods can enhance image brightness effectively, they may concurrently introduce undesired noise and distortion. In addition, traditional methods have some limitations [18,19,20]. While the favorable notion of adopting the reflection component as the enhancement outcome may hold inconsistently, particularly when factoring in diverse lighting attributes, it could potentially result in impractical improvements, such as the omission of details and distortion of color. Additionally, the model overlooks noise, allowing it to persist or amplify within the enhanced output [21,22,23,24,25].
In recent years, deep learning methods have made significant advancements in various fields, particularly in image processing tasks. Unlike traditional approaches, deep learning techniques place a stronger emphasis on capturing the spatial features in an image, allowing for better preservation of details and increased resistance to noise [26,27,28,29]. Deep learning has demonstrated remarkable achievements in various fields, including enhancing low-light images. In comparison with conventional approaches, deep learning-based solutions for improving low-light image quality have gained substantial attention due to their enhanced precision, robustness, and efficiency. The existing deep learning techniques for low-light image enhancement establish a relationship between an output image and a correspondingly enhanced input image under low-light conditions by designing a network structure. However, this approach can result in a high dependency on the training data, which limits its effectiveness to some extent.
This paper focuses on employing deep learning techniques for enhancing low-light images while offering an extensive assessment and analysis of current methods within this domain. The noteworthy contributions of this study encompass the following:
(1) We systematically classify and summarize the deep learning-based low-light image enhancement methods proposed in recent years, introduce the core ideas of the mentioned algorithms in detail, and provide an insightful analysis of the problems of and possible solutions for the existing methods.
(2) We summarize the datasets in the field of low-light image enhancement in detail, including the sources, characteristics, and application scenarios of the datasets. We also provide a comprehensive comparison of different datasets and discuss their respective advantages and disadvantages.
(3) We analyze the advantages and disadvantages of some existing methods through experimental comparisons, present possible problems, and look forward to future research directions.

2. Low-Light Image Enhancement Method Based on Deep Learning

Deep learning has gained widespread popularity and is extensively utilized in various vision processing tasks due to its efficiency and convenience. In recent years, it has emerged as a prominent topic in machine learning and has found significant applications in the field of image enhancement. Deep learning-based approaches have demonstrated exceptional performance in enhancing low-light images, making them a prominent trend in current image processing research. The purpose of low-light image enhancement is to improve the visibility and quality of images captured in low-light conditions. Nowadays, deep convolutional neural networks (CNNs) and generative adversarial networks (GANs) have been applied in several directions and have proven to be effective solutions in low-light image enhancement. In Table 1, we summarize the basic characteristics of representative methods based on deep learning. These methods utilize the power of deep learning to enhance image quality and improve visibility to provide impressive results.

2.1. CNN-Based Methods

Utilized for low-light image enhancement, the convolutional neural network (CNN) functions as a supervised learning approach. It effectively addresses the challenge of improving low-light images by acquiring the mapping correlation between the input and output images. This approach involves utilizing a multi-layer convolutional network to obtain higher-quality enhanced images. The standard CNN architecture consists of three key elements: the convolutional layer, the pooling layer, and the fully connected layer. These components play distinct roles, with the convolutional layer focused on extracting local features from the input image. Subsequently, the pooling layer plays a role in parameter reduction, enhancing network efficiency. Ultimately, the fully connected layer generates the desired output outcomes. By integrating these layers, a CNN effectively enhances low-light images and achieves outstanding results.

2.1.1. Physical Model-Based Methods

Lore et al. [30] introduced the pioneering Low Light Net (LLNet), a deep learning-based technique for enhancing low-light images. They employed a densely layered sparse denoising autoencoder to execute contrast enhancement and denoising procedures. This groundbreaking research set the foundation for comprehensive web applications in low-light image enhancement (LLIE). Building upon this foundation, Lv et al. [33] introduced the MBLLEN, an innovative end-to-end multi-branch enhancement network. This architecture markedly improves low-light image enhancement performance, achieving this through the extraction of impactful feature representations across a feature extraction module, enhancement module, and fusion module. Furthermore, Ren et al. [56] developed a sophisticated end-to-end architecture involving an encoder-decoder network dedicated to enhancing image content alongside a recurrent neural network specifically for improving image edge enhancements. Their approach offers comprehensive enhancement capabilities, addressing both image content and edges. These studies highlight the continuous advancements in low-light image enhancement through deep learning techniques, demonstrating the potential for improving image quality in challenging lighting conditions. Tao et al. [57] introduced a learning framework based on a low-light CNN (LLCNN) to address the gradient disappearance problem in low-light image enhancement. They utilized multi-scale feature maps to mitigate this issue and incorporated SSIM loss to preserve the image texture during model training. This approach enables the adaptive enhancement of low-light images by effectively enhancing contrast while maintaining image details. Similarly, Xu et al. [58] introduced a decomposition enhancement network. Their method is centered on restoring image content by addressing noise within the low-frequency layer while highlighting intricate details in the high-frequency layer. By leveraging this approach, they achieved enhanced image quality while effectively preserving the image details. Zhu et al. [41] introduced the Edge-Enhanced Multi-Exposure Fusion Network (EEMEFN) [42], a method structured into two primary stages: multi-exposure fusion (MEF) and edge enhancement. The multi-exposure fusion network is tailored to enhancing low-light images and calculates a transfer function in both branches to produce a pair of improved images. These images are then fused using a simple averaging scheme and further refined using refinement units, resulting in improved enhancement results. These works highlight various innovative approaches in low-light image enhancement, showcasing the ongoing advancements in the field. Figure 2 provides a flow chart of CNN method combined with physical model. These methods employ different strategies, such as multi-scale feature maps, frequency-based decomposition, and fusion techniques, to achieve superior results in enhancing low-light images.

2.1.2. Non-Physical Model-Based Methods

Gharbi et al. [59] introduced a novel real-time image enhancement technique that seamlessly integrates deep learning and the bilateral filter. This inventive approach reimagines the conventional bilateral filter as a deep neural network, facilitating end-to-end learning. This methodology enables real-time image enhancement but also harnesses the capabilities of deep learning for enhanced performance. Similarly, Shen et al. [60] introduced an innovative approach known as the multi-scale Retinex network (MSR-net), which fuses CNN technology with Retinex theory. This technique parallels the concept of feedforward convolutional neural networks, using diverse Gaussian convolutional kernels to create a multi-scale Retinex framework. By directly learning the mapping between low-light images and standard-brightness images, this method effectively enhances low-light images. These studies showcase innovative methods that combine deep learning techniques with well-established image enhancement theories. By integrating deep neural networks and Retinex theory, researchers have made significant strides in developing end-to-end networks that effectively enhance low-light images. Wei et al. [32] introduced a deep neural network that draws on Retinex theory to enhance images taken in low-light settings. Utilizing the principles of Retinex theory rooted in physics, this technique involves image decomposition into reflection, intermediate reflection, and shadow components, followed by individualized enhancement of each constituent element. In a similar vein, Li et al. [31] introduced LightenNet, a trainable convolutional neural network designed specifically for enhancing low-light images. Through training, the network is fed dimly lit images as the input and produces corresponding light maps as the output. The output image is then processed by a Retinex-based model to obtain the enhanced image. Although this method can achieve desirable enhancement results for some images, it still performs poorly in some challenging real-life scenes. It is worth noting that while the proposed methods show promise in improving image enhancement under low-light conditions, there is still room for further improvement, particularly in challenging real-world scenes. Cai et al. [34] introduced a method for learning a contrast enhancer for individual images using multiple exposure images. Their approach employs a deep neural network to learn the mapping from low-contrast images to high-contrast images. By leveraging the information from multiple exposure images, this method enables effective enhancement of contrast in individual images. In a similar vein, Wu et al. [61] proposed a rapid end-to-end trainable image enhancement method based on the guided filter. This approach performs enhancement while preserving high-frequency information by incorporating the concept of the guided filter. By leveraging this technique, this method achieves efficient and effective image enhancement. Additionally, Wei et al. [62] introduced an image denoising and enhancement method based on recurrent neural networks (RNNs). Their approach incorporates the concept of nonlocal mean filtering by integrating nonlocal mean filters into the RNN architecture. By utilizing nonlocal mean filters to model the spatial information of images, this method achieves effective denoising and enhancement of images. These methods highlight different approaches to image enhancement, including learning-based contrast enhancement, guided filter-based enhancement, and RNN-based denoising and enhancement. Figure 3 provides a flow chart of CNN method for non-physical model. By leveraging various techniques and models, researchers are continuously advancing the field of image enhancement, addressing challenges related to contrast, noise, and preserving image details.
Overall, the advantages of CNN-based low-light image enhancement methods lie first in the fact that local feature learning is more robust, and more image information can be extracted. Secondly, no additional processing steps are required, as it is usually an end-to-end process. In addition, the structure of CNNs is controllable and can be adapted according to the needs. Lastly, as CNNs are driven by a large amount of data, it is possible to be able to learn from the data how to perform low-light image enhancement.
Although there are many advantages to CNN-based low-light image enhancement methods, some disadvantages that cannot be hidden also need to be overcome. The first is that the global information of the image may be ignored, resulting in enhanced images that are not natural enough globally. The second disadvantage is that if the training data are insufficient, this may lead to overfitting. Finally, the performance of the CNN receives the influence of a variety of factors, which need to be debugged in a comprehensive manner.

2.2. GAN-Based Methods

Generative adversarial networks (GANs) constitute a class of unsupervised deep learning models composed of two primary components: a generator and a discriminator. GANs operate on the foundational concept of a competitive interplay between the generator and discriminator, fostering mutual learning to produce high-quality outcomes. The generator’s role involves creating simulated realistic samples aimed at deceiving the discriminator, which in turn specializes in discerning real samples from synthetically generated ones. In the context of low-light image enhancement, the generator focuses on extracting enhanced features from low-light images, while the discriminator evaluates the quality and authenticity of the generated images. During an iterative procedure, the generator strives to enhance its capability to generate images that closely resemble authentic ones. Meanwhile, the discriminator hones its proficiency in discerning between actual and synthetic samples. The objective is pursued until the discriminator reaches a point where it can no longer differentiate between real and generated samples. Figure 4 provides a flow chart of GAN-based method. Compared with other generative models, GANs offer the advantage of generating clear and realistic samples without relying on complex Markov chains. GANs utilize back propagation as the primary mechanism for learning, allowing for efficient training and the generation of high-quality samples.

2.2.1. Condition-Based Methods

Jiang et al. [36] introduced EnlightenGAN, a GAN-based method that addresses the issue of overfitting and the limited generalization ability when training deep models on paired data. EnlightenGAN employs a U-Net architecture for generation and a composite global-local discriminator. This amalgamation involves global and local adversarial losses accompanied by self-feature preservation losses, ensuring fidelity between the augmented and authentic images. The objective behind these loss functions is to preserve coherence in the image content prior to and after enhancement, a vital aspect for maintaining structural stability during training. By combining these elements, EnlightenGAN achieves stable training and enhances the quality of low-light images. The attention-guided U-Net generator, along with the global-local discriminator and self-feature retention losses, contribute to generating more realistic and visually pleasing enhanced results, mitigating the limitations associated with overfitting and improving the generalization ability of the model. Meng et al. [63] introduced a GAN-based nighttime image enhancement framework that utilizes the properties of the GAN to generate pseudo-real images from real image distributions. The results prove its effectiveness and signify that a GAN applied to nighttime image enhancement is viable. In 2017, Ignatov et al. [64] introduced an image enhancement technique utilizing the GAN model. While this approach enhances image quality, its broader applicability is constrained by the pronounced correspondence-matching association between the initial and improved images, resulting in a heavily supervised procedure. To address this limitation, the authors later introduced an improvement using a weakly supervised network model called WESPE [65], reducing the reliance on strong supervision and yielding a more generalized algorithm. The VGG19 network calculates the content loss to maintain image content consistency, mitigating the risk of an excessively prominent correspondence match between the original and enhanced images. Furthermore, Chen et al. [66] proposed an image enhancement method based on adversarial generative networks (GANs). This approach utilizes a generator and a discriminator to achieve the mapping from ordinary photos to high-quality photos, enabling effective image enhancement. These studies demonstrate the utilization of GAN-based models for image enhancement, particularly for nighttime images. They address challenges such as strong supervision, generalizability, and quality improvement, showcasing the potential of GANs in advancing image enhancement techniques.

2.2.2. Circular Consistency-Based Methods

Zhu et al. [67] introduced a generative adversarial network (GAN) framework that utilizes a recurrent consistency loss function for unpaired image transformation. This method enables learning to transform images from one domain to another while preserving the semantic information of the images. In a similar vein, Fu et al. [68] introduced LE-GAN, an innovative unsupervised low-light image enhancement network founded on generative adversarial networks. This network is trained using disparate pairs of low-light and normal-light images. To improve the quality of vision and address issues such as noise and color bias, they added an illumination-aware attention module to improve feature extraction. Additionally, a new invariant loss is introduced to tackle overexposure problems, allowing the network to adaptively enhance low-light images. These methods highlight the use of GANs for unpaired image transformation and low-light image enhancement. By incorporating recurrent consistency loss and attention modules, researchers have made significant advancements in preserving semantic information, reducing noise and color bias, and improving the visual quality of transformed and enhanced images. Yan et al. [69] introduced a low-light image enhancement method that leverages an optimization-enhanced enhancement network module within the generative adversarial network (GAN) framework. This method utilizes an enhancement network to input images into a generator, generating similar images in a new space. Subsequently, a loss function is constructed and minimized to train a discriminator, which then compares the generated images with real images to enhance the network’s performance. Similarly, You et al. [70] introduced Cycle-CBAM, a retinal image enhancement technique built upon the foundation of Cycle-GAN. This method aims to elevate the quality of fundus images from lower to higher levels without necessitating paired training data. To tackle challenges posed by texture information loss and detail degradation due to unpaired image training, Cycle-GAN is augmented by the integration of Convolutional Block Attention Module (CBAM). These strategies highlight the utilization of GAN-based methodologies in enhancing both low-light and retinal images. By optimizing the enhancement network module and incorporating attention mechanisms, researchers strive to enhance the quality and fidelity of the resulting enhanced images.
Overall, the advantages of GAN-based low-light image enhancement methods lie first in the fact that they are good at generating more realistic enhanced images. Secondly, they can learn more advanced image features, which can be used to recover image details and information. In addition, GANs are able to perform complex nonlinear modeling and capture complex features. Lastly, GANs are able to generate diversity images, which makes the results more diverse.
Although there are many advantages of GAN-based low-light image enhancement methods, there are some significant drawbacks for the direction of our future work. The first is that the training process of a GAN is complex, and the problem of unstable training may occur. The second is that the training may suffer from the problem of pattern collapse, which may lead to a lack of diversity in the results. In addition, a GAN also requires a certain amount of data to support the training of the model. Finally, the performance of a GAN receives the influence of various hyperparameters, which need to be carefully adjusted to keep it stable.

3. Low-Light Image Quality Evaluation

Commonly used image quality assessment (IQA) methods can be categorized into two main categories: reference-based and referenceless. Reference-based algorithms rely on having both the original (considered to be of high quality) and distorted images in order to calculate quality scores. These algorithms are commonly employed to measure the quality of images after undergoing processes such as image compression, image transfer, or image stitching. Referenceless methods, on the other hand, do not require a reference image. They estimate image quality by analyzing the characteristics and features of the distorted image itself. These methods are particularly useful when a reference image is not available or when evaluating image quality in real-time scenarios. By categorizing IQA methods into reference-based and referenceless approaches, researchers have been able to develop techniques suitable for various scenarios, enhancing our ability to assess and quantify image quality accurately.

3.1. Full-Reference Metrics

Commonly used reference-based image quality evaluation methods include the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) [71]. These methods evaluate image quality by comparing the target image with the original image. These evaluation methods have found widespread application in tasks such as image compression and evaluating image quality after processes like image transmission and stitching. However, it is important to note that if the reference image itself has inherently poor quality, then the credibility of these evaluation indices is significantly compromised. The accuracy of the evaluation depends on the assumption that the reference image represents a high-definition image. On the other hand, due to the complexity of the signal perceived by the human eye, different types of images, such as those with varying texture complexity or different image attributes, may yield similar PSNR or SSIM scores but can be judged differently by the human eye. This highlights the limitations of relying solely on objective evaluation metrics, as subjective human perception may vary even when objective metrics suggest similar quality. Overall, while the PSNR and SSIM serve as widely used reference-based evaluation metrics, the credibility of their results depends on the quality of the reference image, and the subjective perception of image quality can differ based on various image attributes and human interpretation.
In 2004, a research paper published by the University of Texas at Austin presented SSIM, a metric employed to gauge the similarity between two images. The metric assigns a value between 0 and 1, where higher values signify increased image similarity and lower values indicate greater dissimilarity. SSIM is a comprehensive reference tool to assess the likeness of x and y images. The calculation of SSIM is as follows:
l ( m , n ) = 2 μ m μ n + C 1 μ m 2 + μ n 2 + C 1 ,
c ( m , n ) = 2 σ m σ n + C 2 σ m 2 + σ n 2 + C 2 ,
s ( m , n ) = σ m n + C 3 σ m σ n + C 3 .
In the above equation, l ( m , n ) is the mean to estimate the brightness, c ( m , n ) is the variance to estimate the contrast, and s ( m , n ) is the covariance to estimate the structural similarity. Meanwhile, μ m and μ n represent the mean of m and n, respectively, σ m and σ n represent the standard deviation of m and n, respectively, σ m n represents the covariance of m and n, and C 1 , C 2 , and C 3 all represent constants. Therefore, SSIM is defined as
SSIM ( m , n ) = ( 2 μ m μ n + C 1 ) ( 2 σ m n + C 2 ) ( μ m 2 + μ n 2 + C 1 ) ( σ m 2 + σ n 2 + C 2 ) ,
where SSIM takes a value between 0 and 1. When two images are identical, the SSIM value is one.
The PSNR is a widely used objective criterion for evaluating images in engineering projects. It measures the ratio between the maximum signal and the background noise and is commonly employed to assess the amount of information loss in a compressed image compared with the original. The PSNR is measured in decibels (dB), with higher values indicating better image quality (i.e., less noise). The range of PSNR values is from 0 to positive infinity.
The PSNR is often calculated using the mean squared error (MSE). Considering a pair of monochrome images labeled as I and J, where I signifies an unadulterated original image and J denotes an angry rendition of I (e.g., I as the uncompressed source image and J as its compressed iteration), the calculation of their mean squared error involves the following:
MSE = 1 x y i = 1 x j = 1 y ( I ( i , j ) J ( i , j ) ) 2 .
The PSNR is defined as
PSNR = 10 · log 10 ( MAX I 2 MSE ) = 20 · log 10 ( MAX I MSE ) ,
where MAX is the maximum pixel value of the image and the PSNR is measured in decibels (dB). A higher PSNR value indicates better image quality. Generally speaking, PSNR values above 40 dB are considered excellent, meaning the image quality is very close to the original. PSNR values between 30 dB and 40 dB typically indicate good image quality, with detectable but acceptable distortion. On the other hand, PSNR values between 20 dB and 30 dB signify poor image quality. PSNR values below 20 dB indicate unacceptable image quality. It is worth noting that PSNR values below −30 dB also indicate poor image quality and are considered unacceptable.

3.2. Non-Reference Metrics

In evaluating non-reference images, image clarity is a crucial indicator for measuring image quality. It closely aligns with the subjective perception of individuals, as low image clarity indicates a blurred image. Traditional reference-free methods for image quality evaluation, such as the Brenner gradient function, Tenengrad gradient function, and image information entropy function (Information entropy), can provide insights into the level of image clarity to a certain extent. These methods contribute to the assessment of image quality by quantifying the clarity of the image, thereby facilitating objective evaluations of image quality in the absence of reference images.
The Brenner gradient function serves as a fundamental gradient assessment approach involving the computation of the squared difference between adjacent pixel grayscale values. This function is defined as follows:
B ( f ) = n m f ( m + 2 , n ) f ( m , n ) 2 ,
where f ( m , n ) denotes the grayscale value of the image f corresponding to the pixel ( m , n ) and B ( f ) is the result of the image sharpness calculation.
The Tenengrad gradient function applies the Sobel operator to capture gradients both horizontally and vertically. The definition for image sharpness within the context of the base and Tenengrad gradient functions is outlined as follows:
B ( f ) = n m E ( m , n ) ( E ( m , n ) > Q ) .
E ( m , n ) takes the following form:
E ( m , n ) = E m 2 ( m , n ) + E n 2 ( m , n ) ,
where Q is the given edge detection threshold and E m and E n are the convolution of the Sobel horizontal and vertical edge detection operators at pixel point ( m , n ) , respectively.
The information entropy function stands as a crucial metric for assessing the information abundance within an image. As established by information theory, the information held within an image f is quantified through the information entropy B ( f ) associated with said image:
B ( f ) = a = 0 G 1 P a ln ( P a ) ,
where P a signifies the probability of encountering a pixel with the gray value a within the image and G represents the total count of gray levels (typically set at 256). As per Shannon’s information theory, maximal information is attained when entropy reaches its peak. Applying this principle to image focusing, heightened B ( f ) values correlate with sharper images. It is noteworthy that the entropy function’s sensitivity is not particularly pronounced, and due to image content variations, outcomes might occasionally deviate from actual scenarios.

4. Benchmark Dataset

Deep learning relies on deep neural networks that require extensive training with substantial data samples to achieve generalizability in the final model. Consequently, the dataset size significantly influences deep learning endeavors. To cater to the needs of low-light image enhancement research, there are several publicly available datasets with varying sizes and diverse scenarios. Table 2 provides summary of different low-light image data. These datasets include naturalness preserved enhancement (NPE), Vasileios Vonikakis (VV), the low-light dataset (LOL), and multi-exposure image fusion (MEF), Figure 5 and Figure 6 provide examples of different datasets. Researchers can utilize these datasets to train and evaluate their low-light image enhancement models, providing a range of options to suit different research requirements.
(1)
NPE dataset
In 2013, Wang et al. [72] introduced the NPE dataset, which comprises two components: a natural image dataset and a low-light image dataset. The natural image dataset consists of a diverse collection of indoor and outdoor scenes captured under normal lighting conditions. In contrast, the low-light image dataset is derived from natural images through specific processing techniques that simulate low-illumination conditions. The low-light image dataset further categorizes the images into two levels: “low” and “high”, which represent different degrees of low-light situations.
(2)
MEF dataset
In 2015, Ma et al. [73] introduced the MEF dataset, comprising images capturing multiple low-light scenes. The dataset encompasses a diverse range of scenes and shooting conditions, including indoor and outdoor environments, various lighting scenarios, candlelight, and daytime and nighttime shots. It has found extensive usage in both academic and industrial settings, serving a wide array of applications. These include the research and development of image enhancement algorithms, testing, and evaluation of digital cameras, among other related areas. The MEF dataset has proven to be a valuable resource for advancing image processing techniques, enabling advancements in the field of low-light photography, and supporting various imaging technology-related endeavors.
(3)
VV dataset
The VV dataset was explicitly designed to explore images that are significantly underexposed or overexposed, exhibiting a key characteristic of having areas that are properly exposed alongside regions that are grossly underexposed or overexposed. This dataset serves as a valuable resource for evaluating algorithms focused on local exposure correction and enhancement. It provides an ideal testing ground for techniques aiming to address the challenges posed by varying exposure levels within an image.
(4)
SID dataset
In 2018, Chen et al. [74] introduced the SID dataset to facilitate the advancement of deep learning approaches in low-light image processing. The dataset encompasses 5094 original short-exposure images, each with a matching long-exposure reference image. These image pairs were obtained using two distinct camera models: the Sony α 7SII and the Fujifilm X-T2. Raw data were collected under low-light conditions with short exposure times, typically around 0.1 seconds or 0.04 seconds. The ground truth images were captured with long exposure times of 10 or 30 seconds, ensuring minimal noise presence. The SID dataset offers a valuable resource for training and evaluating deep learning models in the domain of low-light image processing, providing paired images with well-defined exposure characteristics.
(5)
LOL dataset
In 2018, Wei et al. [32] introduced the LOL dataset, a paired collection comprising 500 low-light and normal-light image pairs. These pairs were further divided into 485 training pairs and 15 test pairs for evaluation purposes. The low-light images in the dataset accurately represent the noise typically encountered during the process of capturing photographs. The majority of the images depict indoor scenes and are sourced from diverse scenes and devices, including cell phones and digital cameras. As a result, the dataset encompasses a wide variety of objects captured within the images. Additionally, the LOL dataset covers various low-light conditions, including twilight and indoor low-light scenarios, providing a comprehensive representation of challenging lighting situations. To ensure consistency and comparability, all raw images were adjusted to a standardized resolution of 400 × 600 pixels and converted to the portable web graphics format. Its diverse range of images and realistic representation of low-light conditions make it a valuable tool for research and development in the field.
(6)
SICE dataset
In 2019, Cai et al. [34] presented the SICE dataset, an extensive collection of multi-exposure images. The dataset creation process involved deploying multi-exposure fusion (MEF) and high dynamic range (HDR) techniques to reconstruct reference images, yielding heightened contrast and visibility improvements. To create the SICE dataset, the authors employed 1200 sequences and applied 13 MEF and HDR algorithms, resulting in a total of 15,600 fusion results (1200 sequences × 13 algorithms). The dataset comprises 589 meticulously selected high-resolution multiple-exposure sequences, consisting of a total of 4413 images. For each sequence, a set of contrast-enhanced images was produced using 13 diverse multiple-exposure image fusion techniques and a stack-based high dynamic range imaging algorithm. Following this, subjective evaluations were carried out to determine the optimal reference images for each scene.
(7)
ExDark dataset
In 2019, Loh et al. [75] introduced the Exclusively Dark (ExDark) dataset, consisting of 7363 low-light images. The dataset covers a range of low-light conditions from very low-light environments to twilight, encompassing 10 different lighting conditions. The images were captured in various real-world scenarios using a diverse array of devices and cameras, providing a wide representation of different scenes and shooting conditions. These low-light images are subject to multiple factors, including insufficient lighting, noise, and blur, which further contribute to the challenges associated with low-light photography. The dataset includes annotations at both the image class level and the local object bounding box level for 12 object classes, such as bicycle, car, cat, dog, chair, and cup. The ExDark dataset serves as a valuable resource for researchers and practitioners in the field of low-light image analysis. Its comprehensive collection of low-light images, diverse lighting conditions, and detailed object annotations provide an excellent foundation for developing and evaluating algorithms related to object recognition, localization, and other tasks in challenging low-light environments.
(8)
RELLISUR dataset
In 2021, Aakerberg et al. [76] introduced the RELLISUR dataset, a large-scale paired dataset specifically designed for low-light and low-resolution image enhancement tasks. The RELLISUR dataset comprises 12,750 paired images, consisting of real low-light and low-resolution images paired with high-resolution reference images captured under normal lighting conditions. The dataset covers a wide range of resolutions and low-light levels, enabling the development and training of deep learning-based models. It enables the exploration and development of deep learning models that can effectively enhance the quality and resolution of low-light images, bridging the gap between these two important image enhancement tasks.
(9)
LLIV-Phone dataset
In 2021, Li et al. [77] introduced the LLIV-Phone dataset, a comprehensive and challenging dataset specifically designed for low-illumination image and video analysis. The LLIV-Phone dataset comprises 120 videos and 45,148 images captured using 18 different cell phone cameras. The dataset covers a wide range of indoor and outdoor scenes with diverse lighting conditions, including low light, underexposure, moonlight, dusk, darkness, extreme darkness, backlighting, non-uniform lighting, and colored lighting. These real-world scenes present various challenges associated with low-light conditions. It offers a wide variety of low-light images and videos collected from real scenes, making it suitable for testing and comparing the performance of different enhancement algorithms. By providing a comprehensive collection of real low-illumination images and videos, the LLIV-Phone dataset significantly contributes to the advancement of research in low-light image and video enhancement, enabling the development and evaluation of robust algorithms in this domain.

5. Experimental Evaluation and Analysis

5.1. Performance Comparison

To evaluate and analyze different methods, experiments were conducted using various datasets, including NPE [72], MEF [73], VV, and ExDark [75]. The comparison methods considered in the study included CERL [78], Zero-DCE [39], Zero-DCE++ [45], and SCI [49]. For quantitative analysis, the researchers selected the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) as evaluation metrics. These metrics were used to measure and analyze the performance of the selected methods. The researchers utilized the pretrained models provided in the original papers for validation and comparison. The comparison results were divided into two main parts: qualitative and quantitative evaluation. Qualitative evaluation involves human perception and judgment of the visual quality of the enhanced images, due to the unique capabilities and sensitivities of the human visual system in image processing and perception. The quantitative evaluation focuses on quantitative metrics, such as the PSNR and SSIM, to provide a more systematic analysis of the performance of the methods. By conducting qualitative and quantitative evaluations, the researchers aimed to comprehensively assess and compare the selected methods by using multiple criteria to elucidate their strengths and weaknesses in enhancing low-light images.

5.2. Qualitative Evaluation

Qualitative evaluation involves comparing the visual differences between the enhanced image and the original low-light image based on human perception. It typically entails presenting the same low-illumination image processed by different algorithms to individuals, who are then asked to evaluate and select the best algorithm based on their visual judgment. However, qualitative evaluation is susceptible to various external factors that can introduce subjectivity and make it challenging to establish a fixed standard. Factors like individual aesthetics, color preferences, and variations in observation angles can influence the evaluation. These subjective factors contribute to the difficulty of achieving a standardized and scientific subjective evaluation. While qualitative evaluation provides valuable insights into the visual effect and overall preference of the enhanced images, it is important to acknowledge its limitations and potential biases. Therefore, it is often combined with quantitative evaluation methods, such as quantitative metrics like the PSNR and SSIM, which provide more standardized and measurable criteria for assessing the performance of image enhancement algorithms.
Figure 7 presents a qualitative comparison of several different methods using four datasets. Upon examining the qualitative comparison images in Figure 7, certain observations can be made regarding the visual effects of the algorithms. The Zero-DCE [39] algorithm model demonstrated superior subjective visual effects. The enhanced image exhibited appropriate brightness without overexposure, and it effectively extracted detailed information from the image, resulting in a more visually pleasing outcome. In contrast, the CERL [78] algorithm model did not appear to be particularly effective in terms of overall image enhancement. The quantitative evaluation suggests that it may not achieve satisfactory results for improving the overall visual quality of the images. The SCI-easy [49] algorithm model showed room for improvement, particularly in addressing local darkness. The images revealed uneven enhancement in certain areas, indicating the need for further refinement to achieve more consistent results. On the other hand, the SCI-medium [49] algorithm model performed well in enhancing extremely dark images. However, for certain specific images, there was an issue of overexposure, where certain areas of the image appeared to be excessively bright. These subjective observations highlight the strengths and weaknesses of each algorithm in terms of their visual effects.

5.3. Quantitative Evaluation

Quantitative evaluation involves assessing image enhancement algorithms using specific criteria and quantitative measurements. In image enhancement, this evaluation method quantifies the differences between the original and enhanced images using mathematical models and calculates the indicators to determine the quality of the image. Quantitative evaluation methods are characterized by their simplicity, computational efficiency, and ability to provide quantitative assessments based on established models, ensuring high stability and reproducibility. The evaluation indicators used in the compared methods in this paper include the PSNR and SSIM. These quantitative evaluation metrics enable a quantitative assessment of image enhancement algorithms, allowing for direct comparisons and providing a basis for performance analysis.
Table 3, Table 4 and Table 5 provide a quantitative analysis of several different methods, using the PSNR, SSIM, and NIQE as evaluation metrics. In these three metrics, the higher the values of the PSNR and SSIM, the better the result, reflecting an improvement in image quality, while the opposite is true for the NIQE; the smaller the value of the NIQE, the better the perceived quality. Upon reviewing the tables, it can be observed that the SCI-easy [49] algorithm consistently achieved excellent results for the PSNR, SSIM, and NIQE. The algorithm’s performance surpassed that of the other methods evaluated, demonstrating better quantitative metrics in terms of image quality. These quantitative assessments provide valuable insights into the algorithms’ performance and their ability to enhance image quality. It is important to consider these metrics along with qualitative evaluations and other factors to gain a comprehensive understanding of the algorithms’ effectiveness in different aspects.

6. Summary and Outlook

Low-light image enhancement plays a crucial role in various applications, both in everyday life and in industries. Deep learning technology has emerged as a powerful tool for tackling this challenge. However, there are still certain limitations associated with deep learning-based approaches. One notable drawback is the heavy reliance on a substantial amount of training data, which can be time-consuming to collect and annotate. Moreover, the selection of the dataset needs to be carefully considered to ensure it represents a wide range of scenes and categories, which adds complexity to the training process. Additionally, some existing deep learning methods for low-light image enhancement primarily focus on improving model performance, potentially overlooking the practicality and generalizability of the results. It is crucial to address these limitations and propose algorithms that benefit higher-level image processing tasks. Efforts should be made to develop techniques that require less training data and have shorter training times while still achieving high-quality results. Furthermore, it is important to prioritize the practical application and universality of the enhanced images. By addressing these challenges, researchers can pave the way for more efficient and effective low-light image enhancement methods that align with the needs of various image processing tasks in real-world scenarios.
Finally, this paper will introduce some possible future research directions. First of all, integrating models from different tasks to achieve multi-task image enhancement is an important area for exploration. By combining the strengths of various models, researchers can develop more comprehensive and versatile solutions. Secondly, addressing the generalization problem of deep learning models is crucial. Enhancing the models’ ability to perform well in diverse scenarios and adapt to different image characteristics will be a key focus. This can involve techniques such as transfer learning, domain adaptation, or developing robust feature representations. Lastly, improving the speed and efficiency of deep learning models is a prominent concern. Developing algorithms that can achieve real-time or near-real-time processing for low-light image enhancement tasks is essential for practical applications.
In summary, deep learning-based low-light image enhancement techniques hold great potential in computer vision and image processing domains. Future research should focus on exploring more efficient and accurate deep-learning models to achieve superior image enhancement effects. By addressing these research directions, we can advance the state of the art in low-light image enhancement and facilitate its broader applications in various fields.

Author Contributions

Conceptualization, W.Z. and P.Q.; methodology, Z.T.; software, Z.T. and J.L.; validation, Z.T. and Y.S.; formal analysis, G.L. and Z.L.; investigation, Z.T., J.L. and Y.S.; writing—original draft preparation, Z.T. and P.Q.; writing—review and editing, Z.T. and W.Z.; visualization, Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Henan Province under Grants 232300420428 and 212300410345, and the Major Special Project of Xinxiang City under Grants 21ZD003, and the Key Specialized Research and Development Program of Science and Technology of Henan Province under Grants 232102111127, 232102210058, and 232102210018 and the Innovation Training Program for college Students of Henan Province under Grants 202310467015, 202310467007, 202310467031.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are publicly available data (sources stated in the citations). Please contact the corresponding author regarding data availability.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, W.; Zhuang, P.; Sun, H.-H.; Li, G.; Kwong, S.; Li, C. Underwater Image Enhancement via Minimal Color Loss and Locally Adaptive Contrast Enhancement. IEEE Trans. Image Process. 2022, 31, 3997–4010. [Google Scholar] [CrossRef]
  2. Zhang, Q.; Yuan, Q.; Song, M.; Yu, H.; Zhang, L. Cooperated Spectral Low-Rankness Prior and Deep Spatial Prior for HSI Unsupervised Denoising. IEEE Trans. Image Process. 2022, 31, 6356–6368. [Google Scholar]
  3. Liu, Y.; Yan, Z.; Ye, T.; Wu, A.; Li, Y. Single Nighttime Image Dehazing Based on Unified Variational Decomposition Model and Multi-Scale Contrast Enhancement. Eng. Appl. Artif. Intell. 2022, 116, 105373. [Google Scholar]
  4. Sun, H.-H.; Lee, Y.H.; Dai, Q.; Li, C.; Ow, G.; Yusof, M.L.M.; Yucel, A.C. Estimating Parameters of the Tree Root in Heterogeneous Soil Environments via Mask-Guided Multi-Polarimetric Integration Neural Network. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  5. Xiong, J.; Liu, G.; Liu, Y.; Liu, M. Oracle Bone Inscriptions Information Processing Based on Multi-Modal Knowledge Graph. Comput. Electr. Eng. 2021, 92, 107173. [Google Scholar] [CrossRef]
  6. Zhang, W.; Dong, L.; Xu, W. Retinex-Inspired Color Correction and Detail Preserved Fusion for Underwater Image Enhancement. Comput. Electron. Agric. 2022, 192, 106585. [Google Scholar]
  7. Sun, H.-H.; Cheng, W.; Fan, Z. Learning to Remove Clutter in Real-World GPR Images Using Hybrid Data. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar]
  8. Liu, Y.; Yan, Z.; Wu, A.; Ye, T.; Li, Y. Nighttime Image Dehazing Based on Variational Decomposition Model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 640–649. [Google Scholar]
  9. Zhang, W.; Li, Z.; Sun, H.-H.; Zhang, Q.; Zhuang, P.; Li, C. SSTNet: Spatial, Spectral, and Texture Aware Attention Network Using Hyperspectral Image for Corn Variety Identification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar]
  10. Zhou, J.; Li, B.; Zhang, D.; Yuan, J.; Zhang, W.; Cai, Z.; Shi, J. UGIF-Net: An Efficient Fully Guided Information Flow Network for Underwater Image Enhancement. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4206117. [Google Scholar] [CrossRef]
  11. Zhang, W.; Jin, S.; Zhuang, P.; Liang, Z.; Li, C. Underwater Image Enhancement via Piecewise Color Correction and Dual Prior Optimized Contrast Enhancement. IEEE Signal Process. Lett. 2023, 30, 229–233. [Google Scholar] [CrossRef]
  12. Pan, X.; Cheng, J.; Hou, F.; Lan, R.; Lu, C.; Li, L.; Feng, Z.; Wang, H.; Liang, C.; Liu, Z. SMILE: Cost-Sensitive Multi-Task Learning for Nuclear Segmentation and Classification with Imbalanced Annotations. Med. Image Anal. 2023, 116, 102867. [Google Scholar]
  13. Liu, Y.; Teng, Q.; He, X.; Ren, C.; Chen, H. Multimodal Sensors Image Fusion for Higher Resolution Remote Sensing Pan Sharpening. IEEE Sens. J. 2022, 22, 18021–18034. [Google Scholar] [CrossRef]
  14. Zhang, W.; Zhou, L.; Zhuang, P.; Li, G.; Pan, X.; Zhao, W.; Li, C. Underwater Image Enhancement via Weighted Wavelet Visual Perception Fusion. IEEE Trans. Circuits Syst. Video Technol. 2023. [Google Scholar] [CrossRef]
  15. Zhou, S.; Li, C.; Change Loy, C. Lednet: Joint Low-Light Enhancement and Deblurring in the Dark. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 573–589. [Google Scholar]
  16. Chen, H.; He, X.; Yang, H.; Wu, Y.; Qing, L.; Sheriff, R.E. Self-Supervised Cycle-Consistent Learning for Scale-Arbitrary Real-World Single Image Super-Resolution. Expert Syst. Appl. 2023, 212, 118657. [Google Scholar] [CrossRef]
  17. Liu, Y.; Wang, A.; Zhou, H.; Jia, P. Single Nighttime Image Dehazing Based on Image Decomposition. Signal Process. 2021, 183, 107986. [Google Scholar] [CrossRef]
  18. Yue, H.; Guo, J.; Yin, X.; Zhang, Y.; Zheng, S.; Zhang, Z.; Li, C. Salient Object Detection in Low-Light Images via Functional Optimization-Inspired Feature Polishing. Knowl.-Based Syst. 2022, 257, 109938. [Google Scholar] [CrossRef]
  19. Zhuang, P.; Wu, J.; Porikli, F.; Li, C. Underwater Image Enhancement With Hyper-Laplacian Reflectance Priors. IEEE Trans. Image Process. 2022, 31, 5442–5455. [Google Scholar]
  20. Liu, Y.; Yan, Z.; Tan, J.; Li, Y. Multi-Purpose Oriented Single Nighttime Image Haze Removal Based on Unified Variational Retinex Model. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1643–1657. [Google Scholar] [CrossRef]
  21. Zhang, W.; Wang, Y.; Li, C. Underwater Image Enhancement by Attenuated Color Channel Correction and Detail Preserved Contrast Enhancement. IEEE J. Ocean. Eng. 2022, 47, 718–735. [Google Scholar] [CrossRef]
  22. He, J.; He, X.; Zhang, M.; Xiong, S.; Chen, H. Deep Dual-Domain Semi-Blind Network for Compressed Image Quality Enhancement. Knowl.-Based Syst. 2022, 238, 107870. [Google Scholar]
  23. Liu, Q.; He, X.; Teng, Q.; Qing, L.; Chen, H. BDNet: A BERT-Based Dual-Path Network for Text-to-Image Cross-Modal Person Re-Identification. Pattern Recognit. 2023, 141, 109636. [Google Scholar] [CrossRef]
  24. Huang, S.-C.; Cheng, F.-C.; Chiu, Y.-S. Efficient Contrast Enhancement Using Adaptive Gamma Correction with Weighting Distribution. IEEE Trans. Image Process. 2012, 22, 1032–1041. [Google Scholar] [CrossRef] [PubMed]
  25. Wang, Q.; Fu, X.; Zhang, X.-P.; Ding, X. A Fusion-Based Method for Single Backlit Image Enhancement. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 4077–4081. [Google Scholar]
  26. Li, M.; Liu, J.; Yang, W.; Sun, X.; Guo, Z. Structure-Revealing Low-Light Image Enhancement via Robust Retinex Model. IEEE Trans. Image Process. 2018, 27, 2828–2841. [Google Scholar] [CrossRef] [PubMed]
  27. Fu, G.; Duan, L.; Xiao, C. A Hybrid L2−Lp Variational Model for Single Low-Light Image Enhancement with Bright Channel Prior. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1925–1929. [Google Scholar]
  28. Ren, X.; Yang, W.; Cheng, W.-H.; Liu, J. LR3M: Robust Low-Light Enhancement via Low-Rank Regularized Retinex Model. IEEE Trans. Image Process. 2020, 29, 5862–5876. [Google Scholar] [CrossRef]
  29. Ueda, Y.; Moriyama, D.; Koga, T.; Suetake, N. Histogram specification-based image enhancement for backlit image. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 958–962. [Google Scholar] [CrossRef]
  30. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A Deep Autoencoder Approach to Natural Low-Light Image Enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  31. Li, C.; Guo, J.; Porikli, F.; Pang, Y. LightenNet: A Convolutional Neural Network for Weakly Illuminated Image Enhancement. Pattern Recognit. Lett. 2018, 104, 15–22. [Google Scholar] [CrossRef]
  32. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
  33. Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-Light Image/Video Enhancement Using CNNs; BMVC: Newcastle, UK, 2018; Volume 220, p. 4. [Google Scholar]
  34. Cai, J.; Gu, S.; Zhang, L. Learning a Deep Single Image Contrast Enhancer from Multi-Exposure Images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef]
  35. Zhang, Y.; Zhang, J.; Guo, X. Kindling the Darkness: A Practical Low-Light Image Enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  36. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep Light Enhancement without Paired Supervision. IEEE Trans. Image Process. 2021, 31, 2340–2349. [Google Scholar] [CrossRef]
  37. Zhang, L.; Zhang, L.; Liu, X.; Shen, Y.; Zhang, S.; Zhao, S. Zero-Shot Restoration of Back-Lit Images Using Deep Internal Learning. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1623–1631. [Google Scholar]
  38. Wang, R.; Zhang, Q.; Fu, C.-W.; Shen, X.; Zheng, W.-S.; Jia, J. Underexposed Photo Enhancement Using Deep Illumination Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6849–6857. [Google Scholar]
  39. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  40. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. From Fidelity to Perceptual Quality: A Semi-Supervised Approach for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3063–3072. [Google Scholar]
  41. Zhu, M.; Pan, P.; Chen, W.; Yang, Y. Eemefn: Low-Light Image Enhancement via Edge-Enhanced Multi-Exposure Fusion Network. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13106–13113. [Google Scholar]
  42. Lu, K.; Zhang, L. TBEFN: A Two-Branch Exposure-Fusion Network for Low-Light Image Enhancement. IEEE Trans. Multimed. 2020, 23, 4093–4105. [Google Scholar] [CrossRef]
  43. Lim, S.; Kim, W. DSLR: Deep Stacked Laplacian Restorer for Low-Light Image Enhancement. IEEE Trans. Multimed. 2020, 23, 4272–4284. [Google Scholar] [CrossRef]
  44. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-Inspired Unrolling with Cooperative Prior Architecture Search for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
  45. Li, C.; Guo, C.; Chen, C.L. Learning to Enhance Low-Light Image via Zero-Reference Deep Curve Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4225–4238. [Google Scholar] [CrossRef] [PubMed]
  46. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. Band Representation-Based Semi-Supervised Low-Light Image Enhancement: Bridging the Gap between Signal Fidelity and Perceptual Quality. IEEE Trans. Image Process. 2021, 30, 3461–3473. [Google Scholar] [CrossRef]
  47. Zhao, Z.; Xiong, B.; Wang, L.; Ou, Q.; Yu, L.; Kuang, F. RetinexDIP: A Unified Deep Framework for Low-Light Image Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1076–1088. [Google Scholar] [CrossRef]
  48. Li, J.; Feng, X.; Hua, Z. Low-Light Image Enhancement via Progressive-Recursive Network. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4227–4240. [Google Scholar] [CrossRef]
  49. Ma, L.; Ma, T.; Liu, R.; Fan, X.; Luo, Z. Toward Fast, Flexible, and Robust Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5637–5646. [Google Scholar]
  50. Huang, H.; Yang, W.; Hu, Y.; Liu, J.; Duan, L.-Y. Towards Low Light Enhancement with Raw Images. IEEE Trans. Image Process. 2022, 31, 1391–1405. [Google Scholar] [CrossRef] [PubMed]
  51. Wang, K.; Cui, Z.; Wu, G.; Zhuang, Y.; Qian, Y. Linear Array Network for Low-Light Image Enhancement. arXiv 2022, arXiv:2201.08996. [Google Scholar]
  52. Panagiotou, S.; Bosman, A.S. Denoising Diffusion Post-Processing for Low-Light Image Enhancement. arXiv 2023, arXiv:2303.09627. [Google Scholar]
  53. Zhang, Y.; Di, X.; Wu, J.; FU, R.; Li, Y.; Wang, Y.; Xu, Y.; YANG, G.; Wang, C. A Fast and Lightweight Network for Low-Light Image Enhancement. arXiv 2023, arXiv:2304.02978. [Google Scholar]
  54. Yang, S.; Ding, M.; Wu, Y.; Li, Z.; Zhang, J. Implicit Neural Representation for Cooperative Low-Light Image Enhancement. arXiv 2023, arXiv:2303.11722. [Google Scholar]
  55. Wu, Y.; Pan, C.; Wang, G.; Yang, Y.; Wei, J.; Li, C.; Shen, H.T. Learning Semantic-Aware Knowledge Guidance for Low-Light Image Enhancement. arXiv 2023, arXiv:2304.07039. [Google Scholar]
  56. Ren, W.; Liu, S.; Ma, L.; Xu, Q.; Xu, X.; Cao, X.; Du, J.; Yang, M.-H. Low-Light Image Enhancement via a Deep Hybrid Network. IEEE Trans. Image Process. 2019, 28, 4364–4375. [Google Scholar] [CrossRef]
  57. Tao, L.; Zhu, C.; Xiang, G.; Li, Y.; Jia, H.; Xie, X. LLCNN: A Convolutional Neural Network for Low-Light Image Enhancement. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), IEEE, St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  58. Xu, K.; Yang, X.; Yin, B.; Lau, R.W. Learning to Restore Low-Light Images via Decomposition-and-Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2281–2290. [Google Scholar]
  59. Gharbi, M.; Chen, J.; Barron, J.T.; Hasinoff, S.W.; Durand, F. Deep Bilateral Learning for Real-Time Image Enhancement. ACM Trans. Graph. (TOG) 2017, 36, 1–12. [Google Scholar] [CrossRef]
  60. Shen, L.; Yue, Z.; Feng, F.; Chen, Q.; Liu, S.; Ma, J. Msr-Net: Low-Light Image Enhancement Using Deep Convolutional Network. arXiv 2017, arXiv:1711.02488. [Google Scholar]
  61. Wu, H.; Zheng, S.; Zhang, J.; Huang, K. Fast End-to-End Trainable Guided Filter. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1838–1847. [Google Scholar]
  62. Wei, K.; Fu, Y.; Huang, H. 3-D Quasi-Recurrent Neural Network for Hyperspectral Image Denoising. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 363–375. [Google Scholar] [CrossRef]
  63. Meng, Y.; Kong, D.; Zhu, Z.; Zhao, Y. From Night to Day: GANs Based Low Quality Image Enhancement. Neural Process. Lett. 2019, 50, 799–814. [Google Scholar] [CrossRef]
  64. Ignatov, A.; Kobyshev, N.; Timofte, R.; Vanhoey, K.; Van Gool, L. Dslr-Quality Photos on Mobile Devices with Deep Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3277–3285. [Google Scholar]
  65. Ignatov, A.; Kobyshev, N.; Timofte, R.; Vanhoey, K.; Van Gool, L. Wespe: Weakly Supervised Photo Enhancer for Digital Cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 691–700. [Google Scholar]
  66. Chen, Y.-S.; Wang, Y.-C.; Kao, M.-H.; Chuang, Y.-Y. Deep Photo Enhancer: Unpaired Learning for Image Enhancement from Photographs with Gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6306–6314. [Google Scholar]
  67. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  68. Fu, Y.; Hong, Y.; Chen, L.; You, S. LE-GAN: Unsupervised Low-Light Image Enhancement Network Using Attention Module and Identity Invariant Loss. Knowl.-Based Syst. 2022, 240, 108010. [Google Scholar] [CrossRef]
  69. Yan, L.; Fu, J.; Wang, C.; Ye, Z.; Chen, H.; Ling, H. Enhanced Network Optimized Generative Adversarial Network for Image Enhancement. Multimed. Tools Appl. 2021, 80, 14363–14381. [Google Scholar] [CrossRef]
  70. You, Q.; Wan, C.; Sun, J.; Shen, J.; Ye, H.; Yu, Q. Fundus Image Enhancement Method Based on CycleGAN. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), IEEE, Berlin, Germany, 23–27 July 2019; pp. 4500–4503. [Google Scholar]
  71. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  72. Wang, S.; Zheng, J.; Hu, H.-M.; Li, B. Naturalness Preserved Enhancement Algorithm for Non-Uniform Illumination Images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef]
  73. Ma, K.; Zeng, K.; Wang, Z. Perceptual Quality Assessment for Multi-Exposure Image Fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef] [PubMed]
  74. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to See in the Dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3291–3300. [Google Scholar]
  75. Loh, Y.P.; Chan, C.S. Getting to Know Low-Light Images with the Exclusively Dark Dataset. Comput. Vis. Image Underst. 2019, 178, 30–42. [Google Scholar] [CrossRef]
  76. Aakerberg, A.; Nasrollahi, K.; Moeslund, T.B. RELLISUR: A Real Low-Light Image Super-Resolution Dataset. In Proceedings of the Thirty-fifth Conference on Neural Information Processing Systems-NeurIPS 2021, Online, 6–14 December 2021. [Google Scholar]
  77. Li, C.; Guo, C.; Han, L.; Jiang, J.; Cheng, M.-M.; Gu, J.; Loy, C.C. Low-Light Image and Video Enhancement Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9396–9416. [Google Scholar] [CrossRef] [PubMed]
  78. Chen, Z.; Jiang, Y.; Liu, D.; Wang, Z. CERL: A Unified Optimization Framework for Light Enhancement With Realistic Noise. IEEE Trans. Image Process. 2022, 31, 4162–4172. [Google Scholar] [CrossRef]
Figure 1. Examples of images under suboptimal lighting conditions.
Figure 1. Examples of images under suboptimal lighting conditions.
Sensors 23 07763 g001
Figure 2. Flow chart of CNN method combined with physical model.
Figure 2. Flow chart of CNN method combined with physical model.
Sensors 23 07763 g002
Figure 3. Flow chart of CNN method for non-physical model.
Figure 3. Flow chart of CNN method for non-physical model.
Sensors 23 07763 g003
Figure 4. Flow chart of GAN-based method.
Figure 4. Flow chart of GAN-based method.
Sensors 23 07763 g004
Figure 5. Example of a partial low-light dataset.
Figure 5. Example of a partial low-light dataset.
Sensors 23 07763 g005
Figure 6. Example of paired low-light dataset for LOL. Here, (a) is a reference image, and (b) is a low light image.
Figure 6. Example of paired low-light dataset for LOL. Here, (a) is a reference image, and (b) is a low light image.
Sensors 23 07763 g006
Figure 7. Qualitative comparison of different low-light image enhancement algorithms. Where, (a) represents the original image, (b) represents the result map after processing by the CERL [78] method, (c) represents the result map after processing by the Zero-DCE [39] method, (d) represents the result map after processing by the Zero-DCE++ [45] method, (e) represents the result map after processing by the SCI-difficult [49] method, (f) represents the result map after processing by the SCI-easy [49] method, and (g) represents the results of the SCI-medium [49] method.
Figure 7. Qualitative comparison of different low-light image enhancement algorithms. Where, (a) represents the original image, (b) represents the result map after processing by the CERL [78] method, (c) represents the result map after processing by the Zero-DCE [39] method, (d) represents the result map after processing by the Zero-DCE++ [45] method, (e) represents the result map after processing by the SCI-difficult [49] method, (f) represents the result map after processing by the SCI-easy [49] method, and (g) represents the results of the SCI-medium [49] method.
Sensors 23 07763 g007
Table 1. Summary of the basic characteristics of representative methods based on deep learning.
Table 1. Summary of the basic characteristics of representative methods based on deep learning.
YearMethodsNetwork StructureTraining DataTest DataEvaluation MetricPlatform
2017LLNet [30]SSDASimulated by gamma correction and Gaussian noiseSimulated self-selectedPSNR SSIMTheano
2018LightenNet [31]Four layersSimulated by random illumination valuesSimulated self-selectedPSNR MAE SSIM user studyCaffe MATLAB
Retinex-Net [32]Multi-scale networkLOL simulated by adjusting histogramSelf-selected-TensorFlow
MBLLEN [33]Multi-branch fusionSimulated by gamma correction and Poisson noiseSimulated self-selectedPSNR SSIM AB VIF LOE TOMITensorFlow
SICE [34]Frequency decompositionSICESICEPSNR FSIM runtime FLOPsCaffe MATLAB
2019KinD [35]Three subnetworks U-NetLOLLOL LIME NPE MEFPSNR SSIM LOE NIQETensorFlow
EnlightenGAN [36]U-Net-like networkUnpaired real imagesNPE LIME MEF DICM VV BBD-100K ExDARKUser study NIQE classificationPyTorch
ExCNet [37]Fully connected layersReal imagesIEpxDUser study CDIQA LODPyTorch
DeepUPE [38]Illumination mapRetouched image pairsMIT-Adobe FiveKPSNR SSIM user studyTensorFlow
2020Zero-DCE [39]U-Net-like networkSICESICE NPE LIME MEF DICM VV DARK FACEUser study PI PNSR SSIM MAE runtime face detectionPyTorch
DRBN [40]Recursive networkLOL images selected by MOSLOLPSNR SSIM SSIM-GCPyTorch
EEMEFN [41]U-Net-like network edge detection networkSIDSIDPSNR SSIMTensorFlow Paddle
TBEFN [42]Three stages U-Net-like networkSCIE LOLSCIE LOL DICM MEF NPE VVPSNR SSIM NIQE runtime P FLOPsTensorFlow
DSLR [43]Laplacian pyramid U-Net-like networkMIT-Adobe FiveKMIT-Adobe FiveK self-selectedPSNR SSIM NIQMC NIQE BTMQI CaHDCPyTorch
2021RUAS [44]Neural architecture searchLOL MIT-Adobe FiveKLOL MIT-Adobe FiveKPSNR SSIM runtime P FLOPsPyTorch
Zero-DCE++ [45]U-Net-like networkSICESICE NPE LIME MEF DICM VV DARK FACEUser study PI PNSR SSIM P MAE runtime face detection FLOPsPyTorch
DRBN [46]Recursive networkLOLLOLPSNR SSIM SSIM-GCPyTorch
RetinexDIP [47]Encoder-decoder networks-DICM, ExDark fusion LIME NASA NPE VVNIQE NIQMC CPCQIPyTorch
PRIEN [48]Recursive networkMEF LOL simulated by adjusting histogramLOL LIME NPE MEF VVPNSR SSIM LOE TMQIPyTorch
2022SCI [49]Self-calibrated illumination networkMIT LOL LSRW DARK FACEMIT LSRW DARK FACE ACDCPSNR SSIM DE EME LOE NIQEPyTorch
LEDNet [15]Encoder-decoder networksLOL-BlurLOL-BlurPSNR SSIM MUSIQ NRQM NIQEPyTorch
REENet [50]Three subnetworksSIDSIDPSNR SSIM VIF NIQE LPIPSTensorFlow
LANNet [51]U-Net-like networkLOL SIDLOL SIDPSNR SSIM GMSD NLPD NIQE DISTSPyTorch
2023LPDM [52]Diffusion modelLOLLIME DICM MEF NPESSIM PSNR MAE LPIPS NIQE BRISQUE SPAQPyTorch
FLW-Net [53]Two-stage networkLOL-V1 LOL-V2LOL-V1 LOL-V2PSNR SSIM NIQEPyTorch
NeRCo [54]Encoder-decoder networksLSRWLOL LSRW LIMEPSNR SSIM NIQE LOEPyTorch
SKF [55]Encoder-decoder networks-LOL LOL-v2 MEF LIME NPE DICMPSNR SSIM LPIPS NIQEPyTorch
Table 2. Summary of different low-light image datasets.
Table 2. Summary of different low-light image datasets.
NameYearQuantityFeaturesType
NPE201384Multi-scene natural imagesReal
MEF201517Fusion imagesReal
VV/24Uneven local exposureReal
SID20185094Combination of long and short exposuresReal
LOL2018500 pairsPaired normal and low-light imagesSynthetic + Real
SCIE20194413Large-scale multi-exposure imagesReal
ExDark20197363Multi-category, multi-sceneReal
RELLISUR202112,750Different resolutions, pairsReal
LLIV-Phone202145,148Large scale, image and videoReal
Table 3. Quantitative comparison of different deep learning low-light image enhancement algorithms in terms of PSNR.
Table 3. Quantitative comparison of different deep learning low-light image enhancement algorithms in terms of PSNR.
MetricsPSNR
MethodsCERLZero-DCEZero-DCE++SCI-difficultSCI-easySCI-medium
NPE17.93214.50913.96313.98918.89212.185
MEF17.53711.811.84111.84218.69310.279
VV18.00615.60613.98414.01617.45911.481
ExDark17.18815.46812.6212.63314.359.898
Table 4. Quantitative comparison of different deep learning low-light image enhancement algorithms in terms of SSIM.
Table 4. Quantitative comparison of different deep learning low-light image enhancement algorithms in terms of SSIM.
MetricsSSIM
MethodsCERLZero-DCEZero-DCE++SCI-difficultSCI-easySCI-medium
NPE0.7290.3220.3230.3260.6770.268
MEF0.7680.4280.4380.4320.7910.417
VV0.6890.5000.5320.5280.7900.488
ExDark0.7020.5670.5800.5830.7970.535
Table 5. Quantitative comparison of different deep learning low-light image enhancement algorithms in terms of NIQE.
Table 5. Quantitative comparison of different deep learning low-light image enhancement algorithms in terms of NIQE.
MetricsNIQE
MethodsInputCERLZero-DCEZero-DCE++SCI-difficultSCI-easySCI-medium
NPE4.3192.9593.0822.5523.0822.8383.052
MEF4.2653.7603.1563.4343.9633.2613.201
VV3.5252.6153.1453.2112.8152.7403.083
ExDark4.4353.6093.6512.7293.8363.2843.882
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tian, Z.; Qu, P.; Li, J.; Sun, Y.; Li, G.; Liang, Z.; Zhang, W. A Survey of Deep Learning-Based Low-Light Image Enhancement. Sensors 2023, 23, 7763. https://doi.org/10.3390/s23187763

AMA Style

Tian Z, Qu P, Li J, Sun Y, Li G, Liang Z, Zhang W. A Survey of Deep Learning-Based Low-Light Image Enhancement. Sensors. 2023; 23(18):7763. https://doi.org/10.3390/s23187763

Chicago/Turabian Style

Tian, Zhen, Peixin Qu, Jielin Li, Yukun Sun, Guohou Li, Zheng Liang, and Weidong Zhang. 2023. "A Survey of Deep Learning-Based Low-Light Image Enhancement" Sensors 23, no. 18: 7763. https://doi.org/10.3390/s23187763

APA Style

Tian, Z., Qu, P., Li, J., Sun, Y., Li, G., Liang, Z., & Zhang, W. (2023). A Survey of Deep Learning-Based Low-Light Image Enhancement. Sensors, 23(18), 7763. https://doi.org/10.3390/s23187763

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop