Next Article in Journal
Influence of MnO2-Birnessite Microstructure on the Electrochemical Performance of Aqueous Zinc Ion Batteries
Previous Article in Journal
Brake-Disc Holes and Slit Shape Design to Improve Heat Dissipation Performance and Structural Stability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Which Features Are More Correlated to Illuminant Estimation: A Composite Substitute

1
Faculty of Light Industry, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
2
Key Laboratory of Paper Science and Technology of Ministry of Education, Faculty of Light Industry, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(3), 1175; https://doi.org/10.3390/app12031175
Submission received: 28 December 2021 / Revised: 14 January 2022 / Accepted: 21 January 2022 / Published: 23 January 2022
(This article belongs to the Topic Applied Computer Vision and Pattern Recognition)

Abstract

:
Computational color constancy (CCC) is to endow computers or cameras with the capability to remove the color bias effect caused by different scene illuminations. The first procedure of CCC is illuminant estimation, i.e., to calculate the illuminant color for a given image scene. Recently, some methods directly mapping image features to illuminant estimation provide an effective and robust solution for this issue. Nevertheless, due to diverse image features, it is uncertain to select which features to model illuminant color. In this research, a series of artificial features weaved into a mapping-based illuminant estimation framework is extensively investigated. This framework employs a multi-model structure and integrates the functions of kernel-based fuzzy c-means (KFCM) clustering, non-negative least square regression (NLSR), and fuzzy weighting. By comparing the resulting performance of different features, the features more correlated to illuminant estimation are found in the candidate feature set. Furthermore, the composite features are designed to achieve the outstanding performances of illuminant estimation. Extensive experiments are performed on typical benchmark datasets and the effectiveness of the proposed method has been validated. The proposed method makes illuminant estimation an explicit transformation of suitable image features with regressed and fuzzy weights, which has significant potential for both competing performances and fast implementation against state-of-the-art methods.

1. Introduction

The human vision system possess an innate “color constancy” capability, i.e., to dismount effects from the illuminant color and perceive the true color of an object [1,2]. Unfortunately, computers and imaging signal processors in vision applications do not have this capability if not to perform specific algorithms. To this end, varieties of computational color constancy (CCC) algorithms have been developed to compensate for the effects of illumination on objects’ color perception in the past decades. Once an estimated illuminant color is obtained for a captured image, the color deviation can be corrected by using the von Kries model [1,2,3]. Therefore, illuminant estimation is the key for CCC.
To date, many approaches have been proposed to estimate illuminant color by many researchers. These approaches can be roughly classified into three groups. Gray world (GW) [4,5], White patch (WP) [6], Shades of gray (SoG) [7], Gray edge (GE) [8], etc., belong to the first group—statistics-based methods, which take account of some image statistical features that are kept consistent in the image captured under canonical illuminating environments. Natural Image Statistics [9], machine learning [10], and deep learning [11,12,13,14,15,16], etc., are included in the second group—learning-based methods, which need a learning procedure utilizing various image information to pre-train models and estimate illuminant color. The third group is the combination version of the first two groups [11,17,18,19]. The unitary algorithms (i.e., the first two groups) just use a single strategy for illuminant estimation by assuming some color distribution features or learning general models from the training dataset, while the combination methods integrate the estimates from unitary algorithms into a single, final estimate. All these three groups of estimation methods usually suffer from either estimation accuracy, or computational efficiency, and result in better performances just for some specific image scenes. The overwhelming method does not appear yet.
Recently, some methods directly mapping image features to illuminant color are given rise to notice due to their simplicity and robustness [20,21,22,23]. These methods usually extract some specific features and then train a regression to capture the relationship between image features and illuminant ground truth. Although with a learning or model regression procedure, they do not need massive computation efforts and complicated model structures like deep learning. However, due to it is not clear which features are more correlated to illuminant estimation, the selection of image features for model regression remains arbitrary in the eye of the beholder. To our knowledge, there have still been no sufficient studies on how to select suitable features for the model regression of illumination estimation.
In this study, some typical image features in illuminant estimation literature are collected into a feature set. A framework with image clustering, multi-model constructing, and illuminant prediction is established to estimate illuminant color and assess the estimation performance of different features. The main contributions of this study are given as follows. (1) We develop an approach to estimate illuminant color, which includes kernel-based fuzzy c-means clustering (KFCM), optimization-based model regression, and fuzzy weighting combination. Using this method, we can easily investigate different features’ influences on illuminant estimation. (2) An optimal combination of image features is obtained by many effective experiments and analyses. Using the chosen composite feature covering a range of image characteristics as wide as possible, the performance of illuminant estimation is improved and the regression model becomes more robust. (3) KFCM and optimization-based model regression enhance the generalization ability of the proposed method; moreover, the multi-model fuzzy weighting scheme employed boosts the proposed approach to be more suitable for various image features and diverse light source characteristics.
The rest of this paper is organized as follows. Section 2 introduces the related background. The proposed method is presented in Section 3, where feature extraction, KFCM clustering, model regression, and fuzzy weighting are described in detail. Afterwards, in Section 4 many experimental results are provided to validate the proposed method, and for some remarkable issues, further discussions are given. Finally, Section 5 concludes the work.

2. Background

2.1. Imaging Model

The most commonly used imaging model for CCC is the one under the assumption of Lambertian reflection [9]. According to this model, the camera’s raw intensity response f c ( x ) is given by [9,16]:
f c ( x ) = Ω I ( λ , x ) R ( λ , x ) ρ c ( λ ) d λ ,
where c is a RGB color channel, c     R , G , B , x is a pixel location of the given image, λ is the wavelength of the light, Ω is the spectrum range of the visible light, λ Ω , I ( λ , x ) denotes the spectral distribution of light source, R ( λ , x ) denotes the surface reflectance of imaging objects, and ρ c ( λ ) denotes the camera sensitivity function of color channel c.
To make the problem straightforward, it is commonly assumed the image is uniformly illuminated by a single light source. Considering the case that the image scene is an ideal white surface, it yields R ( λ , x ) = 1 , then the observed illuminant color ^ = [ ^ R , ^ G , ^ B ] T will be obtained as:
^ = Ω I ( λ ) ρ c ( λ ) d λ .
Once we know ^ , the image can be successfully chromatically adapted under the help of von Kries diagonal transformation [3]. However, with only image pixel responses f c given, but both I ( λ ) and ρ c ( λ ) unknown, the estimation of ^ is an under-determined issue, which needs additional assumptions to be taken.

2.2. Illuminant Estimation

This section briefly introduces the three typical methods (statistics-based, learning-based, and combinational) for illuminant estimation, and additionally, discusses some methods most related to the topic of this paper.
Statistics-based methods. These algorithms take account of certain assumptions or physical observations, and directly compute some statistical metrics of a given image to obtain the illuminant color. The advantages of these methods lie in low computational efforts, although occasionally the estimation accuracy cannot be assured. Weijer et al. [24] developed a GE framework to unify most of these algorithms, which includes n-order partial derivatives and the Minkowski p-norm given as:
n f σ x x n p d x 1 p = k e n , p , σ ,
where f ( x ) σ = f ( x ) G σ ( x ) is defined as convolving the image f ( x ) by a Gaussian filter G ( x ) with scale parameter σ , k is a scaling, and e n , p , σ is the estimated illuminant color. The framework is denoted as G E n , p , σ with the choices of n, p and σ . Some special instantiations of this framework [18,19,24,25,26] include GW ( G E 0 , 1 , 0 ), WP ( G E 0 , , 0 ), SoG ( G E 0 , 6 , 0 ), 1-order gray edge (GE1) ( G E 1 , 1 , 6 ), 2-order gray edge (GE2) ( G E 2 , 1 , 5 ), General gray world (GGW) ( G E 0 , 13 , 2 ), etc.
Learning-based methods. In these methods, a training process is needed and then a pre-trained model is obtained to estimate the illuminant. Commonly, these methods can be further classified into two types: low-level and high-level ones [16]. The typical low-level methods are gamut mapping algorithm [27] and its extensions [28]. Other examples include some methods using probabilistic models [29], Classification-based algorithm selection [17], and machine learning algorithms [10,30], etc. Some regression-based methods [19,20,22,23] are also included in this group. These methods learn a mapping relationship between image features and illuminant ground truths from massive training data. Following the recent blooming of deep learning, high-level methods based on convolutional neural network (CNN) [11,12,14,15,16,31,32,33,34,35] have received state-of-the-art performances on various color constancy datasets. Although these methods commonly are more accurate than statistics-based algorithms, their structural complexities and computational efforts on model training hinder their applications for fast and real-time tasks.
Combination methods. Most unitary algorithms just use a single strategy that usually cannot work well for all types of images. To this end, many algorithms try to use multiple model strategies where the model set of illumination estimation is acquired from some unitary algorithms. Then by weighting the multiple estimates, a final estimate is obtained. According to whether the weights used are fixed or not, combination methods can be grouped into static weight combination and dynamic weight combination. Static weights are not suitable for diverse image features [18,36] and commonly can not ensure the robustness of illuminant estimation. Conversely, dynamic weights are more effective, but they need to tune the combination of unitary estimates against varying image features [11,17,18]. There are many examples [9,17,37,38,39] to produce the dynamic weights and find the best combination.
Related methods. In the paper [40], Zeinab Abedini provided a new explanation to some statistics-based methods from the perspective of the percentage of pixels involved in illuminant estimation. In these methods, all pixels equally attend in illuminant estimation (e.g., GW), or some pixels are radically selected for processing and the others are ignored (e.g., WP, Bright Pixel [41], PCA-based method [21], Gray Pixels [42], Gray Index [43]). Naturally, the explanation can be extended as a weight-based technique, i.e., all pixels are used, but the weights for different pixels are not the same. And, the weights can be implicitly or explicitly presented. GGW, GE, and SoG are included in this explanation since they can be explained to have implicit weights with the structure of the p-th Minkowski norm. Local surface reflectance (LSR) [44] which uses WP to weight GW can also be seen to have implicit weights [40]. According to this explanation, illumination estimation becomes the regression function of the pixels distribution features and the corresponding weights. Similarly, in learning-based methods, some direct regression-based methods have been proposed [20,21,23,40,45] in recent years, which simply map image features to illuminant. One of the remarkable methods is based on the Corrected Moment (CM) [20]. The CM-based method provides an efficient methodology to yield the illuminant estimation using a polynomial regression from multiple low-level-based methods. Other techniques, e.g., regression tree [30], fuzzy logic [19], deep learning [46,47], also can be regarded as regression-based methods with complex structures. Although these algorithms can achieve outstanding performances, their disadvantage is that they make illuminant estimation work in a black box without further considering the specific features of different images. These methods can be improved if the regression or weighting model is adjusted with image intrinsic features. So it is very important to assess the impacts of selected image features on illuminant estimation and determine which features are more correlated to illuminant estimation. The optimal combination of selected features will make illumination estimation regression more efficient.

2.3. Color Correction

For a given image, once the illuminant color is estimated as a illumination vector ^ = ( ^ R , ^ G , ^ B ) , the image should be corrected by the von Kries diagonal transformation [3]. All the colors of the color-biased image under the unknown light source (u) should be mapped to the colors under a canonical illuminant (c), which is demonstrated as follows:
R c G c B c   =   ^ R 1 0 0 0 ^ G 1 0 0 0 ^ B 1 R u G u B u ,
where ( R u , G u , B u ) T and ( R c , G c , B c ) T denote the image colors under the unknown light source and a canonical illuminant, respectively.

3. Proposed Method

In the proposed method, illuminant estimation is achieved by a serial of linear or nonlinear transformation of suitable image features. The proposed method also can be easily used to assess the illuminant estimation performance by utilizing different features to construct more efficient feature substitutes. The flowchart of the proposed framework is illustrated in Figure 1.
In the training phase, the training dataset is classified into different clustering using the KFCM algorithm. For each cluster, the illuminant estimation models M 1 , M 2 , , M K are obtained using a non-negative least square regression (NLSR) method. For a given test image that has been removed dark level and saturation level in the linear RGB space, the corresponding image features will be extracted and then the final estimated illuminant will be produced by fuzzy weighting all outputs of multiple regression models.

3.1. Features Extraction

The image features can be used as representatives of an original image, just with a compact data length. Since it is not clear which features will affect the illuminant estimation much more, in this paper a series of features are gathered into a candidate set to explore the suitable feature selections and better illuminant estimation. The features in the feature set are derived from the up-to-date literature that provides some effective learning-based methods of illuminant estimation. These features are low-level, or middle-level [38]; here we do not use high-level or semantic features, since semantic features are commonly specific to different images and may have small effects on illuminant estimation [38]. Several image features are extracted for each image as follows.
Corrected moments (CM) [20,48]. They are the polynomial expansions of the 2nd, 3rd, and 4th degree, which are described by the following vectors:
C M 9 = r , g , b , r 2 , g 2 , b 2 , r g , g b , r b T ,
C M 19 = r , g , b , r 2 , g 2 , b 2 , r g , g b , r b , r 3 , g 3 , b 3 , r g 2 , g b 2 , r b 2 , g r 2 , b g 2 , b r 2 , r g b T ,
C M 34 = ( r , g , b , r 2 , g 2 , b 2 , r g , g b , r b , r 3 , g 3 , b 3 , r g 2 , g b 2 , r b 2 , g r 2 , b g 2 , b r 2 , r g b , r 4 , g 4 , b 4 , r 3 g , r 3 b , g 3 r , g 3 b , b 3 r , b 3 g , r 2 g 2 , g 2 b 2 , r 2 b 2 , r 2 g b , g 2 r b , b 2 r g ) T .
Here r, g, and b are the illuminant components estimated by some general methods, such as GW, SoG, etc. According to the above expressions, the element numbers recorded in a CM feature are 9, 19, and 34, respectively.
RGB-uv histogram (RGBuv). The paper [49] extracted a histogram from the log-chrominance space to represent the bulk of information in a color image. This feature is called RGB-uv histogram, which describes the color distribution feature of an image as a m × m × 3 tensor. The works in [49] have validated the effectiveness of this feature to illuminant estimation. To short the length of this feature, a PCA-based dimensionality reduction step is applied to get a compact representation for each RGBuv histogram. As a result, the RGBuv of any image I can be expressed by a serial of primary component coefficients v i , ( i = 1 , 2 , , p ) :
R G B u v ( I ) = [ v 1 , v 2 , , v p ] .
where p is the number of primary components, and p m × m × 3 .
Color moments (ClrM) can characterize the color distribution in an image. Commonly, three central moments, i.e., first-order moments (Mean), second-order moments (Standard deviation), and third-order moments (Skewness), are sufficient to express the main color distribution of an image. Considering the size of image, each image is evenly divided into 3 × 3 sub-blocks in this study; therefore, 81 moments for an image I can be obtained if the moments are calculated for each RGB channel:
C l r M ( I )   =   φ 1 , φ 2 , , φ 81 .
Cheng’s simple features (ChengSF) [30]. These features are a group of intensity-invariant features based on normalized chromaticity. There are four color chromaticity features extracted from an image, i.e., average color chromaticity [ r a , g a ] , brightest color chromaticity [ r b , g b ] , dominant color chromaticity [ r d , g d ] , and chromaticity mode of the color palette [ r m , g m ] . Here r and g are the normalized chromaticity components of RGB color [30]. For an image I, ChengSF can be expressed as follows:
C h e n g S F ( I )   =   r a , g a , r b , g b , r d , g d , r m , g m .
Color distribution features (CDF). In paper [19], Luo proposed a compact color distribution feature for illuminant estimation, which is a combination of the number of image colors n c , chromaticity features in Equation (10), and color moments in Equation (9). For an image I, the obtained CDF vector is as follows:
C D F ( I )   =   n c , λ 1 r a , λ 1 g a , λ 1 r b , λ 1 g b , λ 1 r d , λ 1 g d , λ 1 r m , λ 1 g m , λ 2 φ 1 , λ 2 φ 2 , , λ 2 φ 81 ,
where the two adjustable factors, λ 1 and λ 2 , control the influences of different components on the illuminant estimation.
Initial illuminant features (IIF). In paper [19], some unitary algorithms, i.e., GW, WP, SoG, GE1, GE2, GGW, PCA, and LSR, are applied to get eight initial illuminant estimates for an image. These initial estimates are integrated into a vector, i.e., IIF, as a metric of illuminant characteristics. For an image I, the IIF vector is expressed as follows:
I I F ( I )   =   [ γ ( e i , e j ) ] , i = 1 , 2 , , 8 , j = ( i + 1 ) , ( i + 2 ) , , 8 ,
where γ ( e i , e j ) is defined as the Euclidean distance between initial illuminant estimates e i and e j .

3.2. KFCM Clustering

In the training stage, a better classification can achieve the appropriate divisions of the training dataset. In each division the features will be close each other and fit for illuminant modeling. We use some above-mentioned features to classify the training dataset by the KFCM method.
KFCM is derived from fuzzy c-means (FCM). FCM is a classification algorithm based on fuzzy sets, which results in clustering partitions with the maximum similarity among samples in the same cluster, while the minimum similarity among different clusters [50]. As an improved version of the common c-means algorithm, FCM provides flexible fuzzy partitions since the training samples are clustered in terms of the membership functions. For the standard FCM, a cost function based on the Euclidean distance is minimized to solve the cluster centroids and the membership degrees of all samples:
J F C M ( U , V ) = k = 1 n i = 1 c μ k i m x k ν i 2 ,
where | . | is the Euclidean norm to measure the similarity between sample data and cluster centroid; μ k i is the degree of membership for the sample x k in the cluster i, U =   μ k i ; ν i is the centroid of the cluster i, V =   ν i ; n is the total number of samples; m is the exponent of the membership; and c is the cluster number. Furthermore, a kernel function can be introduced into FCM by a nonlinear mapping Φ : x Φ ( x ) , which maps the low dimension vector x into a high dimension space. Then the kernel-based FCM, i.e., KFCM, is to minimize the following cost function:
J K F C M ( U , V ) = k = 1 n i = 1 c μ k i m Φ x k     Φ ν i 2 ,
where
Φ x k     Φ ν i 2 = K x k , x k   +   K ν i , ν i     2 K x k , ν i .
Here K ( x , ν ) is an inner product kernel function, and K ( x , ν ) = Φ ( x ) T Φ ( ν ) . In this study, we adopt the Gaussian function as a kernel function, i.e., K ( x , ν ) = exp | x ν | 2 σ 2 , where σ is the kernel width; then we can get K ( x , x ) = 1 and K ( ν , ν ) = 1 . According to Equation (15), Equation (16) can be rewritten as:
J K F C M ( U , V ) = 2 k = 1 n i = 1 c μ k i m 1 K x k , ν i .
From Equation (16), we can see that the cost function is minimized when the first-order partial derivatives with respect to μ k i and ν i both are zeros. By these two necessary conditions, the final clusters and their centroids are iteratively obtained as follows:
μ k i = 1 1 K x k , ν i 1 m 1 j = 1 c 1 1 K x k , ν j 1 m 1 ,
ν i = k = 1 n μ k i m K x k , ν i x k k = 1 n μ k i m K x k , ν i .
Here i = 1 , 2 , , c , and k = 1 , 2 , , n .
It is worthy to mention that the clustering performance of KFCM will be influenced by different kernel functions and their parameters. Although the results of Equations (17) and (18) are deduced using the Gaussian kernel function, other kernel functions just satisfying K ( x , x ) = 1 also can be used here. In the Gaussian kernel function, the kernel width is commonly used as the key tuning parameter. Due to the Gaussian kernel being a typical local kernel function, the function value is largely affected by the data closer to the test point. Hence we can determine the kernel width according to the average distance between the sample data and the cluster centroid. This will ensure that the data in the same cluster play a dominant role in the corresponding kernel function values, while the data out of this cluster make a slight impact on the corresponding kernel function values. In additions, different from single-model methods, multi-model ones used in this study can select different kernel parameters for different clusters.

3.3. Model Regression

In this study, to avoid complicated structures of deep neural networks, we just employ the least-square-based method since we focus on finding more suitable features for illuminant estimation rather than aiming for high performance. The precision of model regression is sacrificed somewhat, but the speed and robustness of illuminant estimation can be attained substantially.
We construct the regression relationship directly from image features to illuminant ground truths as follows:
F M = L ,
where L = e 1 e 2 e N T is the vector of illumination ground truths of the training samples, M is the regression model matrix, and F is the feature matrix of all sampling images. In general, we obtain the solution of Equation (19) as M = F 1 L , where F 1 is the Moore-Penrose inversion of the feature matrix F . But some elements in the resulting M might be negative, which may deteriorate the robustness of the model. To address this problem, we seek the solution to minimize M with each element in M no less than zero. Therefore, we can train a regression model to minimize the L 2 -norm loss between the reconstructed and ground truth illumination by solving the non-negative least square problem:
min M F M L 2 2 ,
Therefore, for all partitions of the training set from KFCM, we can get a series of models and construct a model set as follows:
M s = { M 1 M 2 M K } ,
where K is the number of the obtained multiple models for the training dataset.

3.4. Fuzzy Weighting

For a test image I i n , the final output of the illuminant estimation is synthesized by fuzzy weighting the outputs of all models, which is given as:
^ = i = 1 K ( F I i n M i ) w i ,
where w i is the weight of the i-th model M i , and F I i n denotes the extracted features of the test image I i n . For an image sample in the training set, w i is known (it is the membership of degrees for this sample); while for a new test image not included in the training dataset, w i is needed to be computed by determining what degree the test sample belongs to each cluster obtained via KFCM. That is to say, all possibilities that an image falls into each cluster should be acquired by calculation. To this end, we define a variable d f , i = f ν i 2 as the squared Euclidean distance between the test image’s feature vector f and the cluster centroid ν i , ( i = 1 , 2 , , K ) . Then a radial basis function is constructed to represent the probability for the test image belonging to the cluster i:
w i = e d f , i d f , i 2 σ f 2 2 σ f 2 l = 1 K e d f , l d f , l 2 σ f 2 2 σ f 2 , i = 1 , 2 , , K ,
where σ f denotes the radial fall-off factor. In this way the weights are obtained, and the illuminant color of the test image can be predicted by Equation (22).

4. Experimental Results

4.1. Experimental Set-Up

Dataset. The proposed framework should be verified on a mass of image examples that have a wide scale of color distributions and diverse illuminant characteristics. In this study, the two datasets (Gehler-Shi [51] and Cube+ [15,52]) are employed as the training and testing datasets. Both datasets are suitable for the evaluation of illumination estimation algorithms due to they include typical real-world images with varieties of illumination. In each image of the Gehler-Shi dataset, there is a color-checker board added into the lower right corner for calculating the ground truth illumination, while in the Cube+ dataset, a SpyderCube color target is used in each scene for the same purpose. These color-checker boards and color targets are masked when the images are utilized in the procedures of our experiments. The proposed method performs the standard 3-fold cross-validation for the training and testing procedures, which is commonly used in existing learning-based illuminant estimation literature.
Implementation details. We conducted three types of experiments to test our proposed approach, i.e., (1) Experiment A, training and testing on an individual image dataset from different cameras; (2) Experiment B, training and testing on an image set from an identical camera; and (3) Experiment C, training and testing on different image datasets or a combination of several image datasets. Since an image from a certain camera or dataset might contain some unique information of this camera or dataset, it is necessary to perform Experiment A and Experiment B. Both experiments will help to investigate on selecting suitable features of illuminant estimation for our method under the circumstances ruling out the influence of camera or dataset. On the other hand, considering the training/testing across datasets is the most common situation for learning-based illuminant estimations, Experiment C is employed to validate our method’s effectiveness on a combination of multiple datasets.
Our MATLAB implementation is shown in https://github.com/yunhuiluo/if2ic (accessed on 28 December 2021). In the training stage, the training dataset will be clustered according to the selected image features and the regression models will be constructed for each cluster. Since the classification algorithm based on KFCM is not complex mathematically, it takes a very short time compared with other learning methods like deep learning. For example, it requires approximately 12.6 seconds for classification and training 6 models for the Cube+ dataset. Once the training is finished, the illuminant estimation process costs an average time of 0.93 seconds for a new test image with 2601 × 1732 pixels, 48 bits depth, and PNG format; this process includes feature extraction, multiple model estimation, weight computation, and fuzzy weighting output. All the runtimes above-mentioned were obtained from a laptop computer with an Intel Core i5-2450M @ 2.50 GHz CPU.
By default, we set the cluster number K = 4 when using the combination of Gehler-Shi and Cube+, and we set the fall-off factor σ f = 0.25 mentioned in Section 3.4. We utilize the KFCM algorithm to perform image clustering based on the features of IIF, and set the kernel width σ to be 140. Our method requires 6.7 MB to store the classification parameters, multiple model matrix, and some configurations. In our experiments, each PCA feature vector was also represented by 53 primary component scores (the same with [49]) for any RGBuv feature. Since each time the cluster iteration of KFCM algorithm begins from some random initial center positions, the performances for each experiment with the same configurations might be tinily different.
Evaluation Metrics. The two performance metrics widely used in color constancy literature are employed in this paper. The first one is recovery angular error (recovery AE) [9,18], which is defined as the angle in degrees between the illumination’s actual 3D-chromaticity e a and its estimated 3D-chromaticity e e : γ e a , e e   = cos 1 e a · e e e a e e . Here the symbol of · denotes the scalar product (dot product), and the symbol of · indicates the Euclidean norm of a vector. The second metric is called reproduction angular error (reproduction AE) [53,54], which usually is more efficient than the traditional recovery AE. The reproduction AE is defined as the angle in degrees between the true color of white surface and the reproduced color of white surface: γ e c , e u   = cos 1 e c · e u e c e u . Here e c denotes the white light source (true color of white surface), e c = [ 1 , 1 , 1 ] T , and e u indicates the reproduced color of white surface, e u = e a / e e , which is the actual 3D-chromaticity e a divided by the estimated 3D-chromaticity e e .
Since both recovery and reproduction AEs are not normally distributed, the Mean and Median values along with the Trimean value are used to assess the statistical performance of color constancy algorithms in this study. Trimean is defined as a weighted average T r i m e a n = ( Q 1 + 2 Q 2 + Q 3 ) / 4 , where Q 1 ( B e s t 25 % ) , Q 2 , and Q 3 ( W o r s t 25 % ) are the first, second, and third quantiles of recovery or reproduction AE, respectively. In addition, the maximums of both AEs are used as extra metrics to assess the robustness of the illuminant models along with the metric of Q 3 ( W o r s t 25 % ) .

4.2. Experiment A: Individually Testing on Different Datasets (Within Dataset)

In this experiment, our method is independently applied to different datasets. Table 1 and Table 2 compare some performance metrics between our method and some conventional algorithms on the Gehler-Shi dataset and the Cube+ dataset, respectively. In order to illustrate the impacts of different features on illuminant estimation performances, all types of features given in Section 3.1 are investigated using the proposed method.
Table 1 shows that our results (except for the one using CDF features) outperform most unitary algorithms on the Gehler-Shi dataset. For example, the Mean of recovery AE is 4.66 for GGW, and 4.17 for our method using ChengSF; The Trimean of recovery AE is 3.81 for GGW, and 3.41 for our method using ChengSF. There is a significant performance improvement near 10% from GGW to our method using ChengSF. Table 1 also shows that our method with some combined features (i.e., RGBuv+ChengSF, RGBuv+ChengSF+IIF, and RGBuv+ChengSF+IIF+ C M 34 ) is superior to some learning-based methods, e.g., Bayesian [29], Natural Image Statistic [46], and WCS with partitioning [40], etc. What’s more, we can notice that the proposed method with the combined features is better than most previous methods in terms of the metrics of Best 25% and Worst 25%. This validates that the proposed method can achieve a outstanding illuminant estimation regardless of image features.
Table 2 gives the results for the Cube+ dataset. Most metrics of our results in Table 2 are better than those in Table 1, since the Cube+ dataset contains more images than the Gehler-Shi dataset. It can be seen that the statistics of our method with the combined features are almost all better than one of learning-based methods—Attention CNN [56]. These results further indicate the proposed method is effective to achieve a better regression accuracy using optimally selected feature combinations. It should be noticed that the composite features of RGBuv+ChengSF+IIF can obtain outstanding performances using our method. More comparisons of statistical metrics are provided in our GitHub repositories.

4.3. Experiment B: Testing on Image Set from Identical Camera (Within Camera)

This experiment uses some images from an identical camera as the training and testing data, which is to verify the estimation accuracy within the camera using the proposed method. Table 3 and Table 4 compare the performance metrics for some conventional methods and our method with different features within an identical camera. Here Canon 5D is one of the cameras used by the Gehler-Shi dataset, and Canon EOS 550D used by the Cube+ dataset, respectively. For Canon 5D, 482 images are picked out from the Gehler-Shi dataset, and for Canon EOS 550D, 853 images are randomly selected from the Cube+ dataset. The standard 3-fold cross validation is independently applied to the constructed datasets to train and test the proposed method.
Table 3 and Table 4 show that the proposed method using image feature regressions achieves better performances than some traditional methods such as WP, GW, PCA-based, LSR, etc. Using the individual feature in the proposed method, the feature of RGBuv receives the best results. According to the discussion in the paper [49], the RGBuv feature can effectively cover the color distribution of an image. The results in Table 3 and Table 4 also validate this viewpoint. It can be seen that compositing several features will provide better results, e.g., RGBuv+ChengSF, RGBuv+ChengSF+IIF, and RGBuv+ChengSF+IIF+ C M 34 . For the camera of Canon 5D, the composite of RGBuv+ChengSF+IIF has a 25.9% performance improvement in recovery AE Mean, and a 25.4% performance improvement in recovery AE Trimean, compared with WP. For the camera of Canon EOS 550D, the same improvements reach 42.3% and 32.3% (ours versus GGW), respectively. In our experiments, the optimal feature combination is RGBuv+ChengSF+IIF, which obtains outstanding performances compared with other feature selections.
The combination of image features affects the illuminant estimation since the proposed method is to directly map the extracted features to illuminant color. Except for the results listed in Table 3 and Table 4, we conducted many other experiments for the proposed method with more feature combinations. We find that it is not the more combined features the better performance. Just the correlated features are effective for illuminant estimation. Thus, it is significant to find which features are more correlated to the illuminant estimation. In this study, we just investigated several features and their combinations. From the experimental results given, there are reasons to believe that if we find more effective features, the performance can be even more improved by the proposed method.

4.4. Experiment C: Cross-Dataset, Cross-Camera Testing

To investigate the robustness of the proposed method, some cross-dataset testing should be performed on multiple datasets. In this experiment, the combination dataset of Gehler-Shi and Cube+ is used by these testing through the 3-fold cross-validation technique. Table 5 provides the statistical metrics of the experimental results.
Since the combined dataset can cover a wider range of image features and illuminant regions, the classification and models obtained by the proposed method commonly receive better generality on illuminant estimation. Table 5 validates this viewpoint. It can be seen that the obtained statistical metrics for all images in the Gehler-Shi and Cube+ datasets reach the good performances, e.g., the Mean and Maximum of recovery AE respectively are 2.13 and 14.83 for the selected optimal features of RGBuv+ChengSF+IIF. Figure 2 shows the recovery and reproduction AEs’ distribution for some images from the Gehler-Shi and Cube+ datasets using some conventional methods and the proposed method. These results indicate the proposed method can work well and reach excellent performances. It should be pointed out that the selected optimal features of RGBuv+ChengSF+IIF are prerequisites for these results.
In real life, photographers usually cannot capture images under perfect or standard illumination. A complex lighting environment results in changing brightness and contrast. A robust color constancy algorithm should be able to deal with these challenges. Our algorithm essentially is a weight-based method in which the features more related to illuminant color are highlighted and used to deduce dynamic weights. Figure 3 shows some visual results to face some challenging image scenes. As shown in this figure, our algorithm with the feature combination of RGBuv+ChengSF+IIF reaches the better performance compared with some conventional methods. Consequently, the proposed method has the great potential to color constancy of a wide range of images. For more examples, please refer to the related materials at our GitHub repositories.

4.5. Additional Experiments

In the proposed method, many factors will affect the performances of illuminant estimation. Some experiments are conducted to explore these influences.
Clustering methods and cluster numbers. For the proposed method, we perform KFCM clustering by default to classify the training data. For comparisons, the other two clustering algorithms, i.e., K-means and the two-step method [19], are also added to the proposed framework. The reason for selecting K-means and the two-step method is that the former is one of the most commonly used clustering methods, and the latter is an effective algorithm of color constancy based on cluster models. For each clustering algorithm, we can manually set different cluster numbers in order to obtain better results. Table 6 gives some statistical metrics using different clustering methods and cluster numbers. Comparing the results using different configurations, it can be seen that the proposed method along with KFCM seems to be more effective than using K-means and the two-step method. One of possible reason might be that KFCM can provide flexible division for the training set, which enhances the robustness and generalization of the proposed method. And we find K = 4 or 5 is the appropriate selection for our method under the KFCM clustering on the combination dataset of Gehler-Shi and Cube+. It is natural that the cluster number should increase for better modeling performance if we use another dataset with more image samples.
Kernel width of KFCM. In the proposed method, kernel width will affect clustering divisions, which will result in different performances. Table 7 provides some statistical metrics on the Gehler-Shi and Cube+ datasets to show the influences of kernel width in KFCM clustering. We can find that the optimal choice of kernel width for our experiments is σ = 100 ~180. Commonly, the optimal values are obtained by trial and tried in various testing. It should be noticed that the optimal kernel width is varying with the training dataset. Therefore, as an option, Particle Swarm Optimization can be used to search for the optimal kernel width.

4.6. Discussion and Analysis

In this paper, the experimental results show that the performances of the proposed method are on par with many statistics-based and learning-based approaches. And we have tentatively identified the feature combinations that are more correlated to illuminant estimation based on a specific feature set. In the meantime, we can obtain some useful observations.
(1) The methods based on composite features generally work better than most methods based on unitary features. The key to the combination of features is to find a compact presentation to cover a much wider range of image features and illuminant characteristics. The dimension of the features is not an obstacle; if the feature dimension is low with useful information compressed, we can use different kernel functions to map them into a higher dimension space [20,48]; on the contrary, if the dimensionality is high with much redundant information, we can use the PCA technique to reduce the dimensionality of the features [49]. To date, as it were, the trial-try method by experiments is still the major route to construct artificial features for regression models of illuminant estimation. Thus, choosing an appropriate feature combination is significant to accurate and robust illuminant estimation.
(2) The methods based on feature regression clearly depend upon not only the optimal feature combinations but also the scale and characteristics of the training sets. If the training examples are from an identical camera or a benchmark dataset using multiple cameras while the testing example is taken from the same camera or dataset, the illuminant estimation accuracy is usually very high. However, if training images from a certain camera or dataset using multiple cameras while the testing example from other camera or dataset, i.e., cross cameras or cross datasets, the illuminant estimation accuracy is commonly worse than expected. As the scale of the training dataset increases and the illuminant range of the training sets extends, a single model is not a preferred option for regression modeling. So a multiple model strategy with high efficient classification (or clustering) becomes preferred, like KFCM clustering and regression modeling associated with fuzzy weighting in this paper.
(3) The type of feature, i.e., features from different cues, generally affects and limits the performances of illuminant estimation. In this paper, we just considered two types of features (low-level CDF and middle-level IIF [18,38,58]. We found the performance improvement is limited although we seem to have an optimal feature combination. High-level scene category or semantic features, e.g., Weibull parameters with content-based features [17], 3D scene geometry [37], indoor/outdoor classification [59], etc., can also be exploited in the proposed method.

5. Conclusions

For the regression-based illuminant estimation methods, it is uncertain which features are more efficient to regress illuminant color. In this research, a flexible framework is developed to integrate KFCM, NLSR, and fuzzy weighting, which can result in outstanding performances of illuminant estimation. By extensive experiments with a series of artificial features weaved into this framework, to some extent it is clarified which features are more correlated to illuminant estimation. Moreover, we exploited an optimal feature combination based on the given feature set, which obtains a significant performance improvement for illuminant estimation.
The proposed method and its framework can be used to test and compare the influences of image features on illuminant estimation. From this, we can optimally deploy the appropriate image features for illuminant estimation in a more objective way, rather than in an empirical, arbitrary manner. If the feature sets contain more feature candidates, the performance of illuminant estimation is expected to be further enhanced.
This study focuses on determining which features are more correlated to illuminant estimation in a limited feature set. Therefore, more related image features, including high-level semantic features, can be further taken into account. Besides, considering there are different intrinsic and implicit cues for different camera sensors, pre-selecting images in terms of camera sensor characteristics and then constructing multiple models may be another pathway to improve the proposed method.

Author Contributions

Conceptualization, Y.L. and X.W.; methodology, Y.L.; software, Y.L.; validation, Y.L., X.W. and Q.W.; formal analysis, Y.L.; investigation, Y.L.; resources, Y.L.; data curation, Y.L. and X.W.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L. and X.W.; visualization, X.W.; supervision, Y.L.; project administration, X.W.; funding acquisition, Q.W. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

The Project Supported by the Foundation (No. KF201603 and ZR202001) of Key Laboratory of Paper Science and Technology of Ministry of Education of China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

CCCComputational color constancy
GWGray world
WPWhite patch
SoGShades of gray
GEGray edge
GGWGeneral gray world
PCAPrimary component analysis
LSRLocal surface reflection
CMCorrected moments
CDFColor distribution feature
IIFInitial illumination feature
KFCMKernel fuzzy c-means clustering
NLSRNon-negative least square regression

References

  1. Barnard, K.; Cardei, V.C.; Funt, B.V. A comparison of computational color constancy algorithms. I: Methodology and experiments with synthesized data. IEEE Trans. Image Process. 2002, 11, 972–984. [Google Scholar] [CrossRef] [PubMed]
  2. Barnard, K.; Martin, L.; Coath, A.; Funt, B.V. A comparison of computational color constancy algorithms. II. Experiments with image data. IEEE Trans. Image Process. 2002, 11, 985–996. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. von Kries, J. Influence of adaptation on the effects produced by luminous stimuli. In Sources of Color Vision; MacAdam, D.L., Ed.; The MIT Press: Cambridge, MA, USA, 1970; pp. 109–119. [Google Scholar]
  4. Buchsbaum, G. A spatial processor model for object colour perception. J. Frankl. Inst. 1980, 310, 1–26. [Google Scholar] [CrossRef]
  5. Provenzi, E.; Gatta, C.; Fierro, M.; Rizzi, A. A spatially variant white-patch and gray-world method for color image enhancement driven by local contrast. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1757–1770. [Google Scholar] [CrossRef]
  6. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–128. [Google Scholar] [CrossRef]
  7. Finlayson, G.D.; Trezzi, E. Shades of gray and colour constancy. In Proceedings of the Twelfth Color Imaging Conference: Color Science and Engineering Systems, Technologies, Applications, CIC 2004, Scottsdale, AZ, USA, 9–12 November 2004; pp. 37–41. [Google Scholar]
  8. van de Weijer, J.; Gevers, T.; Gijsenij, A. Edge-based color constancy. IEEE Trans. Image Process. 2007, 16, 2207–2214. [Google Scholar] [CrossRef] [Green Version]
  9. Gijsenij, A.; Gevers, T. Color constancy using natural image statistics and scene semantics. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 687–698. [Google Scholar] [CrossRef]
  10. Agarwal, V.; Gribok, A.V.; Abidi, M.A. Machine learning approach to color constancy. Neural Netw. 2007, 20, 559–563. [Google Scholar] [CrossRef]
  11. Oh, S.W.; Kim, S.J. Approaching the computational color constancy as a classification problem through deep learning. Pattern Recognit. 2017, 61, 405–416. [Google Scholar] [CrossRef] [Green Version]
  12. Hu, Y.; Wang, B.; Lin, S. FC^4: Fully convolutional color constancy with confidence-weighted pooling. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 330–339. [Google Scholar] [CrossRef]
  13. Afifi, M.; Brown, M.S. Sensor-independent illumination estimation for DNN models. In Proceedings of the 30th British Machine Vision Conference 2019, BMVC 2019, Cardiff, UK, 9–12 September 2019; p. 282. [Google Scholar]
  14. Afifi, M.; Brown, M.S. Deep white-balance editing. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 1394–1403. [Google Scholar] [CrossRef]
  15. Koscevic, K.; Subasic, M.; Loncaric, S. Deep learning-based illumination estimation using light source classification. IEEE Access 2020, 8, 84239–84247. [Google Scholar] [CrossRef]
  16. Qiu, J.; Xu, H.; Ye, Z. Color constancy by reweighting image feature maps. IEEE Trans. Image Process. 2020, 29, 5711–5721. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Bianco, S.; Ciocca, G.; Cusano, C.; Schettini, R. Automatic color constancy algorithm selection and combination. Pattern Recognit. 2010, 43, 695–705. [Google Scholar] [CrossRef]
  18. Li, B.; Xiong, W.; Hu, W.; Funt, B.V. Evaluating combinational illumination estimation methods on real-world images. IEEE Trans. Image Process. 2014, 23, 1194–1209. [Google Scholar] [CrossRef] [PubMed]
  19. Luo, Y.; Wang, X.; Wang, Q.; Chen, Y. Illuminant estimation using adaptive neuro-fuzzy inference system. Appl. Sci. 2021, 11, 9936. [Google Scholar] [CrossRef]
  20. Finlayson, G.D. Corrected-moment illuminant estimation. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2013, Sydney, Australia, 1–8 December 2013; pp. 1904–1911. [Google Scholar] [CrossRef]
  21. Cheng, D.; Prasad, D.K.; Brown, M.S. Illuminant estimation for color constancy: Why spatial-domain methods work and the role of the color distribution. J. Opt. Soc. Am. A 2014, 31, 1049–1058. [Google Scholar] [CrossRef]
  22. Afifi, M.; Punnappurath, A.; Finlayson, G.; Brown, M.S. As-projective-as-possible bias correction for illumination estimation algorithms. J. Opt. Soc. Am. A 2019, 36, 71–78. [Google Scholar] [CrossRef]
  23. Gao, S.B.; Zhang, M.; Li, Y.J. Improving color constancy by selecting suitable set of training images. Opt. Express 2019, 27, 25611. [Google Scholar] [CrossRef]
  24. Gijsenij, A.; Gevers, T.; van de Weijer, J. Computational color constancy: Survey and experiments. IEEE Trans. Image Process. 2011, 20, 2475–2489. [Google Scholar] [CrossRef]
  25. Faghih, M.M.; Moghaddam, M.E. Multi-objective optimization based color constancy. Appl. Soft Comput. 2014, 17, 52–66. [Google Scholar] [CrossRef]
  26. Cepeda-Negrete, J.; Sánchez-Yáñez, R.E. Automatic selection of color constancy algorithms for dark image enhancement by fuzzy rule-based reasoning. Appl. Soft Comput. 2015, 28, 1–10. [Google Scholar] [CrossRef]
  27. Finlayson, G.D.; Hordley, S.D.; Hubel, P.M. Color by correlation: A simple, unifying framework for color constancy. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 1209–1221. [Google Scholar] [CrossRef] [Green Version]
  28. Gijsenij, A.; Gevers, T.; van de Weijer, J. Generalized gamut mapping using image derivative structures for color constancy. Int. J. Comput. Vis. 2010, 86, 127–139. [Google Scholar] [CrossRef] [Green Version]
  29. Gehler, P.V.; Rother, C.; Blake, A.; Minka, T.P.; Sharp, T. Bayesian color constancy revisited. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2008), Anchorage, AK, USA, 24–26 June 2008. [Google Scholar] [CrossRef]
  30. Cheng, D.; Price, B.L.; Cohen, S.; Brown, M.S. Effective learning-based illuminant estimation using simple features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, 7–12 June 2015; pp. 1000–1008. [Google Scholar] [CrossRef]
  31. Bianco, S.; Cusano, C.; Schettini, R. Single and multiple illuminant estimation using convolutional neural networks. IEEE Trans. Image Process. 2017, 26, 4347–4362. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Barron, J.T.; Tsai, Y.T. Fast fourier color constancy. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6950–6958. [Google Scholar] [CrossRef] [Green Version]
  33. Afifi, M.; Barron, J.T.; LeGendre, C.; Tsai, Y.; Bleibel, F. Cross-camera convolutional color constancy. arXiv 2020, arXiv:2011.11890. [Google Scholar]
  34. Choi, H.H.; Kang, H.S.; Yun, B.J. CNN-based illumination estimation with semantic information. Appl. Sci. 2020, 10, 4806. [Google Scholar] [CrossRef]
  35. Xiao, J.; Gu, S.; Zhang, L. Multi-domain learning for accurate and few-shot color constancy. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 3255–3264. [Google Scholar] [CrossRef]
  36. Bianco, S.; Gasparini, F.; Schettini, R. Consensus-based framework for illuminant chromaticity estimation. J. Electron. Imaging 2008, 17, 023013. [Google Scholar] [CrossRef] [Green Version]
  37. Elfiky, N.M.; Gevers, T.; Gijsenij, A.; Gonzàlez, J. Color constancy using 3D scene geometry derived from a single image. IEEE Trans. Image Process. 2014, 23, 3855–3868. [Google Scholar] [CrossRef]
  38. Huang, X.; Li, B.; Li, S.; Li, W.; Xiong, W.; Yin, X.; Hu, W.; Qin, H. Multi-cue semi-supervised color constancy with limited training samples. IEEE Trans. Image Process. 2020, 29, 7875–7888. [Google Scholar] [CrossRef]
  39. van de Weijer, J.; Schmid, C.; Verbeek, J.J. Using high-level visual information for color constancy. In Proceedings of the IEEE 11th International Conference on Computer Vision, ICCV 2007, Rio de Janeiro, Brazil, 14–20 October 2007; pp. 1–8. [Google Scholar] [CrossRef] [Green Version]
  40. Abedini, Z.; Jamzad, M. Weight-based colour constancy using contrast stretching. IET Image Process. 2021, 15, 2424–2440. [Google Scholar] [CrossRef]
  41. Joze, H.R.V.; Drew, M.S.; Finlayson, G.D.; Rey, P.A.T. The role of bright pixels in illumination estimation. In Proceedings of the 20th Color and Imaging Conference, CIC 2012, Los Angeles, CA, USA, 12–16 November 2012; pp. 41–46. [Google Scholar]
  42. Yang, K.; Gao, S.; Li, Y. Efficient illuminant estimation for color constancy using grey pixels. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, 7–12 June 2015; pp. 2254–2263. [Google Scholar] [CrossRef]
  43. Qian, Y.; Nikkanen, J.; Kämäräinen, J.; Matas, J. On finding gray pixels. arXiv 2019, arXiv:1901.03198. [Google Scholar]
  44. Gao, S.; Han, W.; Yang, K.; Li, C.; Li, Y. Efficient color constancy with local surface reflectance statistics. In Proceedings of the Computer Vision—ECCV 2014—13th European Conference, Zurich, Switzerland, 6–12 September 2014; Volume 8690, pp. 158–173. [Google Scholar] [CrossRef]
  45. Ciurea, F.; Funt, B.V. A large image database for color constancy research. In Proceedings of the Eleventh Color Imaging Conference: Color Science and Engineering Systems, Technologies, Applications, CIC 2003, Scottsdale, AZ, USA, 4–7 November 2003; pp. 160–164. [Google Scholar]
  46. Bianco, S.; Cusano, C.; Schettini, R. Color constancy using CNNs. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2015, Boston, MA, USA, 7–12 June 2015; pp. 81–89. [Google Scholar] [CrossRef] [Green Version]
  47. Lou, Z.; Gevers, T.; Hu, N.; Lucassen, M.P. Color constancy by deep learning. In Proceedings of the British Machine Vision Conference 2015, BMVC 2015, Swansea, UK, 7–10 September 2015; pp. 76.1–76.12. [Google Scholar] [CrossRef] [Green Version]
  48. Finlayson, G.D.; Mackiewicz, M.; Hurlbert, A.C. Color correction using root-polynomial regression. IEEE Trans. Image Process. 2015, 24, 1460–1470. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Afifi, M.; Price, B.L.; Cohen, S.; Brown, M.S. When color constancy goes wrong: Correcting improperly white-balanced images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 1535–1544. [Google Scholar] [CrossRef]
  50. Ding, Y.; Fu, X. Kernel-based fuzzy c-means clustering algorithm based on genetic algorithm. Neurocomputing 2016, 188, 233–238. [Google Scholar] [CrossRef]
  51. Hemrit, G.; Finlayson, G.D.; Gijsenij, A.; Gehler, P.; Bianco, S.; Drew, M.S.; Funt, B.; Shi, L. Providing a single ground-truth for illuminant estimation for the ColorChecker dataset. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 1286–1287. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Banic, N.; Loncaric, S. Unsupervised learning for color constancy. In Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2018)—Volume 4: VISAPP, Funchal, Portugal, 27–29 January 2018; pp. 181–188. [Google Scholar] [CrossRef]
  53. Finlayson, G.D.; Zakizadeh, R. Reproduction angular error: An improved performance metric for illuminant estimation. In Proceedings of the 25th British Machine Vision Conference, BMVC 2014, Nottingham, UK, 1–5 September 2014. [Google Scholar] [CrossRef] [Green Version]
  54. Finlayson, G.D.; Zakizadeh, R.; Gijsenij, A. The reproduction angular error for evaluating the performance of illuminant estimation algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1482–1488. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Yu, H.; Chen, K.; Wang, K.; Qian, Y.; Zhang, Z.; Jia, K. Cascading convolutional color constancy. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; pp. 12725–12732. [Google Scholar]
  56. Koscevic, K.; Subasic, M.; Loncaric, S. Guiding the illumination estimation using the attention mechanism. In Proceedings of the 2020 2nd Asia Pacific Information Technology Conference, APIT 2020, Bangkok, Thailand, 14–16 January 2022; pp. 143–149. [Google Scholar] [CrossRef] [Green Version]
  57. Koscevic, K.; Banic, N.; Loncaric, S. Color beaver: Bounding illumination estimations for higher accuracy. In Proceedings of the 14th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, VISIGRAPP 2019, Volume 4: VISAPP, Prague, Czech Republic, 25–27 February 2019; pp. 183–190. [Google Scholar] [CrossRef]
  58. Li, B.; Xiong, W.; Hu, W.; Funt, B.V.; Xing, J. Multi-cue illumination estimation via a tree-structured group joint sparse representation. Int. J. Comput. Vis. 2016, 117, 21–47. [Google Scholar] [CrossRef]
  59. Banic, N.; Loncaric, S. Illumination estimation is sufficient for indoor-outdoor image classification. In Proceedings of the Pattern Recognition—40th German Conference, GCPR 2018, Stuttgart, Germany, 9–12 October 2018; Volume 11269, pp. 473–486. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed method. For a given test image, the related image features are extracted and used for estimating candidate illuminant colors and obtaining fuzzy weights, then the final estimated illuminant can be produced by weighting all outputs of multiple models.
Figure 1. The framework of the proposed method. For a given test image, the related image features are extracted and used for estimating candidate illuminant colors and obtaining fuzzy weights, then the final estimated illuminant can be produced by weighting all outputs of multiple models.
Applsci 12 01175 g001
Figure 2. The recovery and reproduction AEs’ distributions for some conventional methods versus the proposed method with the combined features. In the two subfigures (a,b), the images of No. 1~50 refer to the first 50 images in the Gehler-Shi dataset, and the images of No. 51~200 refer to the first 150 images the Cube+ dataset.
Figure 2. The recovery and reproduction AEs’ distributions for some conventional methods versus the proposed method with the combined features. In the two subfigures (a,b), the images of No. 1~50 refer to the first 50 images in the Gehler-Shi dataset, and the images of No. 51~200 refer to the first 150 images the Cube+ dataset.
Applsci 12 01175 g002
Figure 3. Comparison among the proposed, Luo’s [19], and conventional methods. The test images are from the Gehler-Shi and Cube+ datasets, and the resulting images are shown here by column corresponding to: (1) input images; (2) ground truths; (3) the proposed; (4) Luo’s, and (5) conventional methods (GW, WP, GE1, GE2, SOG, PCA-based, GGW, or LSR), respectively. These images include some common lighting scenes such as outdoor natural, outdoor artificial lighting, indoor artificial lighting, indoor natural lighting, combined natural and artificial lighting, etc. For each corrected image, the recovery AE (AE, for short) and the reproduction AE (RPE, for short) are displayed at the top, respectively. It should be noticed that these images displayed here are rendered just with some preliminary post-processing in the sRGB color space.
Figure 3. Comparison among the proposed, Luo’s [19], and conventional methods. The test images are from the Gehler-Shi and Cube+ datasets, and the resulting images are shown here by column corresponding to: (1) input images; (2) ground truths; (3) the proposed; (4) Luo’s, and (5) conventional methods (GW, WP, GE1, GE2, SOG, PCA-based, GGW, or LSR), respectively. These images include some common lighting scenes such as outdoor natural, outdoor artificial lighting, indoor artificial lighting, indoor natural lighting, combined natural and artificial lighting, etc. For each corrected image, the recovery AE (AE, for short) and the reproduction AE (RPE, for short) are displayed at the top, respectively. It should be noticed that these images displayed here are rendered just with some preliminary post-processing in the sRGB color space.
Applsci 12 01175 g003
Table 1. Comparative performance metrics between our method and some statistics-/learning-based methods on the Gehler-Shi dataset. The statistical metrics of the conventional methods are cited from [40] and the related literature.
Table 1. Comparative performance metrics between our method and some statistics-/learning-based methods on the Gehler-Shi dataset. The statistical metrics of the conventional methods are cited from [40] and the related literature.
MethodRecovery AEReproduction AE
MeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
WP [6]7.555.686.351.4516.12-8.16.57.1---
Edge-based Gamut [28]6.525.045.431.9013.58-------
GW [4,5]6.366.286.282.3310.58-7.06.86.9---
GE1 [8]5.334.524.731.8610.03-6.44.95.3---
GE2 [8]5.134.444.622.119.26-6.04.85.2---
SoG [7]4.934.014.231.1410.20-5.84.44.9---
Bayesian [29]4.823.463.881.2610.49-5.63.94.4---
GGW [8]4.663.483.811.0010.09-5.34.04.4---
Natural Image Statistics [9]4.193.133.451.009.22-4.83.53.9---
CART-based Combination [17]3.92.93.3---4.53.53.8---
WCS with partitioning [40]3.662.052.450.449.49-4.832.523.24---
PCA-based [21]3.522.142.470.508.74-4.712.723.28---
LSR [44]3.312.802.871.146.39-------
CNN-based method [46]2.631.982.100.723.90-------
FFCC [32]1.610.861.020.234.27-2.121.071.35---
C4-SqueezeNet-FC4 [55]1.350.880.990.283.21-------
Proposed
ClrM2.591.351.760.007.5319.283.341.572.130.009.8225.43
C M 19 3.332.622.760.697.2619.944.293.303.500.859.5323.42
C M 34 3.622.953.050.807.7220.434.673.553.810.9810.2533.69
CDF8.652.603.110.0029.53177.129.643.053.820.0032.31177.44
IIF4.143.033.370.979.0622.345.403.684.281.1312.3736.04
RGBuv3.252.352.550.677.3524.814.392.983.220.7910.48110.51
ChengSF4.173.033.410.859.2919.045.473.634.300.9812.7230.84
RGBuv+ChengSF3.062.202.400.686.8518.253.982.732.990.799.2729.51
RGBuv+ChengSF+IIF3.012.242.400.676.7019.483.942.752.970.799.1747.80
RGBuv+ChengSF+IIF+ C M 34 2.992.172.360.686.6819.163.902.803.020.809.1245.81
Table 2. Comparative performance metrics between our method and the previous methods on the Cube+ dataset. The statistical metrics of the previous methods are selected from [15,19].
Table 2. Comparative performance metrics between our method and the previous methods on the Cube+ dataset. The statistical metrics of the previous methods are selected from [15,19].
MethodRecovery AEReproduction AE
MeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
WP [6]9.697.488.561.7220.49-------
GW [4,5]7.714.294.981.0120.19-------
Gray pixels [42]6.653.263.950.6818.75-------
Color Tiger [52]3.912.052.530.9810.00-------
SoG [7]2.591.731.930.466.19-------
GE1 [8]2.501.591.780.486.08-------
GE2 [8]2.411.521.720.455.89-------
GGW [8]2.381.431.660.356.01-------
Attention CNN [56]2.051.321.530.424.84-------
Lighting classification DL [15]1.861.271.390.424.31-------
Luo [19]1.691.121.240.314.06-------
Color Beaver (GW) [57]1.490.770.980.213.94-------
Proposed
ClrM2.451.761.950.475.53143.543.362.392.610.617.7376.69
C M 19 2.161.521.710.414.9720.512.881.962.260.526.6925.38
C M 34 2.361.731.920.475.2318.193.362.342.600.617.84121.57
IIF2.281.581.730.465.3221.223.312.132.370.588.17113.33
RGBuv1.951.391.510.354.4815.722.631.862.020.446.1522.12
ChengSF2.241.491.680.415.3633.703.081.972.250.517.4942.90
RGBuv+ChengSF1.921.371.500.384.3715.022.581.852.030.476.0119.88
RGBuv+ChengSF+IIF1.791.311.410.364.0614.222.431.741.900.445.6220.04
RGBuv+ChengSF+IIF+ C M 34 1.801.311.420.364.0814.332.441.731.890.455.6420.18
Table 3. Comparative performance metrics between the proposed method and some conventional methods for Canon 5D.
Table 3. Comparative performance metrics between the proposed method and some conventional methods for Canon 5D.
MethodRecovery AEReproduction AE
MeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
GW [4,5]4.903.593.990.9610.8220.026.234.755.141.1713.6524.81
WP [6]3.822.492.890.509.3723.444.813.103.590.6111.7031.12
SoG [7]6.343.894.720.7615.5635.757.164.895.690.9216.7325.69
GE1 [8]5.894.084.690.7514.0032.436.745.205.690.8915.4127.41
GE2 [8]6.433.834.660.9216.0537.277.184.775.641.0617.1531.32
PCA-based [21]6.944.365.331.0516.6945.977.625.286.191.2017.3932.77
GGW [8]4.112.512.960.5210.2328.605.153.163.820.6412.7540.59
LSR [44]3.942.993.301.367.9617.924.843.884.161.619.6920.08
Proposed
ClrM2.320.681.240.007.2016.463.230.841.660.0010.2344.45
C M 19 2.832.142.340.536.3215.043.672.693.000.658.3221.28
C M 34 3.062.282.450.586.7414.884.142.963.200.709.4935.81
CDF8.491.542.190.0029.92179.089.871.632.660.0034.39178.91
IIF3.002.122.320.736.8025.024.002.562.950.829.5342.86
RGBuv2.972.172.270.586.8451.564.142.672.910.7310.10112.29
ChengSF3.232.212.480.607.5730.314.352.883.230.7110.6238.42
RGBuv+ChengSF2.852.152.230.576.5133.464.082.722.880.6910.05115.51
RGBuv+ChengSF+IIF2.832.082.180.576.5139.054.022.682.860.689.90116.17
RGBuv+ChengSF+IIF+ C M 34 2.781.992.130.616.3132.804.022.512.750.729.91120.15
Table 4. Comparative performance metrics between the proposed method and some conventional methods for Canon EOS 550D.
Table 4. Comparative performance metrics between the proposed method and some conventional methods for Canon EOS 550D.
MethodRecovery AEReproduction AE
MeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
GW [4,5]4.362.733.110.6110.8239.545.453.634.090.7713.1142.57
WP [6]3.622.372.590.558.9320.654.423.023.330.7110.5324.99
SoG [7]4.482.663.200.4711.3922.915.313.364.030.6013.0225.28
GE1 [8]4.022.803.140.559.5820.315.053.614.030.6711.8224.59
GE2 [8]4.402.553.070.6011.2424.585.223.263.940.7812.8627.53
PCA-based [21]4.982.993.550.7512.3526.805.853.884.500.9413.8327.84
GGW [8]3.281.972.260.448.4920.404.072.572.950.5710.2026.87
LSR [44]5.254.154.581.3410.7917.176.515.515.841.5713.0920.80
Proposed
ClrM2.642.132.250.615.5026.983.682.873.060.817.8936.12
C M 19 2.301.691.840.555.0622.163.062.222.440.706.7827.69
C M 34 2.521.982.110.625.3218.273.612.542.830.818.07113.87
CDF49.206.4940.981.36173.44179.6850.288.6241.391.85171.63179.64
IIF2.501.781.950.575.7421.023.642.302.600.718.89110.91
RGBuv2.041.451.590.434.6328.632.791.962.150.546.4244.16
ChengSF2.441.591.830.475.8334.983.342.132.450.588.1247.04
RGBuv+ChengSF2.001.431.570.434.4922.392.711.952.120.536.2032.05
RGBuv+ChengSF+IIF1.881.411.500.414.1620.352.541.822.000.515.7726.24
RGBuv+ChengSF+IIF+ C M 34 1.871.441.530.424.1420.522.531.862.020.515.7427.23
Table 5. Comparative performance metrics between the proposed method and some conventional methods, with training and testing on the combination of the Gehler-Shi and Cube+ datasets.
Table 5. Comparative performance metrics between the proposed method and some conventional methods, with training and testing on the combination of the Gehler-Shi and Cube+ datasets.
MethodRecovery AEReproduction AE
MeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
GW [4,5]4.413.053.360.6810.5339.545.564.004.390.8812.8942.57
WP [6]3.572.302.570.528.8923.444.393.003.300.6610.5831.12
SoG [7]4.852.833.470.5312.3835.755.653.594.290.6713.8126.08
GE1 [8]4.422.883.360.6010.7132.435.403.714.260.7412.6927.41
GE2 [8]4.822.773.360.6512.3737.275.583.464.120.8113.7231.32
PCA-based [21]5.393.153.870.8113.4945.976.194.014.741.0014.6932.77
GGW [8]3.402.002.340.458.7728.604.212.643.040.5710.6140.59
LSR [44]4.803.784.101.3510.0020.485.964.795.171.5912.2221.47
Luo [19]1.971.301.450.364.7525.382.571.641.860.456.3237.14
Proposed
ClrM2.501.691.910.126.2340.803.432.232.480.148.80113.29
C M 19 2.611.902.090.615.7920.873.412.452.700.757.73124.29
C M 34 2.862.232.380.666.1320.673.912.973.170.828.69102.71
CDF45.004.969.270.26167.64179.5445.426.3311.330.32166.02179.40
IIF2.882.062.220.606.6122.853.932.582.860.779.4149.67
RGBuv2.341.701.810.515.3321.963.122.212.380.637.2234.57
ChengSF2.901.972.210.546.8034.914.022.552.900.689.87123.33
RGBuv+ChengSF2.221.601.740.455.0618.222.932.042.260.576.7721.28
RGBuv+ChengSF+IIF2.131.551.690.484.7614.832.802.062.230.596.3518.79
RGBuv+ChengSF+IIF+ C M 34 2.101.521.670.464.7416.912.771.992.180.586.2818.81
Table 6. Statistical metrics on the combination dataset of Gehler-Shi and Cube+ with different clustering methods and cluster numbers.
Table 6. Statistical metrics on the combination dataset of Gehler-Shi and Cube+ with different clustering methods and cluster numbers.
ClusteringRecovery AEReproduction AE
MethodNumberMeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
K-means22.121.531.680.484.7618.322.801.992.210.596.3523.79
KFCM22.101.521.670.474.7015.712.761.982.190.586.2518.18
K-means32.121.511.650.464.8415.152.771.952.160.576.3720.10
KFCM32.131.531.670.464.8318.872.791.972.180.566.4327.39
twoStep2 × 25.984.055.292.4910.3510.988.554.166.813.4715.3615.68
K-means42.151.461.610.455.0651.862.841.862.100.576.7791.14
KFCM42.141.541.690.484.8014.832.822.062.230.586.4118.86
K-means52.111.401.560.434.9931.642.781.792.040.536.6452.73
KFCM52.171.501.690.474.9514.772.851.952.180.586.6118.81
twoStep2 × 36.295.305.002.3712.0217.758.586.427.043.1616.2425.69
K-means62.271.441.600.445.57108.903.081.812.060.557.74119.54
KFCM62.221.531.700.455.1621.962.911.972.200.556.8628.93
Table 7. Statistical metrics on the Gehler-Shi and Cube+ datasets with different kernel width in KFCM clustering algorithm, where the selected feature combination is RGBuv+ChengSF+IIF, and the cluster number K = 5 .
Table 7. Statistical metrics on the Gehler-Shi and Cube+ datasets with different kernel width in KFCM clustering algorithm, where the selected feature combination is RGBuv+ChengSF+IIF, and the cluster number K = 5 .
Kernel WidthRecovery AEReproduction AE
MeanMed.Tri.B25%W25%MaxMeanMed.Tri.B25%W25%Max
302.161.541.690.464.9126.052.841.992.200.576.5327.08
602.141.551.680.484.8218.852.802.032.200.586.3927.35
1002.151.561.700.494.8315.382.832.042.230.596.4318.79
1402.141.551.700.484.8214.832.822.042.220.596.4218.71
1802.141.551.700.484.7915.362.812.032.230.596.3818.64
2402.161.521.690.464.9514.932.851.942.200.576.6219.46
3002.161.521.690.474.9214.932.851.972.210.576.6018.97
4002.191.521.700.465.0219.922.881.972.220.576.7123.64
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Luo, Y.; Wang, X.; Wang, Q. Which Features Are More Correlated to Illuminant Estimation: A Composite Substitute. Appl. Sci. 2022, 12, 1175. https://doi.org/10.3390/app12031175

AMA Style

Luo Y, Wang X, Wang Q. Which Features Are More Correlated to Illuminant Estimation: A Composite Substitute. Applied Sciences. 2022; 12(3):1175. https://doi.org/10.3390/app12031175

Chicago/Turabian Style

Luo, Yunhui, Xingguang Wang, and Qing Wang. 2022. "Which Features Are More Correlated to Illuminant Estimation: A Composite Substitute" Applied Sciences 12, no. 3: 1175. https://doi.org/10.3390/app12031175

APA Style

Luo, Y., Wang, X., & Wang, Q. (2022). Which Features Are More Correlated to Illuminant Estimation: A Composite Substitute. Applied Sciences, 12(3), 1175. https://doi.org/10.3390/app12031175

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop