Next Article in Journal
Proposal of an Alpine Skiing Kinematic Analysis with the Aid of Miniaturized Monitoring Sensors, a Pilot Study
Next Article in Special Issue
Retina-like Computational Ghost Imaging for an Axially Moving Target
Previous Article in Journal
Pseudo-Static Gain Cell of Embedded DRAM for Processing-in-Memory in Intelligent IoT Sensor Nodes
Previous Article in Special Issue
Color Demosaicing of RGBW Color Filter Array Based on Laplacian Pyramid
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Crosstalk Correction for Color Filter Array Image Sensors Based on Lp-Regularized Multi-Channel Deconvolution

School of Electrical and Electronic Engineering, Yonsei University, Seoul 03722, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(11), 4285; https://doi.org/10.3390/s22114285
Submission received: 8 May 2022 / Revised: 1 June 2022 / Accepted: 2 June 2022 / Published: 4 June 2022
(This article belongs to the Collection Computational Imaging and Sensing)

Abstract

:
In this paper, we propose a crosstalk correction method for color filter array (CFA) image sensors based on L p -regularized multi-channel deconvolution. Most imaging systems with CFA exhibit a crosstalk phenomenon caused by the physical limitations of the image sensor. In general, this phenomenon produces both color degradation and spatial degradation, which are respectively called desaturation and blurring. To improve the color fidelity and the spatial resolution in crosstalk correction, the feasible solution of the ill-posed problem is regularized by image priors. First, the crosstalk problem with complex spatial and spectral degradation is formulated as a multi-channel degradation model. An objective function with a hyper-Laplacian prior is then designed for crosstalk correction. This approach enables the simultaneous improvement of the color fidelity and the sharpness restoration of the details without noise amplification. Furthermore, an efficient solver minimizes the objective function for crosstalk correction consisting of L p regularization terms. The proposed method was verified on synthetic datasets according to various crosstalk and noise levels. Experimental results demonstrated that the proposed method outperforms the conventional methods in terms of the color peak signal-to-noise ratio and structural similarity index measure.

1. Introduction

Advancements in image acquisition devices, including camera imaging systems, have continually opened new application areas. In recent years, the rapid development of imaging systems fostered by mobile-camera technology has promoted the application of imaging systems in many fields, such as security, military, aerial imaging, satellite imaging, and autonomous vehicles. Accordingly, the demand for cost-effective high-performance imaging using low computational power has increased. To fulfill this need for efficiency in camera imaging systems, a single sensor based on a color filter array (CFA) is essential, instead of using three sensors or optical beam splitters.
In response to consumer perceptions and expectations of digital color image contents, attempts have been made to increase the spatial resolution engendered by the number of pixels in a limited-sized imaging system. Many signal-processing-based methods for single sensor architectures have been required to achieve high-resolution, high-sensitivity imaging. Studies on CFAs [1,2,3] and demosaicing methods [4,5,6,7,8] have been conducted to maximize CFA performance. To increase the spatial resolution in a limited-size image sensor, the pixel size should be reduced. However, this approach can compromise the sensitivity [9]. As a result, smaller pixels tend to generate more noise. Thus, denoising algorithms [10,11] are needed to solve such problems, which often occur in low-light situations. Although the effect on noise is minimized through signal processing, the issue caused by the pixel size reduction remains.
Another problem caused by the pixel size reduction is the crosstalk phenomenon. Specifically, it is caused by the physical limitations of the image sensor, that is, the leakage of photons and electrons between adjacent pixels [12,13]. As the pixel size continues to shrink, crosstalk between adjacent pixels emerges. In CFA image sensors, photon and electron leakage due to optical refraction and minority carriers inevitably occurs as the geometric distance between adjacent pixels decreases. In the obtained image, crosstalk typically generates two artifacts: desaturation and blurring. As shown in Figure 1, color constancy and preservation of the spatial resolution of the image thus cannot be achieved.
Several attempts have been made to overcome this crosstalk phenomenon. The first approach is based on the hardware design of the sensor or CFA. To achieve better color reproduction, the CFA is designed to optimize the spectral responses of the entire imaging system by compensating for the crosstalk effects [14]. Use of backside illumination technology with a parabolic color filter decreases the crosstalk while increasing the efficiency [15]. New CFA patterns have been proposed to minimize crosstalk in a sensor with a small pixel size using a combination of primary and complementary color filters, such as yellow, cyan, and magenta, as in [12]. Although the crosstalk effect has been reduced through physical improvements, the image sensors experience difficulties as the pixel size decreases.
The second approach is based on signal processing. For example, the crosstalk phenomenon was analyzed by Hirakawa et al. using a spatio-spectral sampling theory, which was formulated by using a convolution operation [13]. This method assumes that the most significant artifact is desaturation; thus, it focuses on color correction. However, the color correction method that employs matrix inversion may have limitations owing to certain issues, such as blurry artifacts and noise amplification under low-light conditions.
To solve the desaturation and blurring problem, joint decrosstalk and demosaicing method was proposed [16]. This method is based on a piecewise autoregressive image model. It performs demosaicing iteratively under crosstalk constraints. However, this method is not practicable because it is difficult to utilize recently studied demosaicing algorithms, such as ARI [8], and it requires considerable computation. In addition, because this method focuses more on estimating the edge directional weight that is required for demosaicing, it does not sufficiently preserve the edges in constraint. Therefore, we herein propose a crosstalk correction method that is based on multi-channel deconvolution with a hyper-Laplacian prior. It restores the image spatial resolution and color fidelity without producing obvious artifacts.
The main contributions of this study are the following:
  • The crosstalk problem is formulated as a multi-channel degradation model;
  • A multi-channel deconvolution method based on the objective function with a hyper-Laplacian prior is designed. The proposed method utilizes L p regularization to achieve the estimated image with sharp edges and details, and it efficiently suppresses noise amplification for each color component. Concurrently, intercolor regularization is employed to smooth the color difference components and to encourage the homogeneity of the edges;
  • An efficient algorithm based on alternating minimization is described. Experimental results validate that the proposed method is more robust than conventional methods.
The remainder of this paper is organized as follows. Section 2 presents the crosstalk analysis and problem formulation on the multi-channel degradation model. In Section 3, the proposed method with a hyper-Laplacian prior is detailed. Section 4 presents the experimental results. Finally, Section 6 concludes this paper.

2. Problem Formulation

An image sensor is based on either a charge-coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS). By using CFA and a microlens array, the image sensor converts the light received from the main lens into digital signals. As each color filter has its own spectrum, only one piece of color information is obtained from one pixel.
To estimate an original color image from a subsampled image according to the CFA pattern, demosaicing is used. It is intended to overcome the physical limitations of a single-sensor imaging system. Consider an original discrete color image, x ( n ) = [ x R ( n ) , x G ( n ) , x B ( n ) ] , where n Z 2 , and x R ( n ) , x G ( n ) , and x B ( n ) are the three color components for respective red (R), green (G), and blue (B) channels. The subsampled image y s ( n ) according to the CFA pattern can be expressed as
y s ( n ) = k { R , G , B } d k ( n ) x k ( n ) + e ( n ) ,
where d k ( n ) is the subsampling function of CFA and k { R , G , B } . This function is periodic, and d R ( n ) + d G ( n ) + d B ( n ) = 1 for each pixel. The term e ( n ) is the corresponding signal-independent additive noise.
During the process of obtaining a subsampled image, desaturation and blurring occur as a result of the crosstalk phenomenon. Figure 2 illustrates the structure of the imaging system. The light entering through the camera’s main lens sequentially passes through the microlens and color filter and reaches the imaging sensor. In this process, interference occurs on account of optical diffraction and minority carriers caused by the surrounding pixels. Therefore, the image acquired by the imaging sensor suffers simultaneously from spatial and spectral degradation. As shown in Figure 2c, the spectral sensitivity shifts owing to crosstalk inside the imaging sensor. The observation model with the crosstalk phenomenon can be represented as follows:
y c t ( n ) = m b ( n , m ) y s ( n ) + e ( n ) = m k { R , G , B } b ( n , m ) d k ( n ) x k ( n ) + e ( n ) ,
where y c t ( n ) is the subsampled image under the crosstalk phenomenon and b ( n , m ) is the crosstalk function. This function represents a combination of optical diffraction and minority carriers. We assume that the crosstalk function is space-invariant. Using matrix notation, the relation between the subsampled image and the original image is rewritten as follows:
y c t = B D x + e ,
where x denotes the three color component vectors of the original color image with M rows and N columns; that is, x T = [ x R T , x G T , x B T ] . Moreover, y denotes the observed image acquired by the sensor according to the CFA, and e is corresponding additive noise. Vector x has dimensions 3 M N , and each vector y and e is a vector of dimensions M N . Matrix B has dimensions M N × M N and denotes a crosstalk function, while D is a subsampling operator with dimensions M N × 3 M N .
In general, it is important to solve the inverse problem according to the degradation model. If the observation model of Equation (3) is a linear shift-invariant (LSI) system, B D can be replaced with D B since the commutative property is established. The special type of matrix associated with LSI systems is known as the Toeplitz matrix. In the above system, B satisfies this matrix; however D is not a Toeplitz matrix. In other words, subsampling operator D violates the commutative property of convolution. Thus, to solve the Equation (3) problem, the inverse problem with respect to matrix D should be solved sequentially after solving the inverse problem in terms of matrix B and considering noise e .
On the other hand, the crosstalk degradation is solved by focusing on color fidelity in [13] after solving the D matrix, which relates to subsampling through demosaicing. As this method performs demosaicing first, it is not suitable for handling the modeling presented in Equation (3). Alternatively, a method of performing deconvolution ( B 1 ) in the subsampled domain should be considered prior to demosaicing ( D 1 ). It is important to solve the problems sequentially according to modeling. However, performing deconvolution in the subsampled domain causes another problem. Although the method can solve the blur phenomenon in Equation (3), it does not sufficiently separate mixing between color channels. Furthermore, an iterative approach is used that repeatedly performs demosaicing and deconvolution in the subsampled domain to overcome blur artifacts and maintain color fidelity in [16].
The crosstalk function is not a simple low-pass filter. Rather, it is a complex process in which the spectral information of three color components is mixed and the spatial information of the surrounding pixels are simultaneously mixed in the subsampled domain, as shown in Figure 3. In particular, as the standard deviation of the Gaussian kernel increases, the image degradation due to the crosstalk phenomenon increases. As the noise increases, it becomes difficult to improve the color fidelity. Therefore, a new formulation is needed to solve the problem based on a multi-channel degradation model. The following multi-channel degradation model is considered:
y = D H x + e ,
where the condition B D = D H should be satisfied, and multi-channel degradation matrix H is defined by
H = H R R H R G H R B H G R H G G H G B H B R H B G H B B .
Here, H k l is the dimension M N × M N for ( k , l ) { R , G , B } . Diagonal matrices H k k represent the within-channel kernel, and off-diagonal matrices H k l represent the cross-channel kernel when k l .
Matrix H can be composed by utilizing the characteristic that each location and channel has a different kernel according to the CFA periodic structure. Figure 4a shows an example of configuring matrix H using matrix B and the CFA structure. As each color filter has its own spectrum, the crosstalk kernel is separated into three different kernels depending on the location of each color filter even though they have the same shape as the low-pass filter. Thereafter, a total of nine kernels can be configured by separating each color channel. Note that multi-channel degradation matrix H changes according to the type of CFA pattern. In the case of a RGBW [3] or multispectral filter array [2], the number of spectral bands and spatial periodicity of CFA sampling patterns are different and the formulation thus must be slightly changed.
Figure 4b shows that different formulations—single-channel and multi-channel degradation models—can yield the same degraded results. By modeling the same phenomenon differently, we propose a method that can simultaneously solve complex spatial and spectral degradation through a multi-channel degradation model.

3. Proposed Method

In this section, we propose an efficient algorithm for multi-channel deconvolution based on the multi-channel degradation model, as mentioned above. To overcome the ill-posed problem, effective regularization terms are employed to restrict the solution space. The proposed method utilizes L p regularization to achieve the estimated image with sharp edges and details. It efficiently suppresses noise amplification for each color component. Concurrently, intercolor regularization is employed to smooth the color difference components and to encourage the homogeneity of the edges. Thereafter, we describe an efficient solver to minimize the cost function regularized by specific prior information.

3.1. Multi-Channel Deconvolution

The multi-channel deconvolution problem is used to obtain an estimate of x given y , D , H , and the prior knowledge of e . Equation (4) can be rewritten as
y k = D k H k x + e k , for k { R , G , B } ,
where matrix H k is given by
H k = [ H k R , H k G , H k B ] ,
with the kth block row matrix of dimension M N × 3 M N . The regularized solution of the multi-channel deconvolution problem in (6) is defined as
x ^ = arg min x k { R , G , B } 1 λ k | | y k D k H k x | | 2 2 + R ( x ) .
Here, the first term, called the data fidelity term, denotes a measure of the Euclidean distance from the original image to the observed image. The second term R ( x ) is the regularization term based on the prior model. Regularization parameters λ k manage the trade-off between the two terms. As each color channel suffers from degradation under the same conditions, it is assumed that λ k for each channel has the same value.
The data fidelity term can generally be defined as the L 2 -norm, which is known as the least-squares approach. Although the L 1 -norm increases the robustness to outliers such as noise or errors, it is employed in this study because the error term is generally modeled as Gaussian noise and the crosstalk kernel is an already known prior. As the crosstalk kernel is not signal-dependent, it can be parameterized through experiments. Moreover, it is generally assumed to be a Gaussian kernel [14]. In addition, L 2 -norm data fidelity enables the finding of solutions with low computational costs [17].
The regularization term is herein adopted to decrease the uncertainty in the inverse problem by constraining the solution space. The regularization term can be expressed according to the combination of several constraints:
R ( x ) = i λ i R i ( x ) ,
where λ i is the regularization parameter for each constraint function R i ( x ) . Tikhonov regularization is a representative smoothing constraint that assumes that the distribution of gradient magnitudes for images is smooth [18]. This strategy decreases the uncertainty using the L 2 -norm and is defined as follows:
R ( x ) = | | C x | | 2 2 ,
where C denotes the second derivative linear operator, which is known as the Tikhonov matrix. Tikhonov regularization smooths the estimated image with a limited high-frequency component. However, this regularization strategy inevitably oversmooth some important information, such as edge sharpness and detail.
To reconstruct the estimated image while preserving the sharp edges, total variation (TV) regularization is introduced in [19]. TV regularization is one of the most successful techniques for reconstructing images. In the field of denoising, various noise types are effectively removed using this regularization strategy [20]. Unlike Tikhonov regularization, TV regularization is employed in the image restoration process under the assumption that the image gradient distribution has a Laplace distribution. Therefore, high-frequency information, such as edge information or details, can be effectively restored. TV regularization is defined as follows:
R ( x ) = | | x | | ,
where denotes the first derivative linear operator. TV regularization is isotropic if the norm | | · | | denotes the L 2 -norm, and it is anisotropic if it denotes the L 1 -norm.
Although TV regularization exhibits excellent performance in various image processing fields, recent works have focused on presenting constraints with consideration of the image characteristics. In research on the properties of natural images [21,22], the statistic of real-world scenes substantially follows a heavy-tailed distribution in their gradients, which is modeled as a hyper-Laplacian prior ( p ( x ) exp ( k | | x | | p p ) ). This prior is incorporated as L p regularization, and it is defined as follows:
R ( x ) = | | x | | p p ,
where the norm | | · | | p p denotes the L p -norm and the value of p is typically set to [ 0.5 , 0.8 ] .

3.2. Constraints

In the spirit of a hyper-Laplacian prior, the proposed constraint consists of spatial regularization to encourage high frequencies within each color channel. It additionally fosters intercolor regularization to enforce the homogeneity of the edges or details in the color channels. The first constraint R 1 ( x ) coerces sharp edges in the estimated image. If the prior knowledge of the image gradient is considered, an optimal solution can be found. Figure 5a presents the probability density of the first-order derivative using 42 images from the Kodak and McMaster datasets [23]. The empirical distribution can be modeled as a hyper-Laplacian distribution with p = 0.66 . This property allows a regularization term to employ a hyper-Laplacian prior in the deconvolution problem. Therefore, R 1 ( x ) is represented as follows:
R 1 ( x ) = | | x | | p p ,
where ∇ denotes the first derivative operator and p = 0.66 .
The second constraint R 2 ( x ) is chosen to encourage smoothness in the differences between the color components. In most demosaicing, color interpolation is performed under the assumption of a high correlation among the three color bands [4,5,6,7,8]. This property is also valid for natural color images. The optimal L p -norm can be selected through the probability density of the color difference gradient, as presented in Figure 5b. Similar to the case of R 1 ( x ) , a hyper-Laplacian distribution with p = 0.62 allows the regularization term to deploy L p -norm prior in the image restoration process. The term R 2 ( x ) is expressed as follows:
R 2 ( x ) = | | ( + 2 x R x G x B ) | | p p + | | ( x R + 2 x G x B ) | | p p + | | ( x R x G + 2 x B ) | | p p .
This regularization term can be expressed in the form R 2 ( x ) = | | M x | | p p , where
M = 2 0 0 2 1 0 0 1 1 0 0 1 1 0 0 1 2 0 0 2 1 0 0 1 1 0 0 1 1 0 0 1 2 0 0 2 ,
and M is an 3 M N × 3 M N dimensional matrix.
Each regularization term of R 1 ( x ) and R 2 ( x ) is incorporated into the objective function of (8), and the overall objective function is represented as follows:
x ^ = arg min x | | y D H x | | 2 2 + λ 1 | | x | | p p + λ 2 | | M x | | p p ,
where λ 1 and λ 2 are regularization parameters for the constraints of color components and color difference components, respectively. As R 1 ( x ) and R 2 ( x ) have similar p values near 2 / 3 , we assume p = 2 / 3 in the following optimization problem for convenience.

3.3. Optimization

In this subsection, we describe how to optimize the objective function of Equation (16). The minimization of an objective function with L p regularization engenders difficulty in obtaining closed-form solutions because of the non-differentiability and non-linearity of the L p regularization. To efficiently recover estimated image x and solve this problem, a strategy based on alternating minimization [24], known as half-quadratic splitting, is employed. Two auxiliary variables w and v are introduced to approximate x and Mx among the non-differentiable terms | | · | | p p , respectively. The approximation model to Equation (16) can be expressed as follows:
x ^ = arg min x , w , v { | | z H x | | 2 2 + θ 1 | | x w | | 2 2 + λ 1 | | w | | p p + θ 2 | | M x v | | 2 2 + λ 2 | | v | | p p } ,
where w and v are the auxiliary variables, θ i denotes the auxiliary regularization parameter, and z represents the estimated full-color image. Alternating minimization guarantees that the minimizer of Equation (17) converges to that of Equation (16) as the value of θ i moves toward . Therefore, the objective function in Equation (17) is convex and differentiable. This formulation allows the minimization problem with respect to the others to obtain a closed-form formula if one of the three variables, w , v , and x , is fixed.
Note that vector z is first estimated based on the demosaicing method that satisfies the condition y = D z . The choice of the demosaicing method is not relevant; however, the state-of-the-art demosaicing method (e.g., ARI [8]) is advantageous in reducing color artifacts. Although the effects of subsampling cannot be completely eliminated, this process enables the objective function to negate the subsampling operator of matrix D .
Since w and v are separable in a given x , the minimizer of the objective function is readily obtained. By fixing x , the formulation of Equation (17) is separable in w and v . Based on the alternative minimization method [24], the minimization problem can be transformed into two subproblems, as follows:
w ^ = arg min w | | w x | | 2 2 + λ 1 θ 1 | | w | | p p ,
v ^ = arg min v | | v M x | | 2 2 + λ 2 θ 2 | | v | | p p .
Each objective function in (18) and (19) is a single-variable problem. Although it is difficult to derive an exact solution for these objective functions, the minimization can be obtained using a numerical root-finder, such as the Newton–Raphson method. In the special case of p = 2 / 3 , the exact analytic solution for non-zero w and v is described in [21]. Specifically, the multidimensional minimizer of α | | s | | p p + | | s t | | 2 2 is extended to the scalar minimizer of α | s | p + | s t | 2 . By calculating the derivative of α | s | p + | s t | 2 with respect to s, the minimizer is given by:
α p | s | p 1 + 2 | s t | = 0 .
For p = 2 / 3 , the equation can be expressed as follows:
α | s | 1 / 3 + 3 | s t | = 0 , α 3 | s | 1 + 27 | s t | 3 = 0 , s 4 3 t s 3 + 3 t 2 s 2 t 3 s + α 3 27 = 0 .
Therefore, the solution of each objective function in (18) and (19) can be solved numerically through the root of the quartic polynomial.
Note that if p = 1 , then the objective functions in (18) and (19) will be equal to the TV regularized problem. Although the proposed method has an optimal solution with approximately p = 2 / 3 and a corresponding solution is derived above, the TV regularization remains a useful strategy in terms of speed and implementation. Each solution of (18) and (19) with p = 1 is given by the one-dimensional shrinkage, as follows:
w ^ = x | | x | | 1 max ( | | x | | 1 λ 1 2 θ 1 , 0 ) ,
v ^ = M x | | M x | | 1 max ( | | M x | | 1 λ 2 2 θ 2 , 0 ) .
The solution to this problem can be achieved component-wise.
For fixed w and v variables, the objective function in (17) can be simplified as a least-squares formulation as
x ^ = arg min x | | z H x | | 2 2 + θ 1 | | x w | | 2 2 + θ 2 | | M x v | | 2 2 .
Moreover, the objective function is convex and differentiable with respect to x . As the objective function is composed of quadratic terms, the corresponding optimal solution of Equation (24) becomes a least-squares problem as follows:
( H T H + θ 1 T + θ 2 ( M ) T ( M ) ) x = H T z + θ 1 T w + θ 2 ( M ) T v .
The solution can be obtained by solving its normal equations.
We summarize the proposed crosstalk correction process based on multi-channel deconvolution in Algorithm 1 with the continuation technique on regularization parameter θ i . The continuous scheme is widely used with the penalty method to speed up the overall convergence. As a small value of θ i is initially set and gradually increases in iterations, the convergence speed can be accelerated. As the value of θ i moves toward inf, the minimizer of Equation (17) transformed by the quadratic penalty method converges to that of Equation (16). In terms of convergence, the objective function in Equation (17) is convex in x , w and v .
Algorithm 1 Crosstalk Correction based on L p -regularized Multi-channel Deconvolution.
Input: The observed image y , the subsamling matrix D , the crosstalk matrix H , and the regularization parameters λ i
Output: The reconstructed image x ^
  Initialization: x 0 = 0
  Solve z based on the demosaicing method ( y = D z )
   θ i λ i
1:
repeat
2:
   Solve w according to Equation (18)
3:
   Solve v according to Equation (19)
4:
   Solve x according to Equation (24)
5:
    θ i 2 θ i
6:
until  θ i < θ m a x
7:
return  x ^
In implementation terms, the computation of each block in Equation (25) can be accelerated by a two-dimensional (2D) fast Fourier transform (FFT) at each iteration. Under the periodic boundary condition for x , T , and all block matrices in H T H are block circulants. Specifically, the matrix on the left-hand side of Equation (25) can be precomputed once before the iterations. At each iteration, six FFTs are applied to w and v . Then, three inverse FFTs are computed to obtain x . Therefore, a total of nine FFTs (including three inverse FFTs) are required to solve Equation (25).

4. Experimental Results

The performance of the proposed method was evaluated using the color peak signal-to-noise ratio (CPSNR) and structural similarity index (SSIM) [25] as objective image quality metrics. The former was used to evaluate the intensity differences between the original image and the estimated image; the latter was employed to evaluate the structural similarity in terms of the human visual system. In particular, SSIM is suitable for comprehensively determining the similarity between luminance and chrominance components.

4.1. Datasets

Several experiments were conducted to verify the performance of the proposed method using artificially degraded images. For the comparisons, we generated crosstalk degradation on four public benchmark datasets: Kodak, McMaster [23], Set5 [26], and Set14 [27]. The point spread function of crosstalk is inspired by the crosstalk kernel given by [14]. Moreover, two levels of degradation were tested with consideration of the crosstalk phenomenon that intensifies in accordance with the pixel size. For crosstalk degradation 1, the synthetic datasets were generated by convolving the ground truth (GT) images within the given dynamic range [ 0 , 1 ] and the crosstalk kernels of the 2D Gaussian kernel with a standard deviation of σ g = 0.45 . They were also generated by adding Gaussian noise with a standard deviation of σ n = 0.01 . For crosstalk degradation 2, the Gaussian crosstalk kernel with a standard deviation of σ g = 0.60 and a Gaussian noise with a standard deviation of σ n = 0.04 were used. We assume that the crosstalk kernel is space-invariant.

4.2. Compared Methods

For performance comparisons, we implemented four conventional methods. The first conventional method (CM1) was explained as the ARI method in [8]. It only performed demosaicing and ignored the crosstalk kernel. The second (CM2), described as Lucy-demosaicing in [16], performed single-channel deconvolution based on the Lucy–Richardson method [28]. The third (CM3) was inspired by CM2 and applied TV-based single-channel deconvolution [24]. For CM2 and CM3, deconvolution was performed in the subsampled domain, followed by demosaicing. The fourth method (CM4) was deemed a color correction method in [13]. It focuses on color correction with consideration of the crosstalk kernel. For CM4, the CFA image was first demosaiced by the various demosaicing methods. The fifth (CM5), sixth (CM6), and seventh (CM7) performed single-channel deconvolution based on deep-learning approach [29,30,31]. Each method uses a convolution neural network (CNN) to solve a constraint optimization problem. In a similar way to CM3 and CM4, deconvolution performed in the subsampled domain for CM5, CM6, and CM7 using pre-trained models. In this experiment, ARI demosaicing [8] was applied equally to various conventional methods requiring demosaicing. The parameters were chosen to achieve the highest CPSNR values for all test images.

4.3. Comparisons

The quantitative evaluation of the Kodak and McMaster datasets [23] for crosstalk degradation 1 in terms of the CPSNR and SSIM values is demonstrated in Table 1. It is observed that the proposed method improved the performance. It achieved the highest CPSNR and SSIM values among all methods, including demosaicing (CM1), deconvolution (CM2 and CM3), color correction (CM4), and deep-learning-based deconvolution approach (CM5, CM6, and CM7). Note that the higher the CPSNR value is, the closer the estimated image is to the original image. Moreover, the higher the SSIM value, the better the perceived quality. The widening performance gap between the conventional methods and the proposed method can be confirmed for crosstalk degradation 2 in Table 2. In other words, the proposed method is robust against both crosstalk degradation and noise. Quantitative assessments for the Set5 [26] and Set14 [27], including the Kodak and McMaster datasets [23], are summarized in Table 3.
A comparison of the qualitative evaluation for crosstalk correction is visualized in Figure 6. In CM1, where only demosaicing was applied, color degradation is clearly observed. When the degree of crosstalk degradation is insignificant, it is defined as crosstalk degradation 1. The single-channel deconvolution methods (CM2 and CM3) show relatively good results in terms of improving color fidelity and producing sharp details. However, noise amplification is not completely considered. In CM4, which focused on color correction, blurring artifacts are not at all improved. The deep-learning-based deconvolution methods (CM5, CM6, and CM7) show usable results in terms of reducing noise and producing sharp details. However, the color fidelity has improved in a limited way. In contrast, the proposed method based on multi-channel deconvolution successfully reconstructed edge sharpness and overcame color degradation without noise amplification.
Furthermore, as the degree of crosstalk degradation increased, the performance difference between the conventional methods and the proposed method became more apparent. Figure 7 and Figure 8 present the results for crosstalk degradation 2. CM1 shows noisy results without overcoming color degradation and blurring artifacts. The majority of single-channel deconvolution methods (CM2 and CM3) exhibit limitations in terms of color fidelity because deconvolution was performed in the subsampled domain. With respect to noise, CM3 shows slightly better results than CM2. In CM4, the color fidelity is improved close to the ground truth; however, the noise is amplified, appearing like ink stains in the pixels with noise. However, as is clearly observed, the proposed method based on multi-channel deconvolution improves the color fidelity and produces much sharper details than the other methods. Specifically, it produces the fewest artifacts, such as noise amplification. In addition, the visual comparison of the difference image shows that the proposed method achieves a low error rate. Therefore, the proposed method yields more natural and credible results than the conventional methods.

4.4. Influence of Parameters

The proposed method involves regularization parameters λ i , which manage the trade-off between the fidelity to the data and smoothness of the solution. To analyze the effects of the values of λ i on the crosstalk correction method, we performed experiments with different parameter settings. The value of λ i should be determined empirically based on experiments so that the restored images have reasonable signal-to-noise ratios. Figure 9 and Figure 10 present examples of the influence of λ 1 and λ 2 in the case of crosstalk degradation 2. The results with small values of λ 1 are much sharper and noisy, whereas the results with large values of λ 1 are seemingly smoothed and have less noise. Smaller values of λ 2 tend to yield more vivid color, whereas large values of λ 2 may be desaturated and have fewer color artifacts.
Figure 11 demonstrates that crosstalk correction can be effectively performed in a range of λ 1 , that is, within [ 0.005 , 0.05 ] and [ 0.05 , 0.01 ] for crosstalk degradation 1 and 2, respectively. It is appropriate to select the values of λ 2 within [ 0.0003 , 0.003 ] . Therefore, choosing the appropriate λ i according to the level of degradation will improve the subjective and objective performances.

4.5. Convergence Analysis

Finally, we analyzed the convergence behavior of multi-channel deconvolution for crosstalk correction. The criterion | | x n + 1 x n | | 2 / | | x n | | 2 < η was used to terminate the iteration, where η was set to 10 4 . The propose method converges for various initial conditions, as plotted in Figure 12a. For the initial estimates, x 0 , a blank image ( x 0 = 0 ), a demosaiced version of the observed image ( x 0 = z ), or a degraded version of the demosaiced image ( x 0 = H T z ) was employed. For the experiment on various levels of crosstalk degradation, the objective function decreased in several iterations, as illustrated in Figure 12b. In addition, the visualization in Figure 12c shows that the residual image converges to the image with no value as the iteration progresses. Therefore, the objective function with a hyper-Laplacian prior converged in several iterations for some variables through the efficient solver.

5. Discussion

The crosstalk problem generated by the crosstalk kernel in the subsampled domain causes not only spatial degradation, but also spectral degradation at the same time. Single-channel deconvolution does not consider the characteristics of CFA, and as a result, there is a limit to solving complex problems. In this study, it is shown that the crosstalk problem can be defined as a multi-channel degradation model. In addition, a method to solve the defined modeling is presented, and an effective objective function and optimization method are proposed. As can be seen from the experimental results, single-channel deconvolution can solve the blurring, but cannot solve the desaturation regardless of various methods. The color correction method is able to improve color fidelity, but the problems regarding blur and noise still remained. On the other hand, the proposed method is able to improve both spatial resolution and color fidelity without noise amplification, and it has been demonstrated that a synergistic effect with the state-of-the-art demosaicing algorithm can be achieved.
One of the important points is that the regularization parameter in the proposed method controls the smoothness. Although we have suggested the effective range of regularization parameters empirically, they will be important variables affecting the results. Overall, the main concept of the proposed multi-channel deconvolution for crosstalk correction is to overcome spatial and spectral degradation simultaneously, and its validity has been proven. The proposed multi-channel deconvolution approach can provide new insights into solving the crosstalk problem that occurs in the image acquisition system.

6. Conclusions

A crosstalk correction method based on L p -regularized multi-channel deconvolution was herein presented. The crosstalk problem with complex spatial and spectral degradation is formulated as a multi-channel degradation model, and an objective function with a hyper-Laplacian prior is designed. The proposed method based on multi-channel deconvolution improves color fidelity and produces sharp edges without obvious artifacts, such as noise amplification. Furthermore, an efficient solver minimizes the objective function for crosstalk correction. The cost function gradually converges toward a minimum value for certain variables. The proposed method was compared with conventional crosstalk correction methods using synthetic datasets according to various crosstalk and noise levels. Experimental results demonstrated that the proposed method outperformed the conventional methods in terms of the CPSNR and SSIM values. However, in terms of speed, the weakness of the proposed method is that the computational complexity is high to perform real-time processing in camera imaging system. In further research, we will implement the proposed method with multiple GPUs for a real-time performance. We believe that the proposed method will be applied to the application of imaging systems in many fields by overcoming the physical limitations of the image sensor.

Author Contributions

Conceptualization, J.K.; methodology, J.K.; software, J.K.; validation, J.K., K.J. and M.G.K.; formal analysis, J.K.; investigation, J.K. and K.J.; resources, J.K. and K.J.; data curation, J.K. and K.J.; writing—original draft preparation, J.K.; writing—review and editing, J.K. and K.J.; visualization, J.K.; supervision, M.G.K.; project administration, M.G.K.; funding acquisition, M.G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. 2022R1A2C200289711).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bayer, B.E. Color Imaging Array. U.S. Patent 3,971,065, 20 July 1976. [Google Scholar]
  2. Hirakawa, K.; Wolfe, P.J. Spatio-Spectral Color Filter Array Design for Optimal Image Recovery. IEEE Trans. Image Process. 2008, 17, 1876–1890. [Google Scholar] [CrossRef]
  3. Condat, L. A New Color Filter Array With Optimal Properties for Noiseless and Noisy Color Image Acquisition. IEEE Trans. Image Process. 2011, 20, 2200–2210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Kimmel, R. Demosaicing: Image reconstruction from color CCD samples. IEEE Trans. Image Process. 1999, 8, 1221–1228. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Pekkucuksen, I.; Altunbasak, Y. Multiscale Gradients-Based Color Filter Array Interpolation. IEEE Trans. Image Process. 2013, 22, 157–165. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, C.; Li, Y.; Wang, J.; Hao, P. Universal Demosaicking of Color Filter Arrays. IEEE Trans. Image Process. 2016, 25, 5173–5186. [Google Scholar] [CrossRef] [PubMed]
  7. Kiku, D.; Monno, Y.; Tanaka, M.; Okutomi, M. Beyond Color Difference: Residual Interpolation for Color Image Demosaicking. IEEE Trans. Image Process. 2016, 25, 1288–1300. [Google Scholar] [CrossRef] [PubMed]
  8. Monno, Y.; Kiku, D.; Tanaka, M.; Okutomi, M. Adaptive residual interpolation for color image demosaicking. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 3861–3865. [Google Scholar] [CrossRef]
  9. Farrell, J.; Xiao, F.; Kavusi, S. Resolution and light sensitivity tradeoff with pixel size. In Proceedings of the Digital Photography II, San Jose, CA, USA, 15–19 January 2006; Volume 6069, p. 60690N. [Google Scholar]
  10. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  11. Khashabi, D.; Nowozin, S.; Jancsary, J.; Fitzgibbon, A.W. Joint Demosaicing and Denoising via Learned Nonparametric Random Fields. IEEE Trans. Image Process. 2014, 23, 4968–4981. [Google Scholar] [CrossRef] [PubMed]
  12. Anzagira, L.; Fossum, E.R. Color filter array patterns for small-pixel image sensors with substantial cross talk. J. Opt. Soc. Am. A 2015, 32, 28–34. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Hirakawa, K. Cross-talk explained. In Proceedings of the 2008 15th IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008; pp. 677–680. [Google Scholar] [CrossRef]
  14. Tian, H.; Sun, Q.; Li, J.; Domine, A. Crosstalk challenges cmos sensor design. Laser Focus World 2005, 41, 119–123. [Google Scholar]
  15. Lee, J.K.; Kim, A.; Kang, D.W.; Lee, B.Y. Efficiency enhancement in a backside illuminated 1.12 μm pixel CMOS image sensor via parabolic color filters. Opt. Express 2016, 24, 16027–16036. [Google Scholar] [CrossRef] [PubMed]
  16. Wu, X.; Zhang, X. Joint Color Decrosstalk and Demosaicking for CFA Cameras. IEEE Trans. Image Process. 2010, 19, 3181–3189. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Zibulevsky, M.; Elad, M. L1-L2 Optimization in Signal and Image Processing. IEEE Signal Process. Mag. 2010, 27, 76–88. [Google Scholar] [CrossRef]
  18. Tikhonov, A.N. On the stability of inverse problems. Dokl. Akad. Nauk SSSR 1943, 39, 195–198. [Google Scholar]
  19. Rudin, L.I.; Osher, S. Total variation based image restoration with free local constraints. In Proceedings of the Proceedings of 1st International Conference on Image Processing, Austin, TX, USA, 13–16 November 1994; Volume 1, pp. 31–35. [Google Scholar] [CrossRef]
  20. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  21. Krishnan, D.; Fergus, R. Fast Image Deconvolution using Hyper-Laplacian Priors. In Advances in Neural Information Processing Systems 22; Bengio, Y., Schuurmans, D., Lafferty, J.D., Williams, C.K.I., Culotta, A., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2009; pp. 1033–1041. [Google Scholar]
  22. Fergus, R.; Singh, B.; Hertzmann, A.; Roweis, S.T.; Freeman, W.T. Removing Camera Shake from a Single Photograph. In Proceedings of the ACM SIGGRAPH 2006 Papers; Association for Computing Machinery: New York, NY, USA, 2006; pp. 787–794. [Google Scholar] [CrossRef]
  23. Zhang, L.; Wu, X.; Buades, A.; Li, X. Color demosaicking by local directional interpolation and nonlocal adaptive thresholding. J. Electron. Imaging 2011, 20, 1–17. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, Y.; Yang, J.; Yin, W.; Zhang, Y. A New Alternating Minimization Algorithm for Total Variation Image Reconstruction. SIAM J. Imaging Sci. 2008, 1, 248–272. [Google Scholar] [CrossRef]
  25. Zhou, W.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  26. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi Morel, M.L. Low-Complexity Single-Image Super-Resolution based on Nonnegative Neighbor Embedding. In Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012; BMVA Press: Surrey, UK, 2012; pp. 135.1–135.10. [Google Scholar] [CrossRef] [Green Version]
  27. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces, Avignon, France, 24–30 June 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 711–730. [Google Scholar]
  28. Richardson, W.H. Bayesian-Based Iterative Method of Image Restoration. J. Opt. Soc. Am. 1972, 62, 55–59. [Google Scholar] [CrossRef]
  29. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning Deep CNN Denoiser Prior for Image Restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Zhang, J.; Pan, J.; Lai, W.S.; Lau, R.W.H.; Yang, M.H. Learning Fully Convolutional Networks for Iterative Non-Blind Deconvolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  31. Gong, D.; Zhang, Z.; Shi, Q.; van den Hengel, A.; Shen, C.; Zhang, Y. Learning Deep Gradient Descent Optimization for Image Deconvolution. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5468–5482. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Problems of imaging under the crosstalk condition in CFA: (a) Original image; (b) degraded image by Gaussian kernel with σ g = 0.45 and noise with σ n = 0.01 ; (c) degraded image by Gaussian kernel with σ g = 0.6 and noise with σ n = 0.01 . As the standard deviation σ g of the Gaussian kernel increases, desaturation and blurring intensify owing to interference with neighboring channels.
Figure 1. Problems of imaging under the crosstalk condition in CFA: (a) Original image; (b) degraded image by Gaussian kernel with σ g = 0.45 and noise with σ n = 0.01 ; (c) degraded image by Gaussian kernel with σ g = 0.6 and noise with σ n = 0.01 . As the standard deviation σ g of the Gaussian kernel increases, desaturation and blurring intensify owing to interference with neighboring channels.
Sensors 22 04285 g001
Figure 2. Bayer CFA and characteristics of crosstalk phenomenon: (a) Bayer CFA structure. The Bayer CFA consists of three filters for respective R, G, and B channels. (b) Schematic image of crosstalk inside the imaging sensor. (c) Example of the spectral sensitivity of the camera imaging system in a crosstalk-free condition (solid lines) and in a crosstalk condition (dashed lines). The spectral sensitivity shift is caused by crosstalk inside the imaging sensor.
Figure 2. Bayer CFA and characteristics of crosstalk phenomenon: (a) Bayer CFA structure. The Bayer CFA consists of three filters for respective R, G, and B channels. (b) Schematic image of crosstalk inside the imaging sensor. (c) Example of the spectral sensitivity of the camera imaging system in a crosstalk-free condition (solid lines) and in a crosstalk condition (dashed lines). The spectral sensitivity shift is caused by crosstalk inside the imaging sensor.
Sensors 22 04285 g002
Figure 3. Crosstalk kernels at G location: (a) Gaussian kernel of σ g = 0.45 and (b) Gaussian kernel of σ g = 0.60 . Owing to the crosstalk kernel, spectral and spatial degradation simultaneously occur.
Figure 3. Crosstalk kernels at G location: (a) Gaussian kernel of σ g = 0.45 and (b) Gaussian kernel of σ g = 0.60 . Owing to the crosstalk kernel, spectral and spatial degradation simultaneously occur.
Sensors 22 04285 g003
Figure 4. Example illustrations of single-channel and multi-channel degradation models: (a) Relationship of B in the single-channel degradation model (left) and H in the multi-channel degradation model (right). (b) Comparison of the single-channel degradation model (top) and the multi-channel degradation model (bottom). ARI demosaicing is applied equally to degraded images. The same degraded results are produced by different formulations of the crosstalk phenomenon, but different results are restored.
Figure 4. Example illustrations of single-channel and multi-channel degradation models: (a) Relationship of B in the single-channel degradation model (left) and H in the multi-channel degradation model (right). (b) Comparison of the single-channel degradation model (top) and the multi-channel degradation model (bottom). ARI demosaicing is applied equally to degraded images. The same degraded results are produced by different formulations of the crosstalk phenomenon, but different results are restored.
Sensors 22 04285 g004
Figure 5. Probability distributions from Kodak and McMaster datasets: (a) Distribution of image gradients and (b) distribution of color difference gradients. The empirical distributions of image gradients and color difference gradients follow the probability distributions of p = 0.66 and 0.62 , respectively.
Figure 5. Probability distributions from Kodak and McMaster datasets: (a) Distribution of image gradients and (b) distribution of color difference gradients. The empirical distributions of image gradients and color difference gradients follow the probability distributions of p = 0.66 and 0.62 , respectively.
Sensors 22 04285 g005
Figure 6. Visual comparison of restored images, enlarged parts, and difference maps from crosstalk degradation 1 (crosstalk kernel with σ g = 0.45 and noise with σ n = 0.01 ): (a) Ground truth, (b) CM1, (c) CM2, (d) CM3, (e) CM4, (f) CM6, and (g) PM. (hn) Same methods as in (ag). The first and second rows present the results of Bike and Parrot in the Kodak dataset, respectively.
Figure 6. Visual comparison of restored images, enlarged parts, and difference maps from crosstalk degradation 1 (crosstalk kernel with σ g = 0.45 and noise with σ n = 0.01 ): (a) Ground truth, (b) CM1, (c) CM2, (d) CM3, (e) CM4, (f) CM6, and (g) PM. (hn) Same methods as in (ag). The first and second rows present the results of Bike and Parrot in the Kodak dataset, respectively.
Sensors 22 04285 g006
Figure 7. Visual comparison of restored images, enlarged parts, and difference maps from crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) Ground truth, (b) CM1, (c) CM2, (d) CM3, (e) CM4, (f) CM6, and (g) PM. (hn) Same methods as in (ag). The first and second rows present the results of Bike and Parrot in the Kodak dataset, respectively.
Figure 7. Visual comparison of restored images, enlarged parts, and difference maps from crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) Ground truth, (b) CM1, (c) CM2, (d) CM3, (e) CM4, (f) CM6, and (g) PM. (hn) Same methods as in (ag). The first and second rows present the results of Bike and Parrot in the Kodak dataset, respectively.
Sensors 22 04285 g007
Figure 8. Visual comparison of restored images, enlarged parts, and difference maps from crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) Ground truth, (b) CM1, (c) CM2, (d) CM3, (e) CM4, (f) CM6, and (g) PM. (hn) Same methods as in (ag). The first and second rows present the results of mcm1 and mcm4 in the McMaster dataset, respectively.
Figure 8. Visual comparison of restored images, enlarged parts, and difference maps from crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) Ground truth, (b) CM1, (c) CM2, (d) CM3, (e) CM4, (f) CM6, and (g) PM. (hn) Same methods as in (ag). The first and second rows present the results of mcm1 and mcm4 in the McMaster dataset, respectively.
Sensors 22 04285 g008
Figure 9. Influence of regularization parameter λ 1 for test image House for crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) λ 1 = 0.01 , (b) λ 1 = 0.03 , (c) λ 1 = 0.05 , and (d) λ 1 = 0.1 .
Figure 9. Influence of regularization parameter λ 1 for test image House for crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) λ 1 = 0.01 , (b) λ 1 = 0.03 , (c) λ 1 = 0.05 , and (d) λ 1 = 0.1 .
Sensors 22 04285 g009
Figure 10. Influence of regularization parameter λ 2 for test image Parrot for crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) λ 2 = 0.0001 , (b) λ 2 = 0.001 , (c) λ 2 = 0.01 , and (d) λ 2 = 0.1 .
Figure 10. Influence of regularization parameter λ 2 for test image Parrot for crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ): (a) λ 2 = 0.0001 , (b) λ 2 = 0.001 , (c) λ 2 = 0.01 , and (d) λ 2 = 0.1 .
Sensors 22 04285 g010
Figure 11. Objective performance comparison for test image Parrot according to regularization parameters λ 1 and λ 2 : (a) CPSNR values versus values of λ 1 ; (b) SSIM values versus values of λ 1 ; (c) CPSNR values versus values of λ 2 ; and (d) SSIM values versus values of λ 2 .
Figure 11. Objective performance comparison for test image Parrot according to regularization parameters λ 1 and λ 2 : (a) CPSNR values versus values of λ 1 ; (b) SSIM values versus values of λ 1 ; (c) CPSNR values versus values of λ 2 ; and (d) SSIM values versus values of λ 2 .
Sensors 22 04285 g011
Figure 12. Convergence for test image Parrot: (a) Value of cost function versus iteration number of various initial conditions for crosstalk degradation 1 (crosstalk kernel with σ g = 0.45 and noise with σ n = 0.01 ); (b) value of cost function versus iteration number of various degradation conditions; and (c) visualization of residual image versus iteration number at crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ).
Figure 12. Convergence for test image Parrot: (a) Value of cost function versus iteration number of various initial conditions for crosstalk degradation 1 (crosstalk kernel with σ g = 0.45 and noise with σ n = 0.01 ); (b) value of cost function versus iteration number of various degradation conditions; and (c) visualization of residual image versus iteration number at crosstalk degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ).
Sensors 22 04285 g012
Table 1. Objective Performance Comparison of the Kodak 24 and McMaster 18 Images in CPSNR (dB) and SSIM for Crosstalk Degradation 1 (crosstalk kernel with σ g = 0.45 and noise with σ n = 0.01 ).
Table 1. Objective Performance Comparison of the Kodak 24 and McMaster 18 Images in CPSNR (dB) and SSIM for Crosstalk Degradation 1 (crosstalk kernel with σ g = 0.45 and noise with σ n = 0.01 ).
Kodak dataset
CPSNRSSIM
No.CM1CM2CM3CM4CM5CM6CM7PMCM1CM2CM3CM4CM5CM6CM7PM
130.5935.1535.2033.7226.4729.5624.5135.670.95270.95530.95680.94940.84970.94830.92230.9628
223.8934.9835.4435.9129.2823.9724.4536.850.89920.88870.89180.90680.84280.78320.84290.9278
327.4736.6336.9437.1327.7525.1925.2238.710.91650.90340.91200.91740.87310.89910.87970.9533
427.1735.4535.7536.4325.8224.2623.0237.150.92590.91120.91530.92650.83550.86530.87450.9411
530.1335.0235.1133.6827.3428.3625.1235.420.95490.95850.96030.95810.89370.94670.92830.9679
630.3735.4935.7035.0325.6227.8524.7236.390.94840.93740.94130.94250.84440.94460.91140.9572
730.7636.6636.8837.1127.1628.5925.3338.720.94170.92470.93060.93730.92260.92920.90730.9658
830.5533.2133.3831.6127.5830.7324.1933.610.95430.95200.95500.95070.89510.95850.92570.9586
933.2036.1936.4736.3231.0130.1424.7838.010.93180.88720.90280.91060.91800.93640.87960.9442
1034.3336.4036.5336.4429.0929.6424.1437.900.93790.90930.91770.92620.91400.93600.88800.9480
1131.5835.3735.6535.2026.6229.0425.2036.440.94060.92550.93200.93230.82410.94420.90140.9482
1230.7136.8636.7237.2924.4627.0224.7438.600.93450.90800.91330.92380.83880.91880.88240.9459
1329.3333.3833.4131.4525.8228.2224.6633.950.94810.95850.95980.94730.81350.95130.92780.9646
1427.9234.4534.7134.3927.8027.1224.9735.200.93910.94000.94410.94430.85380.93610.90850.9518
1528.2134.8735.2035.6026.0825.0024.2736.810.91940.89130.89680.91190.84850.87890.86990.9381
1634.4836.5836.9336.9428.6132.1925.0837.980.94350.92340.93070.93340.88220.94810.90370.9531
1734.8636.1636.5436.2130.7532.6223.2037.630.94190.91860.92860.93340.90250.94610.89050.9539
1830.1834.3134.4533.6227.9929.0423.4434.780.93480.92920.93550.93380.84500.92680.89770.9434
1931.8035.6535.8334.8828.3730.2523.0236.440.93820.92380.93060.93190.85260.93700.89050.9476
2032.1736.6336.7836.5627.7931.1021.7838.190.93670.91530.92180.92780.88610.93650.87300.9521
2131.9835.4135.5934.8228.5732.1224.1236.440.93820.90740.91330.92150.89210.92960.89260.9492
2229.9534.9535.0935.1126.1028.1324.6035.920.93050.91640.92290.92750.81060.92150.89460.9399
2325.7436.8836.9037.3826.7623.0624.9539.360.91350.89650.90440.91530.89020.87380.87060.9522
2431.3433.7133.7932.9028.1329.7723.4134.300.94880.94110.94510.94400.89370.95250.91770.9592
Avg.30.3635.4335.6335.2427.5428.4624.2936.690.93630.92180.92760.93140.86760.92290.89500.9511
McMaster dataset
CPSNRSSIM
No.CM1CM2CM3CM4CM5CM6CM7PMCM1CM2CM3CM4CM5CM6CM7PM
122.3329.2129.1828.9626.3120.1220.8529.120.84600.89680.89670.89200.81950.80170.83290.8992
227.0133.4233.3733.3527.4325.3622.1633.880.87540.89870.89940.90260.79000.81900.82910.9196
328.4033.0733.0932.3627.7225.7322.1733.410.92450.93480.93780.93770.89100.90250.87640.9549
430.8534.7134.9034.6430.3427.1720.3536.210.95160.93720.94250.94740.95040.95830.89930.9713
527.5533.5333.5333.5327.5326.6122.1933.880.90770.91880.92080.92430.85680.87730.87520.9370
628.0835.7735.7135.9328.8327.2523.3136.500.90450.92630.92790.93200.85000.86500.88060.9447
730.0335.7235.9735.5230.0928.7822.9336.110.92540.93130.93420.93310.83480.90790.86100.9458
833.8035.8535.9835.3632.2131.4422.2736.770.91320.91040.91460.91220.88910.90210.76970.9449
926.0634.7734.7334.8728.0924.8921.6835.800.88410.90270.90360.91320.84310.81000.84240.9409
1024.0135.7435.5835.8727.7922.0622.5536.660.85820.91850.91810.92180.81370.76440.83050.9464
1124.4536.2836.0636.4327.1422.8322.3437.360.81500.91180.90750.91610.73360.70520.78100.9426
1225.6336.1236.0936.0330.7727.2722.3637.690.89380.89730.89880.90940.86240.83670.84710.9455
1327.7936.1636.1936.8833.0431.3222.6038.790.90150.86390.86750.88950.88380.85920.84350.9323
1426.2635.6635.5536.0427.5125.0722.3637.040.87400.89060.89240.90070.80840.80050.82480.9296
1524.1635.8535.7136.0428.6122.1222.4137.130.85090.89040.89020.89710.74580.70690.78380.9280
1621.7433.6633.4833.1127.9618.2321.6733.170.80980.93460.93320.93260.81410.72830.82650.9369
1722.4232.7832.6532.7126.9019.0721.5932.920.81370.92180.92020.92390.78960.63460.81470.9326
1825.4633.9933.9134.0027.5622.8121.5434.700.91030.92480.92570.92930.85810.86280.85930.9419
Avg.26.4534.5734.5434.5328.6624.9022.0735.400.88110.91170.91280.91750.83520.81900.83770.9386
Table 2. Objective Performance Comparison of the Kodak 24 and McMaster 18 Images in CPSNR (dB) and SSIM for Crosstalk Degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ).
Table 2. Objective Performance Comparison of the Kodak 24 and McMaster 18 Images in CPSNR (dB) and SSIM for Crosstalk Degradation 2 (crosstalk kernel with σ g = 0.6 and noise with σ n = 0.04 ).
Kodak dataset
CPSNRSSIM
No.CM1CM2CM3CM4CM5CM6CM7PMCM1CM2CM3CM4CM5CM6CM7PM
123.4520.4022.9620.1422.8023.1419.5425.910.72480.44250.63440.49940.62680.79460.46420.7379
217.0621.1419.9818.2221.6416.5118.5429.960.53100.23560.26620.19250.33760.57780.23900.7700
320.8219.6122.3917.5623.3720.4119.3030.910.57600.19690.39850.19170.63790.77230.23640.8465
420.6120.1621.6618.5821.3319.6818.5229.880.60190.24480.41590.25180.42170.73730.27060.7980
523.2321.0923.0120.6023.5723.2420.3625.960.74260.53070.66060.53540.73300.82750.55390.8009
623.5419.9422.8719.0623.1523.1718.8927.070.67320.32320.55410.35990.62410.80060.36220.7546
723.5920.2623.5018.4924.0823.2819.9129.910.65420.31970.50780.32590.71940.83830.35750.8836
823.7920.6323.2020.0524.9324.5819.7924.890.75870.54140.69010.56290.79240.84870.57090.7847
925.6619.8324.4617.2128.0825.8019.5530.530.62070.23490.46720.23560.83670.82850.28100.8570
1026.6319.8524.7018.1127.6926.7519.4630.250.62990.23160.48370.24920.82250.82650.28680.8423
1124.5820.3423.7518.9624.4724.8119.9128.090.64350.30450.52640.31330.70580.79340.35070.7661
1223.9619.5323.1617.6323.0023.3819.1230.740.59950.18220.42790.19200.54840.79950.24050.8114
1322.8220.2322.5820.4220.9122.9319.4124.170.73060.48030.66950.53780.56560.78280.50300.7413
1421.0820.4822.7820.1323.5521.1219.8527.170.67430.37280.56260.40580.64920.76270.40630.7613
1521.8220.7222.4618.6123.1821.3718.6930.040.59250.25470.43250.23100.67550.77840.27410.8176
1626.4119.8824.5618.0926.0327.0119.7429.200.62890.21360.48850.24700.73830.80920.27940.7712
1726.8121.2025.0819.3327.4927.4919.6929.780.64450.30880.50960.29220.81430.82270.33600.8388
1823.3020.8323.4319.8423.1723.1419.5526.950.67130.37870.56390.39330.66480.77470.39630.7646
1924.2720.0023.8218.0623.9123.8518.9927.830.65240.29540.52070.31270.69860.80160.33480.7913
2024.9621.1524.8418.6924.9824.7616.5930.660.66530.27530.52210.24780.73800.83180.27080.8546
2124.5420.1623.7017.7623.6624.1519.4827.940.65790.32530.51180.33150.59430.81700.36060.8237
2223.1519.8623.1318.7122.1022.6519.3328.380.63550.25390.49390.29600.52150.76780.30370.7641
2319.2819.4121.1217.4722.0718.9618.8430.940.57420.20650.37950.18640.57090.75870.24280.8714
2424.7920.3623.9119.6124.4225.2419.0626.500.69240.35850.58610.39630.73590.82280.39590.7950
Avg.23.3420.2923.2118.8123.9023.2319.2628.490.64900.31300.51140.32450.65720.79060.34660.8020
McMaster dataset
CPSNRSSIM
No.CM1CM2CM3CM4CM5CM6CM7PMCM1CM2CM3CM4CM5CM6CM7PM
116.5419.4619.1919.7420.5916.1517.0924.080.59660.48670.53710.45830.57320.62650.46500.7191
220.7321.5822.0719.9522.8720.1019.1828.060.60100.42090.50220.36320.58800.68160.39570.7877
322.2920.4922.3219.8722.8522.2018.9825.750.70100.51530.62050.51290.76280.80700.51750.8282
424.5320.4423.3220.0025.8025.4917.5027.940.71790.44640.61710.47520.91380.88330.48150.9034
521.0620.6422.5019.7822.9420.5218.7827.340.63600.37440.51870.38070.60860.73140.39430.8081
621.2120.8222.6419.5923.6820.6519.3527.600.60420.35920.48700.34280.60100.70540.37050.7908
723.1921.0623.7719.9926.7223.3719.3428.260.65920.37500.55590.38620.74650.76560.37860.7660
826.8623.8625.5219.7829.0326.9620.1929.730.63510.48120.52680.30500.84590.79720.36220.8207
919.4920.9121.7418.7622.2918.8218.4328.670.58370.36750.44950.32570.51740.69150.35590.8365
1017.5820.8420.3719.7121.0817.1518.2828.890.55410.38840.43670.33310.47840.63150.35290.8135
1118.0021.1520.6919.4422.5417.5318.4729.910.49690.35070.38090.28060.48370.55860.30510.7948
1218.6520.6221.4117.7323.6717.9218.1629.780.57280.30180.37510.26000.49560.66000.29400.8558
1320.5520.0222.0916.6724.2219.2218.4432.290.56530.19680.30620.16310.46070.69440.22160.8820
1419.6720.9821.7918.8323.6019.0919.0130.980.55020.30410.39590.23480.58720.66830.28500.8344
1517.8621.1420.4018.8721.9417.1918.1030.980.50470.33220.36830.22470.43920.58590.27770.8161
1615.8718.0618.5021.5320.1015.4917.2326.500.58250.46620.52220.49820.52670.61430.45850.7429
1716.6019.8018.7820.1620.8316.0817.4424.380.50130.43000.45680.40320.50880.53890.40690.7004
1819.1620.6420.9120.3121.4918.7618.0828.170.65600.44170.52720.43330.58500.73350.43820.7988
Avg.19.9920.6921.5619.4823.1219.5918.4528.300.59550.39100.47690.35450.59570.68750.37560.8055
Table 3. Objective Performance Comparison in Average CPSNR (dB) and Average SSIM.
Table 3. Objective Performance Comparison in Average CPSNR (dB) and Average SSIM.
Crosstalk Degradation 1 ( σ g = 0.45 and σ n = 0.01 )
KodakMcMasterSet5Set14
MethodCPSNRSSIMCPSNRSSIMCPSNRSSIMCPSNRSSIM
CM130.360.936326.450.881127.030.906027.090.8974
CM235.430.921834.570.911734.950.916433.170.8972
CM335.630.927634.540.912834.920.918633.210.9001
CM435.240.931434.530.917535.000.923333.060.9034
CM527.540.867628.660.835227.660.857527.510.8534
CM628.460.922924.900.819025.940.880124.610.8654
CM724.290.895022.070.837719.810.835319.900.7395
PM36.690.951135.400.938635.420.929233.590.9098
Crosstalk Degradation 2 ( σ g = 0.6 and σ n = 0.04 )
KodakMcMasterSet5Set14
MethodCPSNRSSIMCPSNRSSIMCPSNRSSIMCPSNRSSIM
CM123.340.649019.990.595520.480.654120.660.6597
CM220.290.313020.690.391020.690.430220.140.4139
CM323.210.511421.560.476921.800.526021.500.5388
CM418.810.324519.480.354519.880.410519.500.4171
CM523.900.657223.120.595722.300.606422.110.5908
CM623.230.790619.590.687519.770.738320.120.7395
CM719.260.346618.450.375617.430.569318.040.4056
PM28.490.802028.300.805528.670.826127.340.7879
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, J.; Jeong, K.; Kang, M.G. Crosstalk Correction for Color Filter Array Image Sensors Based on Lp-Regularized Multi-Channel Deconvolution. Sensors 2022, 22, 4285. https://doi.org/10.3390/s22114285

AMA Style

Kim J, Jeong K, Kang MG. Crosstalk Correction for Color Filter Array Image Sensors Based on Lp-Regularized Multi-Channel Deconvolution. Sensors. 2022; 22(11):4285. https://doi.org/10.3390/s22114285

Chicago/Turabian Style

Kim, Jonghyun, Kyeonghoon Jeong, and Moon Gi Kang. 2022. "Crosstalk Correction for Color Filter Array Image Sensors Based on Lp-Regularized Multi-Channel Deconvolution" Sensors 22, no. 11: 4285. https://doi.org/10.3390/s22114285

APA Style

Kim, J., Jeong, K., & Kang, M. G. (2022). Crosstalk Correction for Color Filter Array Image Sensors Based on Lp-Regularized Multi-Channel Deconvolution. Sensors, 22(11), 4285. https://doi.org/10.3390/s22114285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop