Next Article in Journal
Rapid Identification of Beached Marine Plastics Pellets Using Laser-Induced Breakdown Spectroscopy: A Promising Tool for the Quantification of Coastal Pollution
Next Article in Special Issue
The Use of Terrestrial and Maritime Autonomous Vehicles in Nonintrusive Object Inspection
Previous Article in Journal
Research on Performance of Cooperative FSO Communication System Based on Hierarchical Modulation and Physical Layer Network Code
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Noise-Adaptive Non-Blind Image Deblurring

GM Technical Center Israel—R&D Lab, 13 Arie Shenkar St., Herzliya 4672513, Israel
Sensors 2022, 22(18), 6923; https://doi.org/10.3390/s22186923
Submission received: 10 August 2022 / Revised: 6 September 2022 / Accepted: 9 September 2022 / Published: 13 September 2022
(This article belongs to the Special Issue Artificial Intelligence in Automotive Technology)

Abstract

:
This work addresses the problem of non-blind image deblurring for arbitrary input noise. The problem arises in the context of sensors with strong chromatic aberrations, as well as in standard cameras, in low-light and high-speed scenarios. A short description of two common classical approaches to regularized image deconvolution is provided, and common issues arising in this context are described. It is shown how a pre-deconvolved deep neural network (DNN) based image enhancement can be improved by joint optimization of regularization parameters and network weights. Furthermore, a two-step approach to deblurring based on two DNNs is proposed, with the first network estimating deconvolution regularization parameters, and the second one performing image enhancement and residual artifact removal. For the first network, a novel RegParamNet architecture is introduced and its performance is examined for both direct and indirect regularization parameter estimation. The system is shown to operate well for input noise in a three orders of magnitude range (0.01–10.0) and a wide spectrum of 1D or 2D Gaussian blur kernels, well outside the scope of most previously explored image blur and noise degrees. The proposed method is found to significantly outperform several leading state-of-the-art approaches.

1. Introduction

Video cameras are the most common sensing modality in modern vehicles. The number of cameras per vehicle is constantly growing, both in the retail sector, e.g., in Advanced Driver Assistance Systems (ADAS), and in the Autonomous Vehicle (AV) domain. There is a constant demand to reduce the cost of imaging systems without compromising the quality.
Raw images acquired by digital image sensors pass through a rather long image signal processing (ISP) chain [1] before they are handed over to a perception system for object detection, recognition, tracking, or a variety of other generic or domain-specific tasks [2]. Success of perception systems critically depends on image quality [3], and is therefore constantly pushing the requirements for resolution, frame rate, dynamic range etc. For the last decade or so, we have witnessed a staggering improvement in image quality of mobile phone and automotive-grade cameras, often comparable to image quality of professional-grade cameras. This improvement came mainly from the direction of ISP modules, driven by rapid progress in image processing algorithms, which in turn was possible due to a significant increase in the available processing hardware. Many such algorithms were invented to compensate for artifacts introduced by relatively cheap optics (e.g., aberrations and distortions) or small pixel size (e.g., noise and cross-talk).
In this work, the problem of image blur is addressed. This is a rather old and cross-disciplinary topic—people have been dealing with blurry images more or less since the invention of photography, in most imaging domains and at all scales—from astronomy [4] to microscopy [5]. Typically, image blur is motion-induced [6] or caused by optical aberrations [7,8]. For instance, maybe the most well-known imaging system optimization challenge deals with choosing a correct value for the exposure time. If it is too short, the image appears very noisy, if it is too long—camera motion or intrinsic motions in the scene smear the image and it becomes too blurry. Thus, long-range cameras designed for highway usage and high-speed driving tend to produce noisy images in low-light conditions, whereas short-range surround view cameras are more lenient when it comes to exposure times. In any case, to achieve less noise means using more expensive sensors or longer exposure times, which means image blur.
Until the introduction of digital imaging more than half a century ago, probably the only way to deal with image blur was to avoid it by using high-grade optics and sensitive photographic film. Digital imaging allows us to build image formation models, to isolate degradation stages, and to formulate image restoration as an inverse numerical problem [9]. As the following sections show, such formulation often paves way to using well-known techniques from other, more established fields. The first publications on digital image deblurring appeared in the late 60’s and early 70’s [10,11,12]. Since then, image deblurring has flourished into a rather wide field that can be roughly divided into two main sub-fields. The first kind of deblurring problems are referred to as “non-blind deblurring”. In this class of problems it is assumed that all components of image formation model are known. In particular, the blur kernel (often referred to as the point-spread function, or PSF) is assumed to be either directly measured in some calibration procedure or inferred from auxiliary measurements, e.g., from motion sensors readings [13]. The second class of problems is called “blind deblurring”, which means that the PSF is not known. As often happens, there is also a rather dense spectrum of problems that fall somewhere in between the two classes. For instance, the exact shape of the PSF may not be known, but assumptions can be made regarding its origin, extent, regularity etc. To appreciate the vast progress in the field up until 2014, the reader is referred to a comprehensive review article by Wang and Tao [14].
Naturally, blind deblurring problems are more challenging than non-blind deblurring ones. For many approaches, the latter can be considered as a central block in a generic solution flow for blind deblurring: find a deblurred image given a PSF, estimate the error, refine the PSF etc. Still, as the next sections show, even if one is fortunate enough to know the PSF, finding a reliable solution is a rather non-trivial task. In particular, the solution should be properly regularized to be of any use, which requires inference of regularization parameters. The latter, in turn, are strongly influenced by the level of noise in the blurred image as well as by intricate texture statistics of the image. Often, regularized solutions exhibit common artifacts, such as boundary artifacts or Gibbs phenomena that inevitably accompany most solutions to ill-posed inverse problems. Much effort has been invested in fighting these artifacts, either as a post-processing step or by devising specialized boundary conditions and inverse operators.
As with almost every field in image processing and computer vision, application of deep learning techniques to image deblurring during the last decade, generated a great number of impressive results [15,16]. Most researchers are interested in blind deblurring since this is where most challenges are. Although there was quite impressive progress on that front [17,18,19], existing deep neural networks (DNNs) are able to learn PSFs of a rather limited extent, so that the problem of blind deblurring remains largely unsolved. Non-blind deblurring has benefited from deep learning as well. For instance, there has been much activity around designing suitable architectures and ways to incorporate known PSF kernels into them [20,21,22]. Another approach proposed by Xu et al. [23] is to use classical deconvolution followed by a DNN trained to remove residual deconvolution artifacts. This approach was further explored by Wang and Tao [24], who proposed using very deep architectures to predict residuals between a pre-deconvolved image and the corresponding sharp image. It is argued that this approach is noise-blind in the sense that it is able to handle different blur kernels and various noise levels. However, in both cases it is assumed that the pre-deconvolved image actually exists, which implies knowledge of suitable regularization parameters, or at least, knowledge of input noise [9].
This work addresses the problem of non-blind image deblurring for almost arbitrary levels of input noise. The two-step approach is generalized to the case of arbitrary noise in a three orders of magnitude range (10 2 –10.0) and arbitrary 1D or 2D blur kernels.
This paper is organized as follows. First, a short review of classical approaches to regularized image deconvolution is provided, and common issues arising in this context are described. Then it is demonstrated how a two-step DNN based deblurring works, and how it can be improved by treating regularization parameters as trainable neural network weights. Next, the two-step approach is generalized by introducing RegParamNet, a 1D convolutional network that estimates optimal regularization parameters for a given blurred and noisy input. Finally, the results of the generalized approach are analyzed and possible extensions are discussed.

2. Regularized Deconvolution

2.1. Tikhonov Regularization

In this work it is assumed that the blurred image J B is related to the original image J through a linear convolution with additive noise:
J B = K * J + Ξ ,
where K is the blur kernel and Ξ is a random Gaussian noise matrix. It is also assumed that K is known. To avoid noise amplification, Tikhonov [25] suggested to reformulate (1) as a regularized least-squares (LS) problem, such that
J λ ( L S ) = argmin K * J J B 2 + λ 2 L * J 2
Here, L is a linear operator (usually, an identity operator or a derivative) that is chosen in accordance with the property of the solution that needs to be regularized. For 1D blur, the convolution operation can be expressed as multiplication of the image J by a blur matrix. For instance, the solution for horizontal blur is
J λ ( L S ) = J B K T K T K + λ 2 L T L 1
Throughout this work, the identity operator L = I will be used. In this way, the λ 2 -term controls the magnitude of the solution and prevents it from diverging. Thus (in what follows, the superscript ( L S ) will be omitted),
J λ = J B K T K K T + λ 2 I 1 = J B VS S 2 + λ 2 I 1 U T
where
K = USV T
is the SVD decomposition of the blur matrix. Tikhonov regularization is a very efficient approach for 1D and 2D separable kernels. Non-separable 2D blurring kernels are less friendly to Tikhonov regularization. The blurring is given by
J B [ i , j ] = k , m J [ i + m , j + k ] K [ m , k ]
In principle, this equation can be rewritten in the form:
A x = b
and thus it is amenable to Tikhonov regularization. However, the images should be then converted to 1D vectors such that a W × H image becomes a W H × 1 vector. The blur matrix A becomes a W H × W H sparse matrix. It is clear that the storage and processing requirements are significantly higher than for a separable case (if W = H = 256 , we have 2 16 values in x and 2 32 values in A ).

Wiener Regularization

A well-known alternative way to solve the blurring Equation (6) is to move to Fourier space, where the equation becomes simply
J B ( k ) = J ( k ) K ( k ) + η ( k ) ,
with η ( k ) being the Fourier transform of image noise. Wiener [26] proposed a solution to this problem in the context of stationary time series. His solution applied to the domain of image deblurring has the following form:
J W ( k ) = K * ( k ) K ( k ) 2 + α 2 ( k ) J B ( k ) .
The expression in parentheses is called the Wiener Filter and the term α ( k ) is the inverse of the original signal SNR
α ( k ) = η ( k ) J ( k ) .
Since the α -term is in general unknown, it is usually replaced by a constant, which is denoted by λ 2 , so that
J λ ( r ) = 1 W H k K * ( k ) K ( k ) 2 + λ 2 J B ( k ) e i k r ,
where K * ( k ) is the complex conjugate of K ( k ) . Equation (10) will be henceforth referred to as the Wiener-regularized solution.

2.2. Optimal Regularization Parameter—MSE Approach

For real images, analytical estimation of the optimal value for λ is not possible. Thus, the following experiment is performed:
1.
Take an image
2.
Design a blur kernel; blur the image; add noise
3.
Perform regularized deconvolution using different values of λ
4.
For each value of λ , calculate the mean squared error (MSE) between the deblurred image and the original one
This procedure is repeated for a set of images with varying illumination, content, size and texture. First, a blur kernel is fixed and plot the mean error for various values of λ . Then, the analysis is performed for different blur kernels. In this section, one-dimensional kernels and Tikhonov-regularized deblurring are used. This approach can be readily extended to include 2D kernels and Wiener-regularized deblurring.
For a given image J and a blur kernel K , the mean squared deblurring error Q is given by
Q K ; λ , η = 1 Ω J K + Ξ K λ J 2 ,
where Ω is the number of pixels in the image, Ξ is the random Gaussian noise matrix of the same dimensions as the image
Ξ i j N 0 ; η ,
and
K λ = K T K T K + λ 2 L T L 1
is the regularized deblur matrix. Figure 1a shows a typical dependence of Q on λ calculated for a set of images J . Gaussian blur with σ = 30 is used, and the added noise intensity is η = 0.1 .
First, one can see that the error varies over several orders of magnitude, it is large for small and large λ , and reaches a minimum for some intermediate value. For λ smaller than the optimum value λ opt , the error comes mainly from noise amplification; for larger values, it comes from image residual blur (Figure 2). The actual numbers vary for different images, however, the main qualitative characteristics of the dependency Q ( λ ) remain the same. Now, the above procedure is repeated for different blur kernels. The same Gaussian blur as above is used, with σ uniformly distributed between 5 and 50. The results are shown in Figure 1b. As one can see, both the minimal attainable error and the optimal value λ opt vary as the size of the blur is varied. The variation of λ opt is not large (∼30%), thus, a single value of λ opt can be used for most cases.
In practice, the reference images are not available, so Q ( λ ) cannot be minimized to estimate the optimal value for λ . Correct estimation of λ has been a rather active research topic for quite a few decades [27], and produced several semi-heuristic approaches. Two most popular methods: Generalized cross-validation (GCV) [28] and the L-Curve [29] usually produce values of λ that are within an order of magnitude of Q ( λ ) -minimizing value. The resulting deconvolution is therefore sub-optimal. In Section 3.2 it will be shown how the problem of λ estimation can be solved using machine learning techniques.

2.3. Common Artifacts in Image Deblurring

Regularized image deblurring techniques described above belong to a category of deconvolution algorithms appearing in various signal estimation contexts [25]. More often than not, such regularized solutions exhibit artifacts of varying severity. In addition to random noise and over-smoothing, the most common types of artifacts appearing in deconvolved images are ringing and boundary artifacts. Both types appear as quasi-periodic modulations of image intensity. Sometimes, both types of artifacts are referred to as “ringing”, although their origins are somewhat different.
Ringing artifacts are a manifestation of the well-known Gibbs phenomenon: overshoots of Fourier sums at signal discontinuities. Regularization effectively removes high frequencies from the reconstructed image, thus exacerbating the phenomenon. In practice, ringing is more visible around objects surrounded by non-textured background such as the sky or uniform highway surface.
In a blurred image, pixels close to the image boundary contain information from the area outside of the image borders. Thus, deconvolution will not have this information available for proper reconstruction. As a result, the effect of missing boundary pixels propagates throughout the image, sometimes corrupting it beyond recognition. This phenomenon is referred to as boundary artifacts. Typical modulation length will be of the order of PSF size, since this is the extent of unknown pixels to be filled in. At the same time, intensity modulation frequency associated with ringing depends on PSF spectrum rather than on its size. Deconvolved images in Figure 2 clearly exhibit boundary artifacts of varying intensity.
Many tricks were proposed in the past to properly handle deconvolution artifacts. For example, a standard approach to fight boundary artifacts is to design suitable boundary conditions—from naïve pixel replication to sophisticated variation-minimization tiling [30]. Other approaches [31,32,33,34] incorporate artifact prevention into reconstruction algorithms. Finally, some researchers have proposed to post-process deconvolved images for detection and removal of unwanted spatial modulations [35].

3. Deblurring with Deep Learning

This section describes how deep neural networks can be trained to remove deconvolution artifacts. Following [23], throughout the rest of this work a two-step approach to deblurring is adopted:
1.
Perform regularized deconvolution of the blurred image
2.
Pass the deconvolved image through a deep neural network to remove residual artifacts.
In order for this approach to work, one needs to select a good value for λ for the first step, and a suitable network architecture for the second step.
This section is organized as follows. First, the two-step approach is described for the case of fixed input noise. It is then extended to incorporate joint parameter optimization for overall deblurring performance improvement. Next, it is demonstrated that there can be no single optimal parameter λ that could handle regularized deconvolution of variable-noise input. Finally, as a solution, the two-step approach is generalized to include two deep neural networks: one for estimating the regularization parameter λ from the input images, and another one to enhance the intermediate deconvolved images.

3.1. Known Input Noise

As shown in Section 2.2, for a fixed level of noise in the blurred set, slight variations in the value of λ around λ opt do not increase the deblur error significantly. It is reasonable to assume that these variations can be handled by subsequent processing in the DNN. Therefore, the deblur error is calculated using (11) averaged over the training set:
Q K ; λ , η = 1 N T Ω J J T J K + Ξ K λ J 2 ,
where J T is the training set and N T is its size. The parameter λ is set at the value that minimizes Q K ; λ , η :
λ = λ opt = argmin λ Q K ; λ , η

3.1.1. Neural Network Architectures

Throughout the rest of this paper, the post-processing neural network will be referred to as the image enhancement network (IEN).. Unlike the well-known classification-oriented architectures, input and output of the IEN should have identical dimensions. Two simple choices for building such a network would be either to keep the image dimensions uniform in all the layers, or to use an encoder-decoder architecture [36]. In this section, both options are examined. Also, in both cases, the aim is to estimate the residual between a pre-deconvolved image and the corresponding sharp image, similarly to [24].

Uniform Width CNN

The first tested IEN architecture is the uniform width (UW) CNN. The input image dimensions are halved in the first ( 7 × 7 ) convolutional layer producing N F feature images. The first layer is followed by N L ( 3 × 3 ) convolutional layers with the same number of features N F , and then a ( 7 × 7 ) convolution transpose layer restoring the image dimensions to their original value. The output of the convolution transpose layer is added to the input image so that the network learns to extract the residual image from the input. Every convolution layer is followed by batch normalization and a ReLU activation function.

U-Net

The second examined IEN architecture is a variant of the U-Net [37] proposed by Jin et al. [16] for solving various kinds of ill-posed inverse problems in imaging. In particular, it was used for removing reconstruction artifacts from medical imagery. Network hyper-parameters are chosen such that input image dimensions should be divisible by 2 6 .

3.1.2. Joint Parameter Optimization

It was previously shown that CNNs do a decent job removing artifacts generated by regularized deconvolution for a carefully chosen value of λ . This value strikes the balance between producing a deconvolved image that is too noisy and one that is too blurry. However, it is known that DNNs have a capacity for residual denoising and deblurring of images [38]. Therefore, the following question can be asked: Is it possible to move away from λ opt and still improve the overall performance of the system? In order to do that, the regularized deconvolution (RD) block should be treated as an optimizable module. Then, λ would be initialized at the value found in (15), and would vary together with the IEN weights during the end-to-end performance optimization process.
To proceed with the joint numerical optimization, it must be specified how to calculate the derivatives of the neural network cost with respect to λ . Within a standard gradient-based cost function optimization scheme such as stochastic gradient descent (SGD) optimization, the update relation for any NN parameter θ (from step n 1 to step n) is:
θ n = θ n 1 ϵ L θ ,
where L is the cost function of the NN and ϵ is the learning rate. Typically, since the dependence of the cost function on any NN parameter θ is manifested through a chain-like dependence scheme, the required gradients are calculated by a back-propagation algorithm. Specifically, if
L = g 0 ( g 1 ( g 2 ( ( g N ( θ ) ) ) ) )
Then
L θ = L g N g N θ = L g N 1 g N 1 g N g N θ ,
and so on. The overall cost as a function of the regularization parameter λ is
L = L ( J D ( λ ) )
with J ( λ ) being the input of the CNN, i.e., the regularized deblurred image.
Now, since for the Tikhonov-regularized deconvolution case
J D ( λ ) = J B K ( λ ) = J B V S S 2 + λ 2 I 1 U T
it follows that
L λ = 2 λ L J D J B V S S 2 + λ 2 I 2 U T ,
so that the update scheme for λ will be
Δ λ n = 2 ϵ λ n 1 L J D J B V S S 2 + λ n 1 2 I 2 U T .
The input gradients L / J D are fed back from the CNN part of the overall NN scheme.
The above idea can be readily extended to the Wiener filter case. Since
J D = IFFT J B ( k ) K * ( k ) | K ( k ) | 2 + λ 2 ,
it follows that
Δ λ n = 2 ϵ λ n 1 L J D IFFT J B ( k ) K * ( k ) | K ( k ) | 2 + λ n 1 2 2

3.2. Noise-Adaptive Deblurring

3.2.1. Deblurring Error for Varying Noise

In real life, noise levels of input images can vary quite significantly. A number of researchers tried to address this point previously. For instance, Schmidt et al. [39] proposed a MAP approach to non-blind deconvolution that incorporates noise estimation in the algorithm.
To quantify the effect of noise variation on regularized image deconvolution, Figure 3 depicts the deblur error (11) as a function of λ for a random image from the training set, for noise levels varying over three orders of magnitude: from η = 10 2 to η = 10.0 . As before (Figure 1), the deblur error spans several orders of magnitude. The variation, however, is significantly larger for higher levels of noise, mainly to the left of λ opt ( η ) due to practically unbounded noise amplification. On the other hand, the residual blur does not exhibit strong dependence on the noise, therefore, the graphs tend to converge for larger values of λ . The minima of Q η ( λ ) appear rather flat in the context of overall error variation, however, they are sufficiently pronounced, more so for higher levels of noise. This is to be expected, since for low noise levels, regularized deconvolution is quite tolerant to the value of λ .
The main observation that is rather obvious from Figure 3, is that there can be no single optimal parameter λ opt * that can handle regularized deconvolution of variable-noise input. The deconvolved images will be either buried in noise (if λ opt * < λ opt ( η ) ) or will have excessive residual blur (if λ opt * > λ opt ( η ) ). It can be predicted with a high degree of confidence, that repeating training procedures from the previous section for variable-noise input will strongly bias the value of λ opt * towards higher values therefore producing over-smoothed images. On the other hand, it is also clear that the minimal attainable error Q [ λ opt ( η ) ] does not vary as much for different levels of noise. Thus, even for strong noise, if a corresponding λ opt can be found, reasonable output may still be produced.
Next, the connection between the input noise and the corresponding λ opt calculated from Equation (15) is examined. To quantify the amount of noise in the corrupted image a metric called S N R B is used. It is defined as a ratio between average blurred image intensity and added noise intensity. Note that S N R B is by no means an image quality metric; it is merely a normalized characteristic of noise contamination for input images. Figure 4 shows a log-log plot of S N R B vs λ opt for 100 blurred images corrupted with 50 different noise levels each, with intensities ranging between η = 10 3 and η = 10.0 . Estimated correlation for S N R B and λ opt data is found to be very strong (–0.98), and the linear fit to the log-log data suggests the following empirical dependence:
λ opt ( S N R B ) 3 / 4
However, for this relation to be useful, a reliable estimate for S N R B is needed. Unfortunately, SNR estimation is in itself not a trivial problem. There are a number of ways to estimate SNR of an image without a reference [40,41], but using them will only increase the uncertainty of λ opt estimation.

3.2.2. The General Idea

If one could find a way to generate a proper value of λ for any blurred and noisy image, then, in principle, the problem will be reduced to the two-step solution described above.
It is therefore proposed to train a deep neural net to estimate λ opt from the blurred image itself. Thus, the overall deblurring system will include two DNNs - the first network for estimating λ opt , and the second one for image enhancement after regularized deconvolution, as described in Section 3.1. The first network will be henceforth referred to as regularization parameter estimation network, or RegParamNet. Figure 5 shows the concept of the proposed solution. A blurred image is fed into a network that generates a value for λ . Then, using this value and the known PSF, the image is deconvolved and the result is fed into the IEN.

3.2.3. RegParam Network Architecture

As Hansen points out in [9], blurred images have much faster decaying singular value (SV) spectra than corresponding sharp ones. Adding noise boosts the small SV part of the spectrum so that noisy blurred images decay have SV spectrum decaying slower than clean ones. Following Hansen, it is assumed that the SV spectrum of an image contains enough information regarding the relationship between high-frequency noise and high-frequency features in the image. Thus, the aim is to build a DNN that would take SV spectrum of an image as an input and would output λ opt . The ResNet architecture [42] is taken as a reference, and its 1D analogue is built: while ResNet works with 2D images and 2D convolutions, RegParamNet processes 1D inputs. Figure 6 shows the final architecture that was chosen after several rounds of experimenting with network hyperparameters, such as depth and number of features. The input signal is a 1D vector composed of image SV logarithms. The input is then converted in the first convolutional layer to 64 feature vectors. The next four stages of RegParamNet are cascades of five residual blocks as shown in Figure 6. There is a feature number doubling convolutional layer and a max-pooling layer following each cascade. Finally, 1024 feature vectors are fed into a fully connected layer to produce the output vector.
Here, two options for λ opt estimation using RegParamNet are considered:
1.
Direct regression of λ opt
2.
Selection of λ opt from a set of values.
In both cases, network training consists of two steps. First, RegParamNet is pre-trained to generate λ opt that minimizes the deblur error Q η ( λ ) of (11). Then, it is connected to the IEN as shown in Figure 5 for end-to-end (E2E) system training.

3.2.4. RegParamNet Training Schemes

Training Data Generation

RegParamNet should be able to handle a wide range of noise magnitudes and blur kernels. For this purpose, the image corruption module (ICM) is introduced, generating blur and noise in a random manner for each training iteration. For instance, 1D blur kernels (Figure 7a) are generated according to the following scheme:
K n = p n m p m , p n = e ( n μ ) 2 / 2 σ 2 , 0 n L 1 L U ( 16 , 64 ) σ U L 4 , L μ U L 4 , 3 L 4
In addition, ICM randomly picks a value η from a log-uniform distribution such that 10 2 η 10.0
log 10 η U ( 2 , 1 ) ,
and adds random noise of this magnitude to the blurred image. For 2D Wiener deconvolution, blur functions are generated as 2D Gaussians with random covariance:
K n = p n n p n , p n = e 1 2 n T Γ 1 n ,
where
L x , y 2 n x , y < L x , y 2 .
and
Γ x x = α L x L y , Γ y y = β L x L y Γ x y = Γ y x = 0.9 δ Γ x x Γ y y α , β U ( 0 , 1 ) , δ U ( 1 , 1 )
The factor 0.9 in Γ x y ensures the invertibility of Γ ; L x = 64 and L y = 32 were used.
For each training iteration, ICM provides data to generate blurred image from a sharp one. Using blur data from ICM and a vector of λ -values, an array of deconvolved images (DeconvArray) is created. The entries of λ -vector are incremented exponentially such that
log 10 λ i = 3 + 3 N λ 1 i , 0 i N λ 1
Calculating MSE of DeconvArray relatively to the sharp image gives us the function Q ( λ ) . At the same time, SVD of the blurred image is calculated and the array of SV logarithms is fed to the RegParamNet. The schemes for Tikhonov and Wiener cases are shown in Figure 7.

RegParamNet Modes

The scheme used for pre-training RegParamNet to directly regress λ opt is shown in Figure 8a. The fully connected layer of RegParamNet outputs a single float number that is passed to MSE loss module together with log λ opt . Regression of log λ is much more stable than regression of λ since for all practical purposes λ < 1 and often it is very small.
An additional way to estimate λ opt is to train the network to generate the function Q ( λ ) given the SV spectrum of the blurred image and find its minimum. In this case, the FC layer of RegParamNet produces a real-valued vector, the size of which equals the length of the λ -vector. Since Q ( λ ) can reach extremely large values, it is more practical to work with another function g ( λ ) :
g ( λ ) = softmin Q ( λ ) = exp Q ( λ ) λ exp Q ( λ )
Then,
λ opt = argmin Q ( λ ) = argmax g ( λ )
To train RegParamNet to generate a function similar to g ( λ ) a setup shown in Figure 8b is used. Blurred images and corresponding MSE vector Q ( λ ) are created like in the previous section. Then, Q ( λ ) is converted to g ( λ ) using (32). The output of RegParamNet passes through a softmax layer to produce a vector f ( λ ) . Since both g ( λ ) and f ( λ ) vectors are normalized, they can be treated as probability distributions. A well-known similarity measure for two distributions is their Kullback–Leibler divergence [43]:
D K L g f = λ g ( λ ) log g ( λ ) f ( λ )
D K L g f is always non-negative and is only zero when the two distributions are identical. Thus, setting D K L g f as the RegParamNet loss and minimizing it, the network is trained to approximate g ( λ ) . Then, to use RegParamNet to estimate λ opt , one just needs to find the maximum of f ( λ ) .

4. Experiments

4.1. Known Input Noise

4.1.1. Training

For training, images from Berkeley DeepDrive (BDD) database [44] were used. Original images were converted to grayscale and down-scaled by a factor of 0.5 to produce 360 × 640 images. To make the images compliant with the U-Net architecture described above, they were further cropped to 320 × 640 . Sharp images were blurred using a variety of 1D Gaussian blur kernels with σ varying from 30 to 64. Blurred images were then additionally corrupted by Gaussian noise with RMS intensity η = 0.1 . The optimal regularization factor for η = 0.1 was found by the technique described in Section 2.2 to be λ opt = 7.5 × 10 3 . The blurred images were deconvolved using Tikhonov regularization. The deconvolved images were used as inputs to the CNNs. MSE between CNN output and the original sharp images was used as the network loss.
The networks were implemented in the PyTorch framework [45] and initialized using Kaiming-normal initialization [46]. The networks were trained on 40,000 images and validated on 10,000 images for 30 epochs. SGD algorithm was used, with initial learning rate of 10 3 that was reduced to 10 4 after 15 epochs.

4.1.2. Quality Metrics

Assessment of image quality is not a very well defined task, especially in the absence of reference images  [47]. Depending on the final goal of image processing chain in question, different metrics may be more or less well suited for quantifying algorithm performance. Tasks like image compression, color balancing, or de-noising can be (and usually are) evaluated differently by a group of human testers compared to some objective metric. In our rather reduced scope, deblurring is treated as a standalone module; also, reference images are available for testing. Therefore, deblurred image quality is assessed using the commonly accepted peak signal-to-noise ratio (PSNR) metric. For the sake of completeness, structural similarity (SSIM)  [48] for deblurred images was evaluated as well; SSIM values correlate rather well with PSNR values and strongly support the main findings.

4.1.3. Results

For testing, a separate subset of 10,000 images from the BDD dataset was used. The images were converted to grayscale, resized and cropped as described in Section 4.1.1.
Three CNN configurations were tested: uniform width (UW) architecture with N F = 64 and N F = 128 and residual U-Net architecture. Mean PSNR values are summarized in Table 1.
First of all, it is clear that just using a deep neural net increases the image PSNR by about 4 dB on average. Figure 9 shows that this large PSNR increase can be mainly attributed to deconvolution artifact removal and some image denoising. Secondly, the difference between UW-64 and UW-128 is not significant. It is also clear that residual U-Net outperforms the UW networks. Thus, throughout the rest of this work, the residual U-Net architecture is used.

4.1.4. Joint Parameter Optimization

The regularized deconvolution (RD) module was implemented using PyTorch framework using AutoGrad functionality [49]. The regularization parameter λ was initialized at λ opt = 7.5 × 10 3 . The entire network (RD+residual U-Net) was trained using the procedure outlined in Section 4.1.1.
The statistics for 10,000 test images are outlined in Table 2 below. The new value of the regularization parameter is λ opt * = 3.3 × 10 3 , so that we expect the deconvolved image to be on the noisy side. Indeed, the mean PSNR for deconvolved images drops by 2.5 dB relatively to λ opt = 7.5 × 10 3 case. However, the IEN is able to remove the additional noise rather effectively demonstrating better overall performance than for λ = λ opt .
Finally, Figure 10 demonstrates image quality improvement from deconvolved images, through standalone two-stage image enhancement to the joint end-to-end system optimization.

4.2. Noise-Adaptive Deblurring

4.2.1. RegParamNet Training

Input images from the BDD database were converted and down-scaled as described in Section 4.1.1. The networks were trained on 40,000 images and validated on 10,000 images for 30 epochs. ADAM optimizer algorithm was used with initial learning rate of 10 4 that was adaptively reduced by a ReduceLROnPlateau scheduler. For each mini-batch, a blur kernel was generated as described in Training Data Generation Section and a value of η is randomly picked such that log 10 η U ( 2 , 1 ) ) .

Direct λ Regression

Figure 11 shows the error distribution function of regularization parameter logarithm for the direct regression mode. One can see that RegParamNet is able to directly infer log 10 λ opt for both Wiener and Tikhonov regularization cases, with overall standard deviations for log 10 λ opt of ∼0.1. This corresponds to λ opt accuracy of 20–25%; an error of this magnitude has a relatively insignificant effect on the deconvolved image quality.

λ -Weight Array Generation

Figure 12 shows the results of λ -weight array generation for several test images and several values of noise magnitude η from the range 10 2 η 10.0 . The networks generate λ -weight arrays for N λ = 16 , 32 , 64 . The resulting arrays approximate the target function softmin [ Q ( λ ) ] with a similar degree of precision (height difference stems from f ( λ ) normalization; the peak approximately halves when N λ doubles).

4.2.2. End-to-End System Training

After the pre-training is finished, output of RegParamNet is connected through a regularized deconvolution (RD) module to the input of the image enhancement network. Specifically, a regressed value of λ opt or a value of λ that maximizes f ( λ ) could be taken, used for regularized deconvolution and then the output can be fed into the IEN. However, as was shown in the previous section, λ opt that minimizes Q ( λ ) does not necessarily lead to the best system performance. Thus, it is desirable to connect RegParamNet to IEN and train the two networks jointly. Figure 13 describes the overall training setup for both λ opt regression and λ -weight cases.
Connecting λ -weight RegParamNet to the RD module and, eventually, to the IEN, for the purpose of joint training is less straightforward than in the regression case. The reason for this is that A r g M a x is not an analytical function and thus cannot be easily integrated into the back-prop optimization framework. Thus, the following estimator I ^ for the deconvolved image is proposed:
I ^ = λ f ( λ ) I λ
where I λ is the DeconvArray entry corresponding to parameter λ . Then, joint end-to-end (E2E) training as shown in Figure 13b can reshape f ( λ ) and change the weights of IEN so that the overall system performance improves.
From the practical point of view, a three-step training approach was found to produce best results. First, for each setup in Figure 13, a pre-trained RegParamNet was connected to the RD and the IEN was bypassed. After this step, f ( λ ) became more peaked and narrow around the optimal value. Then, RegParamNet was frozen and the IEN was pre-trained on RegParamNet outputs. Finally, RegParamNet was unfrozen to allow full E2E training. E2E training was implemented using λ -weight approach for both 1D and 2D deblurring; direct λ opt regression was implemented for 1D deblurring only; we are currently working to extend it to the 2D (Wiener) case as well. For each step, the training was performed as described in Section 4.2.1.

4.2.3. Results: Statistics

The system was tested on 10,000 images from the BDD database for randomly generated blur and noise magnitudes as described above. Processed image quality was tested in the following configurations:
(a)
RD output before E2E training
(b)
RD output after E2E training
(c)
IEN output before E2E training
(d)
IEN output after E2E training
PSNR and SSIM were used as the quality assessment metrics. Table 3 summarizes the corresponding mean PSNR/SSIM values. One can see that, for the 1D case, λ -weight and direct regression RegParamNets perform quite well across all levels of input noise and blur. Thus, from the practical point of view, direct λ opt regression could be preferable since it is more lightweight. In addition, since 2D blur is in general much stronger than 1D blur, final PSNR for Wiener deblurring cases is substantially lower than for 1D case. Figure 14 shows IEN E2E result PSNR as a function of input noise.
It is noteworthy that the effect described in the previous section is clearly observable here as well: E2E training moves the RegParamNet away from the point where Q ( λ ) is minimized thus making the deconvolution sub-optimal. However, the IEN learns to compensate for this performance drop. In the 1D λ -weight case, it even improves the overall system performance by 1.3dB on the average. In other RegParamNet configurations, E2E training does not exhibit clear overall performance gain, though the IEN capacity is improved in all cases. Figure 15 clearly shows the effect of E2E training on the λ -weight array in the Tikhonov deconvolution case. The preferred λ values are typically smaller than in the standalone RegParamNet training, thus moving the pre-deconvolved images to the more noisy side of the standalone optimum. The only cases where this effect is not so clear are the ones with a very low level of input noise ( η 10 2 ), where the deconvolution is much less sensitive to the exact value of λ opt . One can therefore conclude that the IEN architecture in use has a better capacity for removing residual noise than for residual deblurring.

4.2.4. Results: Images

To summarize, examples of noise-adaptive image deblurring for test images (best viewed on a soft copy. Additional high-resolution images are provided in the Supplementary Materials). are presented in Figure 16 and Figure 17. First, it can be seen that the system is capable of handling both strong input noise and large blur kernels. The deconvolution step is adequately regularized across the entire range of input noise intensities, and the subsequent image enhancement step removes most artifacts. To control noise amplification, regularization is naturally stronger for high input noise, resulting in less sharp output images. However, additional quality gain of 0.3–0.8 dB coming from E2E training is consistently observed on the high-noise side (rows (d) vs. (c) in Figure 16 and Figure 17). As described above, this gain is accompanied by a quality drop at the deconvolution step (rows (b) vs. (a) in Figure 16 and Figure 17). Second, one can also see that small features are clearly discernible even for η 1.0 . Stronger noise leads to eventual elimination of at least some of these features; however, larger objects are quite well defined and likely allow automatic classification.

4.2.5. Results: Comparison to Other Approaches

In this section, the performance of the proposed method is compared to several previous approaches. The analysis follows the paper by Wang and Tao [24], and corresponding results from two additional leading works [50,51] are cited for reference as they appear there. The system was tested on Berkeley Segmentation Dataset [52] (BSD100). The following test scenarios were used:
  • Gaussian kernel with spatial standard deviation equal to 1.6 and noise η = 0.008 (denoted as GaussianA; η = 0.01 in the proposed method)
  • Gaussian kernel with spatial standard deviation equal to 3 and noise η = 0.04 (denoted as GaussianB; η = 0.05 in the proposed method)
  • Gaussian kernel with spatial standard deviation equal to 5 and noise η = 0.04 (denoted as GaussianC; η = 0.05 in the proposed method)
  • Square kernel with a side size of 19 and noise of η = 0.01 (denoted as SquareA)
  • Square kernel with a side size of 13 and noise of η = 0.04 (denoted as SquareB; η = 0.05 in the proposed method)
In addition, all 32 motion blur kernels from [53] were tested. The kernels were applied to test images from BSD100, followed by addition of noise. Two noise levels were tested: η = 0.01 (MotionA), and η = 0.22 (MotionB, compared to η = 0.06 in [24]) was added. In each case average PSNR/SSIM was calculated over 3200 deblurred images.
Finally, the proposed method was tested at higher levels of input noise for the following Gaussian/Square configurations:
  • Gaussian kernel with spatial standard deviation equal to 3 and noise η = 1.06 (denoted as GaussianD)
  • Gaussian kernel with spatial standard deviation equal to 5 and noise η = 0.22 (denoted as GaussianE)
  • Square kernel with a side size of 19 and noise of η = 0.22 (denoted as SquareC)
  • Square kernel with a side size of 13 and noise of η = 1.06 (denoted as SquareD)
At these noise levels there are no data points from previous publications; for instance, Wang and Tao [24] reported that their experiments started breaking down at η 0.1–0.2.
The results are summarized in Table 4. The current method is referred to as NANDB (Noise-Adaptive Non-Blind Deblurring). For the sake of completeness, the last column shows the performance of NANDB on 100 test images from the BDD set (denoted as NANDB*). It is rather clear that the proposed method outperforms other methods, often by as much as 3–5 dB in PSNR, especially for higher values of input noise. This can be attributed to the flexibility of the current method with respect to input noise intensity, whereas other works use “one-fits-all approach” in this regard. As was shown in Section 3.2, regularization parameter can be under- or overestimated by a factor of 1.5–2 without severe consequences for deconvolution. However, if input noise is allowed to vary over several orders of magnitude, fixed regularization parameters inevitably produce deconvolved images that are either over-smoothed or noisy beyond recovery. This also explains why other methods are applicable only in a narrow range of input noise intensities For instance, in [24] the system broke down already at η ∼ 0.1–0.2, as it was trained to handle noise in the range 0.008–0.06 only.
One can also see that the system performs even better if its usage is confined to automotive scenarios: The system trained on the BDD dataset exhibits significantly higher PSNR/SSIM values when tested on the BDD test images, as shown in the last column of Table 4. This improvement is likely due to a larger diversity of low- and high-level features present in the BSD100 dataset compared to BDD images. Thus, system performance is likely to improve if it is trained on large general datasets.

5. Summary

In this work, a systematic approach to non-blind deblurring is presented. Regularized deconvolution approaches are described as the means to control noise amplification, which is the main image degradation factor. Next, following previous work [23,24], removing deconvolution artifacts by a deep neural net (image enhancement network - IEN) is proposed. For the case of constant input noise, the two-step deblurring is extended by incorporating the regularized deconvolution module into a joint training framework. As a next step, the impact of input noise on the required amount of regularization is explored. To infer the regularization parameter ( λ ) values, RegParamNet, a novel 1D convolutional neural network is proposed. Two approaches for regularization parameter inference are explored (direct regression and λ -weights). Finally, a noise-adaptive non-blind image deblurring system is built by incorporating both RegParamNet and IEN into a common end-to-end training and inference framework.
The proposed system performs rather well on a wide range of large 1D/2D blur kernels, over three orders of magnitude of input noise. It is also found that end-to-end training biases the inferred regularization parameters downwards, so that deconvolved intermediate images are sharper, albeit noisier that for a standalone optimum. However, overall system performance is found to benefit from end-to-end training, especially for high levels of input noise ( η = 0.2–10.0). Compared to other approaches, the system exhibits rather superior PSNR/SSIM performance even without domain adaptation, i.e., networks trained on automotive images (BDD) perform well on general type images (BSD100).
Although the present work suggests a principal solution to the problem of noise-adaptive non-blind deblurring, there are several research directions in which it can be extended. One rather important subject to be explored is the performance of perception algorithms on deblurred images. It has been noted before that some approaches to blind deblurring corrupt low-level features in processed images, thereby impacting detection and classification performance rather severely [3]. It is important to establish that no such effect takes place in our case. In addition, previous work suggests that imprecise knowledge of PSF produces effects akin to boundary artifacts [21]. Since PSF calibration or indirect estimation is always of limited precision, it is important to analyze the impact of PSF uncertainty on system performance, and to adjust the solution if needed. Also, it should be noted that current work presents a solution for gray-scale images only; an extension for color images should not be too difficult. Finally, certain pre- and post-processing stages can be added to the deblurring pipeline, such as preliminary de-noising of input images [54] or super-resolution [55] on output images. Both additions are likely to improve the overall system performance.
There is still much work to be done in order to convert the proposed deblurring scheme into a production-grade software system: architecture optimization, streamlining the processing flow, etc. For instance, it is clear that there is no need to infer regularization parameters for each captured frame, since the SNR changes at a much lower rate. However, automotive scenarios include abrupt scene illumination changes (e.g., when entering or exiting tunnels and underground parking lots). Thus, additional modules, possibly learning-based, are needed to pace λ estimation and to provide other system-wide inputs. Finally, it should be noted that system optimization with respect to computational hardware requirements is beyond the scope of this work. We have a high degree of confidence, however, that the solution can be efficiently implemented using novel mobile-oriented architectures, which have been at the focus of many research and development efforts in recent years [56,57].

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s22186923/s1, Figures S1–S7: Additional examples of deblurred images for various blur kernels and noise intensities.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The author wishes to thank E. Kishon, T. Philipp, E. Mordechai, and N. Garnett for fruitful discussions and helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
1D/2DOne-Dimensional / Two-Dimensional
ADASAdvanced Driver Assistance System(s)
AVAutonomous Vehicle
BDDBerkeley Deep Drive
BSDBerkeley Segmentation Dataset
CNN/DNNConvolutional Neural Network / Deep Neural Network
DSLRDigital Single-Lens Reflex (camera)
E2EEnd-to-End
FCFully Connected (layer)
GCVGeneralized Cross-Validation
ICMImage Corruption Module
IENImage Enhancement Network
ISPImage Signal Processing/Processor
MAPMaximum A Posteriori (estimation)
MSEMean Squared Error
PSFPoint Spread Function
PSNRPeak Signal-to-Noise Ratio
RDRegularized Deconvolution
ReLURectified Linear Unit
RMSRoot Mean Square
SGDStochastic Gradient Descent
SNRSignal-to-Noise Ratio
SSIMStructural Similarity
SVDSingular Value Decomposition
UWUniform Width

References

  1. Nakamura, J. Image Sensors and Signal Processing for Digital Still Cameras; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  2. Janai, J.; Güney, F.; Behl, A.; Geiger, A. Computer vision for autonomous vehicles: Problems, datasets and state-of-the-art. arXiv 2017, arXiv:1704.05519. [Google Scholar]
  3. Pei, Y.; Huang, Y.; Zou, Q.; Zhang, X.; Wang, S. Effects of Image Degradation and Degradation Removal to CNN-based Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1239–1253. [Google Scholar]
  4. Hege, E.K.; Jefferies, S.M.; Lloyd-Hart, M. Computing and telescopes at the frontiers of optical astronomy. Comput. Sci. Eng. 2003, 5, 42–51. [Google Scholar] [CrossRef]
  5. Sage, D.; Donati, L.; Soulez, F.; Fortun, D.; Schmit, G.; Seitz, A.; Guiet, R.; Vonesch, C.; Unser, M. DeconvolutionLab2: An open-source software for deconvolution microscopy. Methods 2017, 115, 28–41. [Google Scholar] [CrossRef]
  6. Maître, H. From Photon to Pixel: The Digital Camera Handbook; John Wiley & Sons: Hoboken, NJ, USA, 2017. [Google Scholar]
  7. Lee, H.C. Review of image-blur models in a photographic system using the principles of optics. Opt. Eng. 1990, 29, 405–422. [Google Scholar] [CrossRef]
  8. Chen, X.; Li, F.; Yang, J.; Yu, J. A theoretical analysis of camera response functions in image deblurring. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 333–346. [Google Scholar]
  9. Hansen, P.C.; Nagy, J.G.; O’Leary, D.P. Deblurring Images: Matrices, Spectra, and Filtering; SIAM: Philadelphia, PA, USA, 2006; Volume 3. [Google Scholar]
  10. Harris, J.L. Image evaluation and restoration. JOSA 1966, 56, 569–574. [Google Scholar] [CrossRef]
  11. Richardson, W.H. Bayesian-based iterative method of image restoration. JOSA 1972, 62, 55–59. [Google Scholar] [CrossRef]
  12. Lucy, L.B. An iterative technique for the rectification of observed distributions. Astron. J. 1974, 79, 745. [Google Scholar] [CrossRef]
  13. Mustaniemi, J.; Kannala, J.; Särkkä, S.; Matas, J.; Heikkila, J. Gyroscope-Aided Motion Deblurring with Deep Networks. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; IEEE: New York, NY, USA, 2019; pp. 1914–1922. [Google Scholar]
  14. Wang, R.; Tao, D. Recent progress in image deblurring. arXiv 2014, arXiv:1409.6838. [Google Scholar]
  15. McCann, M.T.; Jin, K.H.; Unser, M. Convolutional neural networks for inverse problems in imaging: A review. IEEE Signal Process. Mag. 2017, 34, 85–95. [Google Scholar] [CrossRef]
  16. Jin, K.H.; McCann, M.T.; Froustey, E.; Unser, M. Deep convolutional neural network for inverse problems in imaging. IEEE Trans. Image Process. 2017, 26, 4509–4522. [Google Scholar] [CrossRef]
  17. Schuler, C.J.; Hirsch, M.; Harmeling, S.; Schölkopf, B. Learning to deblur. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 1439–1451. [Google Scholar] [CrossRef]
  18. Zhang, K.; Ren, W.; Luo, W.; Lai, W.S.; Stenger, B.; Yang, M.H.; Li, H. Deep image deblurring: A survey. Int. J. Comput. Vis. 2022, 130, 2103–2130. [Google Scholar] [CrossRef]
  19. Nah, S.; Son, S.; Lee, S.; Timofte, R.; Lee, K.M. NTIRE 2021 challenge on image deblurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 149–165. [Google Scholar]
  20. Ren, W.; Zhang, J.; Ma, L.; Pan, J.; Cao, X.; Zuo, W.; Liu, W.; Yang, M.H. Deep non-blind deconvolution via generalized low-rank approximation. In Proceedings of the Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 297–307. [Google Scholar]
  21. Vasu, S.; Reddy Maligireddy, V.; Rajagopalan, A. Non-blind deblurring: Handling kernel uncertainty with CNNs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3272–3281. [Google Scholar]
  22. Hosseini, M.S.; Plataniotis, K.N. Convolutional Deblurring for Natural Imaging. IEEE Trans. Image Process. 2019, 29, 250–264. [Google Scholar] [CrossRef]
  23. Xu, L.; Ren, J.S.; Liu, C.; Jia, J. Deep convolutional neural network for image deconvolution. In Proceedings of the Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 1790–1798. [Google Scholar]
  24. Wang, R.; Tao, D. Training very deep CNNs for general non-blind deconvolution. IEEE Trans. Image Process. 2018, 27, 2897–2910. [Google Scholar] [CrossRef]
  25. Tikhonov, A.N.; Arsenin, V.Y. Solutions of Ill-Posed Problems; Wiley: New York, NY, USA, 1977. [Google Scholar]
  26. Wiener, N. The Interpolation, Extrapolation and Smoothing of Stationary Time Series; MIT: Cambridge, MA, USA, 1942. [Google Scholar]
  27. Engl, H.W.; Hanke, M.; Neubauer, A. Regularization of Inverse Problems; Springer Science & Business Media: Berlin, Germany, 1996; Volume 375. [Google Scholar]
  28. Golub, G.H.; Heath, M.; Wahba, G. Generalized cross-validation as a method for choosing a good ridge parameter. Technometrics 1979, 21, 215–223. [Google Scholar] [CrossRef]
  29. Hansen, P.C. The L-curve and its use in the numerical treatment of inverse problems. In Computational Inverse Problems in Electrocardiology; WIT Press: Southampton, UK, 1999; pp. 119–142. [Google Scholar]
  30. Liu, R.; Jia, J. Reducing boundary artifacts in image deconvolution. In Proceedings of the 2008 15th IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008; IEEE: New York, NY, USA, 2008; pp. 505–508. [Google Scholar]
  31. Reeves, S.J. Fast image restoration without boundary artifacts. IEEE Trans. Image Process. 2005, 14, 1448–1453. [Google Scholar] [CrossRef]
  32. Sorel, M. Removing boundary artifacts for real-time iterated shrinkage deconvolution. IEEE Trans. Image Process. 2011, 21, 2329–2334. [Google Scholar] [CrossRef]
  33. Yuan, L.; Sun, J.; Quan, L.; Shum, H.Y. Progressive inter-scale and intra-scale non-blind image deconvolution. ACM Trans. Graph. (TOG) 2008, 27, 1–10. [Google Scholar] [CrossRef]
  34. Lee, J.H.; Ho, Y.S. High-quality non-blind image deconvolution with adaptive regularization. J. Vis. Commun. Image Represent. 2011, 22, 653–663. [Google Scholar] [CrossRef]
  35. Mosleh, A.; Langlois, J.P.; Green, P. Image deconvolution ringing artifact detection and removal via PSF frequency analysis. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 247–262. [Google Scholar]
  36. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  37. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  38. Lucas, A.; Iliadis, M.; Molina, R.; Katsaggelos, A.K. Using deep neural networks for inverse problems in imaging: Beyond analytical methods. IEEE Signal Process. Mag. 2018, 35, 20–36. [Google Scholar] [CrossRef]
  39. Schmidt, U.; Schelten, K.; Roth, S. Bayesian deblurring with integrated noise estimation. In Proceedings of the 24th IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2011), Colorado Springs, CO, USA, 20–25 June 2011; IEEE: New York, NY, USA, 2011; pp. 2625–2632. [Google Scholar]
  40. Meer, P.; Jolion, J.M.; Rosenfeld, A. A fast parallel algorithm for blind estimation of noise variance. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 216–223. [Google Scholar] [CrossRef]
  41. Li, Z.; Zhang, W.; Lin, W. Adaptive median filter based on SNR estimation of single image. In Proceedings of the 2012 International Conference on Computer Science and Service System, Nanjing, China, 11–13 August 2012; IEEE: New York, NY, USA, 2012; pp. 246–249. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. MacKay, D.J. Information Theory, Inference and Learning Algorithms; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  44. Yu, F.; Xian, W.; Chen, Y.; Liu, F.; Liao, M.; Madhavan, V.; Darrell, T. BDD100K: A Diverse Driving Video Database with Scalable Annotation Tooling. arXiv 2018, arXiv:1805.04687. [Google Scholar]
  45. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Nice, France, 2019; pp. 8024–8035. [Google Scholar]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  47. Liu, Y.; Wang, J.; Cho, S.; Finkelstein, A.; Rusinkiewicz, S. A no-reference metric for evaluating the quality of motion deblurring. ACM Trans. Graph. 2013, 32, 175. [Google Scholar] [CrossRef]
  48. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  49. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic differentiation in PyTorch. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  50. Schuler, C.J.; Christopher Burger, H.; Harmeling, S.; Scholkopf, B. A machine learning approach for non-blind image deconvolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1067–1074. [Google Scholar]
  51. Son, H.; Lee, S. Fast non-blind deconvolution via regularized residual networks with long/short skip-connections. In Proceedings of the 2017 IEEE International Conference on Computational Photography (ICCP), Stanford, CA, USA, 12–14 May 2017; IEEE: New York, NY, USA, 2017; pp. 1–10. [Google Scholar]
  52. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics. In Proceedings of the 8th International Conference on Computer Vision, Vancouver, BC, Canada, 9–12 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  53. Levin, A.; Weiss, Y.; Durand, F.; Freeman, W.T. Understanding and evaluating blind deconvolution algorithms. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 22–24 June 2009; IEEE: New York, NY, USA, 2009; pp. 1964–1971. [Google Scholar]
  54. Romano, Y.; Elad, M.; Milanfar, P. The little engine that could: Regularization by denoising (RED). SIAM J. Imaging Sci. 2017, 10, 1804–1844. [Google Scholar] [CrossRef]
  55. Kim, J.; Kwon Lee, J.; Mu Lee, K. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  56. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 4510–4520. [Google Scholar]
  57. Wang, E.; Davis, J.J.; Zhao, R.; Ng, H.C.; Niu, X.; Luk, W.; Cheung, P.Y.; Constantinides, G.A. Deep Neural Network Approximation for Custom Hardware: Where We’ve Been, Where We’re Going. ACM Comput. Surv. (CSUR) 2019, 52, 1–39. [Google Scholar] [CrossRef]
Figure 1. Deblurring Error for (a) single kernel, different images; (b) single image, various kernels.
Figure 1. Deblurring Error for (a) single kernel, different images; (b) single image, various kernels.
Sensors 22 06923 g001
Figure 2. Deconvolved images: (ac) single blur kernel, different λ values; (df) different blur kernels, λ = λ opt for each kernel.
Figure 2. Deconvolved images: (ac) single blur kernel, different λ values; (df) different blur kernels, λ = λ opt for each kernel.
Sensors 22 06923 g002aSensors 22 06923 g002b
Figure 3. Deblur error for different levels of input noise.
Figure 3. Deblur error for different levels of input noise.
Sensors 22 06923 g003
Figure 4. Dependence of λ opt on the input S N R B .
Figure 4. Dependence of λ opt on the input S N R B .
Sensors 22 06923 g004
Figure 5. The proposed concept for noise-sensitive solution. Blue blocks/arrows describe the method for self-supervised end-to-end training of the system.
Figure 5. The proposed concept for noise-sensitive solution. Blue blocks/arrows describe the method for self-supervised end-to-end training of the system.
Sensors 22 06923 g005
Figure 6. RegParamNet architecture.
Figure 6. RegParamNet architecture.
Sensors 22 06923 g006
Figure 7. RegParamNet training data generation: (a) Tikhonov deconvolution; (b) Wiener deconvolution.
Figure 7. RegParamNet training data generation: (a) Tikhonov deconvolution; (b) Wiener deconvolution.
Sensors 22 06923 g007
Figure 8. RegParamNet training: (a) Direct λ regression; (b) g ( λ ) approximation (“ λ -weights”).
Figure 8. RegParamNet training: (a) Direct λ regression; (b) g ( λ ) approximation (“ λ -weights”).
Sensors 22 06923 g008
Figure 9. Two-step image deblurring for a set of images: (a) Tikhonov Deconvolution with λ = 7.5 × 10 3 ; (b) Initial + UW-64; (c) Initial + UW-128; (d) Initial + residual U-Net.
Figure 9. Two-step image deblurring for a set of images: (a) Tikhonov Deconvolution with λ = 7.5 × 10 3 ; (b) Initial + UW-64; (c) Initial + UW-128; (d) Initial + residual U-Net.
Sensors 22 06923 g009
Figure 10. Effect of joint training for a set of images: (a) Deconvolved image with λ = λ opt ; (b) Deconvolved image after joint training; (c) Two-step deblurring (d) Two-step deblurring after joint training.
Figure 10. Effect of joint training for a set of images: (a) Deconvolved image with λ = λ opt ; (b) Deconvolved image after joint training; (c) Two-step deblurring (d) Two-step deblurring after joint training.
Sensors 22 06923 g010
Figure 11. Direct λ opt regression results.
Figure 11. Direct λ opt regression results.
Sensors 22 06923 g011
Figure 12. λ -weight array generation for various levels of input blur and noise (1D blur). Horizontal axis: λ values. Red Line: reference for 64 array entries. Blue/orange/green lines—generated arrays for 16/32/64 array entries, respectively.
Figure 12. λ -weight array generation for various levels of input blur and noise (1D blur). Horizontal axis: λ values. Red Line: reference for 64 array entries. Blue/orange/green lines—generated arrays for 16/32/64 array entries, respectively.
Sensors 22 06923 g012
Figure 13. E2E system training schemes: (a) Regression RegParamNet; (b) λ -weight RegParamNet.
Figure 13. E2E system training schemes: (a) Regression RegParamNet; (b) λ -weight RegParamNet.
Sensors 22 06923 g013
Figure 14. Mean result PSNR as a function of input noise.
Figure 14. Mean result PSNR as a function of input noise.
Sensors 22 06923 g014
Figure 15. Effect of E2E training on λ -weight array generation; Blue line: reference g ( λ ) , Orange: standalone f ( λ ) , Green: f ( λ ) after E2E training.
Figure 15. Effect of E2E training on λ -weight array generation; Blue line: reference g ( λ ) , Orange: standalone f ( λ ) , Green: f ( λ ) after E2E training.
Sensors 22 06923 g015
Figure 16. Deblurring using regression RegParamNet (Tikhonov): (a) RD output before E2E training; (b) RD output after E2E training; (c) IEN output before E2E training; (d) IEN output after E2E training.
Figure 16. Deblurring using regression RegParamNet (Tikhonov): (a) RD output before E2E training; (b) RD output after E2E training; (c) IEN output before E2E training; (d) IEN output after E2E training.
Sensors 22 06923 g016
Figure 17. Deblurring using Wiener deconvolution: (a) RD output before E2E training; (b) RD output after E2E training; (c) IEN output before E2E training; (d) IEN output after E2E training.
Figure 17. Deblurring using Wiener deconvolution: (a) RD output before E2E training; (b) RD output after E2E training; (c) IEN output before E2E training; (d) IEN output after E2E training.
Sensors 22 06923 g017
Table 1. Typical performance for different test configurations.
Table 1. Typical performance for different test configurations.
Deblurring ConfigurationPSNR [dB]/SSIM
Initial: Tikhonov Deconvolution λ = 7.5 × 10 3 30.13 ± 3.04/0.926 ± 0.025
Initial + UW-6434.11 ±3.14/0.964 ± 0.015
Initial + UW-12834.46 ± 3.55/0.968 ± 0.016
Initial + residual U-Net35.65 ± 3.95/0.974 ± 0.015
Table 2. Typical deblur quality before and after joint optimization.
Table 2. Typical deblur quality before and after joint optimization.
Deblurring ConfigurationPSNR [dB]/SSIM
Tikhonov Deconvolution λ opt = 7.5 × 10 3 30.13 ± 3.04/0.926 ± 0.024
Tikhonov Deconvolution λ opt * = 3.3 × 10 3 (after joint training)27.66 ± 1.46/0.818 ± 0.014
Initial + residual U-Net λ opt = 7.5 × 10 3 35.65 ± 3.95/0.974 ± 0.015
Initial + residual U-Net: Jointly Trained36.90 ± 3.93/0.980 ± 0.012
Table 3. Mean PSNR/SSIM values for deblurring performance, before and after E2E training.
Table 3. Mean PSNR/SSIM values for deblurring performance, before and after E2E training.
RegParamNetRD before E2ERD after E2EIEN before E2EIEN after E2E
ConfigurationPSNR [dB]/SSIMPSNR [dB]/SSIMPSNR [dB]/SSIMPSNR [dB]/SSIM
λ -Weights (T) 28.61 ± 4.66 / 0.874 ± 0.085 31.67 ± 4.35/0.76 ± 0.1831.67 ± 4.34/0.932 ± 0.05432.0 ± 5.0/0.94 ± 0.056
Regression (T)28.58 ± 4.74/0.875 ± 0.0928.0 ± 5.0/0.80 ± 0.1432.84 ± 5.0/0.94 ± 0.05333.0 ± 5.0/0.94 ± 0.054
λ -Weights (W)25.16 ± 3.23/0.734 ± 0.10725.17 ± 3.0/0.742 ± 0.1027.49 ± 3.1/0.865 ± 0.06427.54 ± 3.19/0.87 ± 0.064
Table 4. Comparison of the proposed method (NANBD) to state-of-the-art approaches. Best values are marked in bold.
Table 4. Comparison of the proposed method (NANBD) to state-of-the-art approaches. Best values are marked in bold.
TestDBCNN [24]MLP [50]Son et al. [51]NANBDNANBD* (BDD Set)
ConfigurationPSNR [dB]/SSIMPSNR [dB]/SSIMPSNR [dB]/SSIMPSNR [dB]/SSIMPSNR [dB]/SSIM
GaussianA28.47/0.879027.16/0.864523.18/0.734729.51/0.873236.45/0.9744
GaussianB25.34/0.781124.48/0.776622.88/0.681429.14/0.86534.97/0.9608
GaussianC22.79/0.719422.31/0.675222.17/0.65928.57/0.8533.62/0.95
GaussianD---21.99/0.547727.92/0.803
GaussianE---25.14/0.722229.99/0.8876
SquareA22.90/0.707822.81/0.697517.74/0.413928.57/0.843234.99/0.9637
SquareB24.01/0.756423.52/0.737519.29/0.478828.91/0.858934.45/0.956
SquareC---21.52/0.712729.99/0.8866
SquareD---24.92/0.51927.61/0.8032
MotionA27.93/0.879526.73/0844827.15/0.852530.65/0.891236.27/0.9748
MotionB25.50/0.800924.77/0.772624.49/0.737829.34/ 0.881935.85/0.9716
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Slutsky, M. Noise-Adaptive Non-Blind Image Deblurring. Sensors 2022, 22, 6923. https://doi.org/10.3390/s22186923

AMA Style

Slutsky M. Noise-Adaptive Non-Blind Image Deblurring. Sensors. 2022; 22(18):6923. https://doi.org/10.3390/s22186923

Chicago/Turabian Style

Slutsky, Michael. 2022. "Noise-Adaptive Non-Blind Image Deblurring" Sensors 22, no. 18: 6923. https://doi.org/10.3390/s22186923

APA Style

Slutsky, M. (2022). Noise-Adaptive Non-Blind Image Deblurring. Sensors, 22(18), 6923. https://doi.org/10.3390/s22186923

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop