Next Article in Journal
Salt-Fog Corrosion Behavior of GCr15 Steels Treated by Ultrasonic Strengthening Grinding Process
Next Article in Special Issue
Comparison of Human Intestinal Parasite Ova Segmentation Using Machine Learning and Deep Learning Techniques
Previous Article in Journal
Artificial Intelligence for Multimedia Signal Processing
Previous Article in Special Issue
An Improved Algorithm for Low-Light Image Enhancement Based on RetinexNet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Interpolation with Regional Gradient Estimation

1
School of Information and Electronic Engineering, Zhejiang Gongshang University, Hangzhou 310018, China
2
Control System Laboratory, Graduate School of Engineering, Kogakuin University, Tokyo 163-8677, Japan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(15), 7359; https://doi.org/10.3390/app12157359
Submission received: 29 June 2022 / Revised: 18 July 2022 / Accepted: 20 July 2022 / Published: 22 July 2022
(This article belongs to the Special Issue Advances in Digital Image Processing)

Abstract

:
This paper proposes an image interpolation method with regional gradient estimation (GEI) to solve the problem of the nonlinear interpolation method not sufficiently considering non-edge pixels. First, the approach presented in this paper expanded on the edge diffusion idea used in CGI and proposed a regional gradient estimation strategy to improve the problem of gradient calculation in the CGI method. Next, the gradient value was used to determine whether a pixel was an edge pixel. Then, a 1D directional filter was employed to process edge pixels while interpolating non-edge pixels using a 2D directionless filter. Finally, we experimented with various representative interpolation methods for grayscale and color images, including the one presented in this paper, and compared them in terms of subjective results, objective criteria, and computational complexity. The experimental results showed that GEI performed better than the other methods in an experiment concerning the visual effect, objective criteria, and computational complexity.

1. Introduction

Image scaling is an important element of image processing and is widely used in applications that include aviation, medicine, communication, meteorology, remote sensing, animation production, film composition, and the military [1,2,3]. Image quality can be improved by scaling images using hardware, but there is a cost. As a result, improving the software aspect, i.e., scaling digital images using interpolation techniques, is critical.
Many image scaling methods have been proposed recently [4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28], and these methods can be divided into two categories, one being sample-based super-resolution reconstruction [4,5,6,7,8,9,10,11,12] and the other being sample-free-based interpolation [13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28]. The main difference between the two is that sample-free-based interpolation uses mathematical methods to estimate pixels directly based on the known pixels, whereas super-resolution reconstruction requires training samples to establish a mapping relationship between low-resolution images and high-resolution images before it can use image block-matching and replacement to complete the interpolation.
Compared to super-resolution reconstruction, sample-free interpolation methods have lower time and space complexities. In terms of modeling characteristics, this type of method can be subdivided into linear interpolation methods and nonlinear interpolation methods. The representative methods of linear interpolation include nearest-neighbor interpolation, bilinear interpolation, and bicubic interpolation [3]. Linear interpolation methods do not consider the pixel’s position to be interpolated during the interpolation process, which causes the blurring of the image’s edges and prevents high-definition visual effects from being achieved.
Nonlinear interpolation methods mainly include methods based on wavelet transform [25,26,27,28] and methods based on edge information [13,14,15,16,17,18,19,20,21,22,23,24]. The interpolation is based on wavelet transform: first, wavelet transform is performed on the image; next, the classical image interpolation method is used to interpolate the frequency domain coefficients; then, threshold processing is performed to obtain the required interpolation image. This method effectively combines the interpolation method and the band-pass filtering feature of the wavelet transform; thus, it can effectively maintain high-frequency detail in the image and improve the visual effect of image interpolation. The approach based on edge information uses a non-directional interpolation method to process non-edge pixels, while for edge pixels, a directional interpolation method is used according to the direction of the edge. This paper mainly studies the interpolation method based on edge information.
In general, image edges are regions where the gray level changes dramatically and its derivative exceeds a threshold, usually by only one or two pixels in width. Such pixels are used in 1D directional interpolation along the edge direction. Meanwhile, non-edge pixels are processed using linear interpolation. However, for non-edge pixels near the edge, if the neighborhood involved in their interpolation crosses the border, the boundaries are blurred after interpolation.
This paper proposes a nonlinear interpolation method called “image interpolation with regional gradient estimation”. Compared to existing nonlinear interpretation methods, the GEI method can distinguish the edge pixels and non-edge pixels of an image more effectively. Specifically, the proposed method employs a regional gradient estimation strategy to reduce the error generated by the CGI method when judging the gradient of unknown pixels, thereby enhancing the image’s interpolation effect.
First, the proposed method utilizes the CGI method’s concept of edge diffusion to diffuse the image edges and determine the properties of the diffused image pixel. Second, the gradients of unknown pixels in high-resolution (HR) images are estimated using bicubic interpolation with regional gradients. The estimated gradient is then applied to determine the properties of the unknown pixel. According to the properties, different interpolation methods are selected, including 2D directionless interpolation for non-edge pixels and 1D directional interpolation for edge pixels. We experimented with various representative interpolation methods for grayscale and color images, including the one presented in this paper, and compared them in terms of subjective results, objective criteria, and computational complexity. The experimental results showed that GEI performed better than the other methods in an experiment concerning the visual effect, objective criteria, and computational complexity.
The structure of this paper is as follows. Section 1 is the introduction. Section 2 briefly describes the image interpolation problem and the idea of edge diffusion. Section 3 describes the principle and process of the proposed method. Section 4 mainly describes the experiment’s design process and results. The paper is concluded in Section 5.

2. Fundamental Issues and Ideas

Image interpolation refers to estimating unknown high-resolution values based on known low-resolution pixels. We denote the low-resolution (LR) image as ILR (i, j), with a size of M × N, and the high-resolution (HR) image with a size of 2M × 2N as IHR (2i, 2j). Figure 1 shows the first step of the interpolation process and demonstrates the relationship between the ILR (i, j) pixels and the IHR (2i, 2j) pixels in the 2 × 2 image interpolation case. To obtain HR image pixel values, we duplicate the LR pixels at (i, j) to (2i − 1, 2j − 1) in the HR image, as shown in Equation (1). Next, we need to construct unknown pixels around known pixel points in IHR (2i, 2j).
I LR ( i , j ) = I HR ( 2 i 1 , 2 j 1 ) , 1   i   M , 1     j     N
To evaluate the interpolation method’s effectiveness, we first downsample the original image; then, we interpolate the downsampled image and, finally, compare the interpolated image to the original image. Currently, the most common downsampling methods are direct-extraction downsampling and function downsampling. Direct-extraction downsampling means keeping the odd rows and columns while deleting the remaining pixels from the HR image; this process yields an LR image, as shown in Figure 1a. Its pixel relationship is shown in Equation (1). Function downsampling makes use of MATLAB’s “imresize” function, which includes three methods: “nearest”, “linear”, and “bicubic”. When acquiring low-resolution images, direct-extraction downsampling is frequently used in the literature [18,19,20,21] on image interpolation methods.
The advantage of direct-extraction downsampling is that the obtained image pixels are part of the original image, allowing the original image information to be preserved to the greatest extent possible. This is consistent with the central question of image interpolation, that is, how to create an image as close to the original as possible with only pixels of the original image? The data obtained by function downsampling are calculated from the original image’s pixel values, and this process destroys the original image’s information. We performed bicubic interpolation on LR images obtained through various downsampling methods and used PSNR [29] to objectively evaluate the interpolated images. Table 1 summarizes the findings. We can see from Table 1 that the LR image obtained by direct extraction displayed superior PSNR performance after interpolation compared to the others, demonstrating that this method can preserve the greatest amount of original image information. Figure 2 depicts the results of the interpolation of the four downsampling methods used on the same test image, labeled “house”. The luminance information of direct-extraction downsampling and nearest downsampling can be seen to be relatively complete. In contrast, the luminance information of bilinear downsampling and bicubic downsampling suffered from some loss. In Section 4, the low-resolution image is obtained by direct-extraction downsampling.
Figure 3a shows a relatively simple image, and Figure 3b shows the edges considered by the standard nonlinear image interpolation method. Point C in Figure 3b is located on the edge considered by the standard nonlinear image interpolation method, and the nonlinear interpolation method can ensure that the pixels on the edge appear clearly in the HR image. Because points A and B in Figure 3b are considered non-edge pixels, linear interpolation is used for them. However, the interpolation neighborhood for pixel B crosses the edge, and if the gray values on both sides of the edge change abruptly, the image edges will be blurred after nonlinear interpolation. To ensure the integrity of the interpolation effect, non-edge pixels within a specific range from the edge should be considered separately; thus, in this work, different interpolation strategies were employed for them. Figure 3c depicts the edge after edge diffusion. Because the edge after diffusion covers pixel B, a nonlinear interpolation method is employed to process it. The variational method of image diffusion is used in the CGI method to achieve edge diffusion [19]. This paper focuses solely on images after gradient edge diffusion has already been performed.
Let us consider image interpolation at the pixel level. An image is made up of discrete pixel points. In Figure 4, the solid squares represent known pixels, and the hollow squares represent unknown pixels. After copying the pixels from the LR image into the HR image, there are three unknown pixels around IHR (2i − 1, 2j − 1), namely the diagonal pixel at (2i, 2j), the column pixel at (2i, 2j − 1), and the row pixel at (2i − 1, 2j). Image edges can only be formed along these four directions in the discrete state; we have simplified the possible edge directions to four cases, that is, θ = 0°, θ = 45°, θ = 90°, and θ = 135°, as shown in Figure 4.
It has been observed that variations in pixel intensity values along the tangent direction of the contrast boundary are always much smaller than those of the direction normal to the boundary. So, we can calculate the pixel gradient along two orthogonal directions to determine whether the pixel is located on the edge; if so, we need to determine the edge direction. Furthermore, if there is a significant difference between the two orthogonal directions, the pixel is classified as an edge pixel. To calculate the gradient difference, we must select a set of orthogonal edge directions. Let us choose 45° and 135° orthogonal directions to determine whether the pixel is on a diagonal edge and 0° and 90° orthogonal directions to determine whether the pixel is on a horizontal or vertical edge.
We use four 3 × 3 convolution masks to calculate the gradient approximations in these four directions, as shown in the following formula, where Uθ denotes the value of the gradient value of ILR (i, j) in the θ direction.
U 0 ( i , j ) = | [ 1 0 1 1 0 1 1 0 1 ]     I LR ( i , j ) |
U 45 ( i , j ) = | [ 0 0 1 1 0 1 1 1 0 ]     I LR ( i , j ) |
U 90 ( i , j ) = | [ 1 1 1 0 0 0 1 1 1 ]     I LR ( i , j ) |
U 135 ( i , j ) = | [ 1 1 0 1 0 1 0 1 1 ]     I LR ( i , j ) |
Secondly, we need to assess the properties of the pixel in the LR image. Specifically, we need to judge whether a pixel is an edge pixel. If so, what is the direction of the edge on which it is located? The specific method of using the gradient to determine the properties of pixels is as follows.
As shown in Figure 1, for diagonal pixels, such as the points at the (2i − 1, 2j − 1) and (2i, 2j) pixel positions, if a pixel in the 45° or 135° direction satisfies |u45°u135°| ≥ T (uθ denotes the θ-directional gradient for pixels), the pixel can be regarded as an edge point, and the direction of the edge on which the pixel is located can then be determined by Equation (6), where u 45 ( h ) denotes the pixel’s 45°-directional gradient in the HR image.
{ θ = 135 ° , u 45 ( h )     u 135 ( h ) θ = 45 ° ,     u 45 ( h ) < u 135 ( h )
Similarly, for a pixel in the horizontal or vertical direction, e.g., the pixels at the (2i − 1, 2j) and (2i, 2j − 1) positions, if the pixel gradient u in the 0° or 90° direction satisfies |uu90°| ≥ T, the pixel can be determined as an edge point, and the edge where the pixel is located can be determined by Equation (7). According to the experiment in [19], the value of T is about 0.01.
{ θ = 90 ° , u 0 ( h )     u 90 ( h ) θ = 0 ° ,     u 0 ( h ) < u 90 ( h )
To obtain the HR image, we first copy the pixel values and properties at the (i, j) position in the LR image to the (2i − 1, 2j − 1) position in the HR image, as shown in Equation (8).
I HR ( 2 i 1 , 2 j 1 ) = I LR ( i , j )
To interpolate the remaining unknown pixels in the HR image, we need to judge the properties of the unknown pixels according to the known pixel properties. As shown in Equation (8) [12], the CGI method uses the nearest-neighbor strategy to determine the properties of unknown pixels.
{ u θ ( h ) ( 2 i 1 , 2 j ) = u θ ( h ) ( 2 i 1 , 2 j 1 ) u θ ( h ) ( 2 i , 2 j 1 ) = u θ ( h ) ( 2 i   1 , 2 j 1 ) u θ ( h ) ( 2 i , 2 j ) = u θ ( h ) ( 2 i 1 , 2 j 1 )
The CGI method is not very precise in judging the properties. When the image’s edge information is rich, this strategy may incorrectly judge the properties of unknown pixels, affecting the final interpolation effect. The arrows in Figure 5a represent the actual gradient directions of the a and b pixels in the test image. Because the CGI method uses the nearest-neighbor strategy, the gradient and direction of pixel b in Figure 5b are consistent with pixel a, and the direction is 135°. At this point, pixel b is interpolated as an edge pixel in the 135° direction. However, Figure 5a shows that the gradient direction of pixel b is 0°. So, we propose a region-based gradient estimation strategy to effectively judge the properties of pixels. The details are discussed in the following section.

3. Image Interpolation with Regional Gradient Estimation

3.1. Regional Gradient Estimation

We introduced a bicubic interpolation method to calculate the gradient value of unknown pixels. Bicubic interpolation uses the gradient values of 16 points around the point to be interpolated for cubic interpolation; this approach not only considers the influence of the gradient of the four directly adjacent points but also considers the influence of the gradient value change rate between adjacent points. However, if bicubic interpolation is performed on all unknown pixels in an image, the image’s local characteristics are ignored, resulting in a large computational burden. In reality, images have both flat areas and areas with rich texture details. The image gradient changes smoothly in the flat area. For such areas, the effect obtained by a complex method is the same as that obtained by a simple method, but the calculation requirements increase significantly. Therefore, by using an interpolation method with lower computational requirements for flat areas and using bicubic interpolation in areas with intricate details, the amount of computation can be reduced while maintaining the quality of the enlarged image. Based on this consideration, we need to improve the bicubic interpolation method according to the region.
To obtain an HR image, as shown in Equation (8), we first copied the pixel values and properties at the (i, j) position in the LR image to the (2i − 1, 2j − 1) position in the HR image. Next, we proposed a gradient estimation strategy. As shown in Figure 6, for the pixel at position (i + v, j + u), we calculated the gradient variance Var of the four-pixel in the (i + v, j + u) position neighborhood in the original image. If the variance was less than the threshold T, the average of these four gradient values was taken as the gradient value of the pixel at (i + v, j + u); otherwise, we used bicubic interpolation to calculate the gradient. Equation (10) is the formula for calculating the variance Var.
V a r = ( U u 11 ) 2 + ( U u 12 ) 2 + ( U u 21 ) 2 + ( U u 22 ) 2
In Equation (11), u11, u12, u21, and u22 are the gradient values of the four pixels around position (i + u, j + v) in the LR image, and U is the mean value of the gradient of the four pixels. Calculating U requires one multiplication and three additions. From Equation (10), we need four multiplications and seven additions to calculate the variance. In total, five multiplications and ten additions are required, much less than the 70 multiplications and 45 additions required to calculate the pixel gradients by bicubic interpolation. According to the experiments, the reduction in computation and the maintenance of image quality remain relatively reasonable when the threshold T = 18 [30]. Although the variance is calculated once for each image point, the total number of operations is still significantly reduced compared to bicubic interpolation.
Figure 7 shows the visualization results of gradient calculations for the same test image, using both the method in this paper and the CGI method. Figure 7a shows the test image, and Figure 7b shows the result of the gradient visualization of the test image. The gradient visualization of the HR image that was estimated from the LR image using the CGI method is shown in Figure 7c. The result of the GEI method is shown in Figure 7d. The gradient calculated by the CGI method produced significant artifacts, especially at edge intersections, with significant errors when compared to the original image gradient, as shown in the yellow box in Figure 7c. In contrast, the gradient visualization estimated by the proposed method was close to the original image.

3.2. Image Interpolation

After estimating the gradient of the unknown pixel in the high-resolution image, as mentioned in Section 2, we used the gradient to judge the properties of the unknown pixel and then calculate the pixel value according to the following steps. First, we calculated the gray value IHR (2i, 2j) at position (2i, 2j) in the HR image according to the gray value ILR (i, j) at position (i, j) in the LR image. Second, IHR (2i − 1, 2j) and IHR (2i, 2j − 1) were calculated from ILR (i, j) and IHR (2i, 2j), respectively. We adopted this order of calculation because the calculation of IHR (2i − 1, 2j) and IHR (2i, 2j − 1) requires ILR (i, j) and IHR (2i, 2j). Finally, each non-edge pixel was interpolated using bicubic interpolation. For edge pixels, we used Equation (11) for interpolation [20].
I = ω ( I a + I b ) + ( 0.5 ω ) ( I c + I d )
ω is an adjustable parameter. We set up an interpolation experiment to estimate the optimal value of ω. Figure 8 shows PSNR results of our proposed method averaged for 12 test images using different values of ω. In this work, we obtained the highest PSNR value when ω = 0.575; therefore, we regarded this value as the optimal value of ω and applied it to all our experiments.
I represents the gray value of the pixel to be interpolated; Ia, Ib, Ic, and Id represent four known pixel values around the unknown pixel, and their values depend on the properties of the unknown pixel. The specific process is as follows [20].
Let us first address pixels on the diagonal in the HR image. If θ = 135°,
I a = I HR ( i 1 , j 1 ) , I b = I HR ( i + 1 , j + 1 )
I c = I HR ( i 3 , j 3 ) , I d = I HR ( i + 3 , j + 3 )
if θ = 45°,
I a = I HR ( i + 1 , j 1 ) , I b = I HR ( i 1 , j + 1 )
I c = I HR ( i + 3 , j 3 ) , I d = I HR ( i 3 , j + 3 )
Let us now address pixels in the horizontal or vertical direction in the HR image. If θ = 90°,
I a = I HR ( i 1 , j ) , I b = I HR ( i + 1 , j )
I c = I HR ( i 3 , j ) , I d = I HR ( i + 3 , j )
if θ = 0°,
I a = I HR ( i , j 1 ) , I b = I HR ( i , j + 1 )
I c = I HR ( i , j 3 ) , I d = I HR ( i , j + 3 )
where θ is the edge direction of the edge where the pixel is located and can be derived from Equations (2)–(5).
So far, this paper has only addressed 2 × 2 interpolation. For example, when the image enlargement factor is four, 2 × 2 interpolation is performed on the original image twice. For the interpolation factor M that is not a power of two, one can first apply N × N image interpolation (where N is a power of two and smaller than but as close as possible to M), followed by 2D bicubic image interpolation with a rational number as the image enlargement factor. For example, if an image requires 9 × 9 image interpolation, we can first use the proposed method to interpolate the image by 8 × 8, followed by a 9/8 × 9/8 bicubic image interpolation.
Figure 9 shows the structure of the proposed method. First, we identified and diffused the LR image’s edge pixels. Second, the gradient and pixel properties of the HR image were estimated. Finally, the pixels of the HR image were separated into edge and non-edge pixels. Multiple interpolation methods were then used to interpolate the classified pixels.

4. Experiments

In order to evaluate the method presented in this paper, seven representative current methods were selected for experimentation: IBI [16], IEDI [17], CGI [18], CED [19], PGI [20], PCI [21], and WTCGI [25]. The source code of each method was either written by the method’s proponent or implemented by the authors of this paper, and the individual parameter configurations used were the default parameters recommended by each method’s proponent.
The proposed method (GEI) is based on the concept of CGI edge diffusion and uses CED to replace the iterative process in the CGI method to reduce time complexity. In contrast, to improve upon the CGI method, PCI and PGI employ the prediction-correction and gradient-prediction strategies, respectively. WTCGI combines wavelet transformation and the contrast-guided approach from CGI. Our experiment included the above five methods, as well as IEDI and IBI. IEDI is an improvement of the edge-direction interpolation method, and IBI is an improvement of the bicubic method, so comparing them to the method proposed in this paper is meaningful.
The twelve test images used in the image experiment are shown in Figure 10. The labels and sizes are as follows: cameraman (256 × 256 pixels), house (256 × 256 pixels), butterfly (512 × 521 pixels), bike (500 × 500 pixels), boats (512 × 512 pixels), wheel (512 × 512 pixels), airplane (256 × 256 pixels), stars (600 × 600 pixels), Barbara (512 × 512 pixels), fence (500 × 500 pixels), peppers (256 × 256 pixels), and baboon (256 × 256 pixels). These twelve images are frequently used in image processing and have a high reference value. We downsampled the reference images by direct extraction to generate LR images and then performed a 2 × 2 image interpolation on the LR images to generate HR images of the same size as the reference images, enabling us to compare the experimental outcomes of each approach.
Table 2 displays the PSNR results obtained from all of the interpolation methods under consideration, demonstrating that PCI and our proposed method outperformed all other methods. Furthermore, when comparing the methods for all of the test images, the average PSNR gain produced by the proposed method was close to that of PCI and better than all methods other than PCI. The proposed method provided the best PSNR performance for images with sharper and more distinct edges, such as the test images “baboon” and “wheel”.
Aside from the common practice of measuring the PSNR for objective performance evaluation, another image quality assessment metric known as structural similarity (SSIM) [29] was used to assess image quality because it correlates well with human visual perception. Table 3 reports the SSIM results (in dB) for all of the interpolation methods under consideration, demonstrating that the proposed method provided better SSIM performance for images with sharper and more distinct edges, such as the test images “butterfly”, ”wheel”, and “baboon”. Although the PCI method was significantly better than all other methods, the proposed method was close to PCI, CED and better than all methods other than PCI and CED in terms of average SSIM values.
The ability of a method to preserve the edges of an image is measured by the EPI (edge preservation index) [31]. Table 4 compares the edge preservation of each method when processing 12 images. The closer the EPI value is to 1, the better the method’s edge preservation ability. The average EPI gain produced by the proposed method was better than other methods. Furthermore, the proposed method achieved the highest EPI values for the images “cameraman”, “Barbara”, ”butterfly”, “baboon”, “airplane”, and “house”. These results show that the proposed method was effective in preserving the edges of the images.
The low computational complexity of our proposed method is an additional attractive feature. The average runtimes of each method for the interpolation of the 12 test images are documented in Table 5. Because the IBI method is based on the bicubic method, it does not have to consider image structure as much, making it the fastest. The CED method, which had the lowest runtime after IBI, replaces the edge-diffusion iterative process in the CGI method with thermal diffusion. Because of the inclusion of the regional gradient estimation strategy, the time complexity of the method proposed in this paper was slightly higher than that of PCI. IEDI was the slowest of all methods, with an average runtime nearly 11 times that of the method in this paper. The original author of the WTCGI method only provided C++ code, so it was not included in this comparison.
For the subjective performance evaluation, we compared the performance of the methods in terms of visual effect. We conducted interpolation by 4 × 4 for the Wheel via 8 methods respectively, and select upper right part of “wheel” for subjective evaluation, as shown in Figure 11. The CGI and CED methods generated false edges at the junction edges, as shown in the yellow anchor boxes in Figure 11b,c; this was due to an ineffective gradient selection strategy, as shown in Equation (9). Although the CED method replaces the edge-diffusion iterative process in the CGI method with thermal diffusion, which increases running speed, false edges remain a problem. The PCI and PGI methods improve on the CGI method by using prediction correction and gradient prediction, respectively, to avoid the generation of false edges and produce sharper edges than the CGI method; however, their results yielded some edge loss in the red sight frames, as shown in Figure 11e,f. Although IEDE yielded more complete edges than the above-mentioned methods, ‘speckle’-like noise was present, as shown in Figure 11g. WTCGI and IBI blurred more edges and generated more distinct artifacts, such as jagged edges, as seen in Figure 11h,i. The proposed method produced clear edges, avoided false and missing edges, and caused minimal artifacts, as shown in Figure 11d.
For the color images, CGI and the proposed method were applied to interpolate the R, G, and B channels of the color images and combine them into a single image. Figure 12 shows the interpolation results obtained by the CGI method and the proposed method for the color images ‘butterfly’ and ‘bike’. Figure 12b shows that the CGI method produced a ringing effect and artifacts in the local area. In contrast, GEI maintained the edges better in rich textural detail, as Figure 12c shows. Due to the introduction of a regional gradient estimation strategy, the proposed method had a slight advantage over the CGI method in interpolating color images.

5. Conclusions

The processing of edges is essential for high-resolution image interpolation. The standard nonlinear interpolation methods tend to blur edges. To address the shortcomings of traditional nonlinear interpolation methods for non-edge pixels, this paper proposes an image interpolation method based on a region gradient estimation strategy.
To begin, the proposed method used the CGI method’s edge diffusion idea to perform edge diffusion on low-resolution images. Then, this paper proposed a gradient estimation strategy to estimate the gradient of unknown pixels in high-resolution images to improve the shortcomings of the CGI in calculating the gradient, namely, using region-based bicubic interpolation to estimate the gradient of the pixel and judging its properties. This approach can determine the properties more effectively because this strategy takes into account the image’s local characteristics. Following that, a 1D directional filter was used to process edge pixels, while a 2D directionless filter interpolated non-edge pixels. To validate the efficacy of the proposed method, experiments were conducted using a variety of representative interpolation methods for both grayscale and color images; we compared the results in terms of visual effect, objective criteria, and computational complexity. Extensive simulation results from both grayscale and color test images showed that our proposed image interpolation approach using regional gradient estimation outperformed various representative image interpolation methods with respect to both objective criteria and subjective visual quality assessments. Furthermore, when compared to other methods, the low computational complexity of the proposed method was a clear advantage.

Author Contributions

Conceptualization, Z.J.; methodology, Z.J.; software, Z.J.; validation, Z.J.; formal analysis, Z.J.; investigation, Z.J.; resources, Z.J.; data curation, Z.J.; writing—original draft preparation, Z.J.; writing—review and editing, Z.J. and Q.H.; visualization, Z.J. and Q.H.; supervision, Q.H.; project administration, Q.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

Data sharing is not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Karwowska, K.; Wierzbicki, D. Using Super-Resolution Methods for Small Satellite Imagery: A Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3292–3312. [Google Scholar] [CrossRef]
  2. Prakoso, W.S.; Soesanti, I.; Wibirama, S. Enhancement Methods of Brain MRI Images: A Review. In Proceedings of the 2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE), Yogyakarta, Indonesia, 6–8 October 2020. [Google Scholar]
  3. Singh, A.; Singh, J. Review and Comparative Analysis of Various Image Interpolation Techniques. In Proceedings of the 2019 2nd International Conference on Intelligent Computing, Instrumentation and Control Technologies (ICICICT), Kannur, India, 5–6 July 2019. [Google Scholar]
  4. Jin, Y.; Zhang, Y.; Cen, Y.G.; LI, Y.D.; Mladenovic, V.; Voronin, V. Pedestrian detection with super-resolution reconstruction for low-quality image. Pattern Recognit. 2021, 115, 107846. [Google Scholar] [CrossRef]
  5. Yang, W.; Zhang, X.; Tian, Y.; Wang, W.; Xue, J.; Liao, Q. Deep Learning for Single Image Super-Resolution: A Brief Review. IEEE Trans. Multimed. 2019, 21, 3106–3121. [Google Scholar] [CrossRef] [Green Version]
  6. Barman, T.; Deka, B.; Prasad, A.V.V. GPU-Accelerated Adaptive Dictionary Learning and Sparse Representations for Multispectral Image Super-Resolution. In Proceedings of the 2021 IEEE 18th India Council International Conference (INDICON), Guwahati, India, 19–21 December 2021. [Google Scholar]
  7. Dong, W.; Zhang, L.; Shi, G. Enhanced Non-Local Total Variation Model and Multi-Directional Feature Prediction Prior for Single Image Super-Resolution. IEEE Trans. Image Process. 2019, 28, 3778–3793. [Google Scholar]
  8. Shi, J.; Qi, C. Low-rank sparse representation or single image super-resolution via self-similarity learning. In Proceedings of the 23rd IEEE International Conference on Image Processing, Settle, WA, USA, 25–28 September 2016. [Google Scholar]
  9. Li, J.; Wu, J.; Deng, H.; Liu, J. A self-learning image super-resolution method via sparse representation and non-local similarity. Neurocomputing 2016, 184, 196–206. [Google Scholar] [CrossRef]
  10. Deng, C.; Xu, J.; Zhang, K.; Tao, D.; Gao, X.; Li, X. Similarity constrains-based structured output regression machine: An approach to image super-resolution. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 2472–2485. [Google Scholar] [CrossRef] [PubMed]
  11. Wang, H.; Gao, X.; Zhang, K.; Li, J. Single image super-resolution using Gaussian process regression with dictionary-based sampling and Student-t likelihood. IEEE Trans. Image Process. 2017, 26, 3556–3568. [Google Scholar] [CrossRef] [PubMed]
  12. Mei, D.F.; Zhu, X.; Wang, X.X.; Ai, N. Image super-resolution based on structural dissimilarity learning dictionary. In Proceedings of the 2017 International Conference on the Frontiers and Advances in Data Science (FADS), Xi’an, China, 23–25 October 2017. [Google Scholar]
  13. Sánchez-García, E.; Balaguer-Beser, Á.; Almonacid-Caballer, J.; Pardo-Pascual, J.E. A New Adaptive Image Interpolation Method to Define the Shoreline at Sub-Pixel Level. Remote Sens. 2019, 11, 1880. [Google Scholar] [CrossRef] [Green Version]
  14. Zeng, X.Y.; Lu, H.C.; Zhang, C.Y. Super-Resolution Method: Based on Wavelet and Interpolation. In Proceedings of the 2020 3rd World Conference on Mechanical Engineering and Intelligent Manufacturing (WCMEIM), Shanghai, China, 4 December 2020. [Google Scholar]
  15. Liu, Y.; Huang, Q.J.; Sato, K. Differential filtering method for robot welding seam image enhancement. J. Phys. Conf. Ser. 2020, 1607, 012045. [Google Scholar] [CrossRef]
  16. Li, Y.M.; Qi, F.F.; Wan, Y. Improvements on Bicubic Image Interpolation. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 February 2020. [Google Scholar]
  17. Hossain, M.S.; Jalab, H.; Kahtan, H.; Abdullah, A. Image Resolution Enhancement Using Improved Edge Directed Interpolation Method. In Proceedings of the 2019 9th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 29 November–1 December 2019. [Google Scholar]
  18. Zhe, W.; Ma, K. Contrast-Guided Image Interpolation. IEEE Trans. Image Process. 2013, 22, 4271–4285. [Google Scholar]
  19. Zhe, W.; Ma, K. Convolutional edge diffusion for fast contrast-guided image interpolation. IEEE Signal Process. Lett. 2016, 23, 1260–1264. [Google Scholar]
  20. Lu, Z.F.; Zhong, B.J. Image Interpolation with Predicted Gradients. Acta Autom. Sin. 2018, 44, 1072–1085. [Google Scholar]
  21. Zhong, B.J.; Ma, K.K.; Lu, Z.F. Predictor-corrector image interpolation. J. Vis. Commun. Image Represent. 2019, 61, 50–60. [Google Scholar] [CrossRef]
  22. Chao, Y.; Ma, C.X.; Shan, W.T.; Feng, J.P.; Zhang, Z.S. Adaptive Directional Cubic Convolution for Integrated Circuit Chip Defect Image Interpolation. Int. J. Circuits Syst. Signal Process. 2021, 15, 1084–1090. [Google Scholar] [CrossRef]
  23. Gupta, S.; Sharma, D.K.; Ranta, S. A new hybrid image enlargement method using singular value decomposition and cubic spline interpolation. Multimed. Tools Appl. 2022, 81, 4241–4254. [Google Scholar] [CrossRef]
  24. Ge, R. A Design of Optimized Colour Image Interpolation Method Based on Edge Gradient. In Proceedings of the 2022 International Conference on Big Data, Information and Computer Network (BDICN), Sanya, China, 20–22 January 2022. [Google Scholar]
  25. Zhao, Y.Y.; Huang, Q.J. Image enhancement of robot welding seam based on wavelet transform and contrast guidance. Int. J. Innov. Comput. Inf. Control 2022, 18, 149–159. [Google Scholar]
  26. Lama, R.K.; Shin, S.; Kang, M. Interpolation Using Wavelet Transform and Discrete Cosine Transform for High-Resolution Display. In Proceedings of the 2016 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 7–11 January 2016. [Google Scholar]
  27. Cao, M.; Liu, Z.; Huang, X.; Shen, Z. Research for Face Image Super-Resolution Reconstruction Based on Wavelet Transform and SRGAN. In Proceedings of the 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 March 2021. [Google Scholar]
  28. Ye, H.; Su, K.; Huang, S. Image Enhancement Method Based on Bilinear Interpolating and Wavelet Transform. In Proceedings of the 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 March 2021. [Google Scholar]
  29. Sabilla, I.A.; Meirisdiana, M.; Sunaryono, D.; Husni, M. Best Ratio Size of Image in Steganography Using Portable Document Format with Evaluation RMSE, PSNR, and SSIM. In Proceedings of the 2021 4th International Conference of Computer and Informatics Engineering (IC2IE), Depok, Indonesia, 14–15 September 2021. [Google Scholar]
  30. Wang, H.P.; Zhou, L.L.; Zhang, J. Region-based Bicubic Image Interpolation Method. Comput. Eng. 2010, 36, 216–218. [Google Scholar]
  31. Chen, L.M.; Jiang, F.; Zhang, H.F.; Wu, S.B.; Yu, S.D.; Xie, Y.Q. Edge preservation ratio for image sharpness assessment. In Proceedings of the 2016 12th World Congress on Intelligent Control and Automation (WCICA), Guilin, China, 12–15 June 2016. [Google Scholar]
Figure 1. Illustration of the interpolation. The dots denote the known pixels, and the circles denote the missing pixels. (a) low-resolution (LR) image; (b) high-resolution (HR) image.
Figure 1. Illustration of the interpolation. The dots denote the known pixels, and the circles denote the missing pixels. (a) low-resolution (LR) image; (b) high-resolution (HR) image.
Applsci 12 07359 g001
Figure 2. Interpolation results of different downsampling methods. (a) test image; (b) direct extraction; (c) nearest; (d) bilinear; (e) bicubic.
Figure 2. Interpolation results of different downsampling methods. (a) test image; (b) direct extraction; (c) nearest; (d) bilinear; (e) bicubic.
Applsci 12 07359 g002
Figure 3. Two different ways of interpreting the edges of an image. (a) test image; (b) canny edge; (c) edge using the CGI method.
Figure 3. Two different ways of interpreting the edges of an image. (a) test image; (b) canny edge; (c) edge using the CGI method.
Applsci 12 07359 g003
Figure 4. Illustration of edge directions in a digital image. The orange arrows represent the 45° and 135° directions to form a set of orthogonal groups, and the blue arrows represent the 90° and 0° directions to form a set of orthogonal groups.
Figure 4. Illustration of edge directions in a digital image. The orange arrows represent the 45° and 135° directions to form a set of orthogonal groups, and the blue arrows represent the 90° and 0° directions to form a set of orthogonal groups.
Applsci 12 07359 g004
Figure 5. Comparison of local gradient calculation results. The red arrow represents the direction in which the pixel intensity value changes the most, which can also be called the gradient value. (a) the gradient of pixel a and pixel b in the original image; (b) the gradient calculation result of the CGI method.
Figure 5. Comparison of local gradient calculation results. The red arrow represents the direction in which the pixel intensity value changes the most, which can also be called the gradient value. (a) the gradient of pixel a and pixel b in the original image; (b) the gradient calculation result of the CGI method.
Applsci 12 07359 g005
Figure 6. Bicubic interpolation mapping.
Figure 6. Bicubic interpolation mapping.
Applsci 12 07359 g006
Figure 7. Comparison of visualization results for gradient estimation. (a) test image; (b) gradient calculation result of the test image; (c) gradient estimation result of the CGI method; (d) gradient estimation result of the GEI method.
Figure 7. Comparison of visualization results for gradient estimation. (a) test image; (b) gradient calculation result of the test image; (c) gradient estimation result of the CGI method; (d) gradient estimation result of the GEI method.
Applsci 12 07359 g007
Figure 8. PSNR results averaged for 12 test images for different values of ω.
Figure 8. PSNR results averaged for 12 test images for different values of ω.
Applsci 12 07359 g008
Figure 9. Structure of the proposed method.
Figure 9. Structure of the proposed method.
Applsci 12 07359 g009
Figure 10. Test images. (a) cameraman; (b) house; (c) butterfly; (d) bike; (e) boats; (f) wheel; (g) airplane; (h) baboon; (i) stars; (j) barbara; (k) peppers; (l) fence.
Figure 10. Test images. (a) cameraman; (b) house; (c) butterfly; (d) bike; (e) boats; (f) wheel; (g) airplane; (h) baboon; (i) stars; (j) barbara; (k) peppers; (l) fence.
Applsci 12 07359 g010
Figure 11. Comparison of interpolation results for top right part of the "wheel” image by different interpolation methods. (a) Upper right part of the "wheel” after interpolation; (b) CGI; (c) CED; (d) GEI; (e) PGI; (f) PCI; (g) IEDI; (h) WTCGI; (i) IBI.
Figure 11. Comparison of interpolation results for top right part of the "wheel” image by different interpolation methods. (a) Upper right part of the "wheel” after interpolation; (b) CGI; (c) CED; (d) GEI; (e) PGI; (f) PCI; (g) IEDI; (h) WTCGI; (i) IBI.
Applsci 12 07359 g011
Figure 12. The comparison of color images interpolated by CGI and our method. (a) test image; (b) interpolation result by CGI; (c) interpolation result by GEI.
Figure 12. The comparison of color images interpolated by CGI and our method. (a) test image; (b) interpolation result by CGI; (c) interpolation result by GEI.
Applsci 12 07359 g012
Table 1. A comparison of different downsampling methods with respect to the PSNR (dB).
Table 1. A comparison of different downsampling methods with respect to the PSNR (dB).
MethodInterlaced Row/ColumnNearestBilinearBicubic
PSNR31.1327.5626.0925.17
Table 2. A comparison of different interpolation methods with respect to the PSNR (dB).
Table 2. A comparison of different interpolation methods with respect to the PSNR (dB).
Test ImagesCGI
[18]
CED
[19]
PCI
[20]
IEDI
[17]
PGI
[21]
WTCGI
[25]
IBI
[16]
GEI
(Proposed)
Bike25.8225.8225.9025.1725.9225.2125.3225.85
Wheel21.0120.9821.2220.3120.8120.5721.0921.32
Boats29.5129.5629.7729.2429.4129.3229.3529.42
Butterfly29.2729.2429.3128.9729.2528.9729.1329.26
House32.8332.7132.8832.3132.2731.8732.5932.84
Cameraman25.8625.9025.8125.4825.8525.7625.4525.83
Baboon22.5022.4122.5322.4122.5122.3522.3422.59
Peppers30.8830.7730.8730.4730.7930.1930.6530.81
Fence25.7025.6325.8425.6125.7625.6925.4925.75
Airplane26.5426.4926.5926.6026.5226.1026.4326.61
Barbara23.7523.6423.8223.5423.6823.4123.3924.01
Stars34.1333.9434.3833.3634.2333.7133.5434.33
Average27.3227.2627.4126.9627.2526.9327.0627.39
Table 3. A comparison of different interpolation methods with respect to the SSIM (dB).
Table 3. A comparison of different interpolation methods with respect to the SSIM (dB).
Test ImagesCGI
[18]
CED
[19]
PCI
[20]
IEDI
[17]
PGI
[21]
WTCGI
[25]
IBI
[16]
GEI
(Proposed)
Bike0.88080.88120.88030.87510.87850.87910.87830.8798
Wheel0.86210.86260.86680.86440.86320.86490.86540.8665
Boats0.87630.88010.87940.87710.88120.87440.87910.8796
Butterfly0.97210.97320.97200.97180.97250.96980.97080.9758
House0.87810.87780.87890.87830.87790.87750.87660.8780
Cameraman0.87110.87320.87150.87040.87100.86920.87020.8732
Baboon0.91250.91110.91300.91210.91740.91120.91140.9165
Peppers0.90320.90410.90350.90290.89120.90260.90260.9025
Fence0.77520.77800.77850.77630.77570.77650.77820.7723
Airplane0.94050.94100.94010.93890.94380.94220.94090.9412
Barbara0.91250.91280.91300.91140.91190.91050.91260.9118
Stars0.95840.96030.96170.96080.96050.96100.95890.9608
Average 0.89520.89630.89660.89500.89540.89490.89540.8965
Table 4. A comparison of different interpolation methods with respect to the EPI (dB).
Table 4. A comparison of different interpolation methods with respect to the EPI (dB).
Test ImagesCGI
[18]
CED
[19]
PCI
[20]
IEDI
[17]
PGI
[21]
WTCGI
[25]
IBI
[16]
GEI
(Proposed)
Bike0.82580.82430.83250.82740.83130.82650.82360.8302
Wheel0.81860.81890.82950.81110.82560.82740.81210.8259
Boats0.79630.78650.79110.78950.79010.78980.78830.7916
Butterfly0.87290.87750.87210.87030.87050.87490.86120.8753
House0.75410.75330.76110.76150.76050.75960.75660.7624
Cameraman0.75540.75250.75440.74940.75400.75560.75020.7601
Baboon0.80290.80350.80510.80460.80690.80150.80140.8072
Peppers0.76570.76540.76820.76250.76800.76520.77260.7677
Fence0.70980.70850.71210.70740.71120.70980.70120.7091
AirPlane0.81090.81120.82250.81120.82350.81650.81190.8254
Barbara0.74500.74650.75190.74850.75150.75010.75260.7535
Stars0.79830.79900.80130.79960.80120.80560.79890.8026
Average 0.78920.78730.79180.78690.79120.79020.78590.7926
Table 5. Average runtime (in seconds) for interpolating 12 test images.
Table 5. Average runtime (in seconds) for interpolating 12 test images.
CGI
[18]
CED
[19]
PCI
[20]
IEDI
[17]
PGI
[21]
WTCGI
[25]
IBI
[16]
GEI
(Proposed)
3.100.891.3117.961.54-0.171.45
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jia, Z.; Huang, Q. Image Interpolation with Regional Gradient Estimation. Appl. Sci. 2022, 12, 7359. https://doi.org/10.3390/app12157359

AMA Style

Jia Z, Huang Q. Image Interpolation with Regional Gradient Estimation. Applied Sciences. 2022; 12(15):7359. https://doi.org/10.3390/app12157359

Chicago/Turabian Style

Jia, Zuhang, and Qingjiu Huang. 2022. "Image Interpolation with Regional Gradient Estimation" Applied Sciences 12, no. 15: 7359. https://doi.org/10.3390/app12157359

APA Style

Jia, Z., & Huang, Q. (2022). Image Interpolation with Regional Gradient Estimation. Applied Sciences, 12(15), 7359. https://doi.org/10.3390/app12157359

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop