Next Article in Journal
New Marginal Spectrum Feature Information Views of Humpback Whale Vocalization Signals Using the EMD Analysis Methods
Next Article in Special Issue
A Heart Image Segmentation Method Based on Position Attention Mechanism and Inverted Pyramid
Previous Article in Journal
Peripheral Blood Leukocyte Detection Based on an Improved Detection Transformer Algorithm
Previous Article in Special Issue
A Salient Object Detection Method Based on Boundary Enhancement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast-MFQE: A Fast Approach for Multi-Frame Quality Enhancement on Compressed Video

1
College of Information Science and Engineering, Huaqiao University, Xiamen 361021, China
2
College of Engineering , Huaqiao University, Quanzhou 362021, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(16), 7227; https://doi.org/10.3390/s23167227
Submission received: 18 July 2023 / Revised: 11 August 2023 / Accepted: 12 August 2023 / Published: 17 August 2023
(This article belongs to the Special Issue Machine Learning Based 2D/3D Sensors Data Understanding and Analysis)

Abstract

:
For compressed images and videos, quality enhancement is essential. Though there have been remarkable achievements related to deep learning, deep learning models are too large to apply to real-time tasks. Therefore, a fast multi-frame quality enhancement method for compressed video, named Fast-MFQE, is proposed to meet the requirement of video-quality enhancement for real-time applications. There are three main modules in this method. One is the image pre-processing building module (IPPB), which is used to reduce redundant information of input images. The second one is the spatio-temporal fusion attention (STFA) module. It is introduced to effectively merge temporal and spatial information of input video frames. The third one is the feature reconstruction network (FRN), which is developed to effectively reconstruct and enhance the spatio-temporal information. Experimental results demonstrate that the proposed method outperforms state-of-the-art methods in terms of lightweight parameters, inference speed, and quality enhancement performance. Even at a resolution of 1080p, the Fast-MFQE achieves a remarkable inference speed of over 25 frames per second, while providing a PSNR increase of 19.6% on average when QP = 37.

1. Introduction

Nowadays, there is a surplus of ultra-high-definition (UHD) videos accessible for online viewing, imposing substantial strain on communication bandwidth. To transmit videos within the constraint of limited network bandwidth, video compression is vital for reducing the bit rate. However, highly efficient video coding standards, such as H.264/AVC [1] and H.265/HEVC [2], introduce artifacts when using de-correlation and predictive coding techniques, degrading the quality of the video to some extent [3]. As illustrated in Figure 1, after being transmitted in low bandwidth, the reconstructed video is of low quality. The artifacts (i.e., blurring, ringing, blocking and distortion in motion, etc.) are obvious. When these videos are used for subsequent visual tasks, such as object recognition, object detection, and object tracking, etc., the low quality affects the performance dramatically [4,5]. Therefore, quality enhancement for compressed video is crucial for video applications and has emerged as a crucial area of research.
For image or single-frame quality enhancement, traditional methods [6,7,8,9,10,11] aimed to enhance the quality of compressed JPEG images by optimizing the transform coefficients of a specific compression standard. Specifically, refs. [8,9] proposed Shaped-Adaptive DCT (SADCT) and Regression Tree Fields (RTF) to reduce JPEG image blocking artifacts, respectively. Nevertheless, it is challenging to apply these methods to other compression tasks due to the limited generalization ability. With the advance of the deep learning method, an expanding range of methods have embraced convolutional neural network (CNN) approaches [12,13,14,15,16] to improve the compressed image quality. In [12], a four-layer AR-CNN was first introduced to deal with various artifacts in JPEG images. Based on this, Zhang et al. [13] proposed a deep DnCNN for multi-image restoration. Then, based on residual non-local attention, a method named RNAN [17] was proposed to eliminate the image noise. Subsequent methodologies included the use of recursive units and gate units to remove JPEG artifacts [18], as well as the implementation of a dual-stream multi-path recursive residual network [19]. Later on, Lin et al. [20] proposed a multiscale image fusion approach to remove JPEG artifacts effectively, and achieved exceptional objective quality. But these methods cannot be extended to compressed video directly, since they treat frames independently and thus fail to exploit temporal information.
To enhance the quality of the compressed video, a 10-layer CNN automatic decoder (DCAD) [21] was the first work to mitigate distortion in compressed videos. In [22], two sub-networks of DS-CNN were introduced to address both intra-frame and inter-frame artifacts. Their main purpose was to enhance the target frame by leveraging the spatial correlation between video frames. There are many multi-frame compressed video enhancement methods [23,24,25,26]. Yang et al. [23] introduced a multi-frame quality enhancement network, named MFQE1.0, which leveraged adjacent high-quality frames to enhance the target frame. MFQE2.0 [24] was an improved version. And then, QG-ConvLSTM [25], a method utilizing bidirectional recurrent convolution, to capture the extensive temporal information. Based on the deformable convolution (DCN), Deng et al. [26] introduced patio-temporal deformable convolution (STDF) to extract temporal information from multiple frames by expanding the input frames to 7 or even 9 frames effectively. These methods aimed to enhance the target frame by leveraging the temporal relationships among multiple video frames primarily. Generally, multi-frame compressed video enhancement tended to achieve better results compared to single-frame enhancement due to its utilization of richer temporal and spatial information. However, these methods for compressed video enhancement faced the following challenges:
(1) The parameters of networks are excessively large, which challenges the efficiency of the training and real-time tasks.
(2) Existing methods tend to prioritize enhanced results at the expense of inference speed.
Therefore, it is necessary to explore lightweight and high-performance models for compressed video quality enhancement.
The term “lightweight model” refers to compressing the model size to maximize computational speed while preserving the accuracy. Researchers have been paying increasing attention to developing lightweight models in the field of image classification to enable deployment on mobile devices [14,15,16,27,28]. Among the pioneer endeavors in developing lightweight models, SqueezeNet [29] emerged, replacing 3 × 3 convolutions with 1 × 1 convolutions, resulting in a parameter reduction of approximately one-fiftieth compared to AlexNet [30]. Subsequently, Xception [31] further reduced the parameters by decoupling the Inception structure [32]. The ResNeXt [33] introduced group convolutions and reduced the parameters by integrating the residual network and Inception structure [32] effectively. In 2017, Google introduced MobileNet [34], which pioneered the concept of depthwise separable convolution (DSC) to effectively reduce the parameters in neural networks. Subsequently, MobileNet V2 [35] surpassed the previous performance benchmarks by implementing inverted residual structures and linear bottleneck layers. After that, ShuffleNet [36,37] reduced the model parameters by successfully employing group convolutions and enabling inter-channel interaction via channel shuffling operations. More recently, Huang et al. introduced CondenseNet [38], a novel approach that combines model pruning and group convolutions to effectively reduce the number of model parameters. However, these methods were not used for video quality enhancement (VQE).
To enhance the quality of compressed video and achieve superior inference performance, an end-to-end CNN-based method for VQE task, named Fast-MFQE, is proposed. The main contributions of the method are as follows:
(1) A novel IPPB module is designed to reduce the multi-frame information redundancy and fasten the inference speed;
(2) STFA and FRN modules are proposed to effectively extract the temporal features and the multi-frame correlation.
More intuitively, the parameters and the performance of diverse VQE methods are shown in Figure 2. It can be seen that compared to state-of-the-art VQE methods, the proposed Fast-MFQE method demonstrates smaller parameters and superior inference performance, as well as improving the quality of the compressed video, such as Δ PSNR and Δ SSIM ×   10 2 . The structure of the remaining sections in this paper is as follows: Section 2 provides a detailed exposition of the proposed method. Section 3 presents the experimental results and provides an analysis of the superior performance of the proposed method. Section 4 concludes the paper.

2. The Proposed Fast-MFQE

The architecture of the proposed Fast-MFQE is shown in Figure 3, where Depthwise Separable Convolution (DSC) [34] is employed in place of traditional convolution to decrease the computational complexity and enhance the inference speed of the neural network. The primary objective of Fast-MFQE is to generate an enhanced video frame O ^ t that closely resembles the Ground-truth frame in the pixel domain. The Ground truth refers to the original uncompressed video frame. To leverage temporal information from adjacent frames, Fast-MFQE takes the target frame V t and its neighboring frame V t ± n n = 1 N as the input of the network. There are three main models in Fast-MFQE; each will be illustrated in the following subsections.

2.1. Image Pre-Processing Building Modules (IPPB)

There are prevalent approaches in compressed video enhancement that utilize multiple video frames as input to effectively incorporate temporal information from diverse frames. However, these networks encounter challenges in achieving rapid inference when processing high-resolution video frames due to the substantial increase in computational complexity. Consequently, pre-processing of the input frames becomes imperative to facilitate fast inference in high-resolution video.
To leverage the information from adjacent frame V t ± n n = 1 N to enhance the target frame V t , Fast-MFQE utilizes both V t ± n n = 1 N and V t as inputs to the network. To decrease the data volume of the input frames and enhance the model’s inference speed, Fast-MFQE introduces the Image Pre-Processing Building Modules (IPPB) inspired by [34,35]. As depicted in Figure 3, IPPB consists of two primary components: Mean Shift and Pixel Unshuffle.

2.1.1. Mean Shift

In general, video frames exhibit substantial spatial redundancy, and mitigating this redundancy can reduce input data effectively. In a groundbreaking work, Zhang et al. [39] first introduced the Mean Shift operation to image super-resolution tasks and presented the RCAN network, which achieved remarkable results. The Mean Shift operation serves to normalize data by emphasizing individual differences by subtracting the statistical mean value from each image sample. Drawing inspiration from [37,39], Fast-MFQE employs the Mean Shift operation to diminish redundant information in images, enhancing model training speed and consequently reducing the inference time.
Let Fast-MFQE take the adjacent frame V t ± n n = 1 N and the target frame V t as the network input (n = 3), with Mean Shift operation denoted as M S · . Then, the feature after Mean Shift operation F M S can be expressed as:
F M S = M S V t ± n n = 1 N , V t

2.1.2. Pixel Unshuffle

While reducing redundancy in individual video frames through mean shift operations is effective, downsampling the input data is necessary to further alleviate the computational burden on the network.
Inspired by the FFDnet network [40], Fast-MFQE employs a reversible downsample (R-Downsample) operation to divide the input frames into four sub-frames, aiming to reduce the input data volume within the network. This operation decreases the model’s computational cost and inference time while effectively preserving more detailed information. Consequently, it facilitates improved model performance and enhances the generalization ability. It is worth noting that the inverse operation of R-Downsample is denoted as R-Upsample.
Let the four sub-frames generated by the R-Downsample operation be denoted as I l (l = 1, 2, 3, 4), with the R-Downsample denoted as R D · . Then, the I l can be expressed as:
I l = R D F M S

2.2. Spatio-Temporal Attention Fusion (STAF)

Spatio-temporal information of video frames is essential for quality enhancement. To enhance the fusion of image information from different temporal and spatial contexts, Fast-MFQE introduces the Spatio-Temporal Attention Fusion (STAF) module. This module consists of two 3 × 3 convolutions that extract spatial information from the video frames. Subsequently, temporal attention extraction is performed to capture the temporal characteristics of the frames, as described in Figure 4. Then, the spatial and temporal information is fused by concatenating in the manner of channel dimension. Finally, the concatenated information undergoes fusion through a 1 × 1 convolutional layer. This process ensures the effectiveness of the spatial and temporal integration for compressed video enhancement. Let the spatial information extracted by the two 3 × 3 convolutions be denoted as F S and the temporal information extracted by temporal attention be denoted as F T . The fused spatio-temporal feature is denoted as F S T by the 1 × 1 convolution. These features are formulated as follows:
F S = C o n v 3 ( C o n v 3 ( [ I 1 , I 2 , I 3 , I 4 ] ) )
F T = T A ( [ I 1 , I 2 , I 3 , I 4 ] )
F S T = C o n v 1 ( [ F S , F T ] )
where C o n v 3 · and C o n v 1 · are the mapping functions of 3 × 3 convolution and 1 × 1 convolution, respectively. [ · , · ] represents the concatenation operation, and T A ( · ) is the mapping function of temporal attention.

2.3. Feature Reconstruction Network (FRN)

To achieve precise reconstruction of video frames, the Feature Reconstruction Network (FRN) is introduced in Fast-MFQE. As shown in Figure 5, the FRN consists of dense residual blocks, primarily difference learning and residual learning. Difference learning is dedicated to capturing high-frequency information in video frames by calculating element-wise differences in feature maps. And residual learning aims to learn diverse feature information by calculating element-wise additions of feature maps. By incorporating both difference and residual learning, relevant features are captured and integrated effectively, enabling the generation of refined video frames. Let R t denote the reconstructed information generated by the FRN while V t represents the target frame; the reconstructed frames, denoted as R t H Q , can be expressed as follows:
R t H Q = R t + V t

2.4. Loss Function

To encourage the enhanced frame R t H Q to be as close as possible to the original uncompressed frame V r a w in the pixel domain, Fast-MFQE adopts the mean square error between the enhanced frame R t H Q and the original uncompressed frame V r a w as the loss function of the model, which is formulated as follows:
L ( θ ) = 1 H W C R t H Q V r a w 2 2
where H, W, and C represent the height, width, and number of channels of the image under evaluation, respectively. θ can be learned through the gradient descent algorithm [41] to solve Equation (7). Thus, Fast-MFQE can be trained effectively to enhance the quality of compressed videos.

3. Experiments

In this section, the effectiveness of the proposed Fast-MFQE method is demonstrated by extensive experiments. The experimental settings are introduced in Section 3.1, and the performance comparisons of the Fast-MFQE method with state-of-the-art methods for JCT-VC testing sequences [2] are illustrated in Section 3.2.

3.1. Settings

3.1.1. Datasets

The Fast-MFQE model is trained using the dataset introduced in MFQE2.0 [24]. The dataset [24] is divided into two parts. Firstly, 18 sequences from the Joint Collaborative Team on Video Coding (JCT-VC) [2] are commonly utilized as a test set. Secondly, the remaining 142 random sequences are split into non-overlapping training (106 sequences) and validation (36 sequences) sets. All 160 sequences are compressed using HM16.5 [1] in Low-Delay configuration, which is an encoding platform for H.265, with QP set to 27, 32, and 37. The results demonstrate the excellent generalization ability of the proposed method, making it applicable to different QP values.

3.1.2. Quality Enhancement Assessment Metrics

Extensive research [13,42,43,44] has been conducted to develop efficient and accurate methods for assessing the quality of images and video frames. The quality enhancement evaluation metric is used to measure the quality of distorted images or video frames by comparing the corresponding ground truth quantitatively using full-reference evaluation metrics. In this experiment, two widely used evaluation metrics, namely PSNR and SSIM [42], are employed.

3.1.3. Parameter Settings

The basic settings and hyperparameters of the experiments are presented here. The Fast-MFQE model is trained using the PyTorch framework. Specifically, the Fast-MFQE takes three frames (n = 3) as input to the network. The iteration number is 3 × 10 5 and the mini-batch size is 32. The cropped size is reduced to 128 × 128. The learning rate is set to 1 × 10 4 and halved every 1 × 10 5 iterations. Note that the above hyperparameters are tuned over the training set. Finally, the parameters of the Fast-MFQE network are updated utilizing the Adam algorithm [41] until the network converges.

3.2. Performance Comparison

3.2.1. Quantitative Comparison

In this section, the performance of the Fast-MFQE model is evaluated using Peak Signal-to-Noise Ratio (PSNR) and Structure Similarity Index Measure (SSIM) quantitatively, which are used objective quality evaluation metrics widely. The Fast-MFQE model is compared with AR-CNN [12], DnCNN [13], RNAN [17], MFQE1.0 [23], and MFQE2.0 [24]. Among these methods, AR-CNN [12], DnCNN [13], and RNAN [17] are methods for enhancing the quality of compressed images, MFQE1.0 [23] is the first method used for multi-frame compressed video enhancement, and MFQE2.0 [24] is the most advanced method for enhancing the quality of compressed videos. To ensure a fair comparison, all of these methods are trained and tested on the same dataset.
Table 1 reports the Δ PSNR and Δ SSIM results, which are calculated between enhanced and compressed frames averaged over each test sequence. Note that Δ PSNR > 0 and Δ SSIM > 0 indicate improvement in objective quality for the compressed video. Specifically, compared to the most advanced compressed video enhancement method, MFQE2.0 [24], Fast-MFQE achieves an average increase of 19.6% in PSNR and an average increase of 8.2% in SSIM at QP = 37, and an average increase of 12.2% in PSNR and an average increase of 14.2% in SSIM at QP = 27. Although the enhancement effect of Fast-MFQE is close to that of MFQE2.0 [24], the inference speed of Fast-MFQE is faster. Overall, The Fast-MFQE outperforms all compared methods in terms of objective quality enhancement.

3.2.2. Subjective Comparison

In this section, the subjective evaluation of the Fast-MFQE is mainly focused on the following. As shown in Figure 6, the video frames are enhanced as follows: BasketballDrill at QP = 37, BlowingBubbles at QP = 32, Catus and Traffic at QP = 42. It can be observed that the proposed Fast-MFQE method has sharper edges and more vivid details than other methods. For example, the basketball in BasketballDrill, the face in BlowingBubbles, the words in Catus, and the car in Traffic can be restored with fine textures in Fast-MFQE, which is similar to MFQE2.0 [24].

3.2.3. Comparison of Inference Performance

In this section, the inference capability and the degree of lightweight of the Fast-MFQE model are quantitatively evaluated based on the frame rate and amount of parameters (Param). The Fast-MFQE model is compared with DnCNN [7], RNAN [17], MFQE1.0 [23], MFQE2.0 [24], and STDF [26]. It should be noted that STDF [26] is currently the most lightweight model used for compressed video enhancement. All models are tested on the following configurations: 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30GHz 2.30 GHz, Nvidia GeForce GTX 1080Ti GPU, and Ubuntu 20.04 for the sake of fairness.
Table 2 reports the number of parameters (Param) and the frame rate of different models. The Fast-MFQE maintains a speed of over 25 frames per second for all resolution videos. Notably, even when processing 1080p resolution videos, the Fast-MFQE achieves a speed close to 25 frames per second, ensuring smooth and non-stuttering video processing. Additionally, the Fast-MFQE reduces the parameters by 33.4% compared to the current lightest model, STDF [26]. Overall, the Fast-MFQE outperforms the compared methods in terms of model inference speed and parameters.

3.2.4. Ablation Studies

As shown in Table 3, the ablation test was performed at QP = 37. Although Model2 and Model3 have significantly improved the Inference speed, the enhancement effects, such as Δ PSNR and Δ SSIM, are declining sharply. Therefore, in order to better balance the enhancement effect and Inference speed, we choose Model1; that is, the three modules of IPPB, STFA, and FRN are all necessary.

3.2.5. Perceptual Quality Comparison

In this section, the performance of the Fast-MFQE is evaluated quantitatively using the learned perceptual image patch similarity (LPIPS) [45] and perceptual index (PI) [46], which are widely used perceptual quality assessment metrics. Table 4 reports the Δ LPIPS and Δ PI results, which are calculated between enhanced and compressed frames averaged over each test sequence. Note that Δ LPIPS < 0 and Δ PI < 0 indicate improvement in perceptual quality. As shown in this table, the Fast-MFQE is significantly superior to all the compared methods in terms of perceptual quality enhancement.

3.2.6. Subjective Quality and Inference Speed at Different Resolutions

This section focuses on the performance of the proposed Fast-MFQE regarding inference and enhancement across videos with different resolutions. Note that these videos are all tested at QP = 37. As shown in Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11, we test the inference speed and performance of the model at different resolutions and QP = 37. The results demonstrate that the Fast-MFQE is capable of performing fast inference with high quality across different resolutions.

4. Conclusions

This paper presents a fast multi-frame quality enhancement approach, named Fast-MFQE, which facilitates efficient model inference. The Fast-MFQE is the first lightweight model in the field of compressed video enhancement. Extensive experiments demonstrate that the Fast-MFQE outperforms previous methods in terms of its lightweight parameters, fast inference speed, and quality enhancement performance on benchmark datasets. Its remarkable attributes make it an ideal solution for real-time applications such as video streaming, video conferencing, and video surveillance, unlocking a range of possibilities in these domains.

Author Contributions

Conceptualization, K.C. and J.C.; writing—original draft preparation, K.C.; writing—review and editing, J.C., H.Z. and X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key R&D Program of China, under Grant 2021YFE0205400, in part by the National Natural Science Foundation of China under Grant 61976098, in part by the Natural Science Foundation for Outstanding Young Scholars of Fujian Province under Grant 2022J06023, in part by the Natural Science Foundation of Fujian Province under Grant 2022J01294, in part by the Key Science and Technology Project of Xiamen City under grant 3502Z20231005, and in part by the Collaborative Innovation Platform Project of Fuzhou-Xiamen-Quanzhou National Independent Innovation Demonstration Zone under Grant 2021FX03.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Acknowledgments

We sincerely appreciate the anonymous reviewers’ critical comments and valuable suggestions for improving the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IPPBImage pre-processing building module.
STFASpatio-temporal fusion attention.
FRNFeature reconstruction network.
UHDUltra-high-definition.
QoEQuality of experience.
CNNConvolutional neural network.
PSNRPeak Signal-to-Noise Ratio.
SSIMStructural Similarity Index Measure.
DSCDepthwise Separable Convolution.

References

  1. Sullivan, G.J.; Ohm, J.-R.; Han, W.-J.; Wiegand, T. Overview of the High Efficiency Video Coding (HEVC) Standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 12, 1649–1668. [Google Scholar] [CrossRef]
  2. Ohm, J.-R.; Sullivan, G.J.; Schwarz, H.; Tan, T.K.; Wiegand, T. Comparison of the coding efficiency of video coding standards—including high efficiency video coding (hevc). IEEE Trans. Circuits Syst. Video Technol. 2012, 12, 1669–1684. [Google Scholar]
  3. Li, S.; Xu, M.; Deng, X.; Wang, Z. Weight-based R-λ rate control for perceptual high efficiency video coding coding on conversational videos. Signal Process. Image Commun. 2015, 10, 127–140. [Google Scholar] [CrossRef]
  4. Lu, G.; Ouyang, W.; Xu, D.; Zhang, X.; Cai, C.; Gao, Z. An end-to-end deep video compression framework. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 10998–11007. [Google Scholar]
  5. Galteri, L.; Seidenari, L.; Bertini, M.; Bimbo, A.D. Deep generative adversarial compression artifact removal. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4836–4845. [Google Scholar]
  6. Foi, A.; Katkovnik, V.; Egiazarian, K. Pointwise Shape-Adaptive DCT for High-Quality Denoising and Deblocking of Grayscale and Color Images. IEEE Trans. Image Process. 2007, 5, 1395–1411. [Google Scholar] [CrossRef]
  7. Zhang, X.; Xiong, R.; Fan, X.; Ma, S.; Gao, W. Compression artifact reduction by overlapped-block transform coefficient estimation with block similarity. IEEE Trans. Image Process. 2013, 12, 4613–4626. [Google Scholar] [CrossRef]
  8. Sheikh, H.R.; Bovik, A.C.; de Veciana, G. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans. Image Process. 2005, 11, 2117–2128. [Google Scholar] [CrossRef]
  9. Jancsary, J.; Nowozin, S.; Rother, C. Loss-specific training of non-parametric image restoration models: A new state of the art. In Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 112–125. [Google Scholar]
  10. Jung, C.; Jiao, L.; Qi, H.; Sun, T. Image deblocking via sparse representation. Signal Process. Image Commun. 2012, 3, 663–677. [Google Scholar] [CrossRef]
  11. Chang, H.; Ng, M.K.; Zeng, T. Reducing artifacts in JPEG decompression via a learned dictionary. IEEE Trans. Signal Process. 2014, 2, 718–728. [Google Scholar] [CrossRef]
  12. Dong, C.; Deng, Y.; Loy, C.C.; Tang, X. Compression Artifacts Reduction by a Deep Convolutional Network. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 576–584. [Google Scholar]
  13. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 7, 3142–3155. [Google Scholar] [CrossRef]
  14. Han, W.; Zhao, B.; Luo, J. Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios. Sensors 2023, 23, 4789. [Google Scholar] [CrossRef]
  15. Coates, W.; Wahlström, J. LEAN: Real-Time Analysis of Resistance Training Using Wearable Computing. Sensors 2023, 23, 4602. [Google Scholar] [CrossRef]
  16. Xiao, S.; Liu, Z.; Yan, Z.; Wang, M. Grad-MobileNet: A Gradient-Based Unsupervised Learning Method for Laser Welding Surface Defect Classification. Sensors 2023, 23, 4563. [Google Scholar] [CrossRef]
  17. Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. arXiv 2019, arXiv:1903.10082. [Google Scholar]
  18. Tai, Y.; Yang, J.; Liu, X.; Xu, C. MemNet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar]
  19. Jin, Z.; Iqbal, M.Z.; Zou, W.; Li, X.; Steinbach, E. Dual-Stream Multi-Path Recursive Residual Network for JPEG Image Compression Artifacts Reduction. IEEE Trans. Circuits Syst. Video Technol. 2021, 2, 467–479. [Google Scholar] [CrossRef]
  20. Lin, M.-H.; Yeh, C.-H.; Lin, C.-H.; Huang, C.-H.; Kang, L.-W. Deep Multi-Scale Residual Learning-based Blocking Artifacts Reduction for Compressed Images. In Proceedings of the IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS), Hsinchu, Taiwan, 18–20 March 2019; pp. 18–19. [Google Scholar]
  21. Wang, T.; Chen, M.; Chao, H. A novel deep learning-based method of improving coding efficiency from the decoder-end for high efficiency video coding. In Proceedings of the Data Compression Conference (DCC), Snowbird, UT, USA, 4–7 April 2017; pp. 410–419. [Google Scholar]
  22. Yang, R.; Xu, M.; Wang, Z. Decoder-side high efficiency video coding quality enhancement with scalable convolutional neural network. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 817–822. [Google Scholar]
  23. Yang, R.; Xu, M.; Wang, Z.; Li, T. Multi-frame quality enhancement for compressed video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–32 June 2018; pp. 6664–6673. [Google Scholar]
  24. Guan, Z.; Xing, Q.; Xu, M.; Yang, R.; Liu, T.; Wang, Z. Mfqe 2.0: A new approach for multi-frame quality enhancement on compressed video. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 946–963. [Google Scholar] [CrossRef] [PubMed]
  25. Yang, R.; Sun, X.; Xu, M.; Zeng, W. Quality-gated convolutional lstm for enhancing compressed video. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 532–537. [Google Scholar]
  26. Deng, J.; Wang, L.; Pu, S.; Zhuo, C. Spatio-temporal deformable convolution for compressed video quality enhancement. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 2374–3468. [Google Scholar]
  27. Zhang, T.; Zhang, Y.; Xin, M.; Liao, J.; Xie, Q. A Light-Weight Network for Small Insulator and Defect Detection Using UAV Imaging Based on Improved YOLOv5. Sensors 2023, 23, 5249. [Google Scholar] [CrossRef]
  28. Han, N.; Kim, I.-M.; So, J. Lightweight LSTM-Based Adaptive CQI Feedback Scheme for IoT Devices. Sensors 2023, 23, 4929. [Google Scholar] [CrossRef]
  29. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  30. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  31. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. arXiv 2017, arXiv:1610.02357. [Google Scholar]
  32. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. arXiv 2014, arXiv:1409.4842. [Google Scholar]
  33. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  34. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  35. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. arXiv 2019, arXiv:1801.04381. [Google Scholar]
  36. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  37. Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. arXiv 2018, arXiv:1807.11164. [Google Scholar]
  38. Huang, G.; Liu, S.; van der Maaten, L.; Weinberger, K.Q. CondenseNet: An Efficient DenseNet Using Learned Group Convolutions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2752–2761. [Google Scholar]
  39. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. arXiv 2018, arXiv:1807.02758. [Google Scholar]
  40. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible Solution for CNN based Image Denoising. IEEE Trans. Image Process. 2018, 9, 4608–4622. [Google Scholar] [CrossRef]
  41. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  42. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 4, 600–612. [Google Scholar] [CrossRef]
  43. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; pp. 1398–1402. [Google Scholar]
  44. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 2, 430–444. [Google Scholar] [CrossRef]
  45. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. arXiv 2018, arXiv:1801.03924. [Google Scholar]
  46. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Change Loy, C.; Qiao, Y.; Tang, X. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. arXiv 2018, arXiv:1809.00219. [Google Scholar]
Figure 1. Artifacts of compressed video at 240p and QP = 37.
Figure 1. Artifacts of compressed video at 240p and QP = 37.
Sensors 23 07227 g001
Figure 2. Inference speed and performance comparison of VQE methods.
Figure 2. Inference speed and performance comparison of VQE methods.
Sensors 23 07227 g002
Figure 3. Architecture of the proposed Fast-MFQE.
Figure 3. Architecture of the proposed Fast-MFQE.
Sensors 23 07227 g003
Figure 4. Architecture of Spatio-Temporal Attention Fusion (STAF).
Figure 4. Architecture of Spatio-Temporal Attention Fusion (STAF).
Sensors 23 07227 g004
Figure 5. Architecture of Feature Reconstruction Network (FRN).
Figure 5. Architecture of Feature Reconstruction Network (FRN).
Sensors 23 07227 g005
Figure 6. Subjective quality comparison on BasketballDrill at QP = 37, BlowingBubbles at QP = 32, Catus and Traffic at QP = 42. It can be seen that the Fast-MFQE approach achieves a clearer effect than the most advanced compressed video enhancement method MFQE2.0 [24].
Figure 6. Subjective quality comparison on BasketballDrill at QP = 37, BlowingBubbles at QP = 32, Catus and Traffic at QP = 42. It can be seen that the Fast-MFQE approach achieves a clearer effect than the most advanced compressed video enhancement method MFQE2.0 [24].
Sensors 23 07227 g006
Figure 7. Subjective quality and inference speed at 240p.
Figure 7. Subjective quality and inference speed at 240p.
Sensors 23 07227 g007
Figure 8. Subjective quality and inference speed at 480p.
Figure 8. Subjective quality and inference speed at 480p.
Sensors 23 07227 g008
Figure 9. Subjective quality and inference speed at 720p.
Figure 9. Subjective quality and inference speed at 720p.
Sensors 23 07227 g009
Figure 10. Subjective quality and inference speed at 1080p.
Figure 10. Subjective quality and inference speed at 1080p.
Sensors 23 07227 g010
Figure 11. Subjective quality and inference speed at 1600p.
Figure 11. Subjective quality and inference speed at 1600p.
Sensors 23 07227 g011
Table 1. Overall comparison for Δ PSRN(dB) and Δ SSIM ( × 10 2 ) over test sequences at three QPs.
Table 1. Overall comparison for Δ PSRN(dB) and Δ SSIM ( × 10 2 ) over test sequences at three QPs.
QPVideo SequenceAR-CNN [12]DnCNN [13]RNAN [17]MFQE1.0 [23]MFQE2.0 [24]Fast-MFQE
Δ PSNR↑ Δ SSIM↑ Δ PSNR↑ Δ SSIM↑ Δ PSNR↑ Δ SSIM↑ Δ PSNR↑ Δ SSIM↑ Δ PSNR↑ Δ SSIM↑ Δ PSNR↑ Δ SSIM↑
37ATraffic0.270.500.350.640.400.860.500.900.591.020.611.23
PeopleOnStreet0.370.760.540.940.741.300.801.370.921.570.971.67
BKimono0.200.590.270.730.330.980.501.130.551.180.661.23
ParkScene0.140.440.170.520.200.770.391.030.461.230.531.33
Cactus0.200.410.280.530.350.760.440.880.501.000.641.16
BQTerrace0.230.430.330.530.420.840.270.480.400.670.520.86
BasketballDrive0.230.510.330.630.430.920.410.800.470.830.740.91
CRaceHourse0.230.490.310.700.390.990.340.550.390.800.530.93
BQMall0.280.690.380.870.451.150.511.030.621.200.721.23
PartyScene0.140.520.220.690.300.980.220.730.361.180.441.31
BasketballDrill0.230.480.420.890.501.070.480.900.581.200.631.26
DRaceHorses0.260.590.340.800.421.020.511.130.591.430.681.47
BQSquare0.210.300.300.460.320.63-0.010.150.340.650.470.68
BlowingBubles0.160.460.250.760.311.080.391.200.531.700.611.89
BasketballPass0.260.630.380.830.461.080.631.380.731.550.881.67
EFourPeople0.400.560.540.730.700.970.660.850.730.950.870.97
Johnny0.240.210.470.540.560.880.550.550.600.680.710.73
KristenAndSara0.410.470.590.620.630.800.660.750.750.850.860.88
Average0.250.500.360.690.410.620.460.880.561.090.671.18
32Average0.190.170.330.41//0.430.580.520.680.630.69
27Average0.160.090.330.26//0.400.340.490.420.550.48
Table 2. Inference speed at different resolutions and number of parameters(param).
Table 2. Inference speed at different resolutions and number of parameters(param).
Inference Speed(f/s)Param(k)
Res.120p240p480p720p1080p
Method
DnCNN [13]191.854.714.16.12.6556
RNAN [17]5.63.21.40.60.088957
MFQE1.0 [23]34.312.63.81.60.71788
MFQE2.0 [24]56.525.38.43.71.6255
STDF [26]13.2736.49.13.81.6365
Fast-MFQE162.160.343.132.325.7243
Table 3. Ablation studies at QP = 37.
Table 3. Ablation studies at QP = 37.
ModelModel1Model2Model3
IPPBYseNoNo
STFAYesYesNo
FRNYesYesYes
Δ PSNR/ Δ SSIM ( × 10 2 )0.68/1.190.42/0.890.21/0.45
Inference speed(f/s)32.145.373.2
Table 4. Perceptual Quality Comparison.
Table 4. Perceptual Quality Comparison.
QPVideo SequenceAR-CNN [12]DnCNN [13]RNAN [17]MFQE1.0 [23]MFQE2.0 [24]Fast-MFQE
Δ LPIPS↓ Δ PI↓ Δ LPIPS↓ Δ PI↓ Δ LPIPS↓ Δ PI↓ Δ LPIPS↓ Δ PI↓ Δ LPIPS↓ Δ PI↓ Δ LPIPS↓ Δ PI↓
37ATraffic0.0280.7200.0270.6530.0260.6440.0270.5930.0230.5720.0180.569
PeopleOnStreet0.0290.7260.0280.6310.0290.6430.0260.6310.0190.5390.0170.520
BKimono0.0300.7330.0310.6640.0320.6570.0290.6220.0200.6170.0180.639
ParkScene0.0280.7280.0320.6350.0310.6240.0280.5720.0240.6910.0210.701
Cactus0.0310.6990.0260.5860.0250.5900.0270.6300.0300.6160.0270.593
BQTerrace0.0290.7460.0270.6140.0280.5730.0270.6210.0260.5930.0280.582
BasketballDrive0.0300.7320.0310.7200.0290.6800.0300.6750.0250.6230.0210.641
CRaceHourses0.0270.7510.0260.7090.0250.7160.0310.6950.0180.6410.0190.611
BQMall0.0320.6380.0310.6820.0320.6440.0240.6320.0170.6710.0280.576
PartScene0.0270.6990.0280.6270.0270.6090.0250.5940.0260.6170.0300.548
BasketballDrill0.0320.7210.0310.6620.0320.6140.0300.6010.0270.6140.0210.509
DRaceHourse0.0300.7580.0290.6310.0280.6290.0270.6220.0300.6010.0190.627
BQSquare0.0290.7710.0290.6910.0270.6780.0250.6310.0290.5970.0200.561
BlowingBubles0.0320.7120.0310.7860.0320.6450.0310.5910.0220.5960.0230.558
BasketballPass0.0310.7330.0270.6730.0280.5930.0320.5730.0230.6100.0260.606
EFourPeople0.0260.7260.0280.7650.0270.7120.0260.6700.0220.6320.0210.617
Johnny0.0280.7610.0270.6680.0270.6230.0280.6400.0240.6210.0230.618
KristenAndSara0.0310.7150.0300.7190.0310.6390.0300.6270.0260.6150.0240.597
Average0.0290.7260.0280.6730.0280.6390.0270.6230.0230.6140.0220.592
32 Average0.0280.5640.0260.5330.0230.5150.0230.5010.0200.4950.0180.493
27 Average0.0260.3770.0240.3450.0220.3860.0210.3740.0190.3260.0170.314
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, K.; Chen, J.; Zeng, H.; Shen, X. Fast-MFQE: A Fast Approach for Multi-Frame Quality Enhancement on Compressed Video. Sensors 2023, 23, 7227. https://doi.org/10.3390/s23167227

AMA Style

Chen K, Chen J, Zeng H, Shen X. Fast-MFQE: A Fast Approach for Multi-Frame Quality Enhancement on Compressed Video. Sensors. 2023; 23(16):7227. https://doi.org/10.3390/s23167227

Chicago/Turabian Style

Chen, Kemi, Jing Chen, Huanqiang Zeng, and Xueyuan Shen. 2023. "Fast-MFQE: A Fast Approach for Multi-Frame Quality Enhancement on Compressed Video" Sensors 23, no. 16: 7227. https://doi.org/10.3390/s23167227

APA Style

Chen, K., Chen, J., Zeng, H., & Shen, X. (2023). Fast-MFQE: A Fast Approach for Multi-Frame Quality Enhancement on Compressed Video. Sensors, 23(16), 7227. https://doi.org/10.3390/s23167227

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop