Next Article in Journal
An Open-Source System for Generating and Computer Grading Traditional Non-Coding Assignments
Next Article in Special Issue
Learning-Based Text Image Quality Assessment with Texture Feature and Embedding Robustness
Previous Article in Journal
Novel Multi-Vehicle Motion-Based Model of Trolleybus Grids towards Smarter Urban Mobility
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RDNet: Rate–Distortion-Based Coding Unit Partition Network for Intra-Prediction

1
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
2
Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2022, 11(6), 916; https://doi.org/10.3390/electronics11060916
Submission received: 14 February 2022 / Revised: 10 March 2022 / Accepted: 15 March 2022 / Published: 15 March 2022

Abstract

:
High efficiency video coding (HEVC) has been finalized as the most widely utilized video coding standard, jointly developed by ITU-T, VCEG, and MPEG. In HEVC, the quad-tree structure of the coding unit partition is one of the most substantial modules and provides significant coding gains following huge coding time. In this paper, a rate–distortion-based coding unit partition network (RDNet) is proposed to make partition decisions based on the statistical features. RDNet is composed of a prediction sub-network and a target sub-network, where the prediction sub-network is used to predict the CU partition modes of the intra-prediction and the target sub-network is designed to optimize the network parameters by evaluating the rate–distortion cost, respectively. To balance the prediction accuracy and the rate–distortion loss, a parameter-exchanging strategy is applied to control the parameters’ sharing between two networks. Experimental results prove that our model can reduce the encoding time of HEVC by 55.83~71.72% with an efficient BD-BR of 2.876~3.347%, and the ablation study evaluates the ability of our strategy on balancing the trade-off between coding accuracy and inference speed.

1. Introduction

Video coding [1] is a key technique for multimedia applications that aims to reduce the bit-rate while preserving the critical visual information of video signals. In the past 30 years, a series of standards regarding video coding has been developed, such as MPEG2 Video/H.262 [2], H.264/AVC [3], H.265/HEVC [4], and H.266/VVC [5]. A variety of new coding techniques have been adopted, such as the quad-tree structure of coding unit (CU) partition and the affine motion compensation prediction. These techniques have dramatically improved the coding efficiency; however, the coding time of the codec has also been much higher. Specifically for HEVC and VVC, the recursive CU partition method [6] based on quad-tree takes more than 80 % of the coding time; therefore, it is crucial to significantly reduce the coding time of HEVC, while maintaining the desirable coding efficiency.
Many attempts have been made to address the increasing coding time of the HEVC and VVC encoders. In general, these research works can be mainly classified into two categories: heuristic approaches and learning-based approaches. The heuristic strategies [7,8,9,10,11,12] mainly explore some heuristic features to skip the rate–distortion (RD) computation. The features are usually utilized to determine the CU partition in the early stage, for example, some textural homogeneity and spatial correlation are exploited to build statistical models for the CU partition. With these models, the rate–distortion optimization (RDO) computation are saved; however, the heuristic features cannot perfectly replace the role of RD, and the representation of heuristic features need to be also enhanced to assist the decision of CU partition. Alternatively, some approaches are further proposed to improve the efficiency of the RD search by optimizing the traversal prediction mode. Unfortunately, there is still a cost associated with the RD search, so the reduced coding time is only marginal.
In these few years, the learning-based approaches have also been widely used in CU partition prediction. Driven by sufficient data, the planning of the CU partition can be automatically learned. Some machine learning methods, such as SVM [13], decision tree [14], and random forest [15], etc., are applied to analyze statistical features, then off-line pre-partition models are used to make decisions on the partition of CU Blocks. In contrast to heuristic approaches, learning-based approaches can address the drawback of heuristic feature extraction. Some of the latest works [16,17,18,19,20,21] have also introduced convolution neural networks (CNN) to learn CU partition, which can automatically extract deep features from large-scale images. It is noted that these works regard the CU partition as a binary classification problem, ignoring the impact of RD; therefore, the coding performance of these methods is generally limited.
In this paper, we propose a rate–distortion-based coding unit partition network (RDNet) to improve the efficiency of CU partition of intra-coding in HEVC. The proposed approach designs a dual-network structure to jointly learn the statistical features and decide the partition modes of CU based on RD cost, which aims to optimize both coding time and RD performance in CU partition. To implement the optimization of the designed dual network, a parameter exchanging strategy is applied to balance the accuracy of CU partition towards RD performance and the corresponding RD loss. Experiment results show that the coding time of the proposed approach is greatly reduced compared with the standard partition approach in HEVC, with better compression visual effect on BD-BR and BD-PSNR.
To sum up, the main contributions of this paper are as follows:
  • We design an RDNet that integrates a prediction network and a target network, to predict the possible CU splitting modes and the RD cost.
  • We propose a parameters exchanging strategy to balance the accuracy of the CU partition and the RD cost. Meanwhile, a dynamic threshold is optimized to realize the rapid optimization of the network.
  • We achieve a coding time reduction of 55.83~71.72% with an efficient BD-BR of 2.876~3.347%, compared to the HEVC test model (HM16.5).

2. Related Work

In order to develop fast video coding technology, many works have made tremendous efforts to optimize the CU partition.

2.1. Heuristic CU Partition

As shown in Figure 1, a classical coding tree unit (CTU) of one frame in the HEVC codec either contains a single CU or is recursively split into smaller square CUs via the quad-tree. The maximum size of a CTU is 64 × 64 , and the minimal size of a CU can be 8 × 8 . To determine whether a CU should be split up, RD cost is the most common metric. Due to the recursive partition along with the quad-tree, the CU partition operation consumes the largest fraction of the encoding time in the HEVC codec; therefore, the CU partition strategy determines the RD performance and time consumption.
To reduce the time consumption of CU partition, some approaches focus on the early decision of the CU partition, and make a lot of efforts to optimize the decision-making mode. Zhang et al. [7] utilized the entropy model to replace the RD optimization in the HEVC to skip the RD calculation in the encoding part. Zhang et al. [22] proposed to simplify the prediction model to improve the coarse RD search process, then implemented the acceleration of the CU partition. Lu et al. [23] proposed an intra-frame CU splitting method to predict the current CU partition by referring to the partition modes of the adjacent CUs. Jamali et al. [24] presented an RD cost prediction strategy based on a low-complexity statistics in the intra mode of HEVC for partition early termination. Zhang et al. [9] presented a hierarchical binary CU depth decision structure, which could balance the trade-off between the coding time and the RD performance. Similarly, Yang et al. [11] proposed a cascade decision optimized structure for the CU partition in the VVC. Wang et al. [25] optimized the original quad-tree structure and proposed an extended quad-tree to store the partition results. On the other hand, some heuristic features are utilized to assist the decision making, Kim et al. [8] and Fu et al. [12] proposed the online learning method to decrease the consuming time of CU partition, respectively. Qing et al. [10] proposed to predict the CU partitioning depth by calculating the corresponding salient features of the given image, in order to reduce the unnecessary partitioning modes. Kibeya et al. [26] proposed a matrix-based feature learning algorithm for judging the partition results in the early stage. Especially, Ferroukhi et al. [27] provided a method combining bandelets and the SPIHT based on 2nd-generation wavelets for CU partition in medical video coding. All the above works were able to reduce the coding time of HEVC up to a point, with the similar compression performance; however, the RD search is not skipped by using these methods, so that the time consumption of RD computation cannot be saved.

2.2. Learning-Based CU Partition

To tackle this problem, the learning-based approaches [8,28,29,30] introduce some statistical models trained from a large number of data to directly make decisions on the CU partition so that the recursive search of RD can be saved and the calculation time can be further reduced. Du et al. [28] proposed to apply random forest to skip or terminate CU partition in advance. Zhang et al. [29] utilized decision tree to predict the strategy of CU splitting offline. Then, an approach jointly using the online and offline mode was described in [8]. Fu et al. [30] further presented the idea to extract the texture features from the input images, and support vector machines (SVM) were employed to split CU. To sum up, learning-based approaches improve the efficiency of the CU partition, but some limitations due to the representation of heuristic features still exist.
Recently, some approaches based on deep learning [16,17,18,19,20,21] have appeared to address the problem of CU partition. To enhance the representation of heuristic features, convolutional neural networks (CNN) are adopted to automatically extract features related to the CU partition. Liu et al. [16] first applied CNN to realize the optimization of HEVC in hardware and software. Jin et al. [17] utilized CNN to optimize the quad-tree of CU splitting. Wang et al. [19] designed a fast partition decision algorithm with a CNN-based quad-tree structure of CU. Xu et al. [18] designed an ETH-CNN to predict the CU partition, aiming to reduce the coding time of both the intra and the inter-mode in the HEVC. Zhang et al. [20] explored the texture information of the given images by using a CNN model to accelerate the CU partition. A similar work [21] continued to explore by applying the CNN approach on how to utilize the texture information to implement the fast decision of CU partition. Moreover, Feng et al. [6] utilized the depth map prediction for fast block partitioning based on CNN. Tissier et al. [31] trained CNN to predict a probability vector that speeds up coding block partitioning in VVC. Nevertheless, these works commonly regard CU partition as a binary classification problem, which leads to the binary output without information relating to RD cost; therefore, we propose a dual network based on RD, with the purpose of making some attempts introduce the RD cost in the training process.

3. Methodology

To introduce RD cost to balance the coding performance and speed, our RDNet is designed based on the dual network architecture, which is composed of a CU splitting prediction network and an RD target network. The original CTU and its corresponding mask, which is averagely split with 8 × 8 , are utilized as inputs of RDNet. The prediction network consists of 3 convolutional layers and 1 full-connection (fc) layer, which is trained by a binary classification loss. In particular, the output of the prediction network is a hierarchical CU partition map (HCPM) as [18], which represents the CU partition mode. The architecture of the target network is similar to that of the prediction network, and the network is trained by an RD loss. To jointly train the dual network, parameters are shared between the target network and the prediction network. Consequently, we design a parameter exchanging strategy. Following with the fc layer of the target network, an RD memory is constructed to store and evaluate the calculated RD cost. In the parameters exchanging, parameters of the prediction network remain unchanged if and only if the RD cost is less than the previous value, otherwise the parameters are updated by those of the target network. The overall structure of RDNet is shown in Figure 2.

3.1. The RD Memory

The existing CNN-based approaches barely consider the RD cost during deciding on the CU partition, which makes it difficult to optimize the distortion. To resolve this issue, an RD memory is designed to store and evaluate the RD cost of the output HCPM from the prediction network in our approach. Correspondingly, a parameter-sharing strategy is proposed to make a balance between the target network and the prediction network.
As the input of the RD memory, HCPM is employed as the storage architecture to quantify the partition results of prediction network. Compared with the traditional CU partition in HEVC, which needs top-down computation for each layer, HCPM restores the total partition results of current CTU as a whole directly. It is convenient to measure the performance of prediction network and the ground truth. The architecture of ground truth (GT) HCPM and predicted HCPM is shown in the two parts of Figure 3. Specifically, different partition levels in HCPM have corresponding splitting depth k (when k is 1, it means that the CU is split to 32 × 32 , and k { 1 , 2 , 3 } ). The y 1 ( U ) , y 2 ( U i ) , y 3 ( U i , j ) { 0 , 1 } and y ^ 1 ( U ) , y ^ 2 ( U i ) , y ^ 3 ( U i , j ) [ 0 , 1 ] represent the partition confidence of CUs in k level for GT HCPM and predicted one. In GT HCPM, y k ( U ) = 1 represents the current CU will be split to next depth, and if y k ( U ) = 0 , the CU remains unchanged. Moreover, y ^ k ( U ) restores the partition confidence in predicted HCPM, which takes the binary classification threshold t h r k as the line between non-partition and partition.
As shown in Figure 4, the target network is optimized by minimizing the difference between the predicted HCPM and the GT HCPM, where the ground truth HCPMs are obtained from the CU partition map in the HEVC encoder. RD memory of the target network is utilized to compute the RD distortion cost Q R D k and the time cost Q T k when the current CU is split to 64 × 64 , 32 × 32 or 16 × 16 . To balance the RD distortion cost and the time cost in the training duration, we use the total cost Q k to optimize the parameters of the target network, where Q k is the weighted sum of Q R D k and Q T k :
Q k = Q T k + λ Q R D k ,
where λ is a Lagrange constant, which locates in [0.5, 2].
As it is difficult to measure the real RD cost and the coding time during training stage, we adopt a parametric strategy to simulate RD cost and coding time in the RD memory. Given the GT HCPM and the predicted HCPM, a 2 × 2 matrix m a t r i x k [ m ] [ n ] is applied to store the partition decisions, where m represents the GT partition decision from HEVC and n denotes the result of the predicted HCPM, m , n [ 0 , 1 ] . Moreover, the value 1 store in the matrix indicates to perform the partition, and 0 represents non-partition. Then, Q R D k is employed to evaluate the different partition decision between the prediction HCPM and the GT one, as:
Q R D k = m a t r i x k [ 0 ] [ 1 ] + m a t r i x k [ 1 ] [ 0 ]
In addition, the time cost Q T k is used to count the number of CUs, which has to do with the partition, as:
Q T k = m a t r i x k [ 0 ] [ 1 ] + m a t r i x k [ 1 ] [ 1 ]
Some details about an example of applying the defined matrix for calculating the RD cost are shown in Figure 4. First, a predicted HCPM and the corresponding GT HCPM are given as input, m a t r i x k records the different partition situations. The yellow part marks the partition level of k = 3 in both the predicted HCPM and the GT one, this means the current CU should be split to 8 × 8 CUs, where the number of CUs without partition of both HCPMs is 5, which are U 2 , 1 , U 2 , 2 , U 2 , 4 , U 3 , 1 , U 3 , 3 , respectively, and the number of CU with partition for both HCPMs is 1. Further, there are 2 CUs with different results of both HCPMs as U 2 , 3 , U 3 , 2 . Then, Q T 3 and Q R D 3 can be computed as Equations (2) and (3). When λ = 2 in Equation (1), Q 3 = 5 in the RD memory. Similarly, Q 1 = 1 and Q 2 = 2 are obtained by the same strategy.

3.2. Parameters Exchanging Strategy

RDNet integrates two networks to predict the partition maps and compute the RD difference. For the further optimization of these two networks, we design a parameter-exchanging strategy. In order to share the parameters between two networks in the training process, we set that if and only if Q k is less than the previous stored value in the memory, the parameters of the prediction network can remain unchanged, otherwise they are updated by those of the target network. Moreover, to maximize the accuracy of the binary classification and minimize the RD cost, a total loss function is applied to jointly optimize the whole dual network, which includes a classification loss and a defined RD loss, as
L k = L c k + L R D k ,
where L k is the total loss and k represents the partition level, L c k is the classification loss to sum the cross-entropy over all valid elements of GT HCPM and predicted HCPM, the same as the binary classification in [18], as
L c k = H ( y 1 ( U ) , y ^ 1 ( U ) ) k = 1 H ( y 2 ( U i ) , y ^ 2 ( U i ) ) k = 2 i { 1 , 2 , 3 , 4 } H ( y 2 ( U i ) H ( y 3 ( U i , j ) , y ^ 3 ( U i , j ) ) k = 3 i , j { 1 , 2 , 3 , 4 } H ( y 3 ( U i , j )
where H ( · , · ) denotes the cross-entropy between the elements of ground truth and predicted HCPM. L R D k is the RD loss, which is represented as the L 1 loss between the RD of current predicted HCPM and best RD in the memory, as
L R D k = | Q k Q b e s t k |
where Q k is the RD of current predicted HCPM, Q b e s t k is the stored best RD in the RD memory.

3.3. Fast Partition Neural Network

In general, neural networks are utilized to learn the context features and making partition decision, such as [18]; however, to skip the CU partition of the current frame in the HEVC Test Model (HM) [32], it is still necessary to read the existing partition model that has been obtained by neural network in the coding process; therefore, the HCPM introduced in Section 3.1 is employed as an output of the network to skip the recursion computation in HEVC.
In order to further speed up the process of the neural network, a pre-processing layer is adopted to perform down-sampling with average pooling on the CTU original luminance matrix, as shown in Figure 5. Due to the fact that HCPM records the CU partition results of different levels, the corresponding layers in the neural network should be also matched with three branches B 1 , B 2 , B 3 . In our prediction network, each branch uses 3 convolutional layers to extract feature maps, and merges the obtained features into a vector before entering in fc. Towards justifying the spatial detail compression quality of current CU, the vector is further divided into three channels again to connect with the quantization parameter (QP) [33], where lower value means the finer details preserving and higher value denotes the lost about some details. Then, L k is utilized in corresponding branches to optimize the different branches of the network and deal with the different levels CU partition. Following with the network, it can obtain a value that belongs to [0,1] for the partition decision. The current CU will be split if the value is greater than 0.5. Since the RD memory and parameter-exchanging strategy change the optimization direction of network, a dynamic threshold strategy is also designed.
To make the threshold approach the appropriate value, it needs to reduce the possible mistakes in the partition results. Similarly with Q R D k , m a t r i x k [ 0 ] [ 1 ] and m a t r i x k [ 1 ] [ 0 ] , which represents the wrong partition, are utilized in the threshold seeking. The m a t r i x k [ 1 ] [ 0 ] indicates the excessive partition of the predicted HCPM, which means it need to raise the current threshold for stopping this partition. On the contrary, m a t r i x k [ 0 ] [ 1 ] indicates the threshold of the network is higher than the suitable value. So the threshold requires adjusting after meeting the corresponding mistakes starting from 0.5. The final threshold can be set as:
t h r k = 0.5 + γ ( m a t r i x k [ 0 ] [ 1 ] m a t r i x k [ 1 ] [ 0 ] )
where t h r k represents the final threshold for different levels, γ are the scale factors of m a t r i x k [ 0 ] [ 1 ] and m a t r i x k [ 1 ] [ 0 ] .

4. Experiment Results

In this section, we train the RDNet for the CU partition of the intra-mode, and validate by applying in the HM16.5, which is utilized in previous works [18,34] as the evaluation benchmarks. The ablation experiments are also designed to evaluate each designed component in our framework.

4.1. Configuration and Settings

The training data used in the experiment is CPH-intra [35] data set, which comes from 2000 lossless images and is compressed by four different QPs. Each sample is composed of the luminance matrix of the CU and a ground truth binary label representing whether to partition or not. The test data are five different types of JCT-VC standard test sequences [32]. The data properties used in ablation study are shown in Table 1. In the final experimental evaluation, we implement more test sequences with different resolutions from JCT-VC for fair comparison, all settings follow the default in CTC [32].
For specific optimization details, we use momentum optimizer and the learning rate is set to 0.01 with TensorFlow-1.14. The parameters γ is 0.001 in Equation (7). The neural network trains on one 11GB GTX 1080Ti. The intra-mode encoder used in the experiments is HM 16.5, testing on Intel(R) Core(TM) i7-8750H CPU with 8 GB RAM. Since all the decisions are made in the intra-coding, so in HM 16.5, the AI configuration was applied with the default configuration file encoder_intra_main.cfg for the performance evaluation.
In our experiments, the Bjøntegaard delta bit-rate (BD-BR), Bjøntegaard delta PSNR (BD-PSNR), and Δ T in VCEG-M33 [36] are utilized to objectively evaluate the current experiment results and to represent the coding performance difference between all optimized approaches and standard coding HM16.5. BD-BR represents the bit rate increment of the optimized algorithm compared with the original algorithm under the condition of the same objective video quality, as:
B D B R = 1 D H D L D L D H ( r 2 r 1 ) d D
where D H , D L are high and low ends of the output RD curves range, respectively, and r 1 , r 2 are two corresponding bit rates. BD-PSNR represents the objective video quality improvement of the optimized algorithm compared with the original algorithm at the same bit rate, as:
B D P S N R = 1 r H r L r L r H ( D 2 ( r ) D 1 ( r ) ) d r
where r H = l o g ( R H ) , r L = l o g ( R L ) are high and low ends of the output bit rate range, respectively, and D 1 ( r ) , D 2 ( r ) are two corresponding RD curves. Δ T represents the time saved by the optimized algorithm in the case of coding the same number of frames. The calculation method of Δ T is shown in Equation (10):
Δ T = ( T T ) T × 100 %
where T is the coding time consumed by standard coding, T represents the coding time consumed by the optimized algorithm.

4.2. Ablation Study

In order to evaluate the performance of RDNet and dynamic threshold, we complete four groups of experiments in ablation study whose detail are shown in Table 2. The four groups of ablation experiments use different experiment settings. To be specific, the ETH-CNN with the partition threshold is adopted as the backbone of the ablation study. RDNet- E 1 and RDNet- E 2 are used to represent the ablation of different modules, and RDNet- E 3 means the final proposed strategy. The dynamic threshold and the RD-based dual network are used for ablation analysis. The threshold of experiments without dynamic threshold is maintained as [0.5 0.5 0.5]. The threshold of experiments with dynamic threshold maintains is adjusted by the threshold training strategy in Equation (7) and the results are shown in Table 2. The model results needs to be tested in four different QPs (QP = 22, 27, 32, 37) with the first 20 frames of each sequence. The test results are compared by calculating BD-BR and BD-PSNR in VCEG-AE07.xls [32]. The calculation results are shown in Table 3.
Due to the fact that the fast algorithm is used to skip the recursive RD calculation process in HEVC, the above approaches have a certain impact on the compression performance in the case of large-scale reduction of computation time. The BD-BR values in the table are all positive, which means that the compression performance is not as good as HM16.5; however, the approach with the lower BD-BR has the better coding performance. The BD-PSNR values are all negative in the table, which means that the video coding quality is not as good as HM16.5; in contrast, the approach that has the higher BD-PSNR has the better coding quality. The specific relationship between BD-BR and BD-PSNR from Table 3 is shown in Figure 6.
From the calculation results in the above table, it can be concluded that the threshold judged by the prediction network is affected by the target network due to RDNet. In RDNet- E 1 , the thresholds develop irreversibly, which results in poor coding performance. RDNet- E 2 shows that the dynamic thresholds can be adapted to the process of supervising optimization between dual networks. Without the dynamic thresholds, the CU partition of dual networks is not as good as that of a single network. Although RDNet- E 3 does not perform best in each resolution of the test sequence, it has excellent performance on average, suggesting that the approach adapts to the CU partition for various resolution test sequences. At the same time, it saves 64.06% time compared with HM16.5, improves the video coding compression efficiency in the intra-frame mode level, and proves the validity of the dual networks.

4.3. Performance Evaluation

In order to evaluate the proposed RDNet whose optimization performance for video coding with other classical efficient approaches, we compare our work to a fast algorithm FSD-SVM [34] based on heuristic features and a single network without RD approach PPMAC [16]. The experiment results are shown in Table 4.
From the comparison results in Table 4, it can be concluded that the proposed RDNet is outstanding in compression effect and time with better stability. To test sequence, the RDNet refers to the HEVC standard coding for realizing the supervision on the single network with RD. Compared with the depth modules without RD supervision, the RDNet has the better compression performance in intra-frame video coding.

5. Conclusions

In this paper, we proposed an RDNet to optimize the decision of CU partition based on RD cost. The approach introduces a sub-network to select the best parameters for the RD memory and predict the partition mode with dynamic threshold. Particularly, the RD memory in the target network is designed to evaluate the results of the prediction network with GT HCPM to keep the best parameters. To avoid deviation from the correct optimization direction, the dynamic threshold strategy is presented to adapt to the change regarding RD cost. Experimental results show the proposed method improves the quality and efficiency of the HEVC codec with a significant reduction in coding time.

Author Contributions

Conceptualization, C.Y.; methodology, C.Y. and C.X.; software, C.X.; validation, C.X., C.Y. and M.L.; formal analysis, C.X.; investigation, C.X.; resources, C.Y. and M.L.; data curation, C.X.; writing—original draft preparation, C.Y.; writing—review and editing, C.Y.; visualization, C.X.; supervision, C.X., C.Y. and M.L.; project administration, C.X., C.Y. and M.L.; funding acquisition, C.Y. and M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fundamental Research Funds for the Central Universities (FRF-TP-19-015A1, FRF-IDRY-20-038) and the National Natural Science Foundation of China (61972028, 61902022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, D.; Li, Y.; Lin, J.; Li, H.; Wu, F. Deep learning-based video coding: A review and a case study. ACM Comput. Surv. (CSUR) 2020, 53, 1–35. [Google Scholar] [CrossRef] [Green Version]
  2. Tudor, P. MPEG-2 video compression. Electron. Commun. Eng. J. 1995, 7, 257–264. [Google Scholar] [CrossRef]
  3. Wiegand, T.; Sullivan, G.J.; Bjontegaard, G.; Luthra, A. Overview of the H. 264/AVC video coding standard. IEEE Trans. Circuits Syst. Video Technol. 2003, 13, 560–576. [Google Scholar] [CrossRef] [Green Version]
  4. Sullivan, G.J.; Ohm, J.R.; Han, W.J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  5. Bross, B.; Wang, Y.K.; Ye, Y.; Liu, S.; Chen, J.; Sullivan, G.J.; Ohm, J.R. Overview of the versatile video coding (VVC) standard and its applications. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3736–3764. [Google Scholar] [CrossRef]
  6. Feng, A.; Gao, C.; Li, L.; Liu, D.; Wu, F. Cnn-Based Depth Map Prediction for Fast Block Partitioning in HEVC Intra Coding. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  7. Zhang, M.; Qu, J.; Bai, H. Entropy-based fast largest coding unit partition algorithm in high-efficiency video coding. Entropy 2013, 15, 2277–2287. [Google Scholar] [CrossRef] [Green Version]
  8. Kim, H.S.; Park, R.H. Fast CU partitioning algorithm for HEVC using an online-learning-based Bayesian decision rule. IEEE Trans. Circuits Syst. Video Technol. 2015, 26, 130–138. [Google Scholar] [CrossRef]
  9. Zhang, Y.; Kwong, S.; Wang, X.; Yuan, H.; Pan, Z.; Xu, L. Machine learning-based coding unit depth decisions for flexible complexity allocation in high efficiency video coding. IEEE Trans. Image Process. 2015, 24, 2225–2238. [Google Scholar] [CrossRef] [PubMed]
  10. Qing, A.; Zhou, W.; Wei, H.; Zhou, X.; Zhang, G.; Yang, J. A fast CU partitioning algorithm in HEVC inter prediction for HD/UHD video. In Proceedings of the 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Jeju, Korea, 13–16 December 2016; pp. 1–5. [Google Scholar]
  11. Yang, H.; Shen, L.; Dong, X.; Ding, Q.; An, P.; Jiang, G. Low-complexity CTU partition structure decision and fast intra mode decision for versatile video coding. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 1668–1682. [Google Scholar] [CrossRef]
  12. Fu, T.; Zhang, H.; Mu, F.; Chen, H. Fast CU partitioning algorithm for H. 266/VVC intra-frame coding. In Proceedings of the 2019 IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 55–60. [Google Scholar]
  13. Noble, W.S. What is a support vector machine? Nat. Biotechnol. 2006, 24, 1565–1567. [Google Scholar] [CrossRef] [PubMed]
  14. Song, Y.Y.; Ying, L. Decision tree methods: Applications for classification and prediction. Shanghai Arch. Psychiatry 2015, 27, 130. [Google Scholar] [PubMed]
  15. Pal, M. Random forest classifier for remote sensing classification. Int. J. Remote Sens. 2005, 26, 217–222. [Google Scholar] [CrossRef]
  16. Liu, Z.; Yu, X.; Gao, Y.; Chen, S.; Ji, X.; Wang, D. CU partition mode decision for HEVC hardwired intra encoder using convolution neural network. IEEE Trans. Image Process. 2016, 25, 5088–5103. [Google Scholar] [CrossRef] [PubMed]
  17. Jin, Z.; An, P.; Shen, L.; Yang, C. CNN oriented fast QTBT partition algorithm for JVET intra coding. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  18. Xu, M.; Li, T.; Wang, Z.; Deng, X.; Yang, R.; Guan, Z. Reducing complexity of HEVC: A deep learning approach. IEEE Trans. Image Process. 2018, 27, 5044–5059. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Wang, Z.; Wang, S.; Zhang, X.; Wang, S.; Ma, S. Fast QTBT partitioning decision for interframe coding with convolution neural network. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 2550–2554. [Google Scholar]
  20. Zhang, Y.; Wang, G.; Tian, R.; Xu, M.; Kuo, C.J. Texture-classification accelerated CNN scheme for fast intra CU partition in HEVC. In Proceedings of the 2019 Data Compression Conference (DCC), Snowbird, UT, USA, 26–29 March 2019; pp. 241–249. [Google Scholar]
  21. Galpin, F.; Racapé, F.; Jaiswal, S.; Bordes, P.; Le Léannec, F.; François, E. CNN-based driving of block partitioning for intra slices encoding. In Proceedings of the 2019 Data Compression Conference (DCC), Snowbird, UT, USA, 26–29 March 2019; pp. 162–171. [Google Scholar]
  22. Zhang, M.; Zhai, X.; Liu, Z.; An, C. Fast algorithm for HEVC intra prediction based on adaptive mode decision and early termination of CU partition. In Proceedings of the 2018 Data Compression Conference, Snowbird, UT, USA, 27–30 March 2018; p. 434. [Google Scholar]
  23. Lu, J.; Li, Y. Fast algorithm for CU partitioning and mode selection in HEVC intra prediction. In Proceedings of the 2019 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Suzhou, China, 19–21 October 2019; pp. 1–5. [Google Scholar]
  24. Jamali, M.; Coulombe, S. Fast HEVC intra mode decision based on RDO cost prediction. IEEE Trans. Broadcast. 2018, 65, 109–122. [Google Scholar] [CrossRef]
  25. Wang, M.; Li, J.; Zhang, L.; Zhang, K.; Liu, H.; Wang, S.; Kwong, S.; Ma, S. Extended quad-tree partitioning for future video coding. In Proceedings of the 2019 Data Compression Conference (DCC), Snowbird, UT, USA, 26–29 March 2019; pp. 300–309. [Google Scholar]
  26. Kibeya, H.; Belghith, F.; Ayed, M.A.B.; Masmoudi, N. A fast CU partitionning algorithm based on early detection of zero block quantified transform coefficients for HEVC standard. In Proceedings of the International Image Processing, Applications and Systems Conference, Sfax, Tunisia, 5–7 November 2014; pp. 1–5. [Google Scholar]
  27. Ferroukhi, M.; Ouahabi, A.; Attari, M.; Habchi, Y.; Taleb-Ahmed, A. Medical Video Coding Based on 2nd-Generation Wavelets: Performance Evaluation. Electronics 2019, 8, 88. [Google Scholar] [CrossRef] [Green Version]
  28. Du, B.; Siu, W.C.; Yang, X. Fast CU partition strategy for HEVC intra-frame coding using learning approach via random forests. In Proceedings of the 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Hong Kong, China, 16–19 December 2015; pp. 1085–1090. [Google Scholar]
  29. Zhang, D.; Duan, X.; Zang, D. Decision tree based fast CU partition for HEVC lossless compression of medical image sequences. In Proceedings of the 2017 9th International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 11–13 October 2017; pp. 1–6. [Google Scholar]
  30. Fu, B.; Zhang, Q.; Hu, J. Fast prediction mode selection and CU partition for HEVC intra coding. IET Image Process. 2020, 14, 1892–1900. [Google Scholar] [CrossRef]
  31. Tissier, A.; Hamidouche, W.; Vanne, J.; Galpin, F.; Menard, D. CNN Oriented Complexity Reduction Of VVC Intra Encoder. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 3139–3143. [Google Scholar]
  32. JCT-VC. HM Software. 2014. Available online: https://hevc.hhi.fraunhofer.de/svn/svnHEVCSoftware/tags/HM-16.5/ (accessed on 5 November 2016).
  33. Uddin, K.; Yang, Y.; Oh, B.T. Double compression detection in HEVC-coded video with the same coding parameters using picture partitioning information. Signal Process. Image Commun. 2022, 103, 116638. [Google Scholar] [CrossRef]
  34. Liu, D.; Liu, X.; Li, Y. Fast CU Size Decisions for HEVC Intra Frame Coding Based on Support Vector Machines. In Proceedings of the 2016 IEEE 14th International Conference on Dependable, Autonomic and Secure Computing, 14th Intl Conf on Pervasive Intelligence and Computing, 2nd International Conference on Big Data Intelligence and Computing and Cyber Science and Technology Congress(DASC/PiCom/DataCom/CyberSciTech), Auckland, New Zealand, 8–12 August 2016; pp. 594–597. [Google Scholar]
  35. CPH-Intra. Available online: https://github.com/HEVC-Projects/CPH (accessed on 3 October 2018).
  36. Grellert, M.; Bampi, S.; Correa, G.; Zatt, B.; da Silva Cruz, L.A. Learning-based complexity reduction and scaling for HEVC encoders. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1208–1212. [Google Scholar]
Figure 1. The results of CU partition. CU partition is a recursive process, which means all partitions are made before and RD is recursively traversed for merging.
Figure 1. The results of CU partition. CU partition is a recursive process, which means all partitions are made before and RD is recursively traversed for merging.
Electronics 11 00916 g001
Figure 2. The overall pipeline of RDNet. The average partition mask is realized on the luminance matrix of original CTU. In addition, the convolution in prediction and target network contains one layer of 4 × 4 convolution and two layers of 2 × 2 convolution. The target network judges the best parameters by the RD memory, and optimizes the prediction network with the combination of the RD and classification loss.
Figure 2. The overall pipeline of RDNet. The average partition mask is realized on the luminance matrix of original CTU. In addition, the convolution in prediction and target network contains one layer of 4 × 4 convolution and two layers of 2 × 2 convolution. The target network judges the best parameters by the RD memory, and optimizes the prediction network with the combination of the RD and classification loss.
Electronics 11 00916 g002
Figure 3. Structure of HCPM. The truth partition is shown in Part A. y 1 ( U ) , y 2 ( U i ) , y 3 ( U i , j ) is the ground truth partition of each layer CUs. The subscripts i , j 1 , 2 , 3 , 4 are the indices of the sub-CUs split from U , U i , U i , j . In level 1, If y 1 ( U ) = 1 represents the CU will be split into four 32 × 32 CUs, conversely y 1 ( U ) = 1 represents it will not be split. In Part B, the same CTU will obtain the corresponding HCPM prediction through network process. y ^ 1 ( U ) , y ^ 2 ( U i ) , y ^ 3 ( U i , j ) represents the partition prediction value of each layer, and the partition prediction is obtained by the prediction network.
Figure 3. Structure of HCPM. The truth partition is shown in Part A. y 1 ( U ) , y 2 ( U i ) , y 3 ( U i , j ) is the ground truth partition of each layer CUs. The subscripts i , j 1 , 2 , 3 , 4 are the indices of the sub-CUs split from U , U i , U i , j . In level 1, If y 1 ( U ) = 1 represents the CU will be split into four 32 × 32 CUs, conversely y 1 ( U ) = 1 represents it will not be split. In Part B, the same CTU will obtain the corresponding HCPM prediction through network process. y ^ 1 ( U ) , y ^ 2 ( U i ) , y ^ 3 ( U i , j ) represents the partition prediction value of each layer, and the partition prediction is obtained by the prediction network.
Electronics 11 00916 g003
Figure 4. An example of the RD memory computation. The RD memory aims to record the difference between ground truth HCPM and network results. The corresponding partition results of different levels of CU blocks are recorded in m a t r i x k . Q k can be calculated from m a t r i x k for three partition levels.
Figure 4. An example of the RD memory computation. The RD memory aims to record the difference between ground truth HCPM and network results. The corresponding partition results of different levels of CU blocks are recorded in m a t r i x k . Q k can be calculated from m a t r i x k for three partition levels.
Electronics 11 00916 g004
Figure 5. The structure of CNN. The network mainly includes three parts, pre-process, convolution, and decision making. The CTU are demeaned and down-sampled to reduce the computation time in pre-processing, the convolution includes 3 convolutional layers and an fc layer to extract image features, and the decision-making part judges the CU partition to HCPM by features.
Figure 5. The structure of CNN. The network mainly includes three parts, pre-process, convolution, and decision making. The CTU are demeaned and down-sampled to reduce the computation time in pre-processing, the convolution includes 3 convolutional layers and an fc layer to extract image features, and the decision-making part judges the CU partition to HCPM by features.
Electronics 11 00916 g005
Figure 6. The ablation study results of JCT-VC test set (AI). The closer the data points are to the upper left corner of the graph, the higher the coding quality at a lower coding bit-rate, which means better coding performance. RDNet- E 3 in the figure is closest to the upper left corner, indicating the best coding performance. From the other three curves, it can be concluded that the proposed dual network and dynamic threshold can make the optimization to the traditional approaches.
Figure 6. The ablation study results of JCT-VC test set (AI). The closer the data points are to the upper left corner of the graph, the higher the coding quality at a lower coding bit-rate, which means better coding performance. RDNet- E 3 in the figure is closest to the upper left corner, indicating the best coding performance. From the other three curves, it can be concluded that the proposed dual network and dynamic threshold can make the optimization to the traditional approaches.
Electronics 11 00916 g006
Table 1. Partly sequences of the JCT-VC test set.
Table 1. Partly sequences of the JCT-VC test set.
ClassResolution RatioTest SequenceTotal FramesFPsBit Depth
A2560 × 1600Traffic150308
B1920 × 1080Basketball Drive500508
C832 × 480Basketball Drill500508
D416 × 240Basketball Pass500508
E1280 × 720Johnny600608
Table 2. The detail of the four groups ablation experiments.
Table 2. The detail of the four groups ablation experiments.
AlgorithmDynamic ThresholdRD-Based Dual NetworksPartition Threshold
ETH-CNN [18]--[0.50 0.50 0.50]
RDNet- E 1 -[0.49 0.55 0.63]
RDNet- E 2 -[0.50 0.50 0.50]
RDNet- E 3 [0.48 0.55 0.63]
Table 3. The ablation study results of JCT-VC test set (AI).
Table 3. The ablation study results of JCT-VC test set (AI).
ClassTest
Sequence
AlgorithmBD-PSNR
(dB)
BD-BR
(%)
Δ T (%)
QP = 22QP = 27QP = 32QP = 37
ATrafficETH-CNN [18]−0.1492.771−56.62−65.17−67.79−63.16
RDNet- E 1 −0.1332.480−60.44−66.98−70.71−69.01
RDNet- E 2 −0.1482.757−59.57−66.41−70.27−67.80
RDNet- E 3 −0.1312.429−58.13−63.82−68.29−63.96
BBasketball
Drive
ETH-CNN [18]−0.1194.981−57.90−72.74−79.33−79.18
RDNet- E 1 −0.0943.904−67.05−78.18−81.53−81.63
RDNet- E 2 −0.124.967−61.00−76.26−81.62−82.37
RDNet- E 3 −0.0943.941−61.26−75.76−78.70−81.43
CBasketball
Drill
ETH-CNN [18]−0.1412.934−23.34−36.87−52.66−62.19
RDNet- E 1 −0.1342.796−35.75−50.64−57.38−64.14
RDNet- E 2 −0.1422.969−30.80−47.48−57.37−64.48
RDNet- E 3 −0.1302.738−20.33−47.09−57.90−66.16
DBasketball
Pass
ETH-CNN [18]−0.1382.412−40.54−40.30−54.42−58.00
RDNet- E 1 −0.1162.029−40.90−44.17−56.06−60.99
RDNet- E 2 −0.1352.359−37.33−43.29−50.66−47.70
RDNet- E 3 −0.1071.853−49.72−53.74−59.48−65.29
EJohnnyETH-CNN [18]−0.1463.636−73.62−73.54−74.45−78.55
RDNet- E 1 −0.1363.355−74.98−76.49−77.97−80.58
RDNet- E 2 −0.1413.501−74.37−76.03−80.71−82.81
RDNet- E 3 −0.1383.421−75.25−77.21−76.52−81.23
Std. dev.ETH-CNN [18]0.0111.01619.1317.8111.889.91
RDNet- E 1 0.0180.73516.8815.2911.649.43
RDNet- E 2 0.0111.01318.0415.6613.8414.54
RDNet- E 3 0.0180.82120.4213.269.518.90
BestETH-CNN [18]−0.1192.412−73.62−73.54−79.33−79.18
RDNet- E 1 −0.0942.029−74.98−78.18−81.53−81.63
RDNet- E 2 −0.1202.359−74.37−76.26−81.62−82.81
RDNet- E 3 −0.0941.853−75.25−77.21−78.70−81.43
AverageETH-CNN [18]−0.1383.347−50.40−57.73−65.73−68.22
RDNet- E 1 −0.1232.913−55.83−63.29−68.73−71.27
RDNet- E 2 −0.1373.311−52.61−61.90−68.13−69.03
RDNet- E 3 −0.1202.876−52.94−63.52−68.18−71.62
Table 4. Performance comparison with state-of-the-art approaches used the same strategy.
Table 4. Performance comparison with state-of-the-art approaches used the same strategy.
ClassTest SequenceFSD-SVM [34]PPMAC [16]RDNet- E 3
BD-PSNR
(dB)
BD-BR
(%)
Δ T (%)BD-PSNR
(dB)
BD-BR
(%)
Δ T (%)BD-PSNR
(dB)
BD-BR
(%)
Δ T (%)
APeopleOnStreet−0.9429.627−43.84−0.2093.969−55.60−0.1272.197−57.53
Traffic−0.3046.411−28.87−0.2404.945−60.84−0.1312.429−63.55
BBasketballDrive−0.2448.923−43.40−0.1416.018−69.51−0.0943.941−74.29
BQTerrace−0.2956.627−56.62−0.2674.815−57.89−0.0781.191−47.96
Cactus−0.2487.533−43.51−0.2086.021−63.23−0.0751.945−52.72
Kimono−0.1705.212−47.80−0.0822.382−72.72−0.0511.403−83.53
ParkScene−0.1493.630−52.85−0.1353.417−66.03−0.0761.756−59.25
CBasketballDrill−0.4399.818−53.93−0.53812.205−63.58−0.1302.738−47.87
BQMall−0.4869.646−42.06−0.4688.077−52.14−0.0841.333−33.08
PartyScene−0.4687.383−43.01−0.6729.448−58.75−0.0280.363−33.66
RaceHorses−0.3797.220−44.59−0.2644.422−58.20−0.1071.656−36.28
DBasketballPass−0.54610.054−39.72−0.4578.401−63.53−0.1071.853−57.06
BlowingBubbles−0.3736.178−37.04−0.4638.328−60.78−0.0520.845−37.87
BQSquare−0.87612.342−57.43−0.2112.563−46.72−0.0220.263−38.67
RaceHorses−0.4878.839−40.23−0.3174.593−57.30−0.0680.977−42.99
EFourPeople−0.4809.077−36.22−0.4398.002−61.54−0.1732.905−64.20
Johnny−0.47412.182−63.55−0.3077.956−66.55−0.1383.421−77.55
KristenAndSara−0.62713.351−57.51−0.2655.478−64.72−0.1392.662−74.00
Std. dev.0.1752.5539.020.1582.6036.170.0411.00815.97
Best−0.1493.630−63.55−0.0822.382−72.72−0.0220.263−83.53
Average−0.4198.559−46.23−0.3166.189−61.09−0.0931.883−54.56
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yao, C.; Xu, C.; Liu, M. RDNet: Rate–Distortion-Based Coding Unit Partition Network for Intra-Prediction. Electronics 2022, 11, 916. https://doi.org/10.3390/electronics11060916

AMA Style

Yao C, Xu C, Liu M. RDNet: Rate–Distortion-Based Coding Unit Partition Network for Intra-Prediction. Electronics. 2022; 11(6):916. https://doi.org/10.3390/electronics11060916

Chicago/Turabian Style

Yao, Chao, Chenming Xu, and Meiqin Liu. 2022. "RDNet: Rate–Distortion-Based Coding Unit Partition Network for Intra-Prediction" Electronics 11, no. 6: 916. https://doi.org/10.3390/electronics11060916

APA Style

Yao, C., Xu, C., & Liu, M. (2022). RDNet: Rate–Distortion-Based Coding Unit Partition Network for Intra-Prediction. Electronics, 11(6), 916. https://doi.org/10.3390/electronics11060916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop