Next Article in Journal
Improved YOLOv3 Model for Workpiece Stud Leakage Detection
Next Article in Special Issue
Speech Emotion Recognition Based on Parallel CNN-Attention Networks with Multi-Fold Data Augmentation
Previous Article in Journal
Virtual CANBUS and Ethernet Switching in Future Smart Cars Using Hybrid Architecture
Previous Article in Special Issue
Assessment of Compressed and Decompressed ECG Databases for Telecardiology Applying a Convolution Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adjacency Encoding Information-Based Fast Affine Motion Estimation Method for Versatile Video Coding

Huaihua Polytechnic College, Huaihua 418000, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(21), 3429; https://doi.org/10.3390/electronics11213429
Submission received: 31 August 2022 / Revised: 17 October 2022 / Accepted: 20 October 2022 / Published: 23 October 2022

Abstract

:
Versatile video coding (VVC), a new generation video coding standard, achieves significant improvements over high efficiency video coding (HEVC) due to its added advanced coding tools. Despite the fact that affine motion estimation adopted in VVC takes into account the translational, rotational, and scaling motions of the object to improve the accuracy of interprediction, this technique adds a high computational complexity, making VVC unsuitable for use in real-time applications. To address this issue, an adjacency encoding information-based fast affine motion estimation method for VVC is proposed in this paper. First, this paper counts the probability of using the affine mode in interprediction. Then we analyze the trade-off between computational complexity and performance improvement based on statistical information. Finally, by exploring the mutual exclusivity between skip and affine modes, an enhanced method is proposed to reduce interprediction complexity. Experimental results show that compared with the VVC, the proposed low-complexity method achieves 10.11% total encoding time reduction and 40.85% time saving of affine motion estimation with a 0.16% Bjøontegaard delta bitrate (BDBR) increase.

1. Introduction

The rapid development of new video applications, together with emerging videos in high frame rate (HFR), high dynamic range (HDR), and high resolution, raises an urgent demand to develop a new generation of video coding standards with coding efficiency beyond the high efficiency video coding (HEVC) standard [1]. The latest standard, versatile video coding [2,3,4], is launched by the joint video experts team (JVET) to solve the above issue. The VVC confirmed in July 2020 achieves superior encoding performance to HEVC by adopting a series of coding tools with high computation [5,6,7,8]. For interprediction, the decoder-side motion vector refinement (DMVR) [9], bi-directional optical flow (BDOF) [10,11], and affine motion compensation (AMC) [12,13,14] are used to optimize the accuracy of prediction. Furthermore, the cross-component linear model (CCLM) [15,16] and the position-dependent intraprediction combination (PDPC) [17,18] are adopted to enhance the intraprediction. In order to further eliminate frequency redundancy, the low-frequency non-separable transform (LFNST) [19,20] is employed in VVC.
As an important coding tool in VVC, the affine motion model improves interprediction accuracy. Compared with the previous motion estimation models, the affine motion model takes into account the translational motion and the rotation and scaling of the object, which is more consistent with the trajectory of the object in real life. There are two types of affine motion models in VVC, the four-parameter affine motion model and the six-parameter affine motion model. Under different models, the affine motion vectors of the current coding unit (CU) are calculated from the corresponding control points. Figure 1 shows the affine motion models in VVC. Specifically, Figure 1a,b are the four-parameter affine motion model and the six-parameter affine motion model, respectively.
Here, we successively describe the specific motion estimation process of the four-parameter and six-parameter affine motion models. In Figure 1a, based on the control point motion vectors (CPMV) at the top-left location m v 0 and top-right location m v 1 , the motion vector (MV) of the current block centered on pixel ( x , y ) can be calculated by
m v h ( x , y ) = m v 1 h m v 0 h W x + m v 1 v m v 0 v W y + m v 0 h , m v v ( x , y ) = m v 1 v m v 0 v W x + m v 1 h m v 0 h W y + m v 0 v ,
where m v h ( x , y ) and m v v ( x , y ) represent the horizontal and vertical motion vectors of the current block. m v 0 h , and m v 0 v are the horizontal and vertical vectors of CPMV at the top-left location. m v 1 h and m v 1 v denote the horizontal and vertical vectors of CPMV in the top-right location. W is the width of the current block.
For the six-parameter affine motion model in Figure 1b, it has one more CPMV at the bottom-left location m v 2 than the four-parameter affine motion model. The calculation of the MV of the current block is shown as follows,
m v h ( x , y ) = m v 1 h m v 0 h W x + m v 2 v m v 0 v H y + m v 0 h , m v v ( x , y ) = m v 1 v m v 0 v W x + m v 2 h m v 0 h H y + m v 0 v ,
where m v 2 h and m v 2 v are the horizontal and vertical vectors of CPMV at the bottom-left location. H represents the height of the current block.
Although the introduction of the affine motion model improves the encoding performance of the VVC, it brings a huge computational complexity to the encoder of VVC. Furthermore, a variety of other coding tools are also adopted in VVC to obtain a better coding performance, such as multi-type tree partition and the 67 intraprediction modes. These coding tools with complex computation make the encoding process in VVC more flexible but also more complex than HEVC, which makes VVC difficult to use for real-time applications. Hence, it is necessary to simplify the encoding process in VVC to make it suitable for hardware devices.
In this paper, we proposed an adjacency encoding information-based fast affine motion estimation method to accelerate the interprediction process. The proposed method can also be combined with the fast CU partition method to save more encoding time.
The primary contributions of this work are summarized as follows:
  • Distinguishing from most of the previous fast algorithms that focus on the CU partition, we fully explore the affine motion estimation in interprediction and propose a fast affine motion estimation algorithm based on the adjacency encoding information to achieve the savings of encoding time.
  • We count the proportion of CUs that use affine mode as the best interprediction in test sequences with different resolutions. Then we analyze the trade-off between computational complexity and performance improvement based on statistical information.
  • The affine motion estimation skipping method is proposed by exploring the relationship between affine and skip modes in interprediction.
The rest of this paper is arranged as follows: Section 2 reviews some related studies by focusing on the fast algorithm for VVC. Section 3 introduces the proposed adjacency encoding information-based fast affine motion estimation method in detail. The experimental results and analysis are presented in Section 4. Finally, the conclusion of this paper is given in Section 5.

2. Related Work

Typically, an encoder with low computational complexity speeds up the encoding and transmission of video, resulting in low latency video streams. Therefore, designing video encoders with high coding efficiency and low complexity is a core requirement for real-time applications with limited transmission bandwidth and computing power.
Although some studies have been done to reduce the computational complexity of VVC, most of these works have focused on making early decisions to speed up the partitioning process [21,22,23,24,25,26,27]. Min et al. [22] proposed a fast method to determine CU partitioning by exploiting global and local edge complexities in multiple directions. Zhao et al. [23] extracted the standard deviation and the edge point ratio to speed up the CU partition. The CU splitting information and the temporal location of the coded frame were used in [24] for the low-complexity encoder. In [25], the edge information extracted by the canny operator was utilized for CU partition in the intra- and intercoding. Similarly, the spatial features were used in [26] to reduce the computational complexity of the CU binary tree partition process. Lei et al. [27] introduced a fast method to accelerate the encoding process by exploring the content property. In recent years, some fast methods [28,29,30,31] based on a convolutional neural network (CNN) to extract and utilize features have been proposed. In [32], Wu et al. proposed a hierarchy grid fully convolutional network (HG-FCN) framework for fast VVC intra coding. In addition, there are some studies to speed up the encoding process by reducing the computational complexity of other coding tools. Pan et al. [33] introduced an entropy-based algorithm for rate-distortion optimization (RDO) mode decision. The histogram of oriented gradient features and the depth information were jointed in [34] to reduce the computational complexity in the visual sensor networks (VSNs). However, a few studies focus on fast algorithms for interprediction. In [35], a low-complexity method combining multi-type tree partition was proposed to reduce the computation of multiple transform selection. In [36], Ren et al. proposed an advanced fast interprediction method based on edge detection. Jung et al. [37] proposed a context-based inter mode decision method to accelerate the encoding process. There is still much room for improvement in reducing the coding complexity of VVC.
This paper focuses on reducing the computational complexity of the interprediction in VVC to speed up the encoding process and meet the requirements of real-time applications. It is worth mentioning that the proposed fast affine motion estimation method can also be combined with fast CU partition methods to save the encoding time further.

3. Materials and Methods

To speed up the interprediction in VVC encoding process, an adjacency encoding information-based fast affine motion estimation method is introduced in this paper. First, we trade off the computational complexity and performance improvement of affine motion estimation. Secondly, the adjacency encoding information is used to determine whether to skip the affine mode in the interprediction. The details are described as follows.

3.1. Statistics and Analysis of the Adjacency Encoding Information

In general, the moving objects in the video tend to occupy a small part of the whole frame, whereas the rest is mostly the background area. The trajectory of moving objects in a video is panning, rotating, and zooming, and the background area is in translational movement or stationary. In the interprediction process of VVC, the affine mode is performed for all blocks in the frame. Only a small proportion of the blocks choose the affine mode as the best interprediction. As a result, the process of affine motion estimation for most blocks is redundant, which significantly increases the computational complexity and encoding time. To trade off added computational complexity and encoding performance gains, we have counted the percentage of blocks that used the affine mode as the best interprediction in test videos of different resolutions to obtain a more accurate usage rate. Specifically, we assume that event A represents the current block choosing the affine mode as the optimal interprediction. P ( A ) denotes the probability that blocks in which the affine mode as the optimal interprediction is selected and can be calculated as
P ( A ) = C a f f i n e C t o t a l × 100 % ,
where C a f f i n e represents the blocks selecting the affine mode as the best interprediction. C t o t a l is the total number of the blocks. The statistical results of the proportion of blocks selecting affine intermode are shown in Table 1. We can observe that the proportion of blocks that select affine mode is relatively small. The average value is 12.3%. This illustrates that only a small part of the region finally selects an affine mode in the interprediction of blocks. However, when choosing the optimal interprediction, the affine motion estimation is calculated for each block, which significantly increases the computational complexity. The additional unnecessary calculations create a large waste in terms of encoding time. Combining the above analysis, we propose a method which reduces the encoding time without significantly degrading the encoding performance to trade-off computational complexity and encoding performance.

3.2. Affine Motion Estimation Early Skipping Method

The rate-distortion (RD) cost calculations for interprediction selection are performed recursively in determining the best interprediction for blocks. It considers the strong correlation between the current block and its adjacent CUs in terms of texture and motion information. Furthermore, the best interprediction mode and RD costs for adjacent blocks have been stored in the encoder. Therefore, it is possible to determine whether the current block needs to execute the affine mode based on the encoding information of the previously reconstructed information of adjacent CUs. In addition, the encoding information of the co-located CU in the adjacent frame is also used to determine whether the current block performs affine motion estimation.
Merge mode, affine advanced motion vector prediction (AMVP) mode, and skip mode are included in the interprediction. Among them, skip mode can be regarded as a special case in merge mode. Compared with merge mode, skip mode is a simpler mode. It does not require transmitting the prediction residuals, only the index of the best element in the candidate list. The descriptions of merge mode, AMVP mode, and skip mode are shown in Table 2. Figure 2 shows the CU partition results of the second frame extracted from the sequence “BasketballDrive” under low delay P (LDP) configuration, where the affine mode as the best interprediction is boxed in blue, and the skip mode as the best interprediction is boxed in red. From Figure 2, we can observe that skip mode is mainly used for background areas with a slight texture and slow-moving areas. Affine mode is used chiefly for blocks with dramatic motions and rich textures. Therefore, skip mode, and affine mode are mutually exclusive in most cases. Furthermore, in both affine and skip modes, they are selected with surrounding CUs (e.g., left and up).
To further explore the relationship between the encoding information of the reconstructed adjacent blocks and the affine mode used in the current block, the probability of using the affine mode as the optimal interprediction under different conditions in videos of different resolutions is investigated. Specifically, we assume that the event X s k i p is the adjacent blocks to the left and the top containing the skip mode, whereas the affine mode is not included. Event Y s k i p represents the co-located CU in the adjacent frame by using skip mode. P ( A | X s k i p ) represents the probability that the best interprediction for the current block is the affine mode under the event X s k i p condition as follows,
P ( A | X s k i p ) = P ( A X s k i p ) P ( X s k i p ) ,
where P ( A X s k i p ) is the probability that event A and event X s k i p will occur simultaneously. P ( X s k i p ) represents the probability of event X s k i p occurring. Similarly, P ( A | Y s k i p ) denotes the probability that the best interprediction for the current block is the affine mode under the event Y s k i p condition calculated as
P ( A | Y s k i p ) = P ( A Y s k i p ) P ( Y s k i p ) ,
where P ( A Y s k i p ) is the probability that event A and event Y s k i p will occur simultaneously. P ( Y s k i p ) represents the probability of event Y s k i p occurring. In experiments, when performing open-source VTM-7.0 to encode video sequences, we record the number of using the affine motion mode as the best interprediction under events X s k i p and P ( Y s k i p ) conditions, respectively. The number of events X s k i p and P ( Y s k i p ) are also recorded. The conditional probabilities are obtained by their respective corresponding ratios. Table 3 presents statistics on the probabilities of the five video sequences at different resolutions. From Table 3 we can see that the average value of P ( A | X s k i p ) is not large, at 0.12. This indicates that the probability of the current CU using the affine mode as the best mode is low when the best interprediction for adjacent blocks contains the skip mode and does not include the affine mode. Furthermore, the average value of P ( A | Y s k i p ) is 0.21, which means there is a small probability that the current block will select affine mode as the interprediction when the skip mode is contained in the co-located CU.
Based on the above statistics and analysis, we propose a low-complexity method to skip the affine mode early by utilizing the adjacency encoding information. More specifically, we obtain the best interprediction of three types of CUs, including the co-located CU in the adjacent frame, adjacent left and upper CUs of current block, to determine whether to perform the affine mode after the process of translational motion estimation. Concretely, if the adjacent blocks to the left and above contain skip mode but not affine mode, the affine motion estimation process will be skipped. Furthermore, suppose the co-located CU in the adjacent frame uses skip mode. In that case, the affine motion estimation process for the current CU will also be skipped ahead. The details of the interprediction process, which also included the proposed affine motion estimation early skipping method, are shown in Figure 3.

4. Experiments and Results Analysis

4.1. Experimental Settings

In order to verify the effectiveness of the proposed adjacency encoding information-based fast affine motion estimation method, we implement our method and conduct the experiments based on a VVC test model (VTM-7.0 anchor) under JVET common test conditions (CTC) configurations [38]. The LDP, low delay B (LDB), and random access (RA) configurations are used in experiments, and the quantization parameters (QPs) are set at 22, 27, 32, and 37. We verify the effectiveness of the proposed method through abundant experiments performed on the “computer with Intel core i5-3470 CPU”, including the comparisons with the VVC anchor and state-of-the-art methods. The details of experimental environments are shown in Table 4. Furthermore, the details of open-source test video sequences are displayed in Table 5. The Bjøntegard delta bitrate (BDBR) [39] is used to measure the encoding performance of the introduced adjacency encoding information-based fast affine motion estimation method, and the negative value represents performance gains. Moreover, we also use the Bjøntegard delta peak signal-to-noise rate (BD-PSNR) [39] to evaluate encoding performance, where a positive value denotes improved performance. In addition, the overall encoding time savings S a v T a l l of the introduced method compared to the VVC anchor is calculated as follows,
S a v T a l l = T o r i T p r o T o r i × 100 % ,
where T o r i and T p r o represent the overall encoding time of the original VVC anchor and proposed method. To further demonstrate the effectiveness of the proposed method, we show the time savings S a v T a f of affine motion estimation in the proposed method compared to the original VVC affine motion estimation process, which is calculated as follows,
S a v T a f = T a f o r i T a f p r o T a f o r i × 100 % ,
where T a f o r i and T a f p r o denote the encoding time of affine motion estimation in the original VVC anchor and proposed method.

4.2. Experimental Results and Analyses

First, the VTM-7.0 with AMC is used as a benchmark comparison to represent the performance of the proposed method. The encoding time savings by the adjacency encoding information-based fast affine motion estimation method is shown in Table 6. Concretely, the BDBR and BD-PSNR, which measure the encoding performance of the model, are also included. Table 6 illustrates that compared to the standard anchor, the proposed algorithm achieves a low-complexity encoder to decrease encoding time for all tested sequences. We can observe that the proposed algorithm saves averages of 10.11% of the overall encoding time and 40.85% of the encoding time for the affine motion estimation process. The BD-PSNR only reduces 0.006 dB, and BDBR increases by 0.16%, which means that the proposed algorithm barely reduces the encoding performance of the VVC encoder. The affine motion estimation times for the “BQTerrace”, “BQMall”, “RaceHorses”, “Johnny”, and “SlideEditing” sequences are all saved by over 50%. This is mainly due to the fact that most of these sequences are background regions and contain mostly translational motion, with only a small proportion of the regions using affine mode. As a result, the proposed method reduces the computational complexity of the affine motion estimation in the background and translational motion regions, thereby significantly reducing the encoding time.
To further validate the effectiveness of the proposed method, we also compare the adjacency encoding information-based fast affine motion estimation method with the state-of-the-art fast methods. As displayed in Table 7, the proposed method achieves more encoding time savings compared to Ren et al. without a significant increase in BDBR. As we understand it, the reason may be that although the method proposed by Ren et al. reduces computational complexity by optimizing affine motion estimation, it still requires the execution of affine mode for all CUs. In order to obtain greater encoding time savings, the proposed method is based on the adjacency encoding information to skip most of the affine motion estimation process.
To visualise the impact of the proposed method on VVC compression performance and video quality, the R-D curves for the test sequences “BlowingBubbles” and “BasketballDrill” are presented in Figure 4. The results show that the proposed method almost coincides with the R-D curve of the original VVC, indicating that the adjacency encoding information-based fast affine motion estimation method proposed in this paper does not significantly reduce the compression performance. Furthermore, Figure 5 displays the comparison of the subjective quality from “PartyScene” encoded by the proposed method and the original VVC anchor with QP 27 under RA configuration. We can observe in Figure 5 that the subjective quality difference between the reconstructed frame and the original is virtually invisible to the eyes. In summary, the adjacency encoding information-based fast affine motion estimation method proposed in this paper achieves significant savings in encoding time with negligible subjective quality loss.

5. Conclusions

The VVC is unsuitable for use in real-time applications as a series of computationally complex coding tools added. In order to address the above issue, this paper proposes an adjacency encoding information-based fast affine motion estimation method to save time in video coding. We analyze the trade-off between computational complexity and encoding performance improvement by counting the probability of choosing the affine mode as the best interprediction. Moreover, the affine mode calculation process can be skipped in advance by fully using the adjacency encoding information. The experimental results show that the proposed algorithm reduces 10.24% compared to the anchor encoder without significantly degrading the subjective quality of the encoded video. In future work, we will focus on fast CU partitioning methods in VVC and combine them with the proposed adjacency encoding information-based fast affine motion estimation method to achieve more encoding time savings.

Author Contributions

Conceptualization, X.L.; methodology, X.L.; software, X.L.; validation, X.L. and J.H.; formal analysis, X.L.; investigation, X.L.; writing—original draft preparation, X.L.; writing—review and editing, X.L., J.H., Q.L. and X.C.; visualization, X.L.; supervision, X.L., J.H., Q.L. and X.C.; project administration, X.L. and J.H.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by Natural Science Foundation of Hunan, China (2020JJ7063).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sullivan, G.J.; Ohm, J.R.; Han, W.J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1649–1668. [Google Scholar] [CrossRef]
  2. Bross, B.; Chen, J.; Liu, S. Versatile video coding (Draft 1), document JVET-J1001. In Proceedings of the 10th JVET Meeting, San Diego, CA, USA, 10–20 April 2018. [Google Scholar]
  3. Bross, B.; Chen, J.; Ohm, J.R.; Sullivan, G.J.; Wang, Y.K. Developments in international video coding standardization after avc, with an overview of versatile video coding (vvc). Proc. IEEE 2021, 109, 1463–1493. [Google Scholar] [CrossRef]
  4. Hamidouche, W.; Biatek, T.; Abdoli, M.; Francois, E.; Pescador, F.; Radosavljevic, M.; Menard, D.; Raulet, M. Versatile video coding standard: A review from coding tools to consumers deployment. IEEE Consum. Electron. Mag. 2022, 11, 10–24. [Google Scholar] [CrossRef]
  5. Sidaty, N.; Hamidouche, W.; Déforges, O.; Philippe, P.; Fournier, J. Compression performance of the versatile video coding: HD and UHD visual quality monitoring. In Proceedings of the 2019 Picture Coding Symposium (PCS), Ningbo, China, 12–15 November 2019; pp. 1–5. [Google Scholar]
  6. Li, X.; Chuang, H.C.; Chen, J.; Karczewicz, M.; Zhang, L.; Zhao, X.; Said, A. Multi-type-tree. JVET-D0117. In Proceedings of the 4th JVET Meeting, Chengdu, China, 15–21 October 2016. [Google Scholar]
  7. Schwarz, H.; Nguyen, T.; Marpe, D.; Wiegand, T. Hybrid video coding with trellis-coded quantization. In Proceedings of the 2019 Data Compression Conference (DCC), Snowbird, UT, USA, 26–29 March 2019; pp. 182–191. [Google Scholar]
  8. Zhao, X.; Chen, J.; Karczewicz, M.; Said, A.; Seregin, V. Joint separable and non-separable transforms for next-generation video coding. IEEE Trans. Image Process. 2018, 27, 2514–2525. [Google Scholar] [CrossRef] [PubMed]
  9. Sethuraman, S. CE9: Results of DMVR Related Tests CE9. In 2.1 and CE9.2.2. In Proceedings of the 13th JVET Meeting, JVET-M0147, Marrakech, MA, USA, 9–18 January 2019. [Google Scholar]
  10. Xiu, X.; He, Y.; Ye, Y. CE9-Related: Complexity reduction and bit-width control for bi-directional optical flow (BIO). In Proceedings of the 12th JVET Meeting, JVET-L0256, Macao, China, 3–12 October 2018. [Google Scholar]
  11. Kato, Y.; Toma, T.A.K. Simplification of BDOF. In Proceedings of the 15th JVET Meeting, JVET-O0304, Gothenburg, Sweden, 3–12 July 2019. [Google Scholar]
  12. Lin, S.; Chen, H.; Zhang, H.; Maxim, S.; Yang, H.; Zhou, J. Affine transform prediction for next generation video coding, document COM16-C1016. In Proceedings of the Huawei Technologies, International Organisation for Standardisation Organisation Internationale De Normalisation ISO/IEC JTC1/SC29/WG11 Coding of Moving Pictures and Audio, ISO/IEC JTC1/SC29/WG11 MPEG2015/m37525, Geneva, Switzerland, 1 January 2017. [Google Scholar]
  13. Chen, J.; Karczewicz, M.; Huang, Y.W.; Choi, K.; Ohm, J.R.; Sullivan, G.J. The joint exploration model (JEM) for video compression with capability beyond HEVC. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 1208–1225. [Google Scholar] [CrossRef]
  14. Meuel, H.; Ostermann, J. Analysis of affine motion-compensated prediction in video coding. IEEE Trans. Image Process. 2020, 29, 7359–7374. [Google Scholar] [CrossRef]
  15. Huo, J.; Ma, Y.; Wan, S.; Yu, Y.; Wang, M.; Zhang, K.; Gao, W. CE3-1.5: CCLM derived with four neighbouring samples. In Proceedings of the Joint Video Experts Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, Document JVET N0271, Geneva, Switzerland, 19–27 March 2019. [Google Scholar]
  16. Laroche, G.; Taquet, J.; Gisquet, C.; Onno, P. CE3-5.1: On cross-component linear model simplification. In Proceedings of the Joint Video Exploration Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, JVET-L0191, Macao, China, 8–12 October 2018. [Google Scholar]
  17. Zhao, X.; Seregin, V.; Said, A.; Zhang, K.; Egilmez, H.E.; Karczewicz, M. Low-complexity intra prediction refinements for video coding. In Proceedings of the 2018 Picture Coding Symposium (PCS), San Francisco, CA, USA, 24–27 June 2018; pp. 139–143. [Google Scholar]
  18. Said, A.; Zhao, X.; Karczewicz, M.; Chen, J.; Zou, F. Position dependent prediction combination for intra-frame video coding. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 534–538. [Google Scholar]
  19. Koo, M.; Salehifar, M.; Lim, J.; Kim, S.H. Low frequency non-separable transform (LFNST). In Proceedings of the 2019 Picture Coding Symposium (PCS), Ningbo, China, 12–15 November 2019; pp. 1–5. [Google Scholar]
  20. Salehifar, M.; Koo, M. CE6: Reduced Secondary Transform (RST) (CE6-3.1). In Proceedings of the Document Joint Video Experts Team of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, JVET-N0193, Geneva, Switzerland, 19–27 March 2019. [Google Scholar]
  21. Yoon, Y.U.; Kim, J.G. Activity-Based Block Partitioning Decision Method for Versatile Video Coding. Electronics 2022, 11, 1061. [Google Scholar] [CrossRef]
  22. Min, B.; Cheung, R.C. A fast CU size decision algorithm for the HEVC intra encoder. IEEE Trans. Circuits Syst. Video Technol. 2014, 25, 892–896. [Google Scholar]
  23. Zhao, J.; Wu, A.; Zhang, Q. SVM-Based Fast CU Partition Decision Algorithm for VVC Intra Coding. Electronics 2022, 11, 2147. [Google Scholar] [CrossRef]
  24. Khan, S.N.; Muhammad, N.; Farwa, S.; Saba, T.; Khattak, S.; Mahmood, Z. Early Cu depth decision and reference picture selection for low complexity Mv-Hevc. Symmetry 2019, 11, 454. [Google Scholar] [CrossRef] [Green Version]
  25. Tang, N.; Cao, J.; Liang, F.; Wang, J.; Liu, H.; Wang, X.; Du, X. Fast CTU partition decision algorithm for VVC intra and inter coding. In Proceedings of the 2019 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS), Bangkok, Thailand, 11–14 November 2019; pp. 361–364. [Google Scholar]
  26. Lin, T.L.; Jiang, H.Y.; Huang, J.Y.; Chang, P.C. Fast binary tree partition decision in H. 266/FVC intra Coding. In Proceedings of the 2018 IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), Taipei, Taiwan, 6–8 July 2018; pp. 1–2. [Google Scholar]
  27. Lei, J.; Li, D.; Pan, Z.; Sun, Z.; Kwong, S.; Hou, C. Fast Intra Prediction Based on Content Property Analysis for Low Complexity HEVC-Based Screen Content Coding. IEEE Trans. Broadcast. 2017, 63, 48–58. [Google Scholar] [CrossRef]
  28. Jin, Z.; An, P.; Shen, L.; Yang, C. CNN oriented fast QTBT partition algorithm for JVET intra coding. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  29. Tang, G.; Jing, M.; Zeng, X.; Fan, Y. Adaptive CU split decision with pooling-variable CNN for VVC intra encoding. In Proceedings of the 2019 IEEE Visual Communications and Image Processing (VCIP), Sydney, Australia, 1–4 December 2019; pp. 1–4. [Google Scholar]
  30. Pan, Z.; Zhang, P.; Peng, B.; Ling, N.; Lei, J. A CNN-based fast inter coding method for VVC. IEEE Signal Process. Lett. 2021, 28, 1260–1264. [Google Scholar] [CrossRef]
  31. Li, Y.; Li, L.; Fang, Y.; Peng, H.; Ling, N. Bagged Tree and ResNet-Based Joint End-to-End Fast CTU Partition Decision Algorithm for Video Intra Coding. Electronics 2022, 11, 1264. [Google Scholar] [CrossRef]
  32. Wu, S.; Shi, J.; Chen, Z. HG-FCN: Hierarchical Grid Fully Convolutional Network for Fast VVC Intra Coding. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5638–5649. [Google Scholar] [CrossRef]
  33. Sharabayko, M.P.; Ponomarev, O.G. Fast rate estimation for RDO mode decision in HEVC. Entropy 2014, 16, 6667–6685. [Google Scholar] [CrossRef] [Green Version]
  34. Wang, R.; Tang, L.; Tang, T. Fast Sample Adaptive Offset Jointly Based on HOG Features and Depth Information for VVC in Visual Sensor Networks. Sensors 2020, 20, 6754. [Google Scholar] [CrossRef]
  35. He, L.; Xiong, S.; Yang, R.; He, X.; Chen, H. Low-Complexity Multiple Transform Selection Combining Multi-Type Tree Partition Algorithm for Versatile Video Coding. Sensors 2022, 22, 5523. [Google Scholar] [CrossRef] [PubMed]
  36. Ren, W.; He, W.; Cui, Y. An improved fast affine motion estimation based on edge detection algorithm for VVC. Symmetry 2020, 12, 1143. [Google Scholar] [CrossRef]
  37. Jung, S.; Jun, D. Context-Based Inter Mode Decision Method for Fast Affine Prediction in Versatile Video Coding. Electronics 2021, 10, 1243. [Google Scholar] [CrossRef]
  38. Bossen, F.; Boyce, J.; Li, X.; Seregin, V.; Sühring, K. JVET common test conditions and software reference configurations for SDR video. Jt. Video Expert. Team (JVET) ITU-T SG 2019, 16, 19–27. [Google Scholar]
  39. Bjontegaard, G. Improvements of the BD-PSNR model. In Proceedings of the ITU-T SG16/Q6, 35th VCEG Meeting, Berlin, Germany, 16–18 July 2008. [Google Scholar]
Figure 1. The affine motion models in VVC. (a) Four-parameter affine motion model; (b) six-parameter affine motion model.
Figure 1. The affine motion models in VVC. (a) Four-parameter affine motion model; (b) six-parameter affine motion model.
Electronics 11 03429 g001
Figure 2. The optimal interprediction selected for each CU in “BasketballDrive”.
Figure 2. The optimal interprediction selected for each CU in “BasketballDrive”.
Electronics 11 03429 g002
Figure 3. The overall framework of interprediction which includes the proposed affine motion estimation early skipping method.
Figure 3. The overall framework of interprediction which includes the proposed affine motion estimation early skipping method.
Electronics 11 03429 g003
Figure 4. The R-D curves of sequences “BlowingBubbles” (Class D) and “BasketballDrill” (Class C) under LDP and RA configurations. (a) LDP configuration; (b) RA configuration.
Figure 4. The R-D curves of sequences “BlowingBubbles” (Class D) and “BasketballDrill” (Class C) under LDP and RA configurations. (a) LDP configuration; (b) RA configuration.
Electronics 11 03429 g004
Figure 5. Subjective quality comparison of the 10th decoding frame of “PartyScene” from Class C. (a) original VVC; (b) proposed method.
Figure 5. Subjective quality comparison of the 10th decoding frame of “PartyScene” from Class C. (a) original VVC; (b) proposed method.
Electronics 11 03429 g005
Table 1. Statistical information of the use of affine mode as optimal interprediction.
Table 1. Statistical information of the use of affine mode as optimal interprediction.
Sequences P ( A )
BlowingBubbles    416 × 24010.7%
BQMall    832 × 48010.3%
RaceHorsesC    832 × 4809.8%
FourPeople    1280 × 7209.0%
Cactus    1920 × 108012.2%
Campfire 3840 × 216015.9%
ParkRunning3 3840 × 216018.5%
Average12.3%
Table 2. Details of the three modes in interprediction.
Table 2. Details of the three modes in interprediction.
ModeTransmission
Skipindex
MergeIndex and prediction residuals
AMVPIndex, prediction residuals, and MVD
Table 3. Statistics on the probability of A event in test video sequences at different resolutions under conditions X s k i p and Y s k i p .
Table 3. Statistics on the probability of A event in test video sequences at different resolutions under conditions X s k i p and Y s k i p .
Sequences P ( A | X skip ) P ( A | Y skip )
BlowingBubbles 416 × 2400.140.20
RaceHorsesC 832 × 4800.100.17
BasketballDrill 832 × 4800.090.19
FourPeople 1280 × 7200.130.24
Cactus 1960 × 12800.160.27
Campfire 3840 × 21600.170.30
ParkRunning3 3840 × 21600.230.36
Average0.120.24
Table 4. The environments and conditions of simulation.
Table 4. The environments and conditions of simulation.
ItemsDescriptions
SoftwareVTM-7.0
Configuration Fileencoder lowdelay P vtm.cfg
encoder lowdelay vtm.cfg
encoder randomaccess vtm.cfg
Video Sequence Size416 × 240, 832 × 480,
1280 × 720, 1920 × 1080, 3840 × 2160
Quantization Parameter (QP)22, 27, 32 and 37
Sampling of Luminance to Chrominance4:2:0
Table 5. Detailed characteristics of the experimental video sequences.
Table 5. Detailed characteristics of the experimental video sequences.
ClassSequencesSizeBit-DepthFrame Rate
A1Campfire3840 × 21601030
FoodMarket43840 × 21601060
A2ParkRunning33840 × 21601050
CatRobot3840 × 21601060
BBasketballDrive1920 × 1280850
BQTerrace1920 × 1280860
Cactus1920 × 1280850
RitualDance1920 × 12801060
BasketballDrill832 × 480850
CBQMall832 × 480860
PartyScene832 × 480850
BasketballPass416 × 240850
DBlowingBubbles416 × 240850
RaceHorses416 × 240830
FourPeople1280 × 720860
EJohhny1280 × 720860
KristenAndSara1280 × 720860
Slideshow1280 × 720820
FSlideEditing1280 × 720830
BasketballDrillText832 × 480850
Table 6. The proposed method compared to the original VVC experimental results.
Table 6. The proposed method compared to the original VVC experimental results.
ClassSequencesBDBR/%BD-PSNR/db SavT all /% SavT af /%
A1Campfire0.20−0.01010.3247.62
FoodMarket40.16−0.0079.1737.36
A2ParkRunning30.22−0.0119.4839.28
CatRobot0.17−0.0069.8343.72
BBasketballDrive0.18−0.0079.0435.27
BQTerrace0.12−0.0019.8059.43
Cactus0.17−0.0048.3727.45
RitualDance0.15−0.0039.1335.82
BasketballDrill0.07−0.00510.7537.22
CBQMall0.010.00010.6850.45
PartyScene0.19−0.0099.2140.28
BasketballPass0.21−0.01011.0733.58
DBlowingBubbles0.23−0.0088.7044.34
RaceHorses0.25−0.0109.1350.16
FourPeople0.21−0.0089.2321.76
EJohhny0.030.00015.4754.19
KristenAndSara0.22−0.0059.4227.49
Slideshow0.09−0.00411.8549.71
FSlideEditing0.06−0.00312.2751.86
BasketballDrillText0.24−0.0119.3229.74
Average-0.16−0.00610.1140.85
Table 7. The proposed method compared to the state-of-the-art experimental results.
Table 7. The proposed method compared to the state-of-the-art experimental results.
SequencesRen et al. [36]Proposed
BDBR/% SavT all /%BDBR/% SavT all /%
Cactus0.116.000.178.37
BQTerrace0.048.000.129.80
BasketballDrive0.085.000.189.04
BQMall0.055.000.0110.68
PartyScene0.264.000.199.21
BasketballDrill0.063.000.0710.75
BasketballPass0.082.000.2111.07
BlowingBubbles0.126.000.238.70
RaceHorses0.085.000.259.13
Average0.104.890.169.64
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, X.; He, J.; Li, Q.; Chen, X. An Adjacency Encoding Information-Based Fast Affine Motion Estimation Method for Versatile Video Coding. Electronics 2022, 11, 3429. https://doi.org/10.3390/electronics11213429

AMA Style

Li X, He J, Li Q, Chen X. An Adjacency Encoding Information-Based Fast Affine Motion Estimation Method for Versatile Video Coding. Electronics. 2022; 11(21):3429. https://doi.org/10.3390/electronics11213429

Chicago/Turabian Style

Li, Ximei, Jun He, Qi Li, and Xingru Chen. 2022. "An Adjacency Encoding Information-Based Fast Affine Motion Estimation Method for Versatile Video Coding" Electronics 11, no. 21: 3429. https://doi.org/10.3390/electronics11213429

APA Style

Li, X., He, J., Li, Q., & Chen, X. (2022). An Adjacency Encoding Information-Based Fast Affine Motion Estimation Method for Versatile Video Coding. Electronics, 11(21), 3429. https://doi.org/10.3390/electronics11213429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop