Embodiments of the present invention generally relate to a method and apparatus for parallel context processing techniques for high coding efficiency entropy coding, which may be used in the video coding standard High Efficiency Video Coding (HEVC).
Context-Adaptive Binary Arithmetic Coding (CABAC) is one of two entropy engines used by the existing video coding standard AVC. CABAC is a method of entropy coding that provides high coding efficiency. Processing in CABAC engine is highly serial in nature. Consequently, in order to decode high bit rate video bit-streams in real-time, the CABAC engine needs to be run at extremely high frequencies which consumes a significant amount of power and in the worst case may not be feasible.
Therefore, there is a need for an improved method and/or apparatus for parallel context processing techniques for high coding efficiency entropy coding in HEVC.
Embodiments of the present invention relate to a method and apparatus for parallel context processing for example for high coding efficient entropy coding, such as, HEVC. The method comprising retrieving syntax element relating to a block of an image, grouping at least two bins belonging to similar context based on the syntax element, and coding the grouped bins in parallel.
So that the manner in which the above recited features of the present invention can be understood in detail, a more particular description of the invention, briefly summarized above, may be had by reference to embodiments, some of which are illustrated in the appended drawings. It is to be noted, however, that the appended drawings illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments.
In some embodiments of parallel entropy coding tools, the parallelism proposed may be broadly classified into three categories: (1) Bin-level parallelism, which parallelizes the BAC, (2) Syntax element-level parallelism, which parallelizes the BAC, the context modeler, and the binarizer and (3) Slice-level parallelism.
A N-bins/cycle coding (NBAC) encodes and decodes N-bins/cycle to achieve N-fold improvement in throughput. The contexts for N-bins are calculated through the use of conditional probabilities. In some HEVC embodiment, the binarizer and context modeler were basically the same as in CABAC of AVC. However, coding schemes are determined variable-to-variable length for coding of the bins. There are two flavors of the scheme: (1) PIPE and (2) V2V. The main difference between the two is the context probabilities are quantized to 12 levels in PIPE and to 64 in V2V. In PIPE/V2V coding scheme, the bins are coded using a parallel bin encoding scheme as shown in
Some embodiments that utilize schemes that interleaves the V2V code words from different partial bitstreams into a single bitstream. As a result, a throughput increase of 6× for PIPE in hardware is possible. Such embodiments usually cause an estimated throughput increase of 3× in BAC stage for PIPE hardware implementation for both the parallel and serial versions of PIPE. Since PIPE uses 12 bitstream buffers and V2V uses 64 bitstream buffers, PIPE is usually utilized more often than V2V from a complexity purpose. However in both cases, there is no estimated overall throughput improvement in the entropy coder due to serial bottlenecks in context processing and binarization.
The NBAC, PIPE, V2V schemes reduces serial dependency in the BAC block. However, the serial dependency in the context modeler and binarizer still remain. So, the effective throughput increase that can be achieved in entropy coding is limited. Hence, techniques for parallelization of context processing (PCP) may be utilized.
In syntax element partitioning, syntax elements such as macroblock type, motion vectors, transform coefficients, significant coefficient map etc. are divided into N groups and each group is coded separately. The context selection and adaptation within a group happens in parallel leading to a potential N-fold speed up in context modeler if the various partitions are balanced in terms of the number of bins they process. In practice, the various partitions are not balanced and the throughput improvement is less than a factor of N.
Syntax element partitioning results showed throughput improvement and BD-Rate. In this embodiment, significance map coding is carried out in AVC CABAC. In such an embodiment, the last significant coefficient flag is transmitted when the related coefficient is determined to be significant. The coefficient is the output of a block after transform and quantization. Also, a coefficient is significant when it has value that is non-zero.
This technique introduces serial dependency in decoding of significance map. When throughput improvement is needed, speculative computation are performed at every bin. Such computations leads to complex logic, as shown in
Significance map coding are parallelized by transmitting the last significant coefficient flag once per certain number of bins. For example,
Such an embodiment reduces the number of last bins that need to be transmitted, but it increases the number of significant bins that need to be transmitted. However, there is about a 5% overall reduction in the number of significance map bins that need to be processed. Our algorithm parallelizes about 21.65% of the bins for largest coding unit (LCTB).
Table 1 shows the distribution of bins used by different syntax element types as a percent of total bins for a LCTB. The bin distribution was obtained by measuring bins in bitstreams generated, for example, by TMuC-0.1 using cfg files in cfp-fast directory. Shown in Table 1 is the distribution of bins used by different syntax element type as a percent of total bins for an LCU.
The coefficient coding is usually carried out in AVC CABAC. The context used for the absolute value of the coefficient minus one, known as the coefficient level (1) depends on the position of the bin. Thus, when the binIdx is 0 (i.e. first bin of the coefficient level), then the context is derived by (ctxIdxlnc=((numDecodAbsLevelGt1 !=0) ? 0: Min(4, 1+numDecodAbsLevelEq1))); Otherwise, context is divided by (ctxIdxlnc=5+Min(4−((ctxBlockCat==3) ? 1:0), numDecodAbsLevelGt1)). Context processing for the first bin in the absolute value of the coefficient minus one (i.e. Coeff Level BinIdx 0 in
In one embodiment, the encoding Coeff Level BinIdx 0 occurs in a separate bin-plane as shown in the second row of
In AVC, sign information is interleaved along with level information as shown in
Level=1 occurs with the highest probability, so the most probable path in the context processing tree of SIGN0
L1(0)
SIGN1. For this particular path, the context processing efficiency is 50%, meaning half the context processing is wasteful. On the average, for the context processing tree of
In some embodiment, the first two bins in the coefficient level are context coded. The rest of the bins, such as, coefficient sign bins and Golomb-Rice+Exp-Golomb (GR-EG) binarized bins, are bypass coded. As an extension of “Coeff Level BinIdx 0 PCP”, the second bin in the absolute value of the coefficient minus 1 (i.e. Coeff Level BinIdx 1) is also coded in a separate bin-plane. The Coeff Sign Level can be interleaved or be on a separate bin-plane with GR-EG bins.
Since bypass coding is simpler than context coding, bypass bins can be coded faster than context coded bins. In particular, many bypass bins can be coded in a cycle which can increase the throughput of the CABAC. With Coeff Level BinIdx 1 PCP all bypass coded bins for coefficients in a given TU are grouped together which increases throughput impact of parallel bypass bins processing.
Variants of this approach include separating GR-EG +sign bins from the Coeff Level BinIdx 0 and Coeff Level BinIdx 1, but keeping the GR-EG +sign bins interleaved and keeping the Coeff Level BinIdx 0 and Coeff Level BinIdx 1 bins interleaved as shown in proposal #2 in
While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
This application is a continuation of U.S. patent application Ser. No. 17/936,003, filed Sep. 28, 2022, currently pending and scheduled to grant as U.S. Pat. No. 11,750,826 on Sep. 5, 2023, which is a continuation of U.S. patent application Ser. No. 16/518,407, filed Jul. 22, 2019 (now U.S. Pat. No. 11,490,103), which is a continuation of U.S. patent application Ser. No. 15/295,689, filed Oct. 17, 2016 (now U.S. Pat. No. 10,362,332), which is a continuation of U.S. patent application Ser. No. 13/184,226, filed Jul. 15, 2011 (now U.S. Pat. No. 9,591,320), which claims the benefit of U.S. Provisional Application No. 61/499,852, filed Jun. 22, 2011, and claims the benefit of U.S. Provisional Application No. 61/364,593, filed Jul. 15, 2010, the entireties of all of which are hereby incorporated by reference. This application is related to U.S. patent application Ser. No. 16/891,353, filed Jun. 3, 2020 (now U.S. Pat. No. 10,939,131).
Number | Name | Date | Kind |
---|---|---|---|
5717394 | Schwartz | Feb 1998 | A |
8660177 | Huang et al. | Feb 2014 | B2 |
20090074057 | Marpe et al. | Mar 2009 | A1 |
20090079602 | Sze | Mar 2009 | A1 |
20090096643 | Chang | Apr 2009 | A1 |
20110001643 | Sze | Jan 2011 | A1 |
20110243226 | Choi et al. | Oct 2011 | A1 |
20120027089 | Chien et al. | Feb 2012 | A1 |
20120230417 | Sole Rojals | Sep 2012 | A1 |
Entry |
---|
Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 “Test Model under Consideration” Dresden, DE, Apr. 15-23, 2010. |
V. Sze, M. Budagavi, “Parallel CABAC,” ITU-T SG 16/Q.6 Doc. COM16-C334, Geneva, Switzerland, Apr. 2008. |
M. Budagavi, V. Sze, et al., Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 “Description of Video Coding Technology Proposal by Texas Instruments, Inc.” Dresden, DE, Apr. 15-23, 2010. |
Andrew Segall, et al., Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 “A Highly Efficient and Highly Parallel System for Video Coding” Dresden, DE, Apr. 15-23, 2010. |
T. Weigand, et al., Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 “Description of Video Coding Technology Proposal by Fraunhofer HHI” Dresden, DE, Apr. 15-23, 2010. |
D. He, et al., Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 “Video Coding Technology Proposal by RIM” Dresden, DE, Apr. 15-23, 2010. |
T. Nguyen, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WPE and ISO/IEC JTC1/SC29/WG11 “CE 11: Coding of Transform Coefficient Levels with Golomb-Rice Codes” Geneva, CH, Mar. 16-23, 2011. |
Thomas Wiegand et al., “WD2: Working Draft 2 of High-Efficiency Video Coding”, JCTVC-D503, Jan. 20-28, 2011, 1-555, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Daegu, KR. |
Marpe et al., “Context-Based Adaptive Binary Arithmetic Coding in the H.264/AVC Video Compression Standard”, Jul. 2003, IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, No. 7. (Year: 2003). |
Number | Date | Country | |
---|---|---|---|
20230412826 A1 | Dec 2023 | US |
Number | Date | Country | |
---|---|---|---|
61499852 | Jun 2011 | US | |
61364593 | Jul 2010 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17936003 | Sep 2022 | US |
Child | 18241316 | US | |
Parent | 16518407 | Jul 2019 | US |
Child | 17936003 | US | |
Parent | 15295689 | Oct 2016 | US |
Child | 16518407 | US | |
Parent | 13184226 | Jul 2011 | US |
Child | 15295689 | US |