Embodiments of the present invention relate generally to video coding.
State-of-the-art video-coding methods and standards, for example H.264/MPEG-4 AVC (H.264/AVC), may provide higher coding efficiency than older methods and standards at the expense of higher complexity. Increasing quality requirements and resolution requirements on video coding methods and standards may also increase their complexity. Decoders that support parallel decoding may improve decoding speeds and reduce memory requirements. Additionally, advances in multi-core processors may make encoders and decoders that support parallel decoding desirable.
H.264/MPEG-4 AVC [Joint Video Team of ITU-T VCEG and ISO/IEC MPEG, “H.264: Advanced video coding for generic audiovisual services,” ITU-T Rec. H.264 and ISO/IEC 14496-10 (MPEG4-Part 10), November 2007], which is hereby incorporated by reference herein in its entirety, is a video codec specification that uses macroblock prediction followed by residual coding to reduce temporal and spatial redundancy in a video sequence for compression efficiency.
Some embodiments of the present invention comprise methods, devices and systems for parallel entropy encoding and decoding of a video bitstream based on partitioning of data into entropy slices that may be entropy encoded and decoded independently.
According to one aspect of the present application, a method for decoding a video bitstream is provided. The method comprises: entropy decoding a first portion of a video bitstream, wherein first portion of video bitstream is associated with a video frame, thereby producing a first portion of decoded data; entropy decoding a second portion of video bitstream, wherein second portion of video bitstream is associated with video frame, thereby producing a second portion of decoded data, wherein entropy decoding second portion of video bitstream is independent of entropy decoding first portion of video bitstream; and reconstructing a first portion of video frame associated with video bitstream using first portion of decoded data and second portion of decoded data.
According to another aspect of the present application, a method for decoding a video frame in a video sequence is provided. The method comprises receiving a bitstream; identifying a reconstruction slice in bitstream; identifying a plurality of entropy slices associated with reconstruction slice in bitstream; entropy decoding each of plurality of entropy slices associated with reconstruction slice, thereby producing a plurality of entropy-decoded entropy slices; and reconstructing a portion of a video frame associated with reconstruction slice using plurality of entropy-decoded entropy slices.
According to another aspect of the present application, a method for encoding a video frame in a video sequence is provided. The method comprises: partitioning a first frame in a video sequence into at least one reconstruction slice, thereby producing a first reconstruction slice; and partitioning first reconstruction slice into a plurality of entropy slices.
According to another aspect of the present application, a method for generating a video bitstream for parallel decoding is disclosed. The method comprises: receiving a first video bitstream; identifying a reconstruction slice in video bitstream; entropy decoding a plurality of symbols from reconstruction slice, thereby producing entropy-decoded data associated with reconstruction slice; partitioning entropy-decoded data associated with reconstruction slice into a plurality of entropy slices associated with reconstruction slice; independently entropy encoding the entropy-decoded data of each entropy slice of plurality of entropy slices, thereby producing a plurality of entropy-encoded entropy slices; and generating a second video bitstream comprising plurality of entropy-encoded entropy slices.
In some embodiments of the present invention, a first portion and second portion of an input compressed-video bitstream may be entropy decoded independently. A block of samples of a video frame associated with the second portion of the input compressed-video bitstream may be reconstructed using decoded data from the first portion and the second portion. Thus, the reconstruction neighbor definition and the entropy decoding neighbor definition are not the same.
In some embodiments of the present invention, an encoder may partition input data into entropy slices. The encoder may entropy encode the entropy slices independently. The encoder may form a bitstream comprising entropy-slice headers each of which may indicate the location in the bitstream of the associated data for the entropy slice. In some embodiments of the present invention, a decoder may parse a received bitstream for entropy-slice headers, and the decoder may entropy decode a plurality of entropy slices according to a decoder-defined level of parallelism.
In some embodiments of the present invention, data may be multiplexed at a picture level to form entropy slices. In some embodiments, one, or more, entropy slices may correspond to prediction data, and one, or more, entropy slices may correspond to residual data. In alternative embodiments of the present invention, one, or more, entropy slices may correspond to each of a plurality of color planes.
In some embodiments of the present invention, a bitstream may be transcoded to comprise entropy slices. In these embodiments, a received bitstream may be entropy decoded, a plurality of entropy slices may be constructed, and each of the entropy slices may be independent encoded and written to a transcoded bitstream with an associated entropy-slice header.
The foregoing and other objectives, features, and advantages of the invention will be more readily understood upon consideration of the following detailed description of the invention taken in conjunction with the accompanying drawings.
Embodiments of the present invention will be best understood by reference to the drawings, wherein like parts are designated by like numerals throughout. The figures listed above are expressly incorporated as part of this detailed description.
It will be readily understood that the components of the present invention, as generally described and illustrated in the figures herein, could be arranged and designed in a wide variety of different configurations. Thus, the following more detailed description of the embodiments of the methods, devices and systems of the present invention is not intended to limit the scope of the invention but it is merely representative of the presently preferred embodiments of the invention.
Elements of embodiments of the present invention may be embodied in hardware, firmware and/or software. While exemplary embodiments revealed herein may only describe one of these forms, it is to be understood that one skilled in the art would be able to effectuate these elements in any of these forms while resting within the scope of the present invention.
While any video coder/decoder (codec) that uses entropy encoding/decoding may be accommodated by embodiments of the present invention, exemplary embodiments of the present invention will be illustrated in relation to an H.264/AVC encoder and an H.264/AVC decoder. This is intended for illustration of embodiments of the present invention and not limitation.
State-of-the-art video-coding methods and standards, for example H.264/AVC, may provide higher coding efficiency than older methods and standards at the expense of higher complexity. Increasing quality requirements and resolution requirements on video coding methods and standards may also increase their complexity. Decoders that support parallel decoding may improve decoding speeds and reduce memory requirements. Additionally, advances in multi-core processors may make encoders and decoders that support parallel decoding desirable.
H.264/AVC, and many other video coding standards and methods, are based on a block-based hybrid video-coding approach, wherein the source-coding algorithm is a hybrid of inter-picture, also considered inter-frame, prediction, intra-picture, also considered intra-frame, prediction and transform coding of a prediction residual. Inter-frame prediction may exploit temporal redundancies, and intra-frame and transform coding of the prediction residual may exploit spatial redundancies.
In H.264/AVC, an input picture is partitioned into fixed-size macroblocks, wherein each macroblock covers a rectangular picture area of 16×16 samples of the luma component and 8×8 samples of each of the two chroma components. The decoding process of the H.264/AVC standard is specified for processing units which are macroblocks. The entropy decoder 54 parses the syntax elements of the compressed-video bitstream 52 and de-multiplexes them. H.264/AVC specifies two alternative methods of entropy decoding: a low-complexity technique that is based on the usage of context-adaptively switched sets of variable length codes, referred to as CAVLC, and a computationally more demanding algorithm of context-based adaptively binary arithmetic coding, referred to as CABAC. In both entropy decoding methods, decoding of a current symbol may rely on previously, correctly decoded symbols and adaptively updated context models. In addition, different data information, for example, prediction data information, residual data information and different color planes, may be multiplexed together. De-multiplexing may not be done until elements are entropy decoded.
After entropy decoding, a macroblock may be reconstructed by obtaining: the residual signal through inverse quantization and the inverse transform, and the prediction signal, either the intra-frame prediction signal or the inter-frame prediction signal. Blocking distortion may be reduced by applying a de-blocking filter to every decoded macroblock. No processing may begin until the input signal is entropy decoded, thereby making entropy decoding a potential bottleneck in decoding.
Similarly, in codecs in which alternative prediction mechanisms may be allowed, for example, inter-layer prediction in H.264/AVC or inter-layer prediction in other scalable codecs, entropy decoding may be requisite prior to all processing at the decoder, thereby making entropy decoding a potential bottleneck.
In H.264/AVC, an input picture comprising a plurality of macroblocks may be partitioned into one or several slices. The values of the samples in the area of the picture that a slice represents may be correctly decoded without the use of data from other slices provided that the reference pictures used at the encoder and the decoder are identical. Therefore, entropy decoding and macroblock reconstruction for a slice do not depend on other slices. In particular, the entropy coding state is reset at the start of each slice. The data in other slices are marked as unavailable when defining neighborhood availability for both entropy decoding and reconstruction. In H.264/AVC, slices may be entropy decoded and reconstructed in parallel. No intra prediction and motion-vector prediction are allowed across the slice boundary. De-blocking filtering may use information across slice boundaries.
Flexible macroblock ordering defines a slice group to modify how a picture is partitioned into slices. The macroblocks in a slice group are defined by a macroblock-to-slice-group map, which is signaled by the content of the picture parameter set and additional information in the slice headers. The macroblock-to-slice-group map consists of a slice-group identification number for each macroblock in the picture. The slice-group identification number specifies to which slice group the associated macroblock belongs. Each slice group may be partitioned into one or more slices, wherein a slice is a sequence of macroblocks within the same slice group that is processed in the order of a raster scan within the set of macroblocks of a particular slice group. Entropy decoding and macroblock reconstruction must proceed serially within a slice.
Some embodiments of the present invention may comprise partitioning a picture into one or more reconstruction slices, wherein a reconstruction slice may be self-contained in the respect that values of the samples in the area of the picture that the reconstruction slice represents may be correctly reconstructed without use of data from other reconstruction slices, provided that the references pictures used are identical at the encoder and the decoder. All reconstructed macroblocks within a reconstruction slice may be available in the neighborhood definition for reconstruction.
Some embodiments of the present invention may comprise partitioning a reconstruction slice into more than one entropy slice, wherein an entropy slice may be self-contained in the respect that symbol values in the area of the picture that the entropy slice represents may be correctly entropy decoded without the use of data from other entropy slices. In some embodiments of the present invention, the entropy coding state may be reset at the decoding start of each entropy slice. In some embodiments of the present invention, the data in other entropy slices may be marked as unavailable when defining neighborhood availability for entropy decoding. In some embodiments of the present invention, macroblocks in other entropy slices may not be used in a current block's context model selection. In some embodiments of the present invention, the context models may be updated only within an entropy slice. In these embodiments of the present invention, each entropy decoder associated with an entropy slice may maintain its own set of context models.
Some embodiments of the present invention may comprise CABAC encoding/decoding. The CABAC encoding process includes the following steps:
Binarization: A non-binary-valued symbol (for example, a transform coefficient, a motion vector, or other coding data) is converted into a binary code, also referred to as a bin string.
Binarization is followed, for each bin, also considered bit, of the binarized symbol by:
Context Model Selection: A context model is a probability model for one or more bins of the binarized symbol. The context model comprises, for each bin, the probability of the bin being a “1” or a “0.” The model may be chosen for a selection of available models depending on the statistics of recently coded data symbols, usually based on the left and above neighboring symbols, if available.
Binary Arithmetic Coding: An arithmetic coder encodes each bin according to the selected probability model and is based on recursive interval subdivision.
Probability Update: The selected context model is updated based on the actual coded value.
In some embodiments of the present invention comprising CABAC encoding/decoding, at the decoding start of an entropy slice, all of the context models may be initialized or reset to predefined models.
Some embodiments of the present invention may be understood in relation to
In some embodiments of the present invention, only data from macroblocks within an entropy slice may be available for context model selection during entropy decoding of the entropy slice. All other macroblocks may be marked as unavailable. For this exemplary partitioning, macroblocks labeled 117 and 118 are unavailable for context model selection when decoding symbols corresponding to the area of macroblock labeled 119 because macroblocks labeled 117 and 118 are outside of the entropy slice containing macroblock 119. However, these macroblocks 117, 118 are available when macroblock 119 is reconstructed.
In some embodiments of the present invention, an encoder may determine whether or not to partition a reconstruction slice into entropy slices, and the encoder may signal the decision in the bitstream. In some embodiments of the present invention, the signal may comprise an entropy-slice flag (entropy-slice flag in first entropy slice may be referred to as first flag), which may be denoted “entropy_slice_flag” in some embodiments of the present invention.
Some decoder embodiments of the present invention may be described in relation to
Some alternative decoder embodiments of the present invention may be described in relation to
In some embodiments of the present invention, when there are more than N entropy slices, a decode thread may begin entropy decoding a next entropy slice upon the completion of entropy decoding of an entropy slice. Thus when a thread finishes entropy decoding a low complexity entropy slice, the thread may commence decoding additional entropy slices without waiting for other threads to finish their decoding.
In some embodiments of the present invention which may accommodate an existing standard or method, an entropy slice may share most of the slice attributes of a regular slice according to the standard or method. Therefore, an entropy slice may require a small header. In some embodiments of the present invention, the entropy slice header may allow a decoder to identify the start of an entropy slice and start entropy decoding. In some embodiments, at the start of a picture, or a reconstruction slice, the entropy slice header may be the regular header, or a reconstruction slice header.
In some embodiments of the present invention comprising an H.264/AVC codec, an entropy slice may be signaled by adding a new bit, “entropy_slice_flag” to the existing slice header. Table 1 lists the syntax for an entropy slice header according to embodiments of the present invention, wherein C indicates Category and Descriptor u(1), ue(v) indicate some fixed length or variable length coding methods.
“first_mb_in_slice” specifies the address of the first macroblock in the entropy slice associated with the entropy-slice header. In some embodiments, the entropy slice may comprise a sequence of macroblocks.
“cabac_init_idc” specifies the index for determining the initialization table used in the initialization process for the context mode.
In some embodiments of the present invention, entropy decoding a entropy slice may comprise initializing a plurality of context models; and updating the plurality of context models during entropy decoding the entropy slice.
In some embodiments of the present invention, an entropy slice may be assigned a different network abstraction layer (NAL) unit type from the regular slices. In these embodiments, a decoder may distinguish between regular slices and entropy slices based on the NAL unit type. In these embodiments, the bit field “entropy_slice_flag” is not required.
In some embodiments of the present invention, an entropy slice may be constructed by altering the data multiplexing. In some embodiments of the present invention, the group of symbols contained in an entropy slice may be multiplexed the macroblock level. In alternative embodiments of the present invention, the group of symbols contained in an entropy slice may be multiplexed at the picture level. In other alternative embodiments of the present invention, the group of symbols contained in an entropy slice may be multiplexed by data type. In yet alternative embodiments of the present invention, the group of symbols contained in an entropy slice may be multiplexed in a combination of the above.
Some embodiments of the present invention method comprises encoding a video frame in a video sequence, which includes partitioning a frame in a video sequence into at least one reconstruction slice, thereby producing a reconstruction slice; and partitioning the reconstruction slice into a plurality of entropy slices.
Some embodiments of the present invention comprising entropy slice construction based on picture level multiplexing may be understood in relation to
In some embodiments of the present invention shown in
In some embodiments of the present invention, a compressed-video bitstream may be transcoded to comprise entropy slices, thereby allowing for parallel entropy decoding as accommodated by embodiments of the present invention described above. Some embodiments of the present invention may be described in relation to
The terms and expressions which have been employed in the foregoing specification are used therein as terms of description and not of limitation, and there is no intention in the use of such terms and expressions of excluding equivalence of the features shown and described or portions thereof, it being recognized that the scope of the invention is defined and limited only by the claims which follow.
This application is a continuation of application Ser. No. 17/195,993, filed Mar. 9, 2021, which is a continuation of application Ser. No. 16/852,164, filed Apr. 17, 2020, now issued as U.S. Pat. No. 10,958,943, which is a continuation of application Ser. No. 16/529,109, filed Aug. 1, 2019, now issued as U.S. Pat. No. 10,652,585, which is a continuation of application Ser. No. 16/375,474, filed Apr. 4, 2019, now issued as U.S. Pat. No. 10,484,720, which is a continuation of application Ser. No. 15/897,510, filed Feb. 15, 2018, now issued as U.S. Pat. No. 10,284,881, which is a continuation of application Ser. No. 15/493,093, filed Apr. 20, 2017, now issued as U.S. Pat. No. 9,930,369, which is a continuation of application Ser. No. 15/254,421, filed Sep. 1, 2016, now issued as U.S. Pat. No. 9,681,143, which is a continuation of application Ser. No. 15/040,482, filed Feb. 10, 2016, now issued as U.S. Pat. No. 9,503,745, which is a continuation of application Ser. No. 14/269,755, filed May 5, 2014, now issued as U.S. Pat. No. 9,473,772, which is a division of application Ser. No. 12/934,938, filed Sep. 27, 2010, now issued as U.S. Pat. No. 8,824,541. Application Ser. No. 12/934,938, filed Sep. 27, 2010, is a national phase application under 35 U.S.C. § 371 of PCT application no. PCT/JP2009/056778, filed on Mar. 25, 2009. PCT application no. PCT/JP2009/056778 is a continuation of patent application Ser. No. 12/058,301 filed on Mar. 28, 2008, now issued as U.S. Pat. No. 8,542,748. The entire contents of all the foregoing applications are hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
1809176 | Macgregor | Jun 1931 | A |
5557332 | Koyanagi et al. | Sep 1996 | A |
5809176 | Yajima | Sep 1998 | A |
6351570 | Kobayashi | Feb 2002 | B1 |
6404817 | Saha et al. | Jun 2002 | B1 |
7061410 | Pearson et al. | Jun 2006 | B1 |
7262722 | Jahanghir et al. | Aug 2007 | B1 |
7289673 | Lin et al. | Oct 2007 | B2 |
7305036 | MacInnis et al. | Dec 2007 | B2 |
7336713 | Woo et al. | Feb 2008 | B2 |
7408488 | Sekiguchi et al. | Aug 2008 | B2 |
7460041 | Yang et al. | Dec 2008 | B2 |
7480335 | Payson | Jan 2009 | B2 |
7525456 | Watanabe | Apr 2009 | B2 |
7586924 | Wiegand | Sep 2009 | B2 |
7630565 | Linzer | Dec 2009 | B2 |
7636395 | Yongang et al. | Dec 2009 | B2 |
7656320 | Schneider | Feb 2010 | B2 |
7813431 | MacInnis | Oct 2010 | B2 |
7843997 | Hellman | Nov 2010 | B2 |
7925099 | Ramirez et al. | Apr 2011 | B2 |
7937397 | Pickens | May 2011 | B2 |
8824541 | Zhao et al. | Sep 2014 | B2 |
10284881 | Zhao et al. | May 2019 | B2 |
10652585 | Zhao et al. | May 2020 | B2 |
10958943 | Zhao et al. | Mar 2021 | B2 |
11438634 | Zhao et al. | Sep 2022 | B2 |
20020057739 | Hasebe et al. | May 2002 | A1 |
20030081850 | Karczewicz et al. | May 2003 | A1 |
20030189982 | MacInnis | Oct 2003 | A1 |
20040008766 | Wang et al. | Jan 2004 | A1 |
20040066852 | MacInnis | Apr 2004 | A1 |
20040228410 | Ameres et al. | Nov 2004 | A1 |
20040258162 | Gordon et al. | Dec 2004 | A1 |
20040268329 | Prakasam | Dec 2004 | A1 |
20050047504 | Sung et al. | Mar 2005 | A1 |
20050074176 | Marpe et al. | Apr 2005 | A1 |
20050123207 | Marpe et al. | Jun 2005 | A1 |
20060013318 | Webb et al. | Jan 2006 | A1 |
20060114985 | Linzer | Jun 2006 | A1 |
20060126744 | Peng et al. | Jun 2006 | A1 |
20060176953 | Mohsenian | Aug 2006 | A1 |
20060220927 | Park | Oct 2006 | A1 |
20060233254 | Lee et al. | Oct 2006 | A1 |
20070009047 | Shim et al. | Jan 2007 | A1 |
20070133678 | Sakai | Jun 2007 | A1 |
20070183491 | Pearson et al. | Aug 2007 | A1 |
20070230567 | Wang et al. | Oct 2007 | A1 |
20070242894 | Kautzer et al. | Oct 2007 | A1 |
20070280345 | Tu et al. | Dec 2007 | A1 |
20080001796 | Oshikiri et al. | Jan 2008 | A1 |
20080013622 | Bao et al. | Jan 2008 | A1 |
20080013633 | Ye et al. | Jan 2008 | A1 |
20080056389 | Chiang et al. | Mar 2008 | A1 |
20080063082 | Watanabe et al. | Mar 2008 | A1 |
20080112489 | Malladi et al. | May 2008 | A1 |
20080170615 | Sekiguchi et al. | Jul 2008 | A1 |
20080181298 | Shi et al. | Jul 2008 | A1 |
20080181304 | Sekiguchi et al. | Jul 2008 | A1 |
20080187053 | Zhao et al. | Aug 2008 | A1 |
20080199094 | Choi et al. | Aug 2008 | A1 |
20080219349 | Huang et al. | Sep 2008 | A1 |
20080240233 | Au et al. | Oct 2008 | A1 |
20080262312 | Carroll et al. | Oct 2008 | A1 |
20080310504 | Ye et al. | Dec 2008 | A1 |
20080310512 | Ye et al. | Dec 2008 | A1 |
20080310745 | Ye et al. | Dec 2008 | A1 |
20090079602 | Sze et al. | Mar 2009 | A1 |
20090175349 | Ye et al. | Jul 2009 | A1 |
20090196343 | Cho et al. | Aug 2009 | A1 |
20090196355 | Kao et al. | Aug 2009 | A1 |
20090245349 | Zhao et al. | Oct 2009 | A1 |
20090245664 | Matsumoto et al. | Oct 2009 | A1 |
20090274382 | Lin et al. | Nov 2009 | A1 |
20090279604 | Chono et al. | Nov 2009 | A1 |
20100020866 | Marpe et al. | Jan 2010 | A1 |
20100040138 | Marpe et al. | Feb 2010 | A1 |
20100098155 | Demircin et al. | Apr 2010 | A1 |
20100124286 | Wang | May 2010 | A1 |
20100135416 | Huang et al. | Jun 2010 | A1 |
20100142618 | Pandit et al. | Jun 2010 | A1 |
20100172593 | Chono | Jul 2010 | A1 |
20100238998 | Nanbu et al. | Sep 2010 | A1 |
20100266049 | Hashimoto et al. | Oct 2010 | A1 |
20110026604 | Zhao et al. | Feb 2011 | A1 |
20110200115 | Hayashi et al. | Aug 2011 | A1 |
20110206135 | Drugeon et al. | Aug 2011 | A1 |
20190356932 | Zhao et al. | Aug 2019 | A1 |
20200245001 | Zhao et al. | Jul 2020 | A1 |
20210195249 | Zhao et al. | Jun 2021 | A1 |
Number | Date | Country |
---|---|---|
1100872 | Mar 1995 | CN |
1656793 | May 2006 | EP |
7-23397 | Jan 1995 | JP |
8-116534 | May 1996 | JP |
H 10-174094 | Jun 1998 | JP |
2001-522175 | Nov 2001 | JP |
2002-199392 | Jul 2002 | JP |
2006-014113 | Jan 2006 | JP |
2006-246524 | Sep 2006 | JP |
2007-166192 | Jun 2007 | JP |
2007-520912 | Jul 2007 | JP |
2007-201938 | Aug 2007 | JP |
2007-295392 | Nov 2007 | JP |
2007-300455 | Nov 2007 | JP |
2008-011204 | Jan 2008 | JP |
2008-067026 | Mar 2008 | JP |
2008-521265 | Jun 2008 | JP |
2008-172599 | Jul 2008 | JP |
2008-545297 | Dec 2008 | JP |
2009-239565 | Oct 2009 | JP |
2009-544188 | Dec 2009 | JP |
2005111002 | Oct 2006 | RU |
WO 2003092169 | Nov 2003 | WO |
WO 2004102946 | Nov 2004 | WO |
WO 2005027495 | Mar 2005 | WO |
WO 2006013854 | Feb 2006 | WO |
WO 2006048807 | May 2006 | WO |
WO 2006051796 | May 2006 | WO |
WO 2006106935 | Oct 2006 | WO |
WO 2006122313 | Nov 2006 | WO |
WO 2007136093 | Nov 2007 | WO |
WO 2008008714 | Jan 2008 | WO |
WO 2009119888 | Oct 2009 | WO |
WO 2010095528 | Aug 2010 | WO |
Entry |
---|
EP Extended European Search Report in European Appln. No. 20194889.0, dated Oct. 30, 2020, 5 pages. |
EP Extended European Search Report in European Appln. No. 20202810.6, dated Mar. 31, 2021, 7 pages. |
EP Search Report and Written Opinion in European Appln. No. 16167565.7, dated Jul. 13, 2016, 8 pages. |
Extended European Search Report—Application No. 09725979.0-2223—Report dated Feb. 15, 2011. |
Ghandi et al., “The H.264/AVC Video Coding Standard for the Next Generation Multimedia Communication”, IAEEE Journal, 2004, pp. 1-10. |
H.264/MPEG-4 AVG [Joint Video Team of ITU-T VCEG and ISO/IEC MPEG, “H.264: Advanced video coding for generic audiovisual services,” ITU-T Rec. H.264 and ISO/IEC 14496-10 (MPEG4—Part 10), Nov. 2007]. |
Halabach: “Reduced slice headers and bit error resilience,” 3. JVT Meeting; 60. MPEG Meeting; Jun. 5, 2002-Oct. 5, 2002; Fairfax; (Joint Video Team of ISO/IEC JTC1/SC29), May 10, 2002, 12 pages. |
Halabach: Reduced slice headers and bit error resilience, 3. JVT Meeting; 60. MPEG Meeting; Jun. 5, 2002-Oct. 5, 2002; Fairfax; (Joint Video Team of ISO/IEC JTC1/SC29, May 10, 2002 (May 10, 2002), XPO30005241, ISSN: 0000-0442, pp. 3,4,6. |
Ichimura D et al.: “Stream Structure and CABAC for Interactive. ROI scalability,” 17. JVT Meeting; 74. MPEG Meeting; Oct. 14, 2005-Oct. 21, 2005; Nice, FR; (Joint Video Ream of ISO/IEC JTC1/SC29/WG11 and ITU-T SG.16), No. JVT-Q021rl, Oct. 14, 2005, 12 pages. |
Ichimura D et al.: “Stream Structure and CABAC for mult. Inter. ROI scal,” 17. JVT Meeting; 74. MPEG Meeting; Oct. 14, 2005-Oct. 21, 2005; Nice, FR; (Joint Video Ream of ISO/IEC JTC1/SC29/WG11 and ITU-T SG.16), No. JVT-Q021rl, Oct. 14, 2005 (Oct. 14, 2005), XPO30006184, ISSN: 000-0413, sections 2.1, 2.2; figures 1, 2. |
International Preliminary Report on Patentability Chapter II in International Appln. No. PCT/JP2009/056778, dated May 25, 2010, 20 pages. |
International Search Report—International Application No. PCT/JP2010/068541—dated Dec. 7, 2010. |
International Search Report—International Application No. PCT/JP2011/073156—dated Jan. 10, 2012. |
International Search Report—International Patent Application No. PCT /JP2011 /073150—dated Dec. 27, 2011. |
International Search Report—PCT/JP2011/066633—dated Oct. 25, 2011. |
ISO/IEC 14496-10, Information technology—Coding of audio-visual objects—Part 10: Advanced Video Coding, Second edition Oct. 1, 2004, pp. 1-280. |
ISO/IEC JTC1/SC29 WG1, JPEG 2000, “Coding of Still Pictures: JPEG 2000 Part 1 020719 (Final Publication Draft),” Project 1.29.15444 (JPEG 2000), ITU-T Rec. T.800 (2002), N2678, Jul. 19, 2002, 245 pages. |
ITU-T—Telecommunication Standardization Sector of ITU—Recommendation H.264—“Advanced video coding for generic audio visual services,” Mar. 2010. |
JCT-VC, “Test Model under Consideration” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, JCTVC-A205, Apr. 15-23, 2010, 898 pages provided. |
Jie Zhao and Andrew Segall, “Parallel entropy decoding for high resolution video coding,” Proc. of SPIE-IS&T Electronic Imaging, SPIE vol. 7257, Jan. 19, 2009, pp. 725706-1-725706-11. |
Joint Collaborative Team on Video Coding (JCT-VG) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/ WG11, “Test Model under Consideration,” Document: JCTVC-A205, Jun. 6, 2010. |
Marpe, Schwarz and Wiegand, “Context-Based Adaptive Binary Arithmetic Coding in the H.264/AVC Video Compression Standard,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, No. 7, Jul. 2003, pp. 620-636. |
Misra et al., “Entropy Slices for Parallel Entropy Coding,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Jul. 2010, pp. 1-6, 2nd Meeting, Geneva, CH. |
PCT/JP2009/056778—“International Search Report and Written Opinion of the International Searching Authority”—dated Jun. 23, 2009. |
Puri et al., “Video coding using the H.264/MPEG-4 Avc compression standard,” Signal Processing: Image Communication 19, 2004, pp. 793-849 (799-802, 838-839), p. 57. |
Segall et al., “A Highly Efficient and Highly Parallel System for Video Coding,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 JCTVC-A105, Apr. 2010, pp. 25-28, Dresden, DE. |
Sharp, “Entropy slices for parallel entropy decoding,” Telecommunication Standardization Sector, International Telecommunication Union, COM 16-C 405-E, pp. 1-9, Apr. 2008. |
Sullivan et al., “Editors' Text for ISO/IEC 14496-10:2005 (AVC 3rd Edition),” ISO/IEC JTC1/SC291WG11 MPEG05/N7081, Apr. 2005, Section 6.3, Section 7_3_3, Section 9.3, Busan, KR. |
Sullivan et al., “Informal Report of Question 6/16 “Visual Coding” at Apr. 2008 SG 15 meeting,” VCEG-AI04, Rapporteur Q6/16, Berlin, Jul. 16-18, 2008, 35 Meeting, 24 pages. |
Sullivan et al., “Video Compression- From Concepts to the H.264/AVC Standard,” Proceedings of the IEEE, IEEE, Jun. 27, 2005, 14 pages. |
Thomas Wiegand, Editor's Proposed Draft Text Modifications for Joint Video Specification (ITU-T Rec. H.264 | ISO/IEC 14496-10 AVC), Draft 7, Joint Video Team of ISO/IEC MPEG and ITU-T VCEG 5th Meeting: Geneva, Switz, Oct. 9-17, 2002, Oct. 17, 2002, JVT-E022d7 , pp. i, 19, 37. |
Tol et al., “Mapping of H.264 decoding on a multiprocessor architecture,” Dec. 7, 2005, Philips Research Laboratories pp. 1-12. |
Van Der Tol et al., “Mapping of H.264 decoding on a multiprocessor architecture,” Image and video Communications and Processing 2003 (Proceedings volume). Editor(s) haskaran Vasudev; T. Russell Hsing; Andrew G. Tescher; Touradj Ebrahimi, Proceedings of the SPIE, vol. 5022, p. 707-718, May 7, 2003, USA. |
Wiegand et al., “Joint Draft ITU-T Rec. H.264 | ISO/IEC 14496-10/Amd.3 Scalable video coding,” Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG, Document JVT-X201, (ISO/IEC JTC1/SC29/WG11 and ITU-T SG 16 Q.6), 24th Meeting: Geneva, Switzerland, Jun. 29-Jul. 5, 2007, 563 pages. |
Wiegand et al., Draft ITU-T Recommendation and Final Draft international Standard of Join Video Specification (ITU-T REC. H.264 | ISO/IEC 14496-0 AVC), Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG, document JVT-G050r1, (ISO/IEC JTC1SC29/WG11 and ITU-T SG16 Q.6), 8th Meeting: Geneva, Switzerland, May 23-27, 2003, 269 pages. |
Wiegand et al., “Editor's Proposed Draft Text Modifications for Joint Video Specification (ITU-T Rec. H.264 | ISO/IEC 14496-'0 AVC),” Joint Video Team (JVT) of ISO/IEC MPEGP and ITU-T VCEG, Document JVT-E022d1, 4th meeting, Klagenfurt, Austria, Jul. 22-26, 2002, 209 pages. |
Wiegand, “Editor's Proposed Draft Text Modifications for Joint Video Specification (ITU-T Rec. H.264 | ISO/IEC 14496-10 AVC), Draft 2,” Joint Video Team (JVT) of ISO/IEC MPEG and ITU-T VCEG, Document JVT-E022d2, 5th Meeting, Geneva, Switzerland, Oct. 9-17, 2002, 211 pages. |
Wiegand, “Overview of the H.264/AVC Video Coding Standard,” IEEE Transactions on Circuits and Systems for Video Technology, Jul. 2003, 560-576. |
Winken et al., “Description of video coding technology proposal by Fraunhofer HHI,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Apr. 2010, pp. 14-17, 1st Meeting, Dresden DE. |
Written Opinion of the International Searching Authority—International Patent Application No. PCT/JP2011/073150—dated Dec. 27, 2011. |
Zhang et al., “Variable-Bin-Rate CABAC Engine for H.264/AVC High Definition Real-Time Decoding”, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 17, No. 3, Mar. 2009, pp. 417-426. |
Zhang et al., “Variable-Bin-Rate CABAC Engine for H.264/AVC High Definition Real-Time Decoding,” IEEE Transactions on Very Large Scale Integration (VLSI) Systems, Feb. 2009, pp. 417-426, vol. 17 No. 3, IEEE Educational Activities Department Piscataway, NJ, USA. |
Zhao et al., “New Results using Entropy Slices for Parallel Decoding,” VCEG-A132, ITU-Telecommunications Standardization Sector, Study Group 16, Question 6, Video Coding Experts Group, Jul. 16-18, 2008, pp. 1-9, Meeting Berlin, Germany. |
Zhao, Jie et al., “Parallel entropy decoding for high-resolution video coding,” Proc. SPIE, vol. 7257, pp. 725705-1 to 725705-11, Jan. 2009, 11 pages. |
Number | Date | Country | |
---|---|---|---|
20230082110 A1 | Mar 2023 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12934938 | US | |
Child | 14269755 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17195993 | Mar 2021 | US |
Child | 17900698 | US | |
Parent | 16852164 | Apr 2020 | US |
Child | 17195993 | US | |
Parent | 16529109 | Aug 2019 | US |
Child | 16852164 | US | |
Parent | 16375474 | Apr 2019 | US |
Child | 16529109 | US | |
Parent | 15897510 | Feb 2018 | US |
Child | 16375474 | US | |
Parent | 15493093 | Apr 2017 | US |
Child | 15897510 | US | |
Parent | 15254421 | Sep 2016 | US |
Child | 15493093 | US | |
Parent | 15040482 | Feb 2016 | US |
Child | 15254421 | US | |
Parent | 14269755 | May 2014 | US |
Child | 15040482 | US | |
Parent | 12058301 | Mar 2008 | US |
Child | 12934938 | US |