Field of the Invention
The present invention generally relates to video encoders. More specifically, the present invention provides the grouping of coded pixel blocks having similar prediction dependencies to more quickly decode a compressed video data sequence.
Background Art
The decoder 104 receives and decodes the compressed bit stream to reproduce the source video data. The decoder 104 can then provide the reproduced source video data to a video display device 110.
The encoder 102 and the decoder 104 can be implemented in hardware, software or some combination thereof. For example, the encoder 102 and/or the decoder 104 can be implemented using a computer system.
As shown in
The computer system 200 can implement a variety of video coding protocols such as, for example, any one of the Moving Picture Experts Group (MPEG) standards (e.g., MPEG-1, MPEG-2, or MPEG-4) and/or the International Telecommunication Union (ITU) H.264 standard. Most coding standards are designed to operate across a variety of computing platforms. Accordingly, many coding standards find application in feature rich computing devices (e.g. personal computers or gaming devices) and also in feature poor computing devices (e.g., single digital signal processing (DSP) devices).
To accommodate the broad variety of computing devices, most coding standards are designed with unsophisticated computer systems in mind. Specifically, many coding standards are designed to be implemented the same way on a feature rich system as they are on a feature poor system. Feature poor systems typically have limited memory and processor capabilities. Accordingly, due to the design of many coding standards, the improved memory and processor capabilities of a feature rich system are not fully exploited. For example, during the coding of an Intra coded frame (I-frame), macroblock data may be coded with reference to other macroblocks in the same frame. A prediction reference (e.g., a prediction vector) can be generated to specify a location from within previously coded macroblocks from which a prediction will be made. This technique can create a long chain of predictions, requiring macroblocks to be retrieved from memory and then decoded in a serial fashion (e.g. one macroblock at a time).
In feature poor computing devices, the decoded reference frames 222 needed by the processor 202 to decode a current frame 224 may take several memory cycles to retrieve since the bandwidth of a memory system 204 in a feature poor computing device is generally limited. Further, due to the serial nature of decoding the coded data 220, different sets of decoded reference frames 222 may be retrieved for each coded video frame 220. This can result in the same decoded reference frames 222 being retrieved several times as the coded video data 220 is decoded.
As previously mentioned, feature rich computing devices typically possess much greater memory capacity and memory bandwidth in comparison to feature poor devices. Accordingly, these devices are able to access, retrieve and process data in quantities much larger than a single macroblock at a time. Theoretically, the performance of a feature rich computing device implementing a coding standard is much greater than what is currently achieved. Performance is limited due to the serial nature of decoding. In particular, performance is mainly limited by (a) the processor being idle when macroblock data is being read in from a memory device or I/O device because the processor needs prediction data contained in data yet to be retrieved; and (b) the memory device or I/O device being idle because the processor is busy decoding data that requires large amounts of processing.
Accordingly, what is needed is a processing system capable of exploiting the improved processor and memory capabilities of feature rich computing devices to more quickly decode video data compressed according to conventional coding techniques. In particular, the processing system should be capable of reading and decoding multiple chunks of video data (e.g., multiple macroblocks or multiple blocks of pixels) whenever possible.
The accompanying drawings illustrate the present invention and, together with the description, further serve to explain the principles of the invention and to enable one skilled the pertinent art to make and use the invention.
Embodiments of the present invention provide systems, apparatuses and methods whereby groups of coded pixel blocks having similar prediction dependencies are defined and efficiently scheduled for decoding. In this regard, the present invention enables a compressed video data sequence to be decoded in less time and with less power by leveraging improved processor and memory capabilities. A group of coded pixel blocks having similar prediction dependencies is understood to mean coded pixel blocks that can share a motion vector or reference frame or that do not depend on a reconstructed pixel block within the group.
At step 302, a processor of a decoder can buffer channel data in a memory. The channel data can be compressed video data. The processor and memory can be components of a computer system operating as decoder. The implemented decoder can operate according to or be operable with various video compression standards such as, for example, any one of the Moving Picture Experts Group (MPEG) standards (e.g., MPEG-1, MPEG-2, or MPEG-4) and/or the International Telecommunication Union (ITU) H.264 standard. Overall, the compression scheme used by the decoder is immaterial to the present discussion unless specifically identified herein.
At step 304, the processor begins a new group of pixel blocks (e.g., macroblocks) at an arbitrary location and examines the pixel blocks within the buffer. Typically, a new group of pixel blocks begins immediately after the previous group.
At step 306, a pixel block can be retrieved. The current pixel block can be considered the ‘ith’ pixel block.
At step 308, the coding parameters of the ith pixel block can be examined to determine if the current pixel block can be decoded in parallel with the other pixel blocks in the group. If the current pixel block does not depend on a reconstructed pixel within the group, then the current pixel block can be added to the current group at step 310. Steps 306 and 308 can then be repeated for additional pixel blocks remaining in the buffer. If the current pixel block does depend on a reconstructed pixel within the group, then the group can be closed at step 312 without including the current pixel block. If the current pixel block does depend on another pixel block within the group (i.e., a reconstructed pixel with the group), then the current pixel block can be considered a non-qualifying pixel block.
Further, at step 312, the group formed can be scheduled for decoding in the manner most efficient for the processor. That is, one or more groups of pixel blocks generated according to the previous steps can be retrieved or read in bulk. Thereafter, the one or more groups of pixel blocks can be processed as additional channel data is fed into the buffer since the one or more groups of pixel blocks do not depend on or require data yet to be retrieved. In this way, the processor and memory system can operate simultaneously, thereby decreasing the time needed to decode a compressed video stream.
In general, the generated groups of coded pixel blocks can be retrieved in any manner. That is, each bulk read operation can include only a portion of a single group (e.g., if a given group is large) or multiple groups (e.g., if some of the included groups are small). This allows any given group of coded pixel block to be retrieved and processed in a chunk size that may be larger, smaller or equal to the group size.
Various groupings of pixel blocks are possible using the method illustrated by the flowchart 300. The groupings can be made on a temporal and/or spatial basis. For example, a group can comprise adjacent intra-frame pixel blocks that can be predicted (and therefore decoded) without using decoded values within the group. A group can also comprise adjacent inter-frame pixel blocks that can be predicted without using decoded values within the group. In this way, it is possible to group together pixel blocks that use multiple reference frames and/or multiple motion vectors. Further, a group can comprise adjacent inter-frame pixel blocks that share one reference frame and/or one motion vector. These grouping decisions can be made, for example, at step 308 using information on the current pixel block retrieved at step 305.
In
In an embodiment of the present invention, a bit stream syntax is introduced to identify a group of pixel blocks having similar prediction dependencies. The bit stream syntax and groupings can be generated by an encoder of the present invention.
The pixel blocks 506-1 through 506-N can be grouped together according to prediction dependencies. The pixel blocks 5064 through 506-N can be grouped together in a number of ways. For example, the payload 504 can be (a) a group of adjacent inter pixel blocks that share a common motion vector and/or reference frame; (b) a group of adjacent inter pixel blocks that can be predicted without using decoded values within the group; (c) a group of adjacent intra pixel blocks that can be predicted without using decoded values within the group; or (d) a group of pixel blocks that do not depend on any reconstructed pixel from the group of pixel blocks. In this way, the present invention allows the formation of a group of pixel blocks for decoding that do not share a motion vector or a reference frame. The payload 502 can include a variety of information on the characteristics of the payload 504 and/or its constituent components. For example, the payload 504 can include information indicating the type of grouping carried by the payload 504 as well as the number of pixel blocks 506 within the payload. A header 502 can be associated with each payload 504 comprising the compressed bit stream and can be considered a group identifier of the payload 504. Further, the compressed bit stream can comprise payloads 504 of different types distributed throughout the compressed bit stream in any manner.
An encoder of the present invention can generate the payload group 504 and the header 502 as shown in
As a two-step process, generation of a formatted coded video sequence can be used to optimize the original coding of the video sequence. The generation of a formatted coded video sequence can also be produced using a one-step process whereby an encoder of the present invention can encode and format a video sequence in a serial fashion. An encoder of the present invention may also adjust the coding and formatting process based on the capabilities of remote decoder that may receive the coded and formatted video data sequence. Further, an encoder of the present invention may be capable of exploiting one or more side channels to send information (e.g., user-defined data and/or encoding/formatting data) to a decoder. In this way, an encoder can provide a decoder with a coded video sequence formatted in such a way (with or without supplemental information) that it can be decoded more easily and more quickly.
While various embodiments of the present invention have been described above, it should be understood that they have been presented by way of example and not limitation. It will be apparent to one skilled in the pertinent art that various changes in form and detail can be made therein without departing from the spirit and scope of the invention. Therefore, the present invention should only be defined in accordance with the following claims and their equivalents.
This application claims priority to U.S. Provisional Patent Application No. 60/737,772, filed Nov. 18, 2005, herein incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
3716851 | Neumann | Feb 1973 | A |
4023110 | Oliver | May 1977 | A |
4131765 | Kahn | Dec 1978 | A |
4217609 | Hatori et al. | Aug 1980 | A |
4394774 | Wildergren et al. | Jul 1983 | A |
4437119 | Matsumoto et al. | Mar 1984 | A |
4670851 | Murakami et al. | Jun 1987 | A |
4698672 | Chen et al. | Oct 1987 | A |
4760446 | Ninomiya et al. | Jul 1988 | A |
4837618 | Hatori et al. | Jun 1989 | A |
4864393 | Harradine | Sep 1989 | A |
4901075 | Vogel | Feb 1990 | A |
5010401 | Murakami et al. | Apr 1991 | A |
5021879 | Vogel | Jun 1991 | A |
5068724 | Krause et al. | Nov 1991 | A |
5091782 | Krause et al. | Feb 1992 | A |
5093720 | Krause et al. | Mar 1992 | A |
5113255 | Nagata et al. | May 1992 | A |
5168375 | Reisch et al. | Dec 1992 | A |
5175618 | Ueda et al. | Dec 1992 | A |
5223949 | Honjo | Jun 1993 | A |
5260783 | Dixit | Nov 1993 | A |
5293229 | Iu | Mar 1994 | A |
5298991 | Yagasaki et al. | Mar 1994 | A |
5317397 | Odaka et al. | May 1994 | A |
5329318 | Keith | Jul 1994 | A |
5343248 | Fujinami | Aug 1994 | A |
5377051 | Lane et al. | Dec 1994 | A |
5412430 | Nagata | May 1995 | A |
RE34965 | Sugiyama | Jun 1995 | E |
5428396 | Yagasaki et al. | Jun 1995 | A |
RE35093 | Wang et al. | Nov 1995 | E |
5469208 | Dea | Nov 1995 | A |
5469212 | Lee | Nov 1995 | A |
RE35158 | Sugiyama | Feb 1996 | E |
5497239 | Kwon | Mar 1996 | A |
5510840 | Yonemitsu et al. | Apr 1996 | A |
5539466 | Igarashi et al. | Jul 1996 | A |
5543843 | Kato | Aug 1996 | A |
5543847 | Kato | Aug 1996 | A |
5557330 | Astle | Sep 1996 | A |
5557684 | Wang et al. | Sep 1996 | A |
5559557 | Kato | Sep 1996 | A |
5565920 | Lee et al. | Oct 1996 | A |
5568200 | Pearlstein et al. | Oct 1996 | A |
5587806 | Yamada et al. | Dec 1996 | A |
5625355 | Takeuo et al. | Apr 1997 | A |
5648733 | Worrell et al. | Jul 1997 | A |
5654706 | Jeong | Aug 1997 | A |
5666461 | Igarashi et al. | Sep 1997 | A |
5668600 | Lee | Sep 1997 | A |
5684534 | Harney et al. | Nov 1997 | A |
5703646 | Oda | Dec 1997 | A |
5711012 | Bottoms et al. | Jan 1998 | A |
5719986 | Kato et al. | Feb 1998 | A |
5831688 | Yamada et al. | Nov 1998 | A |
5841939 | Takahashi et al. | Nov 1998 | A |
5852664 | Iverson et al. | Dec 1998 | A |
5887111 | Takahashi et al. | Mar 1999 | A |
5917954 | Girod et al. | Jun 1999 | A |
5929916 | Legall et al. | Jul 1999 | A |
5946043 | Lee et al. | Aug 1999 | A |
5949948 | Krause et al. | Sep 1999 | A |
5991447 | Eifrig et al. | Nov 1999 | A |
5991503 | Miyasaka et al. | Nov 1999 | A |
6052507 | Nilda et al. | Apr 2000 | A |
6064776 | Kikuchi et al. | May 2000 | A |
6081296 | Fukunaga et al. | Jun 2000 | A |
6081551 | Etoh | Jun 2000 | A |
RE36761 | Fujiwara | Jul 2000 | E |
6088391 | Auld et al. | Jul 2000 | A |
6115070 | Song et al. | Sep 2000 | A |
6125146 | Frencken et al. | Sep 2000 | A |
6141383 | Yu | Oct 2000 | A |
6144698 | Poon et al. | Nov 2000 | A |
6144700 | Kim | Nov 2000 | A |
6167087 | Kato | Dec 2000 | A |
6169821 | Fukunaga et al. | Jan 2001 | B1 |
6188725 | Sugiyama | Feb 2001 | B1 |
6217234 | Dewar et al. | Apr 2001 | B1 |
6256420 | Sako et al. | Jul 2001 | B1 |
6563549 | Sethuraman | May 2003 | B1 |
6591016 | Fert et al. | Jul 2003 | B2 |
6873654 | Rackett | Mar 2005 | B1 |
7158570 | Nagumo et al. | Jan 2007 | B2 |
7409097 | Zhang et al. | Aug 2008 | B2 |
7535959 | Lightstone et al. | May 2009 | B2 |
7693220 | Wang et al. | Apr 2010 | B2 |
20020168012 | Ramaswamy | Nov 2002 | A1 |
20030123552 | Prakash et al. | Jul 2003 | A1 |
20040008766 | Wang | Jan 2004 | A1 |
20040017951 | Koto et al. | Jan 2004 | A1 |
20040233995 | Abe et al. | Nov 2004 | A1 |
20050053295 | Holcomb | Mar 2005 | A1 |
20050105809 | Abe | May 2005 | A1 |
20050169374 | Marpe | Aug 2005 | A1 |
20050201470 | Sievers | Sep 2005 | A1 |
20050265450 | Raveendran et al. | Dec 2005 | A1 |
20050265452 | Miao | Dec 2005 | A1 |
20050286631 | Wu et al. | Dec 2005 | A1 |
20060013298 | Tong et al. | Jan 2006 | A1 |
20060104360 | Gordon | May 2006 | A1 |
20060168284 | Holthe | Jul 2006 | A1 |
20060222085 | Kishore | Oct 2006 | A1 |
20070025447 | Kojokaro | Feb 2007 | A1 |
20080256091 | Kampmann | Oct 2008 | A1 |
Number | Date | Country |
---|---|---|
WO2005076218 | Aug 2005 | WO |
Entry |
---|
He et al. “Low-delay rate control for DCT video coding via ρ-domain source modeling,” IEEE Trans. on Circuits and Systems for Video Technology, Aug. 2001, vol. 11, No. 8. |
He et al., “Optimum bit allocation and accurate rate control for video coding via ρ-domain source modeling,” IEEE Trans. on Circuits and Systems for Video Technology, Oct. 2002, pp. 840-849, vol. 12, No. 10. |
He et al., “A unified rate-distortion analysis framework for transform coding,” IEEE Trans. on Circuits and Systems for Video Technology, Dec. 2001, pp. 1221-1236, vol. 11, No. 12. |
Wei, “Joint encoder and channel rate control of VBR video over ATM networks,” IEEE Trans. on Circuits and Systems for Video Technology, Apr. 1997, pp. 266-278, vol. 7, No. 2. |
Wei et al., “Rate control of MPEG video coding and recording by Rate-Quantization modeling,” IEEE Trans. on Circuits and Systems for Video Technology, Feb. 1996, pp. 12-20, vol. 6, No. 1. |
Pao et al., “Encoding stored video for streaming applications,” IEEE Trans. on Circuits and Systems for Video Technology, Feb. 2001, pp. 199-209, vol. 11, No. 2. |
Ribas-Corbera et al., “A frame-layer bit allocation for H.263+,” IEEE Trans. on Circuits and Systems for Video Technology, Oct. 2000, pp. 1154-1158, vol. 10, No. 7. |
Yang et al., “Rate control for VBR video over ATM: Simplification and implementation,” IEEE Trans. on Circuits and Systems for Video Technology, Sep. 2001, pp. 1045-1058, vol. 11, No. 9. |
Aramvith et al., “A rate-control for video transport over wireless channels,” IEEE Trans. on Circuits and Systems for Video Technology, May 2001, pp. 569-580, vol. 11, No. 5. |
Boroczky et al., “Joint rate control with look-ahead for multi-program video coding,” IEEE Trans. on Circuits and Systems for Video Technology, Oct. 2000, pp. 1159-1163, vol. 10, No. 7. |
Ribas-Corbera et al., “Rate control in DCT video coding for low-delay communications,” IEEE Trans. on Circuits and Systems for Video Technology, Feb. 1999, pp. 172-185, vol. 9, No. 1. |
Cheng et al., “Rate control for an embedded wavelet video coder,” IEEE Trans. on Circuits and Systems for Video Technology, Aug. 1997, pp. 696-702, vol. 7, No. 4. |
Fan et al., “An active scene analysis-based approach for pseudoconstant bit-rate video coding,” IEEE Trans. on Circuits and Systems for Video Technology, Apr. 1998, pp. 159-170, vol. 8, No. 2. |
Jagmohan et al., “MPEG-4 one-pass VBR rate control for digital storage,” IEEE Trans. on Circuits and Systems for Video Technology, May 2003, pp. 447-452, vol. 13, No. 5. |
Vetro et al., “MPEG-4 rate control for multiple object coding,” IEEE Trans. on Circuits and Systems for Video Technology, Feb. 1999, pp. 186-199, vol. 9, No. 1. |
Ronda et al., “Rate control and bit allocation for MPEG-4,” IEEE Trans. on Circuits and Systems for Video Technology, Dec. 1999, pp. 1243-1258, vol. 9, No. 8. |
Lee et al., “Scalable rate control for MPEG-4 video,” IEEE Trans. on Circuits and Systems for Video Technology, Sep. 2000, pp. 878-894, vol. 10, No. 6. |
Pan et al., “A study of MPEG-4 rate control scheme and its improvements,” IEEE Trans. on Circuits and Systems for Video Technology, May 2003, pp. 440-446, vol. 13, No. 5. |
Lee et al., “Bit allocation for MPEG-4 video coding with spatio-temporal tradeoffs,” IEEE Trans. on Circuits and Systems for Video Technology, Jun. 2003, pp. 488-502, vol. 13, No. 6. |
Puri & Chen, editors, Multimedia Systems, Standards, and Networks, Chapter 3, pp. 55-64, Marcel Dekker, Inc., New York, 2000. |
“Coding of Moving Pictures and Associated Audio Information”, ISO/JEC JTC1/SC29WG11 Video Group, Feb. 1996. |
Haskell, Barry G.; Puri, Atul; and Netravali, Arun N.; “Digital Video: An Introduction to MPEG-2”, Digital Multimedia Standards Series, pp. 33-54, 80-182, 369-411, Chapman & Hall, 1997. |
Mitchell, Joan.; Pennebaker, William B.; Fogg, Chad E.; and LeGall, Didier J.; “MPEG Video Compression Standard” Digital Multimedia Standards Series, pp. 17-32 and 81-103, 333-356, Chapman & Hall, 1996. |
Ghanbari, “Standard Codecs: Image Compression to Advanced Video Coding”, pp. 155-156, The Institute of Electrical Engineers, London, UK, Jul. 2003. |
Number | Date | Country | |
---|---|---|---|
20150334391 A1 | Nov 2015 | US |
Number | Date | Country | |
---|---|---|---|
60737772 | Nov 2005 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13562229 | Jul 2012 | US |
Child | 14723264 | US | |
Parent | 11462449 | Aug 2006 | US |
Child | 13562229 | US |