This application claims the benefit, under 35 U.S.C. §365 of International Application PCT/EP2006/068786, filed Nov. 23, 2006, which was published in accordance with PCT Article 21(2) on Jun. 14, 2007 in English and which claims the benefit of French patent application No. 0553713, filed Dec. 5, 2005.
The invention relates to a method for generating, for pictures of a high resolution interlaced sequence, at least one motion predictor and, where appropriate, at least one texture predictor from motion data and, where appropriate, texture data associated with pictures of a low resolution interlaced sequence.
Hierarchical encoding methods with spatial scalability are known. Scalability represents the ability to stagger information to make it decodable at multiple resolution and/or quality levels. More specifically, a data stream generated by this type of encoding method is divided into several layers, in particular a base layer and one or more enhancement layers. These methods are used in particular to adapt a single data stream to variable transport conditions (bandwidth, error ratios, etc.), and to the expectations of the customers and the varying capabilities of their receivers (CPU, specifications of the display device, etc.). In the particular case of spatial scalability, the part of the data stream corresponding to low resolution pictures of the sequence can be decoded independently of the part of the data stream corresponding to high resolution pictures. On the other hand, the part of the data stream corresponding to the high resolution pictures of the sequence can be decoded only from the part of the data stream corresponding to the low resolution pictures.
Hierarchical encoding with spatial scalability makes it possible to encode a first data part called the base layer, relative to the low resolution pictures and, from this base layer, a second data part called the enhancement layer, relative to the high resolution pictures. Normally, each macroblock of the high resolution picture is temporally predicted according to a conventional prediction mode (for example, bidirectional prediction mode, direct prediction method, early prediction mode, etc.) or is predicted according to an inter-layer prediction method. In this latter case, motion data (for example, a partitioning of the macroblock into blocks, possibly motion vectors and reference picture indices) and where appropriate texture data associated with a block of pixels of the high resolution picture, is deduced or inherited from the motion data or respectively texture data, is associated with blocks of pixels of a low resolution picture. However, the known methods do not allow such predictors to be generated in the case where the low resolution sequence and the high resolution sequence are interlaced.
The object of the invention is to overcome at least one of the drawbacks of the prior art.
The invention relates to a method for generating for at least one block of pixels of a picture of a sequence of high resolution interlaced pictures, called high resolution sequence, at least one motion predictor from motion data associated with the pictures of a sequence of low resolution interlaced pictures, called low resolution sequence, of the same temporal frequency as the high resolution sequence. Each interlaced picture comprises a top field interlaced with a bottom field and is able to be coded in frame mode or in field mode. Each field of an interlaced picture has associated with it a temporal reference. The method makes it possible in particular to generate the at least one motion predictor for the at least one block of pixels of the high resolution picture:
According to a first embodiment, the at least one motion predictor is generated for the at least one block of pixels of the high resolution picture by sub-sampling the motion data associated with the at least one block of pixels of the top field of a low resolution picture of the same temporal reference as the top field of the high resolution picture with a horizontal inter-layer ratio in the horizontal direction of the picture and a vertical inter-layer ratio in the vertical direction of the picture.
According to another embodiment, the at least one motion predictor is generated for the at least one block of pixels of the high resolution picture by sub-sampling the motion data associated with the at least one block of pixels of the bottom field of the low resolution picture of the same temporal reference as the bottom field of the high resolution picture, with a horizontal inter-layer ratio in the horizontal direction of the picture, and a vertical inter-layer ratio in the vertical direction of the picture.
According to another embodiment, the at least one motion predictor is generated for the at least one block of pixels of the high resolution picture by sub-sampling the motion data associated with the at least one block of pixels of the low resolution picture which comprises a field with the same temporal reference as the top field or bottom field of the high resolution picture having a horizontal inter-layer ratio in the horizontal direction of the picture and a vertical inter-layer ratio in the vertical direction of the picture.
The method also makes it possible to generate, for the block of pixels of the high resolution picture, at least one texture predictor:
Advantageously, the motion data associated with the low resolution pictures comprises motion vectors.
Preferably, the motion vectors are associated with a low resolution picture coded in frame mode or with each of the top and bottom fields of a low resolution picture coded in field mode have the same parity.
The method is advantageously used by a method of encoding high resolution pictures from low resolution pictures and by a method of decoding high resolution pictures from low resolution pictures.
Preferably, the low resolution pictures are encoded according to the MPEG-4 AVC standard.
The invention will be better understood and illustrated by means of exemplary embodiments and advantageous implementations, by no means limiting, given with reference to the appended figures in which:
The invention relates to an inter-layer prediction method which consists in generating motion predictors and, where appropriate, texture predictors for pictures of an ordered sequence of high resolution interlaced pictures, called high resolution sequence, from pictures of an ordered sequence of low resolution interlaced pictures, called low resolution sequence. The sequences are divided into groups of pictures (GOP). Each low and high resolution picture comprises a top field interlaced with a bottom field. In
The invention therefore consists in generating, for pictures of the high resolution sequence or for at least one block of pixels, at least one motion predictor and, where appropriate, at least one texture predictor. A texture predictor associated with a high resolution picture or with at least one block of pixels of a high resolution picture is a picture or a prediction block associated with each of its pixels' texture data (for example, a luminance value and, where appropriate, chrominance values), which is generated from texture data associated with at least one picture (or field) or at least one block of pixels of a low resolution picture (or at least one block of pixels of a field), according to a method of sub-sampling the texture such as the ESS method applied to the texture (ESS standing for Extended Spatial Scalability) that is described in sections S.8.3.6.4 and S.8.5.14.2 of document ISO/IEC MPEG & ITU-T VCEG, entitled “Joint Scalable Video Model JSVM3 Annex-S”, referenced JVT-P202, J. Reichel, H. Schwarz, M. Wien. This document is referenced as JSVM3 below. A motion predictor associated with a high resolution picture or with at least one block of pixels of a high resolution picture is defined as a prediction picture or a prediction block which is associated with motion data (for example, a type of partitioning, possibly reference picture indices making it possible to identify the reference pictures to which the motion vectors point). The motion predictor is generated from motion data associated with at least one picture (or field) or at least one block of pixels of a low resolution picture (or at least one block of pixels of a field) according to a motion sub-sampling method such as the ESS method applied to the motion which is described in section S.8.4.1.6.3 of JSVM3, or such as the modified ESS method, described below, derived from the ESS method applied to the motion. The modified ESS method, referenced below as MESS in
According to the modified ESS method, an intermediate motion predictor is generated by sub-sampling by 2 the motion data associated with the low resolution picture, more particularly, the motion data associated with each of the macroblocks of the low resolution picture, in the vertical direction of the picture, or in the horizontal direction of the picture, or in both directions. The method of sub-sampling by 2 is repeated in the vertical direction of the picture as long as the height of said intermediate predictor is greater than the height of the high resolution picture and it is repeated in the horizontal direction of the picture as long as the width of said intermediate predictor is greater than the width of the high resolution picture. The sub-sampling consists in particular in dividing by two the coordinates of the motion vectors associated with the blocks of pixels. For example, with reference to
The motion predictor associated with the high resolution picture is generated from the last intermediate motion predictor by applying the ESS method with an inter-layer ratio equal to
in the horizontal direction of the picture and
in the vertical direction of the picture, where w, and 2h, are respectively the width and the height of the last intermediate motion predictor generated. Furthermore, for each prediction macroblock, the motion vector inheritance method is modified so as not to generate invalid motion vectors, i.e. vectors that point to fields or frame pictures that are not available in the temporal breakdown process. In this case, if all the motion vectors associated with a prediction macroblock MB_pred are invalid, then the inter-layer motion prediction is not authorized for this macroblock. Otherwise (i.e. if at least one of the vectors is valid), the ESS prediction method applied to the motion is used.
The method according to the invention, illustrated by
Texture predictors associated with a high resolution picture of index k in
the horizontal direction of the picture and
in the vertical direction of the picture;
in the horizontal direction of the picture and
in the vertical direction of the picture.
in the horizontal direction of the picture and
in the vertical direction of the picture.
If the low resolution picture of index k is coded in field mode, motion predictors associated with a high resolution picture of index k in
in the horizontal direction of the picture and
in the vertical direction of the picture;
in the horizontal direction of the picture, and
in the vertical direction of the picture; and
in the horizontal direction of the picture, and
in the vertical direction of the picture.
In the other cases, i.e. if the low resolution picture of index k is coded in frame mode, motion predictors associated with a high resolution picture of index k in
in the horizontal direction of the picture, and
in the vertical direction of the picture;
in the horizontal direction of the picture, and
in the vertical direction of the picture; and
in the horizontal direction of the picture, and
in the vertical direction of the picture.
In this case, the motion predictors associated with the bottom field and top field of the high resolution picture are identical.
If the method is used by a coding method, all motion predictors (respectively texture predictors) may be generated in order to select the most appropriate one according to a given criteria, e.g. a rate distorsion criteria. If said method is used by a decoding method, then a single motion predictor (respectively a single texture predictor) is generated, the type of predictor being specified in the bitstream.
Of course, the invention is not limited to the above-mentioned exemplary embodiments. In particular, those skilled in the art can apply any variant to the embodiments described and combine them to benefit from their different advantages. For example, the method according to the invention can be applied to a part of the high resolution picture. In practice, it is possible to generate motion and/or texture predictors for blocks of pixels (for example, macroblocks measuring 16 by 16 pixels) of the high resolution picture from motion and/or texture data associated with blocks of pixels of the low resolution pictures. Similarly, the invention has been described in the case where the top field of an interlaced picture is displayed first (“top field first” case) and can be extended directly to the case where the bottom field is displayed first (“bottom field first” case) by reversing the top and bottom fields. Moreover, the invention can also be extended to the case of several high resolution sequences (i.e. several enhancement layers). Furthermore, the invention is advantageously used by a method of encoding or decoding a sequence of pictures or video. Preferably, the sequence of low resolution pictures is encoded according to the MPEG4 AVC encoding standard defined in document ISO/IEC 14496-10 (“Information technology—Coding of audio-visual objects—Part 10: Advanced Video Coding”).
Number | Date | Country | Kind |
---|---|---|---|
05 53713 | Dec 2005 | FR | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP2006/068786 | 11/23/2006 | WO | 00 | 9/23/2009 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2007/065796 | 6/14/2007 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5270813 | Puri et al. | Dec 1993 | A |
5742343 | Haskell et al. | Apr 1998 | A |
6057884 | Chen | May 2000 | A |
6400768 | Nagumo et al. | Jun 2002 | B1 |
6493391 | Komori et al. | Dec 2002 | B1 |
7961785 | Marquant et al. | Jun 2011 | B2 |
7970056 | Soroushian | Jun 2011 | B2 |
8023561 | Garrido et al. | Sep 2011 | B1 |
8155192 | Marquant et al. | Apr 2012 | B2 |
20020041632 | Sato et al. | Apr 2002 | A1 |
20030086622 | Klein Gunnewiek et al. | May 2003 | A1 |
20030185456 | Sato et al. | Oct 2003 | A1 |
20040013195 | Panusopone et al. | Jan 2004 | A1 |
20050089098 | Sato et al. | Apr 2005 | A1 |
20050129122 | Booth et al. | Jun 2005 | A1 |
20050190844 | Kadono et al. | Sep 2005 | A1 |
20110170001 | Francois et al. | Jul 2011 | A1 |
Number | Date | Country |
---|---|---|
1390061 | Jan 2003 | CN |
0577337 | Jan 1994 | EP |
0644695 | Mar 1995 | EP |
0876061 | Nov 1998 | EP |
06078292 | Mar 1994 | JP |
06209468 | Jul 1994 | JP |
07162870 | Jun 1995 | JP |
1118085 | Jan 1999 | JP |
2001045475 | Feb 2001 | JP |
2009517941 | Apr 2009 | JP |
2005064948 | Jul 2005 | WO |
2007063017 | Jun 2007 | WO |
Entry |
---|
Applicant's admitted prior art in the disclosure. |
“Spatial Prediction in Scalable Video Coding,” A. Vincent et al., Communications Research Centre, Canada, Sep. 14-18, 1995, pp. 244-249. |
“Hierarchical Coding of Digital Television,” Tihao Chiang et al., IEEE Communications Magazine, May 1994, pp. 38-45. |
“Spatial Domain Resolution Scalable Video Coding,” A Puri et al., Proceedings of the SPIE, vol. 2094, 1993, pp. 718-729. |
“MPEG-2/ECVG Lookahead Hybrid Quantization and Spatially Scalable Coding” by Sadik Bayrakeri and Russel M. Mersereau, Center for Signal and Image Processing, vol. 3024, 1997, pp. 129-137. |
“Extended Spatial Scalability for Non Dyadic Video Formats: from SDTV to HDTV” by Marquant et al, Thomson Video Compression Lab, vol. 5960 No. 1, Jul. 12, 2005, pp. 547-558. |
J. Reichel, H. Schwarz, M. Wien: “Joint Scalable Video Model JSVM-3 Annex S (JVT-P202)”, Joint Video Team (JVT) of ISP/IEC MPEG & ITU-T VCEG, Jul. 25, 2005, XP002378451, pp. 1-151. |
H. Schwarz et al: “Constrained Inter-Layer Prediction for Single-Loop Decoding in Spatial Scalability” Image Processing, 2005,IEEE Int'l Conference on Genova, Italy Sep. 11-14, 2005, pp. 870-873, XP010851192. |
Gwenaelle Marquant et al: “Extended spatial scalability for non-dyadic video formats: From SDTV to HDTV” Proc Spie Int. Soc. Opt. Eng. Proceedings of SPIE—The Int'l Society for Optical Engineering, vol. 5960, No. 1, Jul. 12, 2005, pp. 547-558, XP002378452. |
Search Report Jun. 14, 2007. |
Number | Date | Country | |
---|---|---|---|
20100034269 A1 | Feb 2010 | US |