The present invention relates to video encoding and decoding methods capable of dividing an input picture in units of blocks and encoding and decoding a division block by simultaneously using intra and inter prediction encoding, and more particularly to video encoding and decoding methods of encoding and decoding a division block through inter and intra predictions in video encoding that can improve encoding efficiency and further improve encoding efficiency by encoding a block video signal by use of a square transform or a non-square transform according to a size of the division block.
International Organization for Standardization/International Electrotechnical Commission (ISO/IEC) 14496-10 (Moving Picture Experts Group (MPEG)-4 Part 10 Advanced Video Coding) or H.264, which is video compression/encoding technology jointly standardized in ISO/IEC and International Telecommunication Union-Telecommunication Standardization (ITU-T), an ISO/IEC 14496-10 Amendment 3 (MPEG-4 Scalable Video Coding) standard, a video codec-1 (VC-1), which is an Society of Motion Picture and Television Engineers (SMPTE) standard, an audio video coding standard (AVS), and the like have achieved significant advance in terms of video data compression efficiency.
There are various factors for improving video compression efficiency. In particular, a process of sub-dividing a macroblock in units of 16×16, 16×8, 8×16, 8×8, 8×4, 4×8, and 4×4, prediction-encoding sub-blocks, and encoding a block for minimizing cost in an optimum block mode in terms of rate-distortion cost is performed, unlike existing video encoding standards (MPEG-1 Video, MPEG-2 Video, MPEG-4 Part 2 Visual, H.261, H.263, and the like) in which prediction encoding is performed after a size of a picture to be encoded is divided in units of macroblocks (each of which has 16×16 pixels).
Thereby, a micro motion or a motion of a complex video may be more effectively predicted, and compression efficiency may be significantly improved by significantly reducing a generated residual signal.
In a block-based prediction encoding method as shown in
When intra encoding of sub-blocks into which a 16×16 macroblock to be encoded is divided is performed, the macroblock is subjected to intra prediction encoding in a size of one 16×16 pixel unit. Alternatively, after the macroblock is divided into sub-blocks, intra prediction encoding of four 8×8 blocks or sixteen 4×4 blocks is performed.
In general, the above-described intra prediction encoding technique efficient in terms of the reduction of the number of cases of various block modes in low definition video encoding, but has a problem in high definition (HD) or ultra high definition (UHD) video encoding. That is, in the case of a super-macroblock having a size of 32×32 or more to which a 16×16 macroblock as an encoding unit block is extended, encoding efficiency is degraded if all divided block modes within the super-macroblock are applied to the same intra prediction based on a 16×16, 8×8, or 4×4 block as in the existing method.
In other words, it should be noted that all divided blocks are encoded by only intra or inter prediction encoding in a prediction encoding method based on division blocks in the related art. That is, only one of the intra prediction encoding and the inter prediction encoding is selected and applied to a division block without applying both the intra prediction encoding and the inter prediction encoding to the division block. This may result in a gain of encoding efficiency in image or video compression at an H D rate or less due to simplicity of a syntax expressing a block encoding mode obtained by applying only one of intra and inter encoding, but may become a factor that degrades the encoding efficiency when a unit of encoding is a super-macroblock that has the same or greater size than a macroblock.
The present invention has been made to solve the above-described problem, and an object of the invention is to provide a more effective prediction encoding method that performs an encoding process by extending a method of selecting intra or inter prediction encoding for sub-division blocks of a divided block during video encoding so that both intra and inter prediction encoding modes can be selected and selectively applying a square or non-square transform kernel to a residual signal according to a block size after motion compensation of the divided block.
Another object of the present invention is to provide a computer-readable recording medium for implementing the method.
According to an example embodiment of the present invention for achieving the above-described object, there is provided a video encoding method, including: dividing an input picture into encoding unit blocks; dividing the encoding unit blocks into sub-blocks; and encoding the sub-blocks by selectively using at least one of intra prediction encoding and inter prediction encoding.
The video encoding method may further include: transforming residual signals through the encoding unit blocks and the sub-blocks by selectively applying a variable block-size transform kernel according to a block size; quantizing the transformed residual signals; and entropy-encoding results of quantization.
The residual signals through the sub-blocks may be encoded by selectively applying one or more de-blocking filters according to the block size and an encoding type.
A size of the encoding unit block may be a square having an N*N size, and the encoding unit block may be divided into one or more square or non-square sub-blocks having any size.
When the square or non-square sub-blocks are encoded using the inter prediction encoding, the sub-blocks may be encoded by selecting one of intra prediction encoding methods.
When the intra prediction encoding or inter prediction encoding of the square or non-square sub-blocks is performed, the entropy encoding may be performed by scanning a quantized transform coefficient selected according to a block size.
The square sub-blocks may be transformed by applying a square transform kernel.
When the square sub-blocks are transformed by applying the square transform kernel, a square transform kernel having a size of the smaller or same number of pixels may be applied by comparing the number of horizontal pixels of the square sub-block with the number of vertical pixels.
The non-square sub-blocks may be transformed by applying a non-square transform kernel.
When the non-square sub-blocks are transformed by applying the non-square transform kernel, a non-square transform kernel having a size of the smaller or same number of pixels may be applied by comparing the number of horizontal pixels of the non-square sub-block with the number of vertical pixels.
According to another example embodiment of the present invention, there is provided a method of encoding a division block in video encoding, including: (a) inputting a picture to be encoded; (b) dividing the input picture into encoding unit blocks; (c) dividing each input encoding unit block into sub-blocks; (d) performing intra prediction encoding and inter prediction encoding of the encoding unit block and the sub-blocks and selecting one block type among the encoding unit block and the sub-blocks; and (e) performing intra prediction encoding and/or inter prediction encoding of the encoding unit block and the sub-blocks by use of a prediction result of the block type.
According to still another example embodiment of the present invention, there is provided a method of encoding a division block in video encoding, including: (a′) inputting a picture to be encoded; (b′) dividing the input picture into encoding unit blocks; (c′) determining whether to perform inter prediction encoding of a current input picture; (d′) initializing an index of a sub-block of an encoding unit block to be encoded in the input picture if the current input picture corresponds to inter prediction; (e′) selecting a block mode of the encoding unit block to be encoded; (f′) determining whether to perform both intra and inter prediction encoding of the selected block mode; (g′) performing the intra and inter prediction encoding of the selected block mode if it is determined to perform both the intra and inter prediction encoding of the selected block mode; (h′) storing a prediction encoding result and a rate-distortion cost value of step (g′); (i′) if the selected block mode is a final mode, determining encoding by comparing rate-distortion costs for respective block modes and selecting a final block mode for the encoding unit block; (j′) determining whether a current encoding unit block is a final block in the current input picture; and (k′) determining whether the current input picture is a final picture if the current encoding unit block is the final block in the current input picture, and iterating steps (a′) to (j′) until the current input picture becomes the final picture.
The method may further include: after step (c′), performing the intra prediction encoding if the current input picture does not correspond to the inter prediction.
The method may further include: after step (f′), performing the inter prediction encoding of the selected block mode if it is not determined to perform both the intra and inter prediction encoding of the selected block mode.
The method may further include: after step (g′), obtaining a residual signal through motion prediction and compensation if the inter prediction encoding of the selected block mode is performed; transforming the selected block by use of the residual signal; quantizing the transformed selected block; and entropy-encoding a result of quantization.
The method may further include: after step (g′), obtaining a residual signal through the intra prediction encoding if the intra prediction encoding of the selected block mode is performed; transforming the selected block by use of the residual signal; quantizing the transformed selected block; and entropy-encoding a result of quantization.
The method may further include: performing a transform by selectively applying a transform kernel according to a block size when the selected block is transformed using the residual signal.
According to still another example embodiment of the present invention, there is provided a method of decoding a division block in video decoding, comprising: (A) inputting a bitstream to be decoded; (B) determining whether the input bitstream corresponds to inter prediction; (C) performing intra prediction encoding if the input bitstream is subjected to intra prediction encoding; (D) analyzing a slice if the input bitstream corresponds to the inter prediction; (E) analyzing a unit encoding block within the slice; (F) decoding an encoding mode of a unit encoding sub-division block; (G) analyzing whether a sub-division encoding block is an inter prediction encoding block; (H) performing inter prediction decoding if the sub-division encoding block corresponds to the inter prediction; (I) performing intra prediction decoding if the sub-division encoding block corresponds to intra prediction; (J) configuring unit decoding block pixels from sub-division decoding results; (K) configuring slice pixels from decoding unit block results; and (K) configuring a picture from slice pixel configuration results. Accordingly, it is possible to perform intra and/or inter prediction decoding.
According to yet another example embodiment of the present invention, there is provided a computer-readable recording medium storing a program for causing a computer to execute the above-described method.
When pixel values of an encoding unit block are encoded in video encoding according to the present invention, a divided block is encoded in an intra or inter prediction encoding mode by applying at least one of intra prediction encoding and inter prediction encoding to a divided sub-block or its sub-division blocks, and prediction encoding is performed using both intra and inter predictions for an encoding unit block or its sub-division blocks, so that the flexibility of encoding mode selection can be increased and the efficiency of encoding can be increased.
In prediction encoding based on division blocks according to the present invention, each division block is encoded by selectively applying a size of a variable block-size transform kernel according to a size of the division block so that both the intra and inter predictions can be applied to sub-division blocks. Therefore, it is possible to significantly improve encoding efficiency.
Hereinafter, example embodiments of the present invention will be described in detail with reference to the accompanying drawings. The same elements are denoted by the same reference numerals throughout the drawings. Detailed description related to well-known functions or configurations will be omitted if they make the subject matter of the present invention unclear.
As shown in
JMODE(s,r,MODE|QP,λMODE)=SSD(s,r,MODE|QP)+λMODER(s,r,MODE|QP) Equation 1
Here, JMODE denotes a rate-distortion function for a block encoding mode, s denotes an original block pixel input to be encoded, r denotes a reference video pixel input, QP denotes a quantization parameter, denotes a Lagrange multiplier dependent on a mode, and MODE denotes a division block mode type.
Also, if transform encoding is applied to a residual signal of a super-macroblock having an increased size, it is possible to increase encoding efficiency by selectively applying a square transform kernel having a size of 16×16 or more, which is greater than existing sizes of 4×4 and 8×8, or a non-square transform kernel having a size of 16×8, 8×16, or more for a non-square transform according to a size of a division block.
If the square transform kernel having the size of 16×16 or more is applied to the super-macroblock, it is possible to carry out a calculation as shown in Equation 2.
Y=AX Equation 2
Here, X denotes an N×N input video signal matrix, A denotes an N×N square transform kernel matrix, and Y denotes a transform coefficient matrix. If a sub-block after division is a non-square block, a transform is performed as shown in Equation 3.
Y=A1XA2 Equation 3
Here, when an input video signal X is an M×(M/2) matrix, A1 denotes an M×M square transform kernel matrix, A2 denotes an (M/2)×(M/2) square transform kernel matrix, and Y denotes a transform coefficient matrix.
When a square or non-square kernel transform is applied according to an example embodiment of the present invention, it is preferable to perform transform encoding by performing a comparison with the smaller number of pixels between the number of horizontal pixels and the number of vertical pixels of a division block and applying a kernel having a size that is equal to or less than the smaller number of pixels.
Referring to
Next, the input picture i is divided into encoding unit blocks (S103). In an example embodiment of the present invention, the encoding unit block may be a macroblock or super-macroblock.
Next, it is checked whether to perform inter prediction encoding of a current picture i (S104). If the current picture i does not correspond to inter prediction, intra prediction encoding is performed (S105). Otherwise, if the current picture i corresponds to inter prediction, an index of an encoding unit block j to be encoded within the one picture i is initialized (j=0) (S106).
Thereafter, the unit block j to be encoded is divided into sub-blocks (S107). An index of a sub-block mode k is initialized (k=0) (S108). One of sub-block modes k is selected (S109).
It is checked whether or not to perform intra and inter predictions for the sub-block mode to be encoded in the encoding unit block (S110). If it is checked to perform the intra and inter predictions, intra and inter encoding are performed (S111). Otherwise, only inter prediction encoding is performed (S112). A prediction encoding result and a rate-distortion cost value are stored as encoding results (S113).
It is checked whether the sub-block mode k is a final block node (S114). If the sub-block mode k is not the final block mode, steps S109 to S113 are iterated for the next block mode. On the other hand, if the sub-block mode k is the final block mode, an optimum division block is determined and corresponding encoding results are finally selected (S115).
It is determined whether a current encoding unit block j is a final block in the current picture i (S116). If the current encoding unit block j is not the final block, steps S107 to S115 are iterated by inputting the next encoding unit block.
If it is determined in step S116 that the current encoding unit block j is the final block in the current picture i, it is checked whether the current picture i is a final picture (S117). If the current picture i is the final picture, the algorithm ends. Otherwise, the algorithm returns to step S102, the next picture is input, and steps S102 to S116 are iterated.
Referring to
Next, it is checked whether an input picture bitstream i corresponds to inter prediction encoding (S203). If a current picture bitstream i does not correspond to the inter prediction encoding, intra prediction decoding is performed (S207). Otherwise, if the input picture bitstream i corresponds to the inter prediction encoding, an index of a slice j to be decoded within one picture i is initialized (j=0) (S204).
Next, slice information regarding the input picture bitstream is analyzed (S205). An index of a unit decoding block j to be decoded within each slice within one picture i is initialized (k=0) (S206). In an example embodiment of the present invention, the decoding unit block may be a macroblock or a super-macroblock.
Next, after information regarding each unit encoding block is analyzed (S208), an index of a division sub-block within the unit encoding block is initialized (m=0) (S209). Thereafter, an encoding mode of a division sub-block within the unit encoding block is decoded (S210). After it is checked whether the division sub-block is an inter prediction encoding block (S211), inter prediction decoding is performed if the sub-block is the inter prediction encoding block (S213), and intra prediction encoding is performed if the sub-block is an intra prediction encoding block or mode (S212).
Thereafter, pixel values of the sub-division block are recovered using sub-block encoding results (S214). After it is checked whether a current sub-division block m is the final block (S215), pixel values of a unit decoding block are configured if the current sub-division block m is the final block (S216). Otherwise, the algorithm returns to step S210 for decoding the next division sub-block, and steps S210 to S214 are performed.
After it is checked whether a current unit encoding block k is a final unit encoding block (S217), slice pixels are configured if the current unit encoding block k is the final unit encoding block (S218). Otherwise, the algorithm returns to step S208, and steps S208 to S216 are performed. After it is checked whether a current slice j is a final slice (S219), picture pixels are configured if the current slice j is the final slice (S220). Otherwise, steps S205 to S218 are performed. After it is determined whether a current picture i is a final picture (S221), the algorithm ends if the current picture i is the final picture. Otherwise, the algorithm returns to step S202, the next bitstream is input, and steps S202 to S220 are performed.
In video encoding according to an embodiment of the present invention, an input video is divided into encoding unit blocks. After the encoding unit block is sub-divided into sub-blocks, each sub-division block is encoded by selectively using at least one of intra prediction and inter prediction.
Thereby, encoding can be performed using both inter and intra prediction sub-block modes in an encoding mode of an encoding unit block and simultaneously a variable block-size transform kernel is selectively applied, so that encoding efficiency can be improved.
In video decoding according to an example embodiment of the present invention, it is possible to decode a compressed bitstream with improved encoding efficiency by performing the reverse process of encoding.
As another example embodiment of the present invention, the above-described method of encoding a division block in video encoding can be implemented as computer-readable codes on a computer-readable recording medium. The computer-readable recording medium includes all types of recording devices storing data readable by a computer system.
Examples of the computer-readable recording medium include a read only memory (ROM), a random access memory (RAM), a compact disk-read only memory (CD-ROM), a magnetic tape, a hard disk, a floppy disk, a mobile storage, a flash memory, an optical data storage, etc. Furthermore, the computer-readable recording medium may be implemented in the form of carrier waves (e.g., transmission over the Internet).
Further, the computer-readable recording medium is distributed to computer systems connected to each other through a computer communication network, so that the computer-readable codes can be stored in a distribution manner and executed thereby.
Although preferred embodiments of an inter prediction encoding method and/or an inter prediction encoding method in video encoding and decoding methods, which are reverse processes thereof, have been described, the example embodiments are illustrative and the present invention is not limited thereto. Those skilled in the art will understand that various modifications and changes may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.
Sequence Listing Free Text
Encoding, decoding, inter prediction, intra prediction, transform kernel, square transform kernel, non-square transform kernel, quantization, MPEG, rate-distortion cost, H.264/MPEG-4 Part 10 Advanced Video Coding.
Number | Date | Country | Kind |
---|---|---|---|
10-2009-0015013 | Feb 2009 | KR | national |
10-2009-0076753 | Aug 2009 | KR | national |
This application is a continuation application of U.S. patent application Ser. No. 17/361,528, filed on Jun. 29, 2021, which is a continuation application of U.S. patent application Ser. No. 16/572,704, filed on Sep. 17, 2019, which is a Continuation application of U.S. application Ser. No. 15/855,019 filed on Dec. 27, 2017, now U.S. Pat. No. 10,462,494 issued on Oct. 29, 2019, which is a Continuation of application Ser. No. 15/258,655 filed on Sep. 7, 2016, now U.S. Pat. No. 9,888,259 issued on Feb. 6, 2018, which is a Continuation of application Ser. No. 13/202,906 having a 371(c) date of Oct. 10, 2011, now U.S. Pat. No. 9,485,512 issued on Nov. 1, 2016, which is a U.S. national stage application of International Application No. PCT/KR2010/001125 filed on Feb. 23, 2010, which claims the benefit of Korean Patent Application No. 10-2009-0076753 filed on Aug. 19, 2009 and Korean Patent Application No. 10-2009-0015013 filed on Feb. 23, 2009, in the Korean Intellectual Property Office, the entire disclosures of which are incorporated herein by reference for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
5241395 | Chen | Aug 1993 | A |
6404814 | Apostolopoulos et al. | Jun 2002 | B1 |
6917648 | Morimatsu et al. | Jul 2005 | B1 |
7088780 | Vetro et al. | Aug 2006 | B2 |
7181072 | Wang et al. | Feb 2007 | B2 |
7362809 | Booth et al. | Apr 2008 | B2 |
7397857 | Romanowski et al. | Jul 2008 | B2 |
7657111 | Nonaka et al. | Feb 2010 | B2 |
8085844 | Holcomb et al. | Dec 2011 | B2 |
8442337 | Zuo et al. | May 2013 | B2 |
8494042 | Park et al. | Jul 2013 | B2 |
8498335 | Holcomb | Jul 2013 | B2 |
8611416 | Hwang et al. | Dec 2013 | B2 |
8817868 | Holcomb | Aug 2014 | B2 |
8929455 | Cohen et al. | Jan 2015 | B2 |
9420293 | Song et al. | Aug 2016 | B2 |
20020172283 | Kawakatsu et al. | Nov 2002 | A1 |
20020176498 | Frojdh et al. | Nov 2002 | A1 |
20020191695 | Irvine et al. | Dec 2002 | A1 |
20040062309 | Romanowski et al. | Apr 2004 | A1 |
20040165765 | Sung et al. | Aug 2004 | A1 |
20040233989 | Kobayashi et al. | Nov 2004 | A1 |
20050053294 | Mukerjee | Mar 2005 | A1 |
20050078750 | Shen et al. | Apr 2005 | A1 |
20050147167 | Dumitras et al. | Jul 2005 | A1 |
20050276323 | Martemyanov et al. | Dec 2005 | A1 |
20060002470 | Sakurai | Jan 2006 | A1 |
20060034532 | Govindaswamy et al. | Feb 2006 | A1 |
20060104357 | Burazerovic et al. | May 2006 | A1 |
20060146941 | Cha et al. | Jul 2006 | A1 |
20060153297 | Boyce | Jul 2006 | A1 |
20060176956 | Ducloux et al. | Aug 2006 | A1 |
20060193388 | Woods et al. | Aug 2006 | A1 |
20070009044 | Tourapis et al. | Jan 2007 | A1 |
20070019729 | Nakagomi et al. | Jan 2007 | A1 |
20070033494 | Wenger et al. | Feb 2007 | A1 |
20070047648 | Tourapis et al. | Mar 2007 | A1 |
20070047656 | Kim et al. | Mar 2007 | A1 |
20070053620 | Mizuno | Mar 2007 | A1 |
20070121731 | Tanizawa et al. | May 2007 | A1 |
20070133676 | Lee et al. | Jun 2007 | A1 |
20070133889 | Horie et al. | Jun 2007 | A1 |
20070140352 | Bhaskaran et al. | Jun 2007 | A1 |
20070198622 | Jou | Aug 2007 | A1 |
20070253484 | Zhao et al. | Nov 2007 | A1 |
20070297506 | Yamanaka | Dec 2007 | A1 |
20070297518 | Han et al. | Dec 2007 | A1 |
20080008250 | Mori | Jan 2008 | A1 |
20080056347 | Chiu et al. | Mar 2008 | A1 |
20080082550 | Ahopelto et al. | Apr 2008 | A1 |
20080084932 | Wang | Apr 2008 | A1 |
20080089597 | Guo et al. | Apr 2008 | A1 |
20080126278 | Bronstein et al. | May 2008 | A1 |
20080151995 | Youn | Jun 2008 | A1 |
20080152005 | Oguz et al. | Jun 2008 | A1 |
20080192824 | Lim et al. | Aug 2008 | A1 |
20090003440 | Karczewicz et al. | Jan 2009 | A1 |
20090016626 | Zhang et al. | Jan 2009 | A1 |
20090046781 | Moriya et al. | Feb 2009 | A1 |
20090219991 | Po et al. | Sep 2009 | A1 |
20090245382 | Ekman | Oct 2009 | A1 |
20090263032 | Tanaka et al. | Oct 2009 | A1 |
20090268810 | Dai | Oct 2009 | A1 |
20090268974 | Takagi | Oct 2009 | A1 |
20090310677 | Shiodera et al. | Dec 2009 | A1 |
20100027686 | Zuo et al. | Feb 2010 | A1 |
20100166069 | Goel | Jul 2010 | A1 |
20100208804 | Yu et al. | Aug 2010 | A1 |
20100208806 | Yu et al. | Aug 2010 | A1 |
20100211537 | Gepperth | Aug 2010 | A1 |
20100208805 | Yu et al. | Sep 2010 | A1 |
20100239015 | Wang et al. | Sep 2010 | A1 |
20110110427 | Teng et al. | May 2011 | A1 |
20110116545 | Zan et al. | May 2011 | A1 |
20120201300 | Kim et al. | Aug 2012 | A1 |
20130136179 | Lim et al. | May 2013 | A1 |
20130202036 | Nagori et al. | Aug 2013 | A1 |
20140079330 | Zuo et al. | Mar 2014 | A1 |
20150334402 | Song et al. | Nov 2015 | A1 |
20160381382 | Oh et al. | Dec 2016 | A1 |
20170041616 | Ramamurthy et al. | Feb 2017 | A1 |
Number | Date | Country |
---|---|---|
2659351 | Jan 2008 | CA |
1819658 | Aug 2006 | CN |
1843038 | Oct 2006 | CN |
1320504 | Jun 2007 | CN |
101360239 | Feb 2009 | CN |
3220887 | Sep 1991 | JP |
2003-533101 | Nov 2003 | JP |
2003-533141 | Nov 2003 | JP |
2004-242318 | Aug 2004 | JP |
2004-528791 | Sep 2004 | JP |
2005-123732 | May 2005 | JP |
2006-516072 | Jun 2006 | JP |
2006-517364 | Jul 2006 | JP |
2006-222968 | Aug 2006 | JP |
2006-287973 | Oct 2006 | JP |
2007-201558 | Aug 2007 | JP |
2008-522537 | Jun 2008 | JP |
2008-219870 | Sep 2008 | JP |
2017-113145 | Jun 2017 | JP |
2021-183506 | Dec 2021 | JP |
10-2001-0062140 | Jul 2001 | KR |
10-2005-0089876 | Sep 2005 | KR |
10-2006-0090749 | Aug 2006 | KR |
10-0631714 | Oct 2006 | KR |
10-2009-0065401 | Jun 2009 | KR |
10-2011-0017719 | Feb 2011 | KR |
WO 03021971 | Mar 2003 | WO |
WO 03026315 | Mar 2003 | WO |
WO 2004036501 | Apr 2004 | WO |
WO 2005022919 | Mar 2005 | WO |
WO 2007034918 | Mar 2007 | WO |
WO 2008012918 | Jan 2008 | WO |
WO 2008027192 | Mar 2008 | WO |
WO 2008088140 | Jul 2008 | WO |
WO 2008120434 | Oct 2008 | WO |
WO 2008126135 | Oct 2008 | WO |
Entry |
---|
Phillips, W., “A Comparison of Four Hybrid block/object Image Coders”, Signal Processing, vol. 54, Issue 1, pp. 103-107. |
Heiko Schwarz, et al. “Tree-Structured Macroblock Partition.” 15th Meeting of the ITU-Telecommunications Standardization Sector Study Group Video Coding Experts Group (VCEG), XP030003350, Pattaya, Thailand, Dec. 2001. (6 pages in English). |
Mathias Wien, et al. “Hybrid Video Coding Using Variable Size Block Transforms.” Visual Communications and Image Processing 2002, XP030080600, San Jose, CA, Jan. 2002: 1040-1051. |
Joint Video Team, “Draft ITU-T Recommendation H.264 (a.k.a. “H.26L”).” 16th Meeting of the ITU-Telecommunications Standardization Sector Study Group Video Coding Experts Group (VCEG), XP030003397, Fairfax, Virginia, USA, May 2002. (142 pages in English). |
Peter List, et al. “Adaptive Deblocking Filter.” IEEE Transactions on Circuits and Systems for Video Technology, XP011221094, 13.7 (2003): 614-619. |
Wien, Mathias. “Variable block-size transforms for H. 264/AVC.” IEEE Transactions on Circuits and Systems for Video Technology 13.7 (2003): 604-613. (10 pages in English). |
Wien, Mathias, “Variable Block-Size Transforms for H.264/AVC”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, No. 7, Jul. 2003 (pp. 604-610). |
Shin et al. “Variable Block-Based Deblocking Filter for H.264/AVC”, IEEE International Conference on Multimedia and Expo, 2007, pp. 436-439, Jul. 2-5, 2007. |
Ma, Siwei, et al. “High-definition video coding with super-macroblocks.” Visual Communications and Image Processing 2007. vol. 6508. International Society for Optics and Photonics, 2007. (12 pages in English). |
Lee, Kyohyuk et al., “Technical considerations for Ad Hoc Group on New Challenges in Video Coding Standardization”, International Organisation for Standardisation Organisation Internationale De Normalisation ISO/IEC JTC1/SC29/WG11 Coding of Moving Pictures and Audio, Jul. 2008, Hannover, Germany (pp. 1-9). |
Jaeil Kim et al., “Enlarging MB Size for High Fidelity Video Coding Beyond HD.” 36th Meeting of the ITU-Telecommunications Standardization Sector Study Group Video Coding Experts Group (VCEG), San Diego, USA, Oct. 2008. (6 pages in English). |
P. Chen, et al., “Video Coding Using Extended Block Sizes”. 36th Meeting of the ITU—Telecommunications Standardization Sector, Study Group 16 (Video Coding Experts Group, VCEG), Question 6, Oct. 2008, San Diego, USA. (3 pages in English). |
Peisong Chenn et al., “Video Coding Using Extended Block Sizes.” ITU—Telecommunications Standardization Sector Study Group, Com 16-C123-E, Jan. 2009. (4 pages in English). |
Japanese Office Action issued on Feb. 5, 2014 in counterpart Japanese Patent Application No. 2011-551018. (2 pages in Japanese). |
Korean Office Action issued on Nov. 22, 2014 in counterpart Korean Application No. 10-2011-0088186 (4 pages in Korean). |
Korean Office Action issued on Apr. 21, 2015 in counterpart Korean Application No. 10- 2015-0010081 (9 pages in Korean). |
Extended European Search Report issued on Nov. 13, 2015 in counterpart European Application No. 10743990.3 (11 pages in English). |
Japanese Office Action issued on Mar. 7, 2016 in counterpart Japanese Patent Application No. 2011-551018. (2 pages in Japanese). |
Japanese Office Action issued on Jun. 27, 2016 in counterpart of Japanese Patent Application No. 2015-040751 (3 pages in Japanese). |
Japanese Office Action issued on Aug. 18, 2016 in counterpart of Japanese Patent Application No. 2011-551018 (3 pages in Japanese). |
United States Office Action issued on Oct. 20, 2016, in corresponding United States U.S. Appl. No. 15/258,518 (14 pages, in English). |
United States Final Office Action issued on May 1, 2017 in corresponding United States U.S. Appl. No. 15/258,460 (9 pages in English). |
United States Final Office Action issued on May 2, 2017 in corresponding United States U.S. Appl. No. 15/258,627 (9 pages in English). |
Japanese Office Action issued on May 17, 2017 in counterpart Japanese Patent Application No. 2015-040751 (3 pages in Japanese). |
Japanese Office Action issued on Jun. 4, 2018 in counterpart Japanese Patent Application No. 2017-113142 (6 pages, in Japanese). |
Chinese Office Action issued on Jun. 5, 2018 in counterpart Chinese Patent Application No. 201610346399.8 (10 pages, in Chinese). |
Korean Office Action issued on Sep. 13, 2018 in corresponding Korean Patent Application No. 10-2018-0051937 (7 pages in Korean). |
Japanese Office Action issued on Dec. 23, 2019 in counterpart Japanese Patent Application No. 2017-113145 (5 pages in Japanese). |
Summons to Attend Oral Proceedings issued on Apr. 23, 2020 in counterpart European Patent Application No. 10 743 990.3 (8 pages in English). |
Decision of Refusal issued on Aug. 10, 2020 in counterpart European Patent Application No. 10 743 990.3-1230 (31 pages in English). |
Korean Notice Of Allowance issued on Feb. 23, 2021 in counterpart Korean Patent Application No. 10-2019-0174503 (3 pages in Korean, 1 page in English). |
Japanese Office Action issued on May 10, 2021 in counterpart Japanese Patent Application No. 2020-076069. (6 pages in English, 5 pages in Japanese). |
Japanese Office Action issued on Oct. 11, 2022 in counterpart Japanese Patent Application No. 2021-183505. (6 pages in English, 5 pages in Japanese). |
Number | Date | Country | |
---|---|---|---|
20230247229 A1 | Aug 2023 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17361528 | Jun 2021 | US |
Child | 18298711 | US | |
Parent | 16572704 | Sep 2019 | US |
Child | 17361528 | US | |
Parent | 15855019 | Dec 2017 | US |
Child | 16572704 | US | |
Parent | 15258655 | Sep 2016 | US |
Child | 15855019 | US | |
Parent | 13202906 | US | |
Child | 15258655 | US |