This patent application claims the benefit and priority of the co-pending Chinese Patent Application No. 201110357809.6, filed on Nov. 14, 2011, by Sien CHEN et al., which is hereby incorporated by reference in its entirety.
As we all know, the principle that we can have a three-dimensional (3D) experience is that the 3D film displays both left-eye image and corresponding right-eye image. In the prior art, taking a 1080p high definition display that progressive scans at the scanning rate of 120 hertz (Hz) with a bit depth of 24 bits for example, the data transmitting rate of the 3D video is up to 1920×1080×120×24 bps (bits per second)=5.6 Gigabits per second (Gbps), i.e. there are 5.6 Gb of data to be transmitted. Thus, when transmitting 3D video data from source, such as a laptop, a handheld and so on, to sink, such as a TV, a large monitor and so on with wireless technology, the 3D data should be compressed before transmission due to the bandwidth limitation of wireless transmission. However, during the compression, the data loss of the 3D video is quite a lot leading to distorted images, which greatly affect the quality of the images.
Accordingly, there is a need in the art for a method and system of prioritized compression for three-dimensional (3D) video wireless display so that there is almost no distortion in the 3D video images.
In order to solve the problem mentioned above, various embodiments of the present invention provide a method and system of prioritized compression for 3D video wireless display.
In one embodiment of the invention, a method of prioritized compression for 3D video wireless display is provided. The method comprising: inputting video data; abstracting scene depth of the video data; foreground and background estimation for each image of the video data, performing different kinds of compressions to the foreground and background in each image, and outputting the processed video data.
Preferably, in an embodiment, the way of the inputting video data or the outputting the processed video data is DMA (direct memory access) and FIFO (first in first out).
Preferably, in an embodiment, the abstracting scene depth of the video data comprises: differentiating the type of the video; and copying the scene depth from a Z-Buffer, if the type of the 3D video is gaming video.
Preferably, in an embodiment, the abstracting scene depth of the video data comprises: differentiating the type of the video; and estimating the motion between the left image and the right image, if the type of the 3D video is stereo stream, namely a movie.
Preferably, in an embodiment, the way for foreground and background estimation for each image is to search the depth threshold in the quantity-depth histogram of each frame so as to estimate the foreground and background.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises: performing high profile stereo coding to the foreground; and performing base profile mono coding to the background.
Preferably, in an embodiment, the performing different kinds of compressions of the foreground and background in each image comprises: the compression ratio of the foreground is less than that of the background.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises: the compression ratio of the foreground is zero (0) and the compression ratio of the background is 85%.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises: the compression rate of the foreground is less than the compression rate of the background.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises: the compression rate of the foreground is 80 Mbps (megabits per second) and the compression rate of the background is 20 Mbps.
In another aspect in accordance with an embodiment of the invention, a system of prioritized compression for 3D video wireless display is provided. The system comprises: a data acquiring unit for acquiring accounting data; an exception management unit for performing exception check on the acquired accounting data according to predetermined exception check rules while or after the data is acquired; and a database unit for storing data and detected exceptions.
Preferably, in an embodiment, the way of inputting video data or outputting the processed video data is DMA (direct memory access) and FIFO (first in first out).
Preferably, in an embodiment, the abstracting scene depth of the video data comprises differentiating the type of the video; and copying the scene depth from a Z-Buffer, if the type of the 3D video is gaming video.
Preferably, in an embodiment, the abstracting scene depth of the video data comprises differentiating the type of the video; and estimating the motion between the left image and the right image, if the type of the 3D video is stereo stream, namely a movie.
Preferably, in an embodiment, the way for foreground and background estimation for each image is to search the depth threshold in the quantity-depth histogram of each frame so as to estimate the foreground and background.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises performing high profile stereo coding to the foreground and performing base profile mono coding to the background.
Preferably, in an embodiment, the performing different kinds of compressions of the foreground and background in each image comprises: the compression ratio of the foreground is less than that of the background.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises: the compression ratio of the foreground is zero (0) and the compression ratio of the background is 85%.
Preferably, in an embodiment, the performing different kinds of compressions to the foreground and background in each image comprises: the compression rate of the foreground is less than the compression rate of the background.
Further, an embodiment in accordance with the present invention provides a computer program product, resident on a computer readable medium, operable for executing instructions for prioritized compression for 3D video wireless display. The computer program product comprising instruction code to input video data, abstract scene depth of the video data, estimate foreground and background for each image of the video data, perform different kinds of compressions to the foreground and background in each image, and output the processed video data.
Thus, the method and system of prioritized compression for 3D video wireless display provided by various embodiments in accordance with the present invention uses a prioritized compression strategy so that the image quality is not affected by the data loss and provides the best 3D experience to users with limited bandwidth during the wireless transmission.
Additional features and advantages of various embodiments in accordance with the invention will be set forth in the description that follows, and in part will be apparent from the description, or may be learned by practice of various embodiments in accordance with the invention. The advantages of various embodiments in accordance with the invention will be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are intended to provide further explanation of various embodiments in accordance with the invention as claimed.
While particular embodiments in accordance with the invention have been specifically described within this Summary, it is noted that the invention and the claimed subject matter are not limited in any way by these embodiments.
The accompanying drawings are included to provide a further understanding of various embodiments in accordance with the invention, and are incorporated in and constitute a part of this specification. The drawings illustrate various embodiments of the invention and, together with the description, serve to explain the principles of various embodiments in accordance with the invention. In the drawings,
Reference will now be made in detail to various embodiments in accordance with the invention, examples of which are illustrated in the accompanying drawings. While the invention will be described in conjunction with various embodiments, it will be understood that these various embodiments are not intended to limit the invention. On the contrary, the invention is intended to cover alternatives, modifications and equivalents, which may be included within the scope of the invention as construed according to the Claims. Furthermore, in the following detailed description of various embodiments in accordance with the invention, numerous specific details are set forth in order to provide a thorough understanding of the invention. However, it will be evident to one of ordinary skill in the art that the invention may be practiced without these specific details or with equivalents thereof. In other instances, well known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the invention.
Various embodiments in accordance with the invention relate generally to a method and system for three-dimensional (3D) video compression, and, in particular, to a method and system of prioritized compression for 3D video wireless display.
Example embodiments are described herein in the context of method and system of prioritized compression for 3D video wireless display. Those of ordinary skill in the art will realize that the following description is illustrative only and is not intended to be in any way limiting. Other embodiments will readily suggest themselves to those skilled in the art having the benefit of this disclosure. Reference will now be made in detail to implementations of the example embodiments as illustrated in the accompanying drawings. The same reference indicators will be used to the extent possible throughout the drawings and the following description to refer to the same or like items.
Reference will now be made in detail to the preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings.
At step 201, the data inputting unit 101 inputs 3D data. For example, the way of inputting video data is DMA (direct memory access) and FIFO (first in first out).
At step 202, the data processing unit 102 differentiates the data type of the 3D video. If it is a rasterized image, e.g., a gaming video, then at step 203, copy the scene depth data from a Z-buffer.
On the other hand, if the data type of the 3D video is a stereo stream, e.g., a 3D movie, then at step 204, motion estimation is performed by comparing the displacement vectors of left image and the right image. The bigger the displacement vector is, the shallower the scene depth is. If the displacement vector is zero (0), the scene depth is indefinite. The scene depth data are relative values and the scene depth data can be abstracted accordingly. For example,
After the scene depth data is abstracted, at step 205, according to the scene depth data, data processing unit 102 estimates the foreground and the background. The foreground is the part that the audience will focus on and it floats from the background. In accordance with one embodiment, the depth threshold of each frame can be searched out in the quantity-depth histogram. Generally speaking, the amount of the foreground pixels is about 20-40% of the total amount. For example,
At step 206, data processing unit 102 performs different compressions with different ratios to the foreground and background in each image. For example, in an embodiment, high profile stereo coding is performed to the foreground and the base profile mono coding is performed to the background. The compression ratio can be 0%, namely, no compression. Compared with the non-prioritized compression, the data loss of the foreground is less and reserves most or all the image details. The compression ratio of the background is 85%, e.g., it is 15% of the original size after compression. The foreground can be compressed with high data compression rate and the background can be compressed with low data rate. For example, in an embodiment, the compression rate of the high profile stereo coding for the foreground can be 80 Mbps (megabits per second) for maximum details, and the base profile mono coding for the background can be 20 Mbps. The background of the two images can be the same or almost the same, so the background data can be provided by only one of the image blocks and the other can restore from it, namely, the background can be shared by the left image 301 and the right image 302 for maximum bandwidth saving.
Finally, the data outputting unit 103 outputs the processed data at step 207. For example, in an embodiment, the way of inputting video data or outputting the processed video data is DMA (direct memory access) and FIFO (first in first out).
Moreover, an embodiment in accordance with the present invention also provides a computer program product, resident on a computer readable medium, operable for executing instructions for prioritized compression for 3D video wireless display. The computer program product comprises instruction code to input video data, abstract scene depth of the video data, estimate foreground and background for each image of the video data, perform different kinds of compressions to the foreground and background in each image, and output the processed video data.
Therefore, the method and system of prioritized compression for 3D video wireless display provided by various embodiments in accordance with the present invention, uses a prioritized compression strategy, so that the image quality is not affected by the data loss and provides the best 3D experience to users with limited bandwidth during the wireless transmission.
It should be appreciated that various modifications, adaptations and alternative embodiments thereof may be made within the scope and spirit of various embodiments in accordance with the invention. The invention is further defined by the following claims.
The foregoing descriptions of various specific embodiments in accordance with the invention have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed, and many modifications and variations are possible in light of the above teaching. The invention is to be construed according to the Claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
2011 1 0357809 | Nov 2011 | CN | national |
Number | Name | Date | Kind |
---|---|---|---|
4864629 | Deering | Sep 1989 | A |
5430464 | Lumelsky | Jul 1995 | A |
5483254 | Powell | Jan 1996 | A |
5500939 | Kurihara | Mar 1996 | A |
5577981 | Jarvik | Nov 1996 | A |
5841447 | Drews | Nov 1998 | A |
5886701 | Chauvin et al. | Mar 1999 | A |
5913727 | Ahdoot | Jun 1999 | A |
5949428 | Toelle et al. | Sep 1999 | A |
6016474 | Kim et al. | Jan 2000 | A |
6057855 | Barkans | May 2000 | A |
6064354 | DeLuca | May 2000 | A |
6141740 | Mahalingaiah et al. | Oct 2000 | A |
6243054 | DeLuca | Jun 2001 | B1 |
6278418 | Doi | Aug 2001 | B1 |
6281903 | Martin et al. | Aug 2001 | B1 |
6308565 | French et al. | Oct 2001 | B1 |
6314493 | Luick | Nov 2001 | B1 |
6438664 | McGrath et al. | Aug 2002 | B1 |
6476807 | Duluk, Jr. et al. | Nov 2002 | B1 |
6492991 | Morein et al. | Dec 2002 | B1 |
6496193 | Surti et al. | Dec 2002 | B1 |
6525725 | Deering | Feb 2003 | B1 |
6545683 | Williams | Apr 2003 | B1 |
6559813 | DeLuca et al. | May 2003 | B1 |
6611253 | Cohen | Aug 2003 | B1 |
6690381 | Hussain et al. | Feb 2004 | B1 |
6750870 | Olarig | Jun 2004 | B2 |
6825847 | Molnar et al. | Nov 2004 | B1 |
6839062 | Aronson et al. | Jan 2005 | B2 |
6891543 | Wyatt | May 2005 | B2 |
6951515 | Ohshima et al. | Oct 2005 | B2 |
7015909 | Morgan, III et al. | Mar 2006 | B1 |
7072477 | Kincaid | Jul 2006 | B1 |
7170515 | Zhu | Jan 2007 | B1 |
7203356 | Gokturk et al. | Apr 2007 | B2 |
7218291 | Abdalla et al. | May 2007 | B2 |
7308115 | Zhang et al. | Dec 2007 | B2 |
7324594 | Lamboray et al. | Jan 2008 | B2 |
7370016 | Hunter et al. | May 2008 | B1 |
7463270 | Vale et al. | Dec 2008 | B2 |
7486290 | Kilgariff et al. | Feb 2009 | B1 |
7616202 | Chen et al. | Nov 2009 | B1 |
7692659 | Molnar et al. | Apr 2010 | B1 |
7719563 | Richards | May 2010 | B2 |
7839803 | Snelgrove et al. | Nov 2010 | B1 |
7856147 | Srinidhi | Dec 2010 | B2 |
7925067 | Bacus et al. | Apr 2011 | B2 |
8019449 | Barzegar et al. | Sep 2011 | B2 |
8024768 | Berger et al. | Sep 2011 | B2 |
8279168 | Glomski et al. | Oct 2012 | B2 |
8363969 | Wang et al. | Jan 2013 | B1 |
8375301 | Nuyttens et al. | Feb 2013 | B2 |
8411966 | Zhang et al. | Apr 2013 | B2 |
8610707 | Chen | Dec 2013 | B2 |
8774267 | Gaddy | Jul 2014 | B2 |
9002125 | Schneider et al. | Apr 2015 | B2 |
9883120 | Adsumilli | Jan 2018 | B2 |
20010038642 | Alvarez, II et al. | Nov 2001 | A1 |
20010043751 | Takahashi et al. | Nov 2001 | A1 |
20020024675 | Foxlin | Feb 2002 | A1 |
20020109701 | Deering | Aug 2002 | A1 |
20030001857 | Doyle | Jan 2003 | A1 |
20030032484 | Ohshima | Feb 2003 | A1 |
20030122820 | Doyle | Jul 2003 | A1 |
20030160798 | Buehler | Aug 2003 | A1 |
20030184468 | Chen et al. | Oct 2003 | A1 |
20040021664 | Takemoto et al. | Feb 2004 | A1 |
20040041822 | Iizuka et al. | Mar 2004 | A1 |
20040102247 | Smoot et al. | May 2004 | A1 |
20040130552 | Duluk, Jr. et al. | Jul 2004 | A1 |
20040205281 | Lin et al. | Oct 2004 | A1 |
20050009583 | Cheung et al. | Jan 2005 | A1 |
20050062738 | Handley et al. | Mar 2005 | A1 |
20050093873 | Paltashev et al. | May 2005 | A1 |
20050140682 | Sumanaweera et al. | Jun 2005 | A1 |
20060170703 | Liao | Aug 2006 | A1 |
20060284792 | Foxlin | Dec 2006 | A1 |
20070018973 | Shih et al. | Jan 2007 | A1 |
20070188444 | Vale et al. | Aug 2007 | A1 |
20070239409 | Alan | Oct 2007 | A1 |
20070257906 | Shimura et al. | Nov 2007 | A1 |
20080049964 | Porwal et al. | Feb 2008 | A1 |
20080143895 | Peterka et al. | Jun 2008 | A1 |
20080252596 | Bell et al. | Oct 2008 | A1 |
20080273755 | Hildreth | Nov 2008 | A1 |
20080293464 | Cheng et al. | Nov 2008 | A1 |
20080293488 | Cheng et al. | Nov 2008 | A1 |
20080312010 | Marty et al. | Dec 2008 | A1 |
20090077504 | Bell et al. | Mar 2009 | A1 |
20090099824 | Falash et al. | Apr 2009 | A1 |
20100073363 | Densham et al. | Mar 2010 | A1 |
20100074489 | Bacus et al. | Mar 2010 | A1 |
20100149372 | Silverstein | Jun 2010 | A1 |
20100177931 | Whytock et al. | Jul 2010 | A1 |
20100194863 | Lopes et al. | Aug 2010 | A1 |
20100296747 | Srinidhi | Nov 2010 | A1 |
20110122224 | Lou | May 2011 | A1 |
20110159885 | Song et al. | Jun 2011 | A1 |
20110165841 | Baek et al. | Jul 2011 | A1 |
20110181622 | Bacus et al. | Jul 2011 | A1 |
20110205389 | Zhang et al. | Aug 2011 | A1 |
20110249076 | Zhou et al. | Oct 2011 | A1 |
20120008672 | Gaddy | Jan 2012 | A1 |
20120014456 | Martinez Bauza et al. | Jan 2012 | A1 |
20120139906 | Zhang et al. | Jun 2012 | A1 |
20120162379 | Dahi et al. | Jun 2012 | A1 |
20120183215 | Van Hook et al. | Jul 2012 | A1 |
20120257065 | Velarde et al. | Oct 2012 | A1 |
20120262592 | Rabii | Oct 2012 | A1 |
20120320232 | Trumbo | Dec 2012 | A1 |
20130009943 | Li et al. | Jan 2013 | A1 |
20130021352 | Wyatt et al. | Jan 2013 | A1 |
20130027521 | DeLuca | Jan 2013 | A1 |
20130027606 | Voss et al. | Jan 2013 | A1 |
20130202191 | Wang | Aug 2013 | A1 |
20140035939 | Schneider | Feb 2014 | A1 |
20140105513 | Schneider et al. | Apr 2014 | A1 |
20140269935 | Gaddy | Sep 2014 | A1 |
20150213640 | Neill et al. | Jul 2015 | A1 |
Number | Date | Country |
---|---|---|
101420618 | Apr 2009 | CN |
101945287 | Jan 2011 | CN |
102006473 | Apr 2011 | CN |
102055982 | May 2011 | CN |
Entry |
---|
Parhami, Computer Arithmetic, Oxford University Press, Jun. 2000, pp. 413-418. city by other. |
gDebugger, graphicRemedy, http://www.gremedy.com, Aug. 8, 2006. |
Duca, et al., A Relational Debugging Engine for Graphics Pipeline, International Conference on Computer Graphics and Ineractive Techniques, ACM SIGGRAPH 2005, pp. 453-463, ISSN: 0730-0301. |
Patent Application As Filed for U.S. Appl. No. 14/604,563; Inventors: Patrick Neill et al.; filed Jan. 23, 2015; “Hybrid Virtual 3D Rendering Approach to Stereovision,”. |
Number | Date | Country | |
---|---|---|---|
20130121563 A1 | May 2013 | US |