1. Field of the Invention
This invention relates to image and video displays, more particularly flat panel displays used as still image and/or video monitors, and methods of generating and driving image and video data onto such display devices.
2. Prior Art
Flat panel displays such as plasma displays, liquid crystal displays (LCD), and light-emitting-diode (LED) displays generally use a pixel addressing scheme in which the pixels are addressed individually through column and row select signals. In general, for M by N pixels—or picture elements—arranged as M rows and N columns, there will be M row select lines and N data lines. When a particular row is selected, N data lines are powered up to the required pixel voltage or current to load the image information to the display element. In a general active-matrix type LCD embodiment, this information is a voltage stored in a capacitor unique to the particular pixel (see
Video and still images are generally converted to compressed forms for storage and transmission, such as MPEG4, H.264, JPEG2000 etc. formats and systems. Image compression methods are based on orthogonal function decomposition of the data, data redundancy, and certain sensitivity characteristics of the human eye to spatial features. Common image compression schemes involve the use of Direct Cosine Transform as in JPEG or motion JPEG, or Discrete Walsh Transform. A video decoder is used to convert the compressed image information, which is a series of orthogonal basis function coefficients, to row and column pixel information to produce the image information, which will be for example at 6 Mbits per frame as in VGA resolution displays. However, from an information content point of view, much of this video information is actually redundant as the image had originally been processed to a compressed form, or it has information content in the higher order spatial frequencies to which the human eye is not sensitive. All these techniques pertain to the display system's components in the software or digital processing domain, and the structure of the actual optical display comprised of M×N pixels is not altered by any of the techniques used for the video format, other than the number of pixels and frame rate.
Spatial Light Modulators (SLM) are devices which alter the amplitude or phase, or both of a transmitted or reflected light beam in two-dimensions, thereby encoding an image to an otherwise uniform light illumination. The image pixels can be written to the device through electrical, or optical addressing means. A simple form of a spatial light modulator is the motion picture film, in which images are encoded on a silver coated film through photo-chemical means. An LCD system is also a particular kind of SLM, such that each pixel's information is encoded through electrical means to a specific position, and the backlit light source's spatial profile, which in general is uniform over the whole display area, is altered by the transmissivity of the pixels.
Prior art in the field generally addresses a single component of the problem at hand. For example, image compression and decompression techniques have not been applied directly on the display element, but only in transmission, storage, and image reconditioning and preparation of data for the display (as in U.S. Pat. No. 6,477,279). Systems incorporating spatial light modulation in which pixels are turned on and off to transmit a backlight to have various degrees of modulation can be implemented (eg. Multiple row select as in U.S. Pat. No. 6,111,560), or both backlight and image modulation can be used to enhance the resolution of the image (as in U.S. Published Application Nos. 2007/0035706 and US 2008/0137990). In especially the latter applications and their relevant disclosures, none of the image construction methods incorporate a temporal dimension in synthesizing the image frame, which is the subject of this disclosure. Thereby both systems, representative of conventional methods of displaying images pixel by pixel on a frame by frame basis, do not benefit from the inherent simplification of the interface and data throughput—which is embedded into the image compression process with which the video is transmitted in.
The present invention may have various modifications and alternative forms from the specific embodiments depicted in the drawings. These drawings do not limit the invention to the specific embodiments disclosed. The invention covers all modifications, improvements and alternative implementations which are claimed below.
An aspect of the invention is a display method and system which constructs an image and/or video through successively displaying a multiple of image components in subframes generated using a coarsely pixelated light array operating at a high frame rate, and a spatial light modulator, which produces certain patterns pertaining to orthogonal basis functions at the same frame rate with a resolution finer than the underlying light source. The image construction system takes advantage of using image compression components whereby the components are distributed in time domain by encoding video images using a spatial light modulator. In each frame, the source image to be driven is first grouped together to a certain size consisting of nx×ny pixels. For example, we can divide the image into rectangular groupings of 4×4 or 8×8 pixels, 4×1, 8×1, or any other arbitrary group size with the provision that we can generate orthogonal basis functions in one or two dimensions. The 1×1 case does not have any compression benefit, and corresponds to methods employed in conventional display systems. The grouping size is limited by the frame rate, which is limited by the switching speed of the components described herein and the image compression ratio. Each image grouping, or coarse pixel as will be referred from here on, is decomposed into components proportional to a series of said orthogonal image basis functions (orthogonal decomposition). These image functions are implemented in display hardware using spatial light modulators, which modulate the amplitude and/or phase of the underlying light, so that it has the desired spatial profile of the orthogonal image basis functions. The image basis functions are shown in
Any image can be decomposed into components, which are found by integrating the image data with the basis functions like those shown in
The invention is based on the inverse transform of EQ. 1, i.e. that an image fc (x,y) can be constructed as a summation of Dcuv*wuv.
The summation is effectively perceived by the human eye in time domain through successively displaying patterns corresponding to the basis functions wuv with a light strength proportional to Dcuv. The human eye would integrate the image patterns and perceive a single image corresponding to fc (x,y).
In orthogonal function implementations used in conventional compression techniques, the basis functions wuv(x,y) take on values of +1 or −1, thereby satisfying orthogonality properties. In this invention, the value of the basis functions are mapped to +1 or 0 instead since we use these functions in the display directly. This creates a non-zero integration component (which is equivalent to the average value of the image Dcuv*wuv). This component is kept track of, and subtracted from the Dc00 component, where Dc00 is the sum of the image over the pixel grouping, or equivalently, the average of the image over the pixel grouping, normalized to 1/(nxny)
Dc00 is also proportional to the light intensity of a single ‘pixel’ (which is the equivalent of a coarse pixel in the definition used herein) if we intend to display the image using the coarsely pixelated display source.
In most cases, Dc00 is greater than or equal to the sum of the rest of the image components derived using the +1 and 0 mapping. Hence, subtracting out each of these non-zero integration components from Dc00 will be greater than or equal to zero. —Consider for example the Dc01 component. Denote wuv(x,y) as the original Walsh function having the values of +1 and −1. Using the new basis functions, w*uv(x,y) =(wuv(x,y)+1)/2, substituting wuv(x,y) which can take on values of 0 and 1 instead of −1 and +1, w*uv(x,y) will transform the image construction equation EQ.2 to
To reproduce the image correctly, the component value to be displayed when the basis function is equal to all 1's (w00) has to be corrected with one half the summation over all Dcuv as in the second term of EQ. 3. Note that if a subset of basis functions are used as in compression, the summation should span only the Dcuv image components that are used. The updated Dc00 component is used in the image construction instead of the original value, since now the total sum of the average components will equal the original Dc00 value.
The image components Dcuv can have positive or negative values. In implementing the display component, the value of Dcuv*w*uv(x,y) can only be positive. In the case of ‘negative’ Dcuv, the image component is generated using the absolute value of Dcuv and the inverse of the basis function pattern w*uv(x,y). The inverse of the function is defined as the two's complement of the binary function w*uv(x,y) in which 0's are mapped to 1's and vice versa.
A block diagram showing the whole system is shown in
For each frame the video image is constructed through:
1. Calculating the image component strength Dcuv related to the image fc (x,y) for each coarse pixel, for each uv component, and for each color.
2. Applying a light intensity mask through the use of a spatial light modulator corresponding to w*uv(x,y).
3. Applying a light proportional to Duv for each coarse pixel. For color displays, three color light elements are used per pixel grouping. The light intensities of the red, green and blue sources are adjusted according to the calculated Dcuv for each color. The light intensities may be adjusted by adjustment of at least one of a voltage, a current and/or the perceived intensity adjusted by the on time of the light source, depending on what light source is used. The Dcuv image components can actually take positive or negative values. In the case of a negative image component, the light intensity is the absolute value of the image component, but in the reconstruction of the image, we use the inverse of the masking pattern.
To arrive at a single frame of the intended image, each image component, which can be defined as a subframe, is displayed sequentially. An observer's eye will integrate the flashed image components to visually perceive the intended image, which is the sum of all flashed image components. Each displayed component, or subframe, duration can be made equal, or the duration can be optimized for bit resolution. The latter case enables one to optimize the spatial light modulator's shutter speed, such that a longer image component duration is allocated to image components which require a higher bit precision, versus shorter image component durations which do not necessarily have to settle to a finer precision. In such a case, when Dcuv components are flashed for shorter durations of time with respect to other components, the light intensity will have to be increased by the same time reduction ratio.
For color images, the red, green and blue light sources can be shined proportional to their respective Dcuv values concurrently, or time-sequentially. In the time-sequential case, where red, green and blue images are flashed separately, the SLM shutter speeds have to be three times faster than the concurrent case. In the concurrent case, one can have either all component values having the same sign, or one of the component values having opposite sign than the other two. For any coarse pixel, we may need both wuv and its inverse pattern to be displayed, since each color component may not necessarily have the same sign. Therefore, the SLM will generate all basis functions, and their inverses for each subframe. If there is no component for the inverse basis function, then the coarse pixel value to be displayed will be equal to zero.
In general, the SLM control will span ideally the whole display, or may be subdivided into smaller sections, so it is expected that both w*uv and its inverse patterns will be required. If the SLM is controlled over each coarse pixel, at the expense of a more complex switching and driving scheme, subframes for unused basis functions need not be included.
Image compression can be either a lossless transformation or a lossy transformation. In lossless transformation, we can construct the image with no fidelity loss from the available image components. In a lossy compression based decomposition, one will neglect certain components, such that, when we construct the image with the unneglected components, the image quality may suffer. In most video and still images, lossy compression is employed to reduce size of the data. In lossy compression, one will usually neglect image components which are below a certain threshold, and image components which the human eyes have reduced sensitivity to. These are generally terms with high order spatial frequencies pertaining to diagonal and off-diagonal tennis. Compression will basically try to describe the image with as few terms as possible, for a given image error bound. In most cases, the terms which are dropped first will be off-diagonal components, followed by diagonal terms, from higher order terms down to lower order terms. Taking the example of 4×4 pixel grouping, which will have 16 image components from D00, D01, D02, D03, D10, D11, etc. up to D33, using the basis functions w*00 through w*33, and the inverses of these components (except for w 00), the original image will be exactly reconstructed if we use all 31 components. In video compression, most images will have the oblique spatial components neglected. A display system which uses only horizontal and vertical image components can be satisfactory in some cases. To improve image accuracy, diagonal spatial frequency components such as D11, D22 and/or D33 can also be added. The oblique components such as D12, D13, D23 etc. may be neglected. In a majority of video sources which use for example MPEG compression, such components have actually been largely eliminated altogether for compressing the video itself for storage and transmission, or turn out to be smaller than a particular threshold which we would deem to be negligible. When image components are neglected, the frame time may be re-proportioned by extending the subframe time for at least one other image component. Even without doing so, a data reduction is achieved. If none of the components are non-negligible, we may resort to lossless operation on the coarse pixel by considering all components. Note also that, in certain embodiments, we can implement a method in which the SLM over a particular coarse pixel can operate independently from other regions. In such a case different coarse pixels can have different levels of compression, from highly compressed to lossless compression. This can be determined from the source video at the same time. Such a case can occur for example in a computer monitor, where during operation, regions of the screen may be stagnant, but require a high accuracy such as a window showing a text and still images, or portions having a fast moving image in which we need a high frame rate to describe the motion more accurately, but not necessarily need a lossless image reproduction scheme. By running the SLM at different rates on different coarse pixels, the image accuracy and power can be optimized. We can decide on which coarse pixel to run which accuracy mode by calculating the Duv components, determining how many are non-negligible, and comparing them to the components in the earlier image frames. A fast moving image vs. slow or stagnant image, and an accurate image vs. a lossy compressed image can be differentiated thus.
Taking the example of a VGA resolution display operating at 30 frames per second, and a 4×4 pixel grouping to define the coarse pixels, the display device to satisfy VGA resolution employing this invention can use
1. 160×100 coarse pixel array whose pixel dimensions are four times larger horizontally and vertically than the intended resolution, and having red, green and blue light elements.
2. A SLM composed of a passive matrix LCD which generates vertical, horizontal and an oblique basis function pattern using horizontal stripes of transparent electrodes in the bottom plane and vertical stripes of transparent electrodes in the top plane of the LCD, or vice versa—such an SLM is capable of generating the sixteen orthogonal basis patterns and their inverses. The electrode widths are equal to the intended pixel resolution size. A total of 640 vertical electrodes and 400 horizontal electrodes exist in the SLM (which may be broken into a multitude of pieces along each direction for faster driving).
3. A computation device which calculates the corresponding Duv components for each color from a VGA resolution image at each frame.
4. Driving the SLM pattern with the macro coarse pixel intensity proportional to Duv, for all non negligible image components. For a compressed video source, using the first 7 or 8 dominant image components will in general be sufficient to reproduce compressed video. This will require the generation of 13 or 15 basis function patterns (out of 31) including the inverse patterns.
5. Other elements may be necessary for light quality, such as a light collimator or diffuser to mix red, green and blue light outputs to produce a uniform light source over the coarse pixel area.
The number of active pixels is reduced from 768000 (for three colors) by a factor of 16 down to 48000 (for three colors). There are 16000 coarse pixels in the display. The raw image data rate depends on the level of image compression desired. For a lossless image reconstruction, there are 16 Duv components per coarse pixel per color. If each Duv is described with 8 bit accuracy, we need 184 Mbps data rate. This corresponds to 128 bits per coarse pixel per color per frame. In reality, only the D00 component needs to have 8 bit accuracy, while the higher order components can have less accuracy. Such component based accuracy assignment is commonly known as a quantization matrix in image compression. In a particular embodiment, one would not need more than 80 bits per coarse pixel per color per frame, which optimizes the data rate down to 120 Mbps. If a medium compression level is used in which we cut off oblique spatial frequency components such as D12, D13, D23 etc. but not D11, D22, D33, we are working with 10 components in total. These components would require a total of 60 bits per coarse pixel per color per frame. The total data rate is reduced to 86 Mbps. For a high compression ratio in which we neglect D11, D22, D33, we would use 46 bits per coarse pixel per color per frame. The total data rate is then 66 Mbps. The SLM pattern needs to be updated 31 times each frame for the lossless compression case, 19 times each frame for the medium level compression case, and 13 times each frame for the high level compression case. The coarse display needs to be updated 8 to 15 times each frame, and will be blank (black) for unused SLM patterns. For 30 frames per second, flashing 13 subframes (for 7 components) results in 390 patterns to be generated per second, or roughly 2.5 msec per subframe. Using 19 subframes for 10 components, we would need to generate 570 SLM patterns per second, or 1.7 msec per subframe. For lossless image reproduction, a total of 31 subframes are needed, which equals 930 patterns per second, requiring 1.1 msec per subframe. The settling speed of conventional LCD's can be made sufficiently fast to be used as spatial light modulators which have only on-off (or black to white) transitions at such speeds by using fast enough liquid crystal material in a smaller geometry. A method to optimize subframe duration for different patterns reflecting the accuracy requirements from the quantization matrix can also be implemented.
For a liquid crystal based SLM, the settling time can be modeled using the liquid crystal materials switching time, and the response time of the voltage applied to a metal line of certain capacitance and resistance. If we have an exponential relationship arising the time constant due to the metal line, when we apply an instantaneous step voltage, the response will be of the form
V(t) =V(0).(1−exp(−t/τ))
where τ is the R.C time constant. Therefore to get an 8-bit accurate voltage applied to the SLM, the minimum time required can be found by taking the natural logarithm of ½8, or 5.5τ. When a 6 bit accurate voltage is sufficient, the time required reduces to 4.15τ, and reduces further to 2.7τ for 4 bit accurate voltages. Therefore, in a particular quantization matrix which employs 6-8 bit accuracy for the low order component terms, and down to 4 bits for high order components, we can allocate down to half the time for the highest order terms which require less accuracy compared to the most significant terms. As illustrated in
The SLM consists of vertical and horizontal electrodes which can span throughout the display. In this case, only 8 drivers, driven by a clock generator is sufficient to generate all patterns which are applied onto coarse pixels. However, for long electrodes, the capacitance of the electrodes may start posing a time-constant limit in addition to the liquid crystal time constant. To speed up the SLM, the electrodes may be broken into smaller pieces, each driven by its dedicated driver or buffers conveying the driver's information, serving a smaller area of the display.
In summary, a video display system which employs image compression techniques based on orthogonal basis function decomposition is disclosed. The system requires a much smaller number of active pixels than a conventional approach, since the image is constructed using coarse pixels, or coarse blocks, which are in essence highly coarse pixelations of the display. The number of rows and columns of the active pixel display is reduced accordingly, hence the interface is simplified. A spatial light modulator operating off a clock generating system is coupled to the active matrix display, such that we do not need to externally supply further data for this system, except to synchronize the images on the active pixel array. Since images are formed using orthogonal image components, a decompression scheme is in effect in which we can truncate the number of components to be used in reconstructing the image in order to reduce the data requirement of the display. The display can be made to generate a lossy decompressed image by truncating image components, or in effect perform a lossless regeneration of a compressed video input. In a particular mode of operation, the display may also regenerate lossless video by displaying all possible orthogonal components.
In a particular embodiment of the invention, a LED based (solid state light source) display system is coupled to a liquid crystal spatial light modulator (see
This application claims the benefit of U.S. Provisional Patent Application No. 61/079,418 filed Jul. 9, 2008.
Number | Name | Date | Kind |
---|---|---|---|
5452024 | Sampsell | Sep 1995 | A |
5508716 | Prince et al. | Apr 1996 | A |
5537492 | Nakajima et al. | Jul 1996 | A |
5675670 | Koide | Oct 1997 | A |
5696524 | Hoshino et al. | Dec 1997 | A |
6111560 | May | Aug 2000 | A |
6229583 | Yasunishi | May 2001 | B1 |
6477279 | Go | Nov 2002 | B2 |
6535195 | Nelson | Mar 2003 | B1 |
6850219 | Aoyama et al. | Feb 2005 | B2 |
7623560 | El-Ghoroury et al. | Nov 2009 | B2 |
7767479 | El-Ghoroury et al. | Aug 2010 | B2 |
7829902 | El-Ghoroury et al. | Nov 2010 | B2 |
20020075217 | Hoshino | Jun 2002 | A1 |
20050128172 | Hoshino | Jun 2005 | A1 |
20060098879 | Kim et al. | May 2006 | A1 |
20070035706 | Margulis | Feb 2007 | A1 |
20070075923 | Beuker et al. | Apr 2007 | A1 |
20080018624 | Cernasov | Jan 2008 | A1 |
20080137990 | Ward | Jun 2008 | A1 |
20090086170 | El-Ghoroury et al. | Apr 2009 | A1 |
20090278998 | El-Ghoroury et al. | Nov 2009 | A1 |
20100003777 | El-Ghoroury et al. | Jan 2010 | A1 |
20100066921 | El-Ghoroury et al. | Mar 2010 | A1 |
20100220042 | El-Ghoroury et al. | Sep 2010 | A1 |
Number | Date | Country |
---|---|---|
1322442 | Nov 2001 | CN |
1348301 | May 2002 | CN |
1666241 | Sep 2005 | CN |
0577258 | Jan 1994 | EP |
0720141 | Jul 1996 | EP |
51-056118 | May 1976 | JP |
2001-350454 | Dec 2001 | JP |
2005-532588 | Oct 2005 | JP |
WO-2004006219 | Jan 2004 | WO |
Entry |
---|
“International Preliminary Examination Report of the International Preliminary Examination Authority Dated Apr. 14, 2011, International Application No. PCT/US2009/050175”. |
“International Search Report and Written Opinion of the International Searching Authority Dated Sep. 9, 2009”, International Application No. PCT/US2009/050175. |
Shirai, T., et al., “RGB-LED Backlights for LCD-TVs with 0D, 1D, and 2D Adaptive Dimming”, 2006 SID International Symposium, Society for Information Display, SD 06 Digest, 44.4, (2006), pp. 1520-1523. |
Poynton, Charles, “Digital Video and HDTV Algorithms and Interfaces”, Morgan Kaufmann Publishers, an Imprint of Elsevier Science, (2003), pp. 447-454. |
“International Search Report and Written Opinion of the International Searching Authority Dated Jun. 10, 2010”, International Application No. PCT/US2010/026325. |
“Office Action Dated Aug. 20, 2012, European Patent Application No. 10710122.2”. |
“Office Action Dated May 31, 2012, Korean Patent Application No. 10-2011-7002953”. |
Lueder, Ernst , “Liquid Crystal Displays, Addressing Schemes and Electro-Optical Effects”, John Wiley & Sons Ltd., (2001), pp. 176-194. |
“Notice of Allowance Dated Nov. 28, 2012, Korean Patent Application No. 10-2011-7002953”. |
“Office Action Dated Apr. 16, 2013; Japanese Patent Application No. 2011-517637”. |
“Office Action Dated Apr. 22, 2013; Chinese Patent Application No. 200980134961.9”. |
Office Action Dated May 28, 2013; U.S. Appl. No. 12/717,365. |
“Office Action Dated Apr. 1, 2013, Korean Patent Application No. 10-2011-7023107”. |
“Office Action Dated Apr. 18, 2013; European Patent Application No. 10710122.2”. |
“Office Action Dated Mar. 4, 2013; European Patent Application No. 09790247.2”. |
“Office Action Dated Mar. 5, 2013, Japanese Patent Application No. 2011-553131”. |
“Office Action Dated Jul. 1, 2014; Chinese Patent Application No. 201080019853.X”. |
“Office Action Dated May 9, 2014; European Patent Application No. 10710122.2”. |
“Notice of Allowance Dated Nov. 26, 2013, Japanese Patent Application No. 2011-553131”. |
“Notice of Allowance Dated Oct. 1, 2013; Japanese Patent Application No. 2011-517637”. |
Notice of Allowance Dated Oct. 28, 2013; U.S. Appl. No. 12/717,365. |
“Office Action Dated Dec. 2, 2013; Korean Patent Application No. 10-2011-7023107”. |
“Office Action Dated Nov. 25, 2013; European Patent Application No. 09790247.2”. |
“Office Action Dated Sep. 4, 2013; Chinese Patent Application No. 201080019853.X”. |
“Office Action Dated Apr. 5, 2014; Korean Patent Application No. 10-2011-7023107”. |
“Office Action Dated Jan. 8, 2014; Chinese Patent Application No. 200980134961.9”. |
“Notice of Allowance Dated Mar. 25, 2014; European Patent Application No. 09790247.2”. |
“Notice of Allowance Dated Oct. 10, 2014; Chinese Patent Application No. 201080019853.X”. |
“Office Action Dated Aug. 12, 2014; Chinese Patent Application No. 200980134961.9”. |
Number | Date | Country | |
---|---|---|---|
20100007804 A1 | Jan 2010 | US |
Number | Date | Country | |
---|---|---|---|
61079418 | Jul 2008 | US |