1. Field of the Invention
The present invention relates to the field of projection displays and more specifically to the automated measurement of such displays.
2. Description of the Related Art
The convergence and focus of projection displays having more than one spatial light modulator (SLM) are typically determined subjectively by an operator. As a result, repeatability and tight tolerances in converging and focusing many projectors are difficult to accomplish. The results often depend on the skill and motivation of the person making the adjustments.
It is clear from the figure that this system needs to be converged, at least in the area of the observed pixel. This is best illustrated by the picture of
Focus is another parameter where the adjustment by an operator is often made subjectively. This parameter is more complicated to properly adjust, with many variables involved. For example, brightness can affect the focus significantly. In a projection system, focus is usually accomplished by means of the projection lens, which can be either a zoom or fixed focal length lens.
What is needed is an objective method for convergence and focus criteria along with a measuring tool for implementing the method. This method needs to reflect the human element since the human eye is the final arbitrator in a display application. The invention disclosed herein addresses this need by means of both a method and a tool.
The method and system disclosed in this invention provide an objective tool for measuring the convergence and focus criteria of a projected image. In addition, lens aberrations caused by lateral color shift are programmatically corrected.
To converge the red, green, and blue images from a projector, snapshots are taken at several locations across the field-of-view. Data from each of these snapshots is separated into primary color images (typically red, green, and blue). The centroid of an identical row and column in each of the three (red, green, and blue) images is measured and the differences in the x and y position between the red (reference) centroid data and the green and blue data indicates the amount of adjustment of the green and blue images that is required to converge the image.
Focus for each Primary color is accomplished by processing the three horizontal data arrays previously chosen by the user. After normalizing the data, a single-sided, scaled power spectrum of the array data is derived. Focus criteria are determined by summing the elements of the power spectrum array to the right of the first relative minima in the spectrum. This power spectrum sum is then maximized for optimal focus.
The included drawings are as follows:
a illustrates the three planes (red, green, and blue) for an out-of-convergence image. (prior art)
b shows a row and column of non-converged pixels. (prior art)
a and 4b are diagrams indicating where test images are taken in the image's field-of-view.
c is a diagram of an un-converged image showing the x and y deltas (Δ).
d is a diagram of a converged image.
a indicates the desired waveform's centroid.
b indicates a false waveform centroid.
c illustrates the method for avoiding false waveform centroid measurements.
a,
9
b and 9c illustrate the method of averaging the waveforms for multiple cuts across a pixel.
a is the Fast Fourier Transform (FFT) for a horizontal pulse with sharp edges.
b illustrates the power sum determined in the tail of the FFT.
a shows an out-of-focus image.
b shows an image focused using the method of this invention.
a and 12b illustrate well-focused and poorly-focused waveforms, respectively.
c and 12d show the waveforms of
a is a block diagram of the automated convergence and focus system of this invention.
b shows typical viewing window locations for the automated convergence and focus system of this invention.
a is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
b is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
c is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
d is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
e is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
The method and system of this invention provide an objective tool for measuring the convergence and focus criteria of a projected image. In addition, lens aberrations caused by lateral color shift are programmatically corrected.
The method for objectively converging the primary color images, typically red, green, and blue, involves capturing a magnified snapshot from several locations across the field-of-view of the picture and separating this data into a separate image for each of the modulators.
While two or three captured images are enough to perform the convergence and focus operations, additional images improve the process and provide better results. Typically five captured images are used. Each captured image typically is 640×480 pixel, 24-bit color image. The captured images are separated into three 8-bit images, one for each modulator and typically are stored in DIB format. The modulators typically each provide a primary color image, such as red, green, and blue images, simplifying the separation process. Although this disclosure is in terms of the use of five 640×480 24-bit images, each dissolved into three 8-bit images, it should be understood this is for purposes of illustration and not for purposes of limitation. Other image resolutions, bit-depths, and numbers of images and modulators are also applicable to the processes taught herein.
After capturing the images, a line and column of interest are chosen from the file and the resulting three horizontal (line) data arrays (Red, Green, and Blue) and three vertical (column) data arrays are used to determine the horizontal and vertical center-points of the three (Red, Green, and Blue) pixels. Using the Red pixel (optional selection) as a reference, the convergence adjustment is calculated by measuring the differences in the x and y dimensions between the Green and Blue pixel's center-points and the Red reference pixel's center-point. The green and blue center-points can then be moved to overlay the red center-point, thereby converge the image.
In the method, a row and column grid pattern is turned ON in the projected image, as shown in
For each of the five snapshots, a 24-bit DIB data file is separated into three 8-bit 640×480 data arrays, one representing each of the three primary colors, red, green, and blue.
In locating the center of a row or column of pixels, there can be several complications involved. First, there is the dip at the center top of the waveform discussed above. Then there is the fact that waveforms representing the three colors each may have a somewhat different shape, as illustrated in
a is a sketch of an ideal pulse 80, which represents the pixel width, with the pulse width being measured at the 90% level 81 and the center of the pixel, indicated by line 82, falling directly through the dip in the waveform. However, as shown in
To this point the discussion has centered around a single scan taken through the center of a pixel. In order to improve the accuracy, multiple sweeps (up to 20 passes) are taken across the pixel in both the horizontal and vertical direction and an average of these pulses is used to make the calculations, as described in
a=1,
a+m=10, and
a+n+1=20.
Table 1 is an overview of the algorithm of this invention, used in converging the three SLM (red, green, and blue).
The method used for the automated focusing of a projected image, under varying illumination conditions, is very difficult. However, it is possible to adjust the focus of the optics to an optimal number during the assembly phase of a projector. The method disclosed in this invention does this and can be used to assure that the focus parameter for shipped projector products are optimally focused and meet specification. The user of the projector can then manually focus the projector to match the brightness and other environmental conditions for a particular application.
In the automated focus method disclosed herein, focus for each color (red, green, and blue) is accomplished by processing the three horizontal data arrays previously used in converging the pixels. After the data is normalized, a single-sided, scaled power spectrum of the data array is derived. Focus criteria are then determined by summing the elements of the power spectrum array to the right of the first relative minima in the spectrum. As the optics are adjusted, the value of the summed power spectrum is observed until a power sum maximum value is found.
aillustrates a typical Fast Fourier Transform (FFT) 10 taken for the horizontal pulses with relative sharp edges, as shown earlier in
As illustrated by
Table 2 is an overview of the algorithm of this invention, used in focusing the image.
a is a system block diagram for carrying out the convergence and focus methods of this invention. Five Cameras 130-134 are used to store data from magnified views at the selected locations across the field-of-view; for example, locations at the upper left (UL) 1300, upper right (UR) 1310, lower left (LL) 1320, lower right (LR) 1330, and center (C) 1340 of the field, as indicated in
a shows the format for storing the data for each selected pixel in the computer's 137 memory. First, the 24-bit (B, G, R) image is stored as a BMP file. This file consists of a header 160 followed by the blue 1601, green 1602, and red 1603 data for horizontal pixel 0 through 639 (161, 162) of line 0 (163). This process is repeated over and over for lines 1 through 479 (164). The 24-bit data is then separated into the three R, G, B 8-bit data files, as shown in
In operation, the data from this system is used to converge and focus the red, green blue images. Aligning the three SLM's to provide proper convergence could be done using fly-in-place robots, or other automated techniques, or even by manual adjustment. The optical focus is adjusted to provide a maximum power spectrum summation value.
a through 17e provide a more detailed listing of the pseudo-code for the convergence and focus algorithm of this invention.
The same techniques described herein for a 3-SLM application apply as well to a 2-SLM system.
While this invention has been described in the context of preferred embodiments, it will be apparent to those skilled in the art that the present invention may be modified in numerous ways and may assume embodiments other than that specifically set out and described above. Accordingly, it is intended by the appended claims to cover all modifications of the invention that fall within the true spirit and scope of the invention.
This application is a Divisional of application Ser. No. 10/054,063, filed Nov. 13, 2001 now U.S. Pat. No. 6,995,810 and Provisional Application No. 60/250,450, filed Nov. 30, 2000.
Number | Name | Date | Kind |
---|---|---|---|
5258830 | Schmidt et al. | Nov 1993 | A |
5345262 | Yee et al. | Sep 1994 | A |
5532765 | Inoue et al. | Jul 1996 | A |
5699440 | Carmeli | Dec 1997 | A |
5835135 | Hamaguri et al. | Nov 1998 | A |
6424412 | Matthews | Jul 2002 | B1 |
6456339 | Surati et al. | Sep 2002 | B1 |
6483555 | Thielemans et al. | Nov 2002 | B1 |
6503195 | Keller et al. | Jan 2003 | B1 |
6717625 | Thielemans | Apr 2004 | B1 |
20020024708 | Lewis et al. | Feb 2002 | A1 |
Number | Date | Country | |
---|---|---|---|
20050168659 A1 | Aug 2005 | US |
Number | Date | Country | |
---|---|---|---|
60250450 | Nov 2000 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10054063 | Nov 2001 | US |
Child | 11096480 | US |