Projection method for reducing interpixel gaps on a viewing surface

Information

  • Patent Grant
  • 8702248
  • Patent Number
    8,702,248
  • Date Filed
    Thursday, June 11, 2009
    15 years ago
  • Date Issued
    Tuesday, April 22, 2014
    10 years ago
Abstract
A display system able to reduce interpixel intensity gaps caused by the use of pulsed laser sources having relatively short optimum pulse duration. The interpixel intensity gaps are reduced by temporally offsetting multiple laser pulses for a display pixel during a corresponding pixel-scan period. The temporally offset pulses for the display pixel are then scanned to different locations on a viewing surface such that the display pixel has an improved intensity distribution. Additional reduction in the interpixel gaps may be accomplished by de-focusing the temporally offset pulses in a scan direction, increasing the duty cycle of the source lasers, and shifting the location of alternating frames on the viewing surface.
Description
STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT

Not Applicable.


BACKGROUND

1. The Field of the Invention


The present disclosure relates generally to display systems, and more particularly, but not necessarily entirely, to display systems that utilize pulsed laser illumination sources.


2. Description of Background Art


Advanced display devices are becoming more prevalent in modern society. Such display devices are used to display information in a wide variety of settings providing, inter alia, education and entertainment. There have been several recent promised enhancements to display technologies including: increased resolution, increased contrast and increased brightness levels as well as other characteristics that improve the overall quality of images produced with dynamic video display systems.


Technologies used to produce advanced video displays include: Texas Instruments' DLP® projector using a digital micromirror device (“DMD”), Sony's SXRD® system and JVC's D-ILA® apparatus both which incorporate liquid crystal on silicon (“LCOS”) technology, Kodak's grating electromechanical system (“GEMS”) as well as systems using grating light valve (“GLV”) technology. All of these particular technologies differ in the devices which are used to modulate the light which is projected, and such light modulation devices are at the core of each system and the component to which the rest of the system components surrounding them are designed.


Previously available display technologies have typically employed either a two-dimensional scan architecture or a column-scan architecture, sometimes referred to as a one-dimensional scan architecture, to form an image on a viewing surface. In a display device employing a two-dimensional scan architecture, the underlying light modulation device includes a two-dimensional array of pixel elements able to generate an entire frame of an image at one time. The two-dimensional array of pixel elements may include micro-electro-mechanical (“MEMS”) structures. Alternatively, the two-dimensional array of pixel elements may include liquid crystals, such as those incorporating LCOS technology. In a display device employing a column-scan architecture, the underlying light modulation device may include a one-dimensional array of MEMS pixel elements able to draw a single column of the image at a time. To generate an entire image on the viewing surface, the single columns of the image are scanned, one-by-one, across the viewing surface, by a scanning device, such as a rotating scanning mirror or oscillating scanning mirror. As used herein, a scanning device may refer to any device having a moving reflective surface operable to scan modulated beams of light onto a viewing surface.


In the past, the previously available display technologies incorporated a variety of different light sources. For example, some of the display technologies utilize an incandescent lamp for generating white light which is passed through a color wheel as the light travels to the surface of the light modulation device. The use of a incandescent lamp in a display system has drawbacks, including, the limited life of the lamp and the need for color filters or a mechanized color wheel to produce different colored light.


Other light sources for light modulation devices have, in the past, included continuous wave lasers. The benefits which accompany the use of the continuous wave lasers include the ability to eliminate the need for separating white light into primary colors and their high power output. However, continuous wave lasers are in some instances disadvantageous due to their high power consumption, complex technical design, and excessive heat output.


Recently, improvements in the operation of semiconductor pulsed lasers have made them more attractive for use as light sources in display devices that utilize light modulators. These improvements eliminate some of the problems associated with the use of lamps and continuous wave lasers. However, even with the benefits provided over the previously available light sources, the use of semiconductor lasers in a display device is still faced with significant challenges of its own.


For example, one drawback to the use of semiconductor lasers is that in order to achieve maximum light intensity, the semiconductor lasers suitable for use in a display device must operate at a relatively low duty cycle. Another drawback to the use of semiconductor lasers is that their power output is relatively low when compared with some of the continuous wave lasers that have been previously available.


Another previous drawback to the use of pulsed light sources, such as a semiconductor laser operating at less than 100% duty cycle, in display devices having a column-scan architecture has been that most such display devices have previously required a light source that produces continuous light, such as a continuous wave laser, in order to generate an acceptable image. That is, past attempts to use pulsed light sources in a display device having a column-scan architecture have been unsuccessful as the pulsed light sources caused noticeable irregularities in the displayed image.


In particular, as the columns of pixels are each scanned across a viewing surface from a display device using pulsed light sources, undesirable vertical stripes and interpixel gaps are visible in the image due to the pulsed nature of the light incident on the light modulator. A primary cause of these vertical stripes and interpixel gaps is believed to be the relatively short pulse time of the pulsed laser sources when compared to the time necessary to scan a column of pixels on a viewing surface. In some instances, the duty cycle of the pulsed lasers, the ratio of the duration of a laser pulse to the time necessary to scan a column or pixel, is less than 50%. This means that, in some cases, the pulsed lasers are only active for an interval which is less than one-half of the time it takes to scan one full column or pixel. The end result of the low duty cycle of the pulsed lasers is that an intensity drop occurs at the edges of the pixels in a column. This problem becomes even more apparent when two columns of pixels in adjacent columns are viewed side-by-side.


It would therefore be an improvement over the previously available technologies and devices to significantly reduce, or eliminate altogether, the irregularities caused by the use of pulsed light sources in a display device. It would be a further improvement over the previously available technologies and devices to significantly reduce, or eliminate altogether, the irregularities caused by the use of pulsed light sources in a display device having a column-scan architecture.


The features and advantages of the disclosure will be set forth in the description which follows, and in part will be apparent from the description, or may be learned by the practice of the disclosure without undue experimentation. The features and advantages of the disclosure may be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims.





BRIEF DESCRIPTION OF THE DRAWINGS

The features and advantages of the disclosure will become apparent from a consideration of the subsequent detailed description presented in connection with the accompanying drawings in which:



FIG. 1 is a diagram of a display system pursuant to an exemplary embodiment of the present disclosure;



FIG. 2 is an overlay of three graphs showing the effects of using a single pulse of light with a low duty cycle to form a display pixel on a viewing surface;



FIG. 3 is a depiction of the interpixel gaps between three adjacent pixels on a viewing surface;



FIG. 4 is an overlay of three graphs showing the effects of using multiple pulses of light to reduce interpixel gaps between display pixels;



FIG. 5 is a depiction of the reduction in interpixel gaps between three adjacent pixels on a viewing surface;



FIGS. 6 and 7 are diagrams showing the effects of de-focusing a beam of light on the surface of a light modulator;



FIG. 8 is an overlay of three graphs showing the reduction in interpixel gaps caused by de-focusing the pulses of light used to form display pixels;



FIG. 9 is an overlay of three graphs showing two alternative frames of an image shifted by one-half of a pixel on a viewing surface;



FIG. 9A is a depiction of two alternate frames of an image shifted by one-half of a pixel on a viewing surface;



FIG. 10 is an overlay of three graphs showing the effects of increasing the duty cycle of pulses of light used to form display pixels on a viewing surface;



FIG. 11 is a diagram of a display system pursuant to an exemplary embodiment of the present disclosure;



FIG. 12 is an overlay of five graphs showing the effects of pulsing each primary color multiple times during each pixel-scan period;



FIG. 13 is a depiction of the spatial separation of different pulses of light on the surface of a light modulator;



FIG. 14 is a diagram of a display system pursuant to an exemplary embodiment of the present disclosure;



FIG. 15 is a block diagram of a projection controller pursuant to an exemplary embodiment of the present disclosure;



FIG. 16 depicts a high-level flow diagram of a process of displaying an image on viewing surface according to an embodiment of the present disclosure;



FIG. 17 depicts a high-level flow diagram of a process of displaying an image on viewing surface according to an embodiment of the present disclosure;



FIG. 18 depicts a high-level flow diagram of a process of displaying an image on viewing surface according to an embodiment of the present disclosure;



FIG. 19 depicts a high-level flow diagram of a process of displaying an image on viewing surface according to an embodiment of the present disclosure; and



FIG. 20 depicts a high-level flow diagram of a process of displaying an image on viewing surface according to an embodiment of the present disclosure.





DETAILED DESCRIPTION

For the purposes of promoting an understanding of the principles in accordance with the disclosure, reference will now be made to the embodiments illustrated in the drawings and specific language will be used to describe the same. It will nevertheless be understood that no limitation of the scope of the disclosure is thereby intended. Any alterations and further modifications of the inventive features illustrated herein, and any additional applications of the principles of the disclosure as illustrated herein, which would normally occur to one skilled in the relevant art and having possession of this disclosure, are to be considered within the scope of the disclosure claimed.


Referring now to FIG. 1, there is depicted an exemplary display system 100 for generating an image on a viewing surface in a single primary color, such as red, green, or blue. It will be appreciated that to generate the image in full-color on the viewing surface, that other systems of similar design for may be combined with, or incorporated into, the system 100. The display system 100 may comprise a plurality of pulsed light sources 102, such as semiconductor lasers. While only two pulsed light sources 102 are shown in FIG. 1, it will be appreciated that any number of pulsed light sources 102 may be utilized in system 100 to generate the desired on screen intensity. For example, in an embodiment of the present disclosure, eight pulsed light sources may be utilized.


The pulsed light sources 102 may each comprise semiconductor lasers that are able to emit pulsed beams of light of a predetermined wavelength. The light emissions of the pulsed light sources 102 may be controlled by a projector controller 104. That is, the projector controller 104 may control the duration and timing of the pulses of light emitted from the pulsed light sources 102. In addition to controlling the pulsed light sources 102, the projector controller 104 may control a light modulator 106 for modulating pulsed beams of light from the pulsed light sources 102. The projection controller 104 may also control a scanning device 116. The light modulator 106 may include a one-dimensional array of pixel elements.


As used herein, the term “pixel element” may refer to one or more MEMS structures that receives light from a light source and modulates the light to form a corresponding display pixel on a viewing surface. Typically, each pixel element is able to vary the intensity of a corresponding display pixel between a dark state and a bright state.


Further, in a display system using a column-scan architecture, each pixel element may be associated with an entire row of display pixels on the viewing surface. That is, each pixel element is functional to form more than one display pixel on the viewing surface. It will be noted that, as used herein, the term “display pixel” may refer to one of a plurality of small spots of light on a viewing surface that are perceived by a human to form an image. A typical image may be comprised of millions of display pixels arranged in columns and rows on a viewing surface.


One type of light modulator 106 suitable for use with the present disclosure includes a plurality of reflective and deformable ribbons suspended over a substrate. These ribbons are arranged in a column of parallel rows and may be deflected, i.e, pulled down, by applying a bias voltage between the ribbons and the substrate. A first group of the ribbons may comprise alternate rows of the ribbons. The ribbons of the first group may be collectively driven by a single digital-to-analog controller (“DAC”) such that a common bias voltage may be applied to each of them at the same time. For this reason, the ribbons of the first group are sometimes referred to as “bias ribbons.” A second group of ribbons may comprise those alternate rows of ribbons that are not part of the first group. Each of the ribbons of the second group may be individually addressable or controllable by its own dedicated DAC device such that a variable bias voltage may be independently applied to each of them. For this reason, the ribbons of the second group are sometimes referred to as “active ribbons.”


The bias and active ribbons may be sub-divided into separately controllable picture elements referred to herein as “pixel elements.” Each pixel element contains, at a minimum, a bias ribbon and an active ribbon. When the reflective surfaces of the bias and active ribbons of a pixel element are co-planar, incident light directed onto the pixel element is reflected. By blocking the reflected light from a pixel element, a dark spot is produced on the viewing surface at a corresponding display pixel. When the reflective surfaces of the bias and active ribbons of a pixel element are not co-planar, incident light may be both diffracted and reflected off of the pixel element. By separating the desired diffracted light from the undesired diffracted light and reflected light, the desired diffracted light produces a bright spot on the corresponding display pixel.


The intensity of the light produced on the viewing surface by a pixel element may be controlled by varying the separation or deflection between the reflective surfaces of its active and bias ribbons. Typically, this is accomplished by varying the voltage applied to the active ribbon while holding the bias ribbon at a common bias voltage. It has been previously determined that the maximum light intensity output for a pixel element may occur in a diffraction based system when the distance between the reflective surfaces its active and bias ribbons is λ/4, where λ is the wavelength of the light incident on the pixel element. The minimum light intensity output for a pixel element may occur when the reflective surfaces of its active and bias ribbons are co-planar. Intermediate light intensities may be output from the pixel element by varying the separation between the reflective surfaces of the active and bias ribbons between co-planar and λ/4.


Referring back to FIG. 1, prior to striking the pixel elements on the light modulator 106, beams of pulsed light from the pulsed light sources 102 may first pass through an input optical assembly 108, which may have different cylindrical elements that are represented by a Y-collimating lens assembly 110 and an X-focus lens 112. The input optical assembly 108 conditions the light for modulation by the light modulator 106 and may include a light tunnel (not shown) as disclosed in U.S. patent application Ser. No. 12/288,577 filed on Oct. 20, 2008, which is hereby incorporated by reference in its entirety. In particular, the output of the input optical assembly 108 is a column of light corresponding in shape to the one-dimensional array of pixel elements on the light modulator 106 that may have a uniform or “top-hat” distribution in the Y-axis and a Gaussian distribution in the X-axis. In addition, the input optical assembly 108 may include an optical assembly for reducing etendue as disclosed in U.S. patent application Ser. No. 12/206,582 filed on Sep. 8, 2008, which is hereby incorporated by reference in its entirety.


The Y-collimating lens assembly 110 may collimate the light in a direction corresponding to a height of the one-dimensional array of pixel elements on the light modulator 106. The Y-collimating lens assembly 110 may include an ashperical Powell lens (not shown) and a collimating lens (not shown). The X-focus lens 112 may focus the light from the input optical assembly 108 in a direction corresponding to a width of the one-dimensional array of pixel elements on the light modulator 106. Thus, it will be appreciated that the Y-collimating lens assembly 110 and the X-focus lens 112 may be anamorphic.


The output optical assembly 114 may be operable to separate light from the light modulator 106 into a desired diffracted portion and an undesired diffracted portion and a reflected portion. The reflected portion and the undesired diffracted portion of the light from the light modulator 106 may be discarded, while the desired diffracted portion may be directed onto the scanning device 116, where it is scanned onto a viewing surface through projection optics 118. In particular, the scanning device 116 is operable to scan columns of display pixels onto a viewing surface column-by-column to thereby form a two-dimensional image. Each column of display pixels is individually modulated by the one-dimensional array of pixel elements on the light modulator 106.


The scanning device 116 may scan an entire image during a sweep of the viewing surface. It will be appreciated that in an embodiment of the present disclosure, the scanning device 116 may comprise an oscillating or rotating reflective surface, such as a mirror. In an embodiment of the present disclosure, the scanning device 116 may be any device able to scan a one-dimensional column of modulated light across a surface to thereby form a two-dimensional image. The time necessary to scan a two-dimensional image on a viewing surface will be referred to herein as a “frame scan period.” The time necessary to scan a single column of display pixels will be referred to herein as a “column-scan period.” The time necessary to scan a single pixel will be referred to herein as a “pixel-scan period.” It will be appreciated that a column-scan period and a pixel-scan period may be the same in the display system 100 since a column is exactly one pixel wide in the scanning direction.


The pulsed light sources 102 may have an optimum pulse duration that is less than a pixel-scan period. (As used herein, the term “optimum pulse duration” may refer to a pulse duration of the pulsed light sources 102 that produces a maximum power output.) As previously explained, this means that the pulsed light sources 102 may be on (or active) for a duration that is less than the time it takes to scan a column or pixel if it is desired to operate at maximum power output for the light sources 102.


Referring now to FIG. 2, there is shown an overlay of three graphs for the system 100 where the pulsed light sources 102 are pulsed simultaneously at their optimum pulse duration for the formation of Display Pixels A, B and C. It will be appreciated that the Display Pixels A, B and C are located in the same row and in adjacent columns on the viewing surface. Because the Display Pixels A, B and C are in the same row of an image, they may be formed by the same pixel element on the light modulator 106 but at different times. The horizontal axes of the top and middle graphs in FIG. 2 represent time, t, while the horizontal axis on the bottom graph represents the width of a display pixel as determined from v*t, where v is the velocity of the scanning device 116 and t the pixel-scan period for the Display Pixels A, B and C.


The vertical axis for the top graph in FIG. 2 represents the output power of the pulsed light sources 102 (FIG. 1). The vertical axis for the middle graph represents the pixel element state for the pixel element on the light modulator 106 associated with the Display Pixels A, B and C. The vertical axis for the bottom graph represents the intensity of the Display Pixels A, B and C on a viewing surface. The dashed vertical lines between the Display Pixels A, B and C, represent the pixel-scan periods in the top and middle graphs, while they represent the relative widths of the Display Pixels A, B and C on the viewing surface in the bottom graph.


Referring now to the top graph in FIG. 2, each of the pulsed light sources 102 is pulsed simultaneously during each pixel-scan period for Display Pixels A, B and C and at the same duty cycle of about 50%. In an embodiment of the present disclosure, the duty cycle may be defined as the pulse duration divided by the pixel or column-scan period. Alternatively, the duty cycles of each of the pulsed light sources 102 may be less than about 50%, less than about 40%, or between about 30% and about 50%. It will be easily observed that the pulsed light sources 102 are not on or active for the entire duration of the pixel-scan periods for Display Pixels A, B and C. The middle graph in FIG. 2 shows that the pixel element state for the pixel element associated with Pixels A, B and C remains unchanged during each pixel-scan period. For example, the pixel element may be configured to produce the maximum display pixel intensity for Display Pixels A, B and C. It will be appreciated, however, that the state of the pixel element may vary for each pixel-scan period in the event that the intensities of the Display Pixels A, B and C are not the same.


The bottom graph in FIG. 2 shows a distribution of the light intensity of each of the Display Pixels A, B and C on the viewing surface. As can be observed, the intensity distributions of the Display Pixels A, B and C may be a Gaussian distribution on the viewing surface. Thus, it can be noted that the light intensities are at a maximum near the center of the Display Pixels A, B and C and at a minimum near the edges of the Display Pixels A, B and C. From the intensity distributions observed in the bottom graph in FIG. 2, it follows that there may be interpixel gaps in the light intensities between the Display Pixels A, B and C on the viewing surface.


As shown in FIG. 3, a representation of the Display Pixels A, B and C on a viewing surface as they would appear to an observer, the interpixel gaps in the intensity distribution between the Display Pixels A, B and C may cause dark lines or dark stripes to be observable in the boundary regions between each of the Display Pixels A, B and C. In real life, these dark lines or stripes can appear between each column of an image and extend between the top and the bottom of the image displayed using system 100.


Turning now to FIG. 4, there is shown an overlay of three graphs for the system 100 having the same axes as described in relation to FIG. 2. However, in order to reduce the interpixel gaps in intensity between Display Pixels A, B and C, the projector controller 104 is configured to temporally offset the pulses of light from the pulsed light sources 102 during each pixel-scan period as can be seen in the top graph in FIG. 4. Further, as observed in the top graph shown in FIG. 4, these temporally offset pulses may be non-overlapping in each pixel-scan period. In an embodiment of the present disclosure, however, the temporally offset pulses may be overlapping in each pixel-scan period. Further, in an embodiment of the present disclosure, each of the pulsed light sources 102 may only pulse once during each pixel-scan period to thereby maintain a maximum power output of each of the pulsed light sources 102. Alternatively, each of the pulsed light sources 102 may pulse multiple times during each pixel-scan period.


As can be observed from the combined intensity distribution in the bottom graph of FIG. 4, the interpixel gaps in intensity between the Display Pixels A, B and C, are significantly reduced through the use of the temporally offset pulses in the pixel-scan periods. In particular, it will be appreciated that the temporally offset pulses from the pulsed light sources 102 are spatially offset from each other on the viewing surface due to the operation of the scanning device 116, which continues to move between each of the pulses in the pixel-scan periods. Thus, the intensity distributions for each of the temporally offset pulses for a given display pixel are centered at different locations on the viewing surface because the scanning device 116 remains in motion between each pulse. Nevertheless, the combined intensity distribution of the temporally offset pulses is a significant improvement over the use of only a single pulse during each pixel-scan period as shown in FIGS. 2 and 3.


It will be noted from the bottom graph of FIG. 4 that the intensity distributions may be overlapping within a display pixel even though their underlying pulses are temporally non-overlapping during the pixel-scan period. Further, the intensity distributions for a display pixel may combine with the intensity distributions of adjacent display pixels near the edges of the display pixels to further reduce the severity of the interpixel gaps between them.



FIG. 5, a representation of the appearance of the Display Pixels A, B and C on a viewing surface, illustrates the reduction in the interpixel gaps between the Display Pixels A, B and C due to the temporal spacing of the pulses within each pixel-scan period as compared to their prevalence as shown in FIG. 3.


It will be noted that it is within the scope of the present disclosure to pulse any number of temporally offset light pulses of the same color onto a pixel element of a light modulator during a pixel-scan period. In an embodiment of the present disclosure, three temporally offset pulses of the same color are pulsed onto a pixel element of a one-dimensional array of pixel elements during a single pixel-scan period. In an embodiment of the present disclosure, four or more temporally offset pulses of the same color are pulsed onto a pixel element of a one-dimensional array of pixel elements during a single pixel-scan period. In an embodiment of the present disclosure, eight or more temporally offset pulses of the same color are pulsed onto a pixel element of a one-dimensional array of pixel elements during a single pixel-scan period.


Further, each temporally offset pulse of light may be overlapping or non-overlapping in time with the other pulses of light pulsed within the same pixel-scan period. Furthermore, each temporally offset pulse may originate from a unique light source during the same pixel-scan period. That is, it is within the scope of the present invention that either a single light source generates all of the temporally offset pulses onto a pixel element of a one-dimensional array of pixel elements during a pixel-scan period; or multiple light sources may generate the temporally offset pulses onto a pixel element of a one-dimensional array of pixel elements during a pixel-scan period. In the case where there are multiple light sources, each light source may pulse only once during each pixel-scan period. Alternatively, the multiple light sources may pulse multiple times during each pixel-scan period.


Referring now back to FIG. 1, as previously discussed, the X-focus lens 112 is an exemplar anamorphic lens able to focus pulses of light onto the one-dimensional array of pixel elements on the light modulator 106. In particular, the X-focus lens 112 focuses light in a column-width direction on the pixel elements of the light modulator 106. That is, in order for the pulsed light to be focused in the X-direction, the reflective surfaces of the pixel elements lie at or very near the focal point of the X-focus lens 112. As will now be described, further reduction in the interpixel intensity gaps between adjacent display pixels on a viewing surface may be accomplished by de-focusing the light in the X-direction from the light sources 102. As will be shown below, the de-focusing of the X-focus lens 112 widens the pulses of light in the X-direction on the light modulator 106.


Referring now to FIGS. 6 and 7, there is shown a pixel element 150 on the light modulator 106. As previously discussed, the pixel element 150 is part of a one-dimensional array of pixel elements extending along the Y-axis. Each of the individual ribbons 152 that make up the pixel element 150 extends parallel to the X-axis. A beam of pulsed light 154 that is focused in both the X-axis and the Y-axis extends perpendicularly across each of the individual ribbons 152. By adjusting the focus of the X-focus lens 112 to thereby de-focus the beam of pulsed light in the X-direction, the width of the beam of pulsed light incident upon the pixel element 150 is wider, as indicated by the de-focused beam 156. Thus, it will be noted that the de-focused beam 156 is wider in the X-direction than that of the focused beam 154.


Referring now to FIG. 8, an overlay of three graphs, arranged similarly to that shown in FIG. 2, it can be observed that by de-focusing the light pulses in the X-direction, the interpixel gaps between Display Pixels A, B and C may be reduced. In particular, the intensity distributions for the Display Pixels A, B and C, as shown in the bottom graph of FIG. 8, are “widened” to thereby compensate for the short duty cycle of the pulses. The combined intensity distribution of the de-focused pulses reduces the interpixel gaps. However, some decrease in the overall intensity of the Display Pixels A, B and C may be observed due to the de-focusing in the X-direction. In an embodiment of the present disclosure, multiple pulses of light within a pixel-scan period may be de-focused was well.


Referring now to FIG. 9, it will now be explained that a further reduction in the interpixel gaps between display pixels on a viewing surface may be accomplished by laterally shifting the position of the display pixels in a frame of an image from the position of the display pixels in a previous frame of the image.


As previously discussed in relation to FIG. 1, a frame of an image is scanned by the scanning device 116 during a frame scan period, with each frame scan period consisting of a plurality of pixel-scan periods or column-scan periods. In order to ensure that each subsequent frame of an image is positioned exactly on the previous frame of the image, the system 100 initiates the scan of each frame when the scanning device 116 is in the same position. Stated another way, because the scanning device 116 is typically operated at a constant frequency, the projector controller 104 starts scanning each new frame at a predetermined time (or based upon feedback from the scan mirror or based upon a video sync signal) such that the scanning device 116 is at the same initial location to begin scanning each frame. However, by purposefully delaying the start of a frame scan period for alternate frames of an image, a reduction in the interpixel gaps may be accomplished, as will now be explained below.



FIG. 9 illustrates the scanning of two frames of an image, namely Frames 1 and 2. Frame 1 comprises Display Pixels A-1, B-1, and C-1 and Frame 2 comprises Display Pixels A-2, B-2 and C-2. Display Pixels A-1 and A-2 in Frames 1 and 2 are related in that they define the same location in the source data for the images in Frames 1 and 2. That is, Display Pixel A-2 is intended to replace Display Pixel A-1 in the same location on the viewing surface as defined by the source data if Frames 1 and 2 were shown as intended. Display Pixels B-1 and B-2 and Display Pixels C-1 and C-2 are similarly related.


The scanning of Frame 1 is commenced when the scanning device 116 is at a predetermined initial position. Once Frame 1 is completely scanned across the viewing screen, the scanning device 116 returns to the initial position at a predetermined frequency. However, the beginning of the scanning of Frame 2 is delayed by a Delay Period such that the scanning device 116 has moved past the initial position used for Frame 1 when Frame 2 actually begins scanning. This delay in the beginning of the scanning of Frame 2 will laterally shift the entire image of Frame 2 on the viewing surface. In an embodiment of the present disclosure, the Delay Period is such that Frame 2 is shifted by approximately one-half of a pixel on the viewing surface with respect to Frame 1. The Delay Period may be repeated for every other frame of the image. Thus, Frame 3 (not shown) would begin scanning when the scanning device 116 is at the same initial position as for Frame 1, while the scanning of Frame 4 (not shown) would be delayed in a manner similar to Frame 2. The delayed frames and the un-delayed frames may be interlaced, i.e., the scanning of every other frame of an image may be delayed.


The delay in scanning between alternate frames of an image may result in a reduction of the interpixel intensity gaps between adjacent pixels and columns as shown in the bottom graph of FIG. 9. In particular, the display pixels for the delayed frames are approximately centered on the interpixel gaps from the un-delayed frames and vice-versa. It will be noted that the frame shift on the viewing surface may be more or less than one-half of a display pixel.


Referring now to FIG. 9A, there is shown the scanning device 116 in relation to a viewing surface 170. A first frame 172 of an image is depicted on the viewing surface 170. For simplicity of representation, a second frame 174 of an image is also depicted on the viewing surface 170, even though in operation they would not be scanned at the same time onto the viewing surface 170. Thus, it will be appreciated that the first frame 172 and the second frame 174 are scanned sequentially onto the viewing surface 170. As previously discussed, the scanning device 116 scans a frame of an image during a sweep of the viewing surface 170 and then returns to its original position to begin the scanning of the next frame. Ideally, the sweeping of the viewing surface 170 by the scanning device 116 occurs at a constant frequency.


The projector controller 104 initiates the scanning of frame 172 at location 176 on the viewing surface 170 during a sweep by the scanning device 116. After the scanning device 116 has returned to sweep the next frame, i.e., frame 174, the projector controller 104 delays initiating the scanning of frame 174 until the scanning device 116 has moved past the position 176 and begins scanning frame 174 at location 178 on the viewing surface 170. The offset distance 180 between the position of frames 172 and 174 on the viewing surface 170 may be less than one-pixel width or approximately one-half of a pixel. The difference in position of the scanning device 116 between the scanning of frames 172 and 174 may be represented angularly by angle 182. Because the scanning device 116 is operating at a known frequency, the projector controller 104 may use a simple time delay to offset frames 172 and 174 on the viewing surface 170.


Referring now to FIG. 10, an overlay of three graphs similar to FIG. 2, the interpixel intensity gaps between adjacent display pixels may be further reduced by increasing the duty cycle of the pulsed light sources 102 such that the pulse duration is longer than the optimum pulse duration for the pulsed light sources 102. In other words, the interpixel gaps may be reduced by pulsing each of the light sources 102 for a longer duration during each pixel-scan period than the optimum pulse duration. A disadvantage to this approach is that the increase in the pulse duration of the light sources 102 may cause a drop in their light intensity output. In an embodiment of the present disclosure, the light sources 102 are operated at more than 10%, 20%, 30% or 40% of their optimum pulse duration. In another embodiment of the present disclosure, the light sources 102 are pulsed at more than 60%, 70%, 80%, or 90% of their optimum pulse duration.


Referring now to FIG. 11, there is depicted a system 200 for forming an image on a viewing surface. The system 200 may be operable to scan, in full-color, the columns of an image as described in U.S. patent application Ser. No. 12/001,771 (“the '771 application”), filed on Dec. 12, 2007, which application is hereby incorporated by reference in its entirety. In particular, the system 200 may include a light modulator 202 for modulating incident light emitted from red pulsed light sources 204, green pulsed light sources 206, and blue pulsed light sources 208. The system 200 may further include a projector controller 210, an input optical assembly 212 having a Y-collimating lens assembly 214 and an X-focus lens 216, an output optical assembly 218, a scanning device 220 and projection optics 222. The function and operation of the input optical assembly 212, Y-collimating lens assembly 214, X-focus lens 216, output optical assembly 218, scanning device 220 and projection optics 222 are similar to those like-named components described above in reference to FIG. 1, and will not be repeated in detail here.


Still referring to FIG. 11, the projector controller 210 may be operable to pulse each of the red pulsed light sources 204, green pulsed light source 206, and blue pulsed light sources 208, in synchronization with the light modulator 202 and scanning device 220, which may also be controlled by the projection controller 210. The system 200 may further include an optical device 224 for directing the light from the red pulsed light sources 204, the green pulsed light sources 206, and the blue pulsed light sources 208 into the input optical assembly 212.


As mentioned, the system described in the '771 application scans full-color columns onto a viewing surface using only a single light modulator. As described in the '771 application, this operation may be primarily accomplished by pulsing light from a red light source, a green light source and a blue light source onto a light modulator during each column-scan period. As further described in '771 application, between each of the colored light pulses during a pixel-scan period or a column-scan period, the light modulator is reconfigured to modulate the next pulse. As further described in '771 application, to account for the temporal separation between each of the light pulses during a column-scan period and the constant movement of the scanning device, each of the different colored light pulses is spatially offset on the pixel elements of the light modulator.


The system described in '771 application, however, does not take into account the use of pulsed laser sources that have short optimum pulse durations. As will now be described, the system 200 provides improvements over the system disclosed in '771 application as system 200 is able to scan each column in full color using light sources that have short optimum pulse durations.


Referring now to both FIGS. 11 and 12, there is depicted an overlay of five graphs in relation to two display pixels, namely Display Pixels A and B. As can be observed in the top graph in FIG. 12, the red pulsed light sources 204 generate two temporally offset pulses onto the pixel element during each pixel-scan period. Likewise, the green pulsed light sources 206 and the blue pulsed light sources 208 each generate two temporally offset pulses onto the pixel element during each pixel-scan period. Further, the red, green and blue pulses generated during each pixel-scan period are also temporally offset from each other. In an embodiment of the present disclosure, it will be appreciated that the same colored light pulses may be overlapping in time during each pixel-scan period. It will be further appreciated that the individual light sources that form each of the pulsed light sources 204, 206 and 208 may be pulsed collectively or individually during each pixel-scan period.


As can be observed in the second graph in FIG. 12, the pulses of light from the pulsed light sources 204, 206 and 208 are pulsed onto the same pixel element of the light modulator 202. In the second graph, the pixel element is depicted as remaining in the same state during each pixel-scan period as may be the case to generate two adjacent white pixels. It will be appreciated, however, that the pixel element may change configuration between each pulse of light in the same pixel-scan period to thereby generate different colored display pixels on the viewing surface.


As may be observed in the third graph of FIG. 12, the intensity distributions for the two red pulses are spatially offset on the viewing surface for each of the Display Pixels A and B. As may be further observed from the third graph, the combined intensity of the two spatially offset red pulses for each of Display Pixels A and B reduce the interpixel intensity gap between the Display Pixels A and B. As can be observed from the fourth and fifth graphs in FIG. 12 (labeled “Green Screen Intensity” and “Blue Screen Intensity”), respectively, the same observation holds true for the green pulses and the blue pulses. As a result of pulsing each colors' light sources more than once during each pixel-scan period, the colors are more evenly distributed in each Display Pixel A and B on the viewing surface.


Referring now to FIG. 13, in order to ensure that each of the separate pulses of light incident on the pixel element during a pixel-scan period is scanned to the correct location on the viewing surface, each of the red, green and blue pulses of light may be spatially offset in the X-direction on the ribbons 230 of the pixel element 232 on the light modulator 202. Although the pulses are shown as non-overlapping in location on the ribbons 230, it will be appreciated that the pulses may be over-lapping in location on the ribbons 230. Further, although all of the pulses are shown together on the ribbons 230 in FIG. 13, it will be appreciated that the pulses are temporally offset from each other such that only one pulse is incident on the ribbons 230 of the pixel element 232 at any one time.


Further, the sequential order in which the pulses are pulsed onto the ribbons 230 may be varied from red, green, blue, red, green, blue as shown in FIG. 12. For example, the light sources 204, 206 and 208 may generate a series of temporally offset pulses during a pixel-scan period that take the sequence of: red, red, green, green, blue, blue. It will therefore be appreciated that any other sequential pulsing of the different colors of light onto a pixel element during a pixel-scan period falls within the scope of the present invention. Likewise, while only two pulses for each color of light are shown in FIG. 12 as being incident on the pixel element 232 during a single pixel-scan period, it will be appreciated that any number of pulses of the same color may be used. Further, each pulse of light incident on the pixel element 232 during a pixel-scan period may be generated by a group of light sources or a single light source.


The system 200 depicted in FIG. 11 may employ the other techniques described herein for reducing the interpixel gaps between adjacent pixels and columns on a viewing surface. For instance, in addition to temporally offsetting pulses of the same color during a column-scan period or a pixel-scan period, the pulses of light generated by the light sources 204, 206 and 208 may be de-focused in the X-direction by adjusting the position of the X-focus lens 216. Also, the duty cycle of the light sources 204, 206 and 208 may be increased with respect to the pixel-scan period such that a pulse duration is longer than an optimum pulse duration for the light sources 204, 206 and 208. Further, the frame scan period between alternating frames of an image may be delayed with respect to the position of the scanning device 220 to thereby offset the frames on the viewing surface.


Referring now to FIG. 14, there is depicted a display system 200A, where like reference numerals depict components similar to those described in connection with system 200 shown in FIG. 11 (and FIG. 1). The system 200A is modified from system 200 in that in place of the red pulsed light sources 204 in FIG. 11, the system 200A includes a red continuous light source 204A and a light filter 240. The red continuous light source 204A may be a continuous laser light source as is known to one having ordinary skill in the art. The use of a red continuous light source 204A in conjunction with the green pulsed light sources 206 and the blue light sources 208 may be desirable in the event that a pulsed red light source 204 as shown in FIG. 11, for example, is not found suitable for use with the system 200.


The light filter 240 may be a device able to cause the red continuous light source 204A to mimic or simulate a pulsed light source in the system 200A. It will be understood that the light filter 240 may include any device able to selectively prevent light from the red continuous light source 204A from striking the light modulator 202.


In an embodiment of the present disclosure, the light filter 240 may include an acousto-optic modulator (“AOM”) operable to diffract light emitted from the red continuous light source 204A. In particular, as light from the red continuous light source 204A passes through the AOM, and the AOM is activated, the light will be diffracted and scattered. The diffracted and scattered light does not reach the light modulator 202 and the red continuous light source 204A is effectively off. When the AOM is not activated, the light from the red continuous light source 204A passes freely through the AOM without obstruction to thereby strike the light modulator 202. In an embodiment of the present disclosure, the light filter 240 may include a spinning structure with holes or spokes to selectively allow light to pass. In an embodiment of the present disclosure, the light filter 240 may include an electro-optical switch, such as a Pockels cell. In an embodiment of the present disclosure, the light filter 240 may include a Galvanometer scanner. In an embodiment of the present disclosure, the light filter 240 may be omitted and the seed laser from the red continuous laser 204A could be pulsed. The system 200A may be operable to reduce interpixel gaps between adjacent pixels using any of the methods described herein.


It will be appreciated that in an embodiment of the present disclosure, the continuous light source used in combination with pulsed light sources may emit light other than red as depicted in FIG. 14. In an embodiment of the present disclosure the continuous light source used in combination with pulsed light sources may emit green light, blue light or light of some other color. In an embodiment of the present disclosure, the pulsed light sources depicted in FIGS. 11 and 14 may emit light other than the colors indicated in the figures and related discussion. Thus, as used herein, the term “light source” may mean a pulsed light source or a continuous light source emitting light of any color.


It will be further appreciated that the concept of forming a display pixel on a viewing surface may mean forming the display pixel on a viewing surface in a single color or in full color. Further, the concept of forming a display pixel on a viewing surface may mean forming the display pixel in a single underlying color of a full color display pixel. For example, the concept of forming a display pixel on a viewing surface may mean forming one of a red, blue, or green component of a full color pixel.


Referring now to FIG. 15, there is depicted a block diagram of an exemplary projection controller 250 pursuant to an embodiment of the present invention. The projection controllers 104 (FIG. 1) and 210 (FIGS. 11 and 14) may take substantially the form of the projection controller 250. The projection controller 250 may comprise a synchronization module 252, a light modulator control module 254, a scanning device control module 256, a light source control module 258 and an optional filter control module 268. The light modulator control module 254 may control the operation of the individual pixel elements on a light modulator 260 in accordance with image data from a video source 260. The scanning device control module 256 may control the operation of a scanning device 262. Further, the scanning device control module 256 may provide position feedback information regarding the position of the scanning device 262 such that the position of the scanning device 262 may determined and controlled. The light source control module 258 may control one or more light sources 264, including pulsed laser light sources and continuous laser light sources. In operation, the video source 260 provides data to the synchronization module 252, which may include a video clock signal. The video source 260 may also provide data, such as image data, to the light modulator control module 258. The image data may be utilized to configure pixel elements on the light modulation device 260 such that display pixels are shown at the proper intensity and color. The filter control module 268, when present, may control a continuous wave laser light filter 266 to thereby provide laser pulses from a continuous wave laser.


The synchronization module 252 assists in generating an image on a viewing surface from the image data provided by the video source 260 by coordinating and synchronizing the operations of the light modulator control module 254, the scanning device control module 256, the light source control module 258, and the filter control module 268 to thereby reduce interpixel gaps as described herein. For example, the synchronization module 252 may ensure that the light sources 264 are correctly pulsed during column or pixel-scan periods and that the light modulator 260 is correctly configured. Thus, the projection controller 250 may be operable to reduce interpixel gaps between columns (or rows) of a display pixels as described in detail herein by causing the light sources 264 to provide a plurality of temporally offset pulses during a column or pixel-scan period or by shifting the start position of a new frame on a viewing surface with respect to a previous frame such that the columns of the new frame and the previous frame are offset from each other. In addition, the projection controller 250 may pulse light sources longer than their optimum pulse duration during a pixel or column-scan period.


It will be appreciated that the structure and apparatus disclosed herein is merely one example of a means for reducing interpixel gaps on a viewing surface and it should be appreciated that any structure, apparatus or system which performs functions the same as, or equivalent to, those disclosed herein are intended to fall within the scope of a means for reducing interpixel gaps on a viewing surface, including those structures, apparatus or systems which are presently known, or which may become available in the future. Anything which functions the same as, or equivalently to, a means for reducing interpixel gaps on a viewing surface is intended to fall within the scope of this element.


Many of the functional units described in this specification have been labeled as modules, in order to more particularly emphasize their implementation independence. For example, a module may be implemented as a hardware circuit comprising custom VLSI circuits or gate arrays, off-the-shelf semiconductors such as logic chips, transistors, or other discrete components. A module may also be implemented in programmable hardware devices such as field programmable gate arrays, programmable array logic, programmable logic devices or the like which may now be available or which may become available in the future.


Modules may also be implemented in hardware in combination with software code, sometimes referred to as computer readable instructions, for execution by various types of processors. An identified block of executable code may, for instance, comprise one or more physical or logical blocks of computer instructions that may, for instance, be organized as an object, procedure, or function. Nevertheless, the executables of an identified module need not be physically located together, but may comprise disparate instructions stored in different locations which, when joined logically together, comprise the module and achieve the stated purpose for the module.


Indeed, a block of executable code may be a single instruction, or many instructions, and may even be distributed over several different code segments, among different programs, and across several memory devices that are coupled to a processor. Similarly, operational data may be identified and illustrated herein within modules, and may be embodied in any suitable form and organized within any suitable type of data structure. The operational data may be collected as a single data set, or may be distributed over different locations including over different storage devices, and may exist, at least partially, merely as electronic signals on a system or network. The projection controller 250 may, in addition to comprising various modules, take the form of a module as described above.


Referring now to FIG. 16, there is depicted a high-level flow diagram of a process 300 for displaying an image on viewing surface according to an embodiment of the present disclosure. At block 302, a projection controller may pulse temporally offset beams of light from one or more laser light sources onto a pixel element of a light modulator during a pixel-scan period associated with a display pixel. The pixel-scan period may be dependent upon the time required by a scanning device to sweep once through an angle that covers the display pixel area on a viewing surface. At block 304, the light modulator may modulate the temporally offset beams of light using the pixel element. For example, the pixel element may be configured based upon image data from an image source to thereby form the required light intensity for the display pixel on a viewing surface. At block 306, a desired portion of the temporally offset beams of light, as modulated by the pixel element, is scanned on the viewing surface to thereby form the display pixel.


Referring now to FIG. 17, there is depicted a high-level flow diagram of a process 350 of displaying an image on viewing surface according to an embodiment of the present disclosure. At block 352, a projection controller may pulse a first group of temporally offset beams of light of a first color onto a pixel element of a light modulator during a pixel-scan period for a display pixel. At block 354, the projection controller may pulse a second group of temporally offset pixel elements of light of a second color onto the pixel element of the light modulator during the pixel-scan period for the display pixel. At block 356, the first and second group of temporally offset beams of light may be modulated by the pixel element. At block 358, a scanning device may scan a desired portion of the first and second groups of temporally offset beams of light, as modulated by the pixel element, onto the viewing surface to thereby form the display pixel.


Referring now to FIG. 18, there is depicted a high-level flow diagram of a process 400 of displaying an image on viewing surface according to an embodiment of the present disclosure. At block 402, a projection controller may pulse temporally offset beams of light onto a pixel element of a light modulator during a pixel-scan period for a display pixel. At block 404, the beams of temporally offset beams of light may be unfocused on the pixel element. At block 406, the unfocused beams of light may be modulated by the pixel element. At block 408, a desired portion of the temporally offset and unfocused beams of light, as modulated by the pixel element, may scanned onto a viewing surface to thereby form the display pixel.


Referring now to FIG. 19, there is depicted a high-level flow diagram of a process 450 of displaying an image on viewing surface according to an embodiment of the present disclosure. At block 452, modulated light may be scanned by a scanning device onto a viewing surface to thereby form a first frame of an image. At block 454, modulated light may be scanned by the scanning device onto the viewing surface to thereby form a second frame of the image. At block 456, the first frame and the second frame of the image, which may be frames of a motion picture, are offset from each other on the viewing surface.


Referring now to FIG. 20, there is depicted a high-level flow diagram of a process 200 of displaying an image on viewing surface according to an embodiment of the present disclosure. At block 502, a projection controller may pulse beams of light from a light source onto a pixel element of a light modulator for a duration longer than an optimum pulse duration associated with the light source. At block 504, the beam of light is modulated by the pixel element. At block 506, a desired portion of the beam of light, as modulated, is scanned onto the viewing surface by a scanning device.


It will be appreciated that a column of display pixels may be arranged in any orientation on a viewing surface, including horizontally and vertically.


In the foregoing Detailed Description, various features of the present disclosure are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed disclosure requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the following claims are hereby incorporated into this Detailed Description by this reference, with each claim standing on its own as a separate embodiment of the present disclosure.


It is to be understood that the above-described arrangements are only illustrative of the application of the principles of the present disclosure. Numerous modifications and alternative arrangements may be devised by those skilled in the art without departing from the spirit and scope of the present disclosure and the appended claims are intended to cover such modifications and arrangements. Thus, while the present disclosure has been shown in the drawings and described above with particularity and detail, it will be apparent to those of ordinary skill in the art that numerous modifications, including, but not limited to, variations in size, materials, shape, form, function and manner of operation, assembly and use may be made without departing from the principles and concepts set forth herein.

Claims
  • 1. A method for generating an image on a viewing surface, said method comprising: (a) pulsing temporally offset beams of light onto a pixel element of a light modulator during a pixel-scan period for a display pixel;(b) causing each of the temporally offset beams of light to be out of focus on said pixel element;(c) modulating said temporally offset and out of focus beams of light with the pixel element; and(d) scanning a desired portion of said temporally offset beams of light, as modulated by the pixel element, onto the viewing surface to thereby form the display pixel.
  • 2. The method of claim 1, wherein step (d) comprises scanning the temporally offset beams of light, as modulated by the pixel element, to spatially offset locations on the viewing surface.
  • 3. The method of claim 1, wherein step (d) comprises scanning the temporally offset beams of light, as modulated by the pixel element, with a moving reflective member.
  • 4. The method of claim 1, wherein step (a) comprises pulsing each of the temporally offset beams of light from a unique light source.
  • 5. The method of claim 1, wherein step (a) comprises pulsing each of the temporally offset beams of light for a duration less than about 50% of the pixel-scan period.
  • 6. The method of claim 1, wherein step (a) comprises pulsing each of the temporally offset beams of light for a duration less than about 40% of the pixel-scan period.
  • 7. The method of claim 1, wherein step (a) comprises pulsing each of the temporally offset beams of light during non-overlapping time intervals of the pixel-scan period.
  • 8. The method of claim 1, wherein step (a) comprises pulsing at least three temporally offset beams of light during the pixel-scan period.
  • 9. The method of claim 1, wherein step (a) comprises pulsing temporally offset beams of light of a same color onto the pixel element of the light modulator during the pixel-scan period for the display pixel.
  • 10. The method of claim 1, wherein step (c) comprises modulating said temporally offset beams of light using diffraction.
  • 11. The method of claim 1, wherein die pixel element comprises at least two deflectable and elongated elements.
  • 12. The method of claim 1, wherein said pixel element forms part of a one-dimensional array of pixel elements on the light modulator.
  • 13. The method a claim 1, further comprising collimating each of the temporally offset beams of light in a second direction prior to said pixel element.
  • 14. A method for generating an image on a viewing surface, said method comprising: (a) scanning modulated light from a one-dimensional array of pixel elements on a light modulator to thereby form a first frame of an image on the viewing surface;(b) scanning modulated light from the one-dimensional array of pixel elements on the light modulator to thereby form a second frame of an image on the viewing surface; and(c) offsetting spatially the first frame and the second frame on the viewing surface, wherein the step of offsetting spatially the first frame comprises initiating a scan of the first frame when a scanning device is at a first position of a sweep of the viewing surface and initiating a scan of the second frame when said scanning device is at a second position of a sweep of the viewing outface.
  • 15. The method of claim 14, wherein step (c) comprises offsetting the first frame and the second frame by less than one pixel width.
  • 16. The method of claim 14, wherein step (c) comprises offsetting the first frame and the second frame by approximately one-half of a pixel width.
  • 17. The method of claim 14, wherein said second position is past said first position in the sweep of the viewing surface.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit of U.S. Provisional Application No. 61/060,744, filed Jun. 11, 2008, which is hereby incorporated by reference herein in its entirety, including but not limited to those portions that specifically appear hereinafter, the incorporation by reference being made with the following exception: In the event that any portion of the above-referenced provisional application is inconsistent with this application, this application supercedes said above-referenced provisional application.

US Referenced Citations (857)
Number Name Date Kind
449435 Brotz Mar 1891 A
1525550 Jenkins Feb 1925 A
1548262 Freedman Aug 1925 A
1702195 Centeno Feb 1929 A
1814701 Ives Jul 1931 A
2415226 Sziklai Feb 1947 A
2688048 Rose Aug 1954 A
2764628 Bambara Sep 1956 A
2783406 Vanderhooft Feb 1957 A
2991690 Grey et al. Jul 1961 A
3201797 Roth Aug 1965 A
3345462 Good et al. Oct 1967 A
3370505 Bryan Feb 1968 A
3418459 Purdy et al. Dec 1968 A
3422419 Mathews et al. Jan 1969 A
3485944 Stephens, Jr. Dec 1969 A
3534338 Christensen et al. Oct 1970 A
3553364 Lee Jan 1971 A
3576394 Lee Apr 1971 A
3577031 Welsh et al. May 1971 A
3600798 Lee Aug 1971 A
3602702 Warnock Aug 1971 A
3605083 Kramer Sep 1971 A
3633999 Buckles Jan 1972 A
3656837 Sandbank Apr 1972 A
3659920 McGlasson May 1972 A
3668622 Gannett et al. Jun 1972 A
3688298 Miller et al. Aug 1972 A
3709581 McGlasson Jan 1973 A
3711826 La Russa Jan 1973 A
3734602 Deck May 1973 A
3734605 Yevick May 1973 A
3736526 Simmons May 1973 A
3737573 Kessler Jun 1973 A
3746911 Nathanson et al. Jul 1973 A
3757161 Kline Sep 1973 A
3760222 Smith Sep 1973 A
3764719 Dell Oct 1973 A
3775760 Strathman Nov 1973 A
3781465 Ernstoff et al. Dec 1973 A
3783184 Ernstoff et al. Jan 1974 A
3785715 Mecklenborg Jan 1974 A
3802769 Rotz et al. Apr 1974 A
3816726 Sutherland et al. Jun 1974 A
3818129 Yamamoto Jun 1974 A
3831106 Ward Aug 1974 A
3846826 Mueller Nov 1974 A
3862360 Dill et al. Jan 1975 A
3886310 Guldberg et al. May 1975 A
3889107 Sutherland Jun 1975 A
3891889 Fazio Jun 1975 A
3896338 Nathanson et al. Jul 1975 A
3899662 Kreeger et al. Aug 1975 A
3915548 Opittek et al. Oct 1975 A
3920495 Roberts Nov 1975 A
3922585 Andrews Nov 1975 A
3934173 Korver Jan 1976 A
3935499 Oess Jan 1976 A
3940204 Withrington Feb 1976 A
3943281 Keller et al. Mar 1976 A
3947105 Smith Mar 1976 A
3969611 Fonteneau Jul 1976 A
3983452 Bazin Sep 1976 A
3991416 Byles et al. Nov 1976 A
4001663 Bray Jan 1977 A
4009939 Okano Mar 1977 A
4016658 Porter et al. Apr 1977 A
4017158 Booth Apr 1977 A
4017985 Heartz Apr 1977 A
4021841 Weinger May 1977 A
4027403 Marsh et al. Jun 1977 A
4028725 Lewis Jun 1977 A
4048653 Spooner Sep 1977 A
4067129 Abramson et al. Jan 1978 A
4077138 Foerst Mar 1978 A
4093346 Nishino et al. Jun 1978 A
4093347 La Russa Jun 1978 A
4100571 Dykes et al. Jul 1978 A
4119956 Murray Oct 1978 A
4120028 Membrino et al. Oct 1978 A
4138726 Girault et al. Feb 1979 A
4139257 Matsumoto Feb 1979 A
4139799 Kureha et al. Feb 1979 A
4149184 Giddings et al. Apr 1979 A
4152766 Osofsky et al. May 1979 A
4163570 Greenaway Aug 1979 A
4170400 Bach et al. Oct 1979 A
4177579 Peters et al. Dec 1979 A
4184700 Greenaway Jan 1980 A
4195911 Bougon et al. Apr 1980 A
4197559 Gramling Apr 1980 A
4200866 Strathman Apr 1980 A
4203051 Hallett et al. May 1980 A
4211918 Nyfeler et al. Jul 1980 A
4222106 Hess et al. Sep 1980 A
4223050 Nyfeler et al. Sep 1980 A
4229732 Hartstein et al. Oct 1980 A
4234891 Beck et al. Nov 1980 A
4241519 Gilson et al. Dec 1980 A
4250217 Greenaway Feb 1981 A
4250393 Greenaway Feb 1981 A
4289371 Kramer Sep 1981 A
4297723 Whitby Oct 1981 A
4303394 Berke et al. Dec 1981 A
4305057 Rolston Dec 1981 A
4318173 Freedman et al. Mar 1982 A
4333144 Whiteside et al. Jun 1982 A
4335402 Holmes Jun 1982 A
4335933 Palmer Jun 1982 A
4338661 Tredennick et al. Jul 1982 A
4340878 Spooner et al. Jul 1982 A
4342083 Freedman et al. Jul 1982 A
4343037 Bolton Aug 1982 A
4343532 Palmer Aug 1982 A
4345817 Gwynn Aug 1982 A
4347507 Spooner Aug 1982 A
4348184 Moore Sep 1982 A
4348185 Breglia et al. Sep 1982 A
4348186 Harvey et al. Sep 1982 A
4349815 Spooner Sep 1982 A
4356730 Cade Nov 1982 A
4360884 Okada et al. Nov 1982 A
4375685 Le Goff et al. Mar 1983 A
4384324 Kim et al. May 1983 A
4390253 Lobb Jun 1983 A
4393394 McCoy Jul 1983 A
4394727 Hoffman et al. Jul 1983 A
4398794 Palmer et al. Aug 1983 A
4398795 Palmer Aug 1983 A
4399861 Carlson Aug 1983 A
4408884 Kleinknecht et al. Oct 1983 A
4422019 Meyer Dec 1983 A
4427274 Pund et al. Jan 1984 A
4431260 Palmer Feb 1984 A
4435756 Potash Mar 1984 A
4437113 Lee et al. Mar 1984 A
4439157 Breglia et al. Mar 1984 A
4440839 Mottier Apr 1984 A
4441791 Hornbeck Apr 1984 A
4445197 Lorie et al. Apr 1984 A
4446480 Breglia et al. May 1984 A
4463372 Bennett et al. Jul 1984 A
4466123 Arai et al. Aug 1984 A
4471433 Matsumoto et al. Sep 1984 A
4472732 Bennett et al. Sep 1984 A
4487584 Allen et al. Dec 1984 A
4492435 Banton et al. Jan 1985 A
4498136 Sproul, III Feb 1985 A
4499457 Hintze Feb 1985 A
4500163 Burns et al. Feb 1985 A
4511337 Fortunato et al. Apr 1985 A
4536058 Shaw et al. Aug 1985 A
4539638 Gaffney Sep 1985 A
4546431 Horvath Oct 1985 A
4566935 Hornbeck Jan 1986 A
4570233 Yan et al. Feb 1986 A
4582396 Bos et al. Apr 1986 A
4583185 Heartz Apr 1986 A
4586037 Rosener et al. Apr 1986 A
4586038 Sims et al. Apr 1986 A
4589093 Ippolito et al. May 1986 A
4590555 Bourrez May 1986 A
4591844 Hickin et al. May 1986 A
4596992 Hornbeck Jun 1986 A
4597633 Fussell Jul 1986 A
4598372 McRoberts Jul 1986 A
4599070 Hladky et al. Jul 1986 A
4609939 Kozawa et al. Sep 1986 A
4616217 Nesbitt et al. Oct 1986 A
4616262 Toriumi et al. Oct 1986 A
4623223 Kempf Nov 1986 A
4623880 Bresenham et al. Nov 1986 A
4625289 Rockwood Nov 1986 A
4630101 Inaba et al. Dec 1986 A
4630884 Jubinski Dec 1986 A
4631690 Corthout et al. Dec 1986 A
4633243 Bresenham et al. Dec 1986 A
4634384 Neves et al. Jan 1987 A
4636031 Schmadel, Jr. et al. Jan 1987 A
4636384 Stolle et al. Jan 1987 A
4642756 Sherrod Feb 1987 A
4642790 Minshull et al. Feb 1987 A
4642945 Browning et al. Feb 1987 A
4645459 Graf et al. Feb 1987 A
4646251 Hayes et al. Feb 1987 A
4647966 Phillips et al. Mar 1987 A
4655539 Caulfield et al. Apr 1987 A
4656506 Ritchey Apr 1987 A
4656578 Chilinski et al. Apr 1987 A
4657512 Mecklenborg Apr 1987 A
4658351 Teng Apr 1987 A
4662746 Hornbeck May 1987 A
4663617 Stockwell May 1987 A
4671650 Hirzel et al. Jun 1987 A
4672215 Howard Jun 1987 A
4672275 Ando Jun 1987 A
4677576 Berlin, Jr. et al. Jun 1987 A
4679040 Yan Jul 1987 A
4684215 Shaw et al. Aug 1987 A
4692880 Merz et al. Sep 1987 A
4698602 Armitage Oct 1987 A
4704605 Edelson Nov 1987 A
4710732 Hornbeck Dec 1987 A
4714428 Bunker et al. Dec 1987 A
4715005 Heartz Dec 1987 A
4720705 Gupta et al. Jan 1988 A
4720747 Crowley Jan 1988 A
4725110 Glenn et al. Feb 1988 A
4727365 Bunker et al. Feb 1988 A
4730261 Smith Mar 1988 A
4731859 Holter et al. Mar 1988 A
4735410 Nobuta Apr 1988 A
4743200 Welch et al. May 1988 A
4744615 Fan et al. May 1988 A
4748572 Latham May 1988 A
4751509 Kubota et al. Jun 1988 A
4760388 Tatsumi et al. Jul 1988 A
4760917 Vitek Aug 1988 A
4761253 Antes Aug 1988 A
4763280 Robinson et al. Aug 1988 A
4766555 Bennett Aug 1988 A
4769762 Tsujido Sep 1988 A
4772881 Hannah Sep 1988 A
4777620 Shimoni et al. Oct 1988 A
4780084 Donovan Oct 1988 A
4780711 Doumas Oct 1988 A
4791583 Colburn Dec 1988 A
4794386 Bedrij et al. Dec 1988 A
4795226 Bennion et al. Jan 1989 A
4796020 Budrikis et al. Jan 1989 A
4799106 Moore et al. Jan 1989 A
4805107 Kieckhafer et al. Feb 1989 A
4807158 Blanton et al. Feb 1989 A
4807183 Kung et al. Feb 1989 A
4811245 Bunker et al. Mar 1989 A
4812988 Duthuit et al. Mar 1989 A
4821212 Heartz Apr 1989 A
4825391 Merz Apr 1989 A
4833528 Kobayashi May 1989 A
4837740 Sutherland Jun 1989 A
4854669 Birnbach et al. Aug 1989 A
4855934 Robinson Aug 1989 A
4855937 Heartz Aug 1989 A
4855939 Fitzgerald, Jr. et al. Aug 1989 A
4855943 Lewis Aug 1989 A
4856869 Sakata et al. Aug 1989 A
4868766 Oosterholt Sep 1989 A
4868771 Quick et al. Sep 1989 A
4873515 Dickson et al. Oct 1989 A
4884275 Simms Nov 1989 A
4885703 Deering Dec 1989 A
4893353 Iwaoka et al. Jan 1990 A
4893515 Uchida Jan 1990 A
4897715 Beamon, III Jan 1990 A
4899293 Dawson et al. Feb 1990 A
4907237 Dahmani et al. Mar 1990 A
4912526 Iwaoka et al. Mar 1990 A
4915463 Barbee, Jr. Apr 1990 A
4918626 Watkins et al. Apr 1990 A
4930888 Freisleben et al. Jun 1990 A
4935879 Ueda Jun 1990 A
4938584 Suematsu et al. Jul 1990 A
4940972 Mouchot et al. Jul 1990 A
4949280 Littlefield Aug 1990 A
4952152 Briggs et al. Aug 1990 A
4952922 Griffin et al. Aug 1990 A
4953107 Hedley et al. Aug 1990 A
4954819 Watkins Sep 1990 A
4955034 Scerbak Sep 1990 A
4959541 Boyd Sep 1990 A
4959803 Kiyohara et al. Sep 1990 A
4969714 Fournier, Jr. et al. Nov 1990 A
4970500 Hintze Nov 1990 A
4974155 Dulong et al. Nov 1990 A
4974176 Buchner et al. Nov 1990 A
4982178 Hintze Jan 1991 A
4984824 Antes et al. Jan 1991 A
4985831 Dulong et al. Jan 1991 A
4985854 Wittenburg Jan 1991 A
4991955 Vetter Feb 1991 A
4992780 Penna et al. Feb 1991 A
4994794 Price et al. Feb 1991 A
5005005 Brossia et al. Apr 1991 A
5007705 Morey et al. Apr 1991 A
5011276 Iwamoto Apr 1991 A
5016643 Applegate et al. May 1991 A
5022732 Engan et al. Jun 1991 A
5022750 Flasck Jun 1991 A
5023725 McCutchen Jun 1991 A
5023818 Wittensoldner et al. Jun 1991 A
5025394 Parke Jun 1991 A
5025400 Cook et al. Jun 1991 A
5035473 Kuwayama et al. Jul 1991 A
5038352 Lenth et al. Aug 1991 A
5043924 Hofmann Aug 1991 A
5047626 Bobb et al. Sep 1991 A
5053698 Ueda Oct 1991 A
5058992 Takahashi Oct 1991 A
5059019 McCullough Oct 1991 A
5061075 Alfano et al. Oct 1991 A
5061919 Watkins Oct 1991 A
5063375 Lien et al. Nov 1991 A
5077608 Dubner Dec 1991 A
5088095 Zirngibl Feb 1992 A
5089903 Kuwayama et al. Feb 1992 A
5095491 Kozlovsky et al. Mar 1992 A
5097427 Lathrop et al. Mar 1992 A
5101184 Antes Mar 1992 A
5103306 Weiman et al. Apr 1992 A
5103339 Broome Apr 1992 A
5111468 Kozlovsky et al. May 1992 A
5113455 Scott May 1992 A
5115127 Bobb et al. May 1992 A
5117221 Mishica, Jr. May 1992 A
RE33973 Kriz et al. Jun 1992 E
5121086 Srivastava Jun 1992 A
5123085 Wells et al. Jun 1992 A
5124821 Antier et al. Jun 1992 A
5132812 Takahashi et al. Jul 1992 A
5134521 Lacroix et al. Jul 1992 A
5136675 Hodson Aug 1992 A
5136818 Bramson Aug 1992 A
5142788 Willetts Sep 1992 A
5155604 Miekka et al. Oct 1992 A
5157385 Nakao et al. Oct 1992 A
5159601 Huber Oct 1992 A
5161013 Rylander et al. Nov 1992 A
5175575 Gersuk Dec 1992 A
5179638 Dawson et al. Jan 1993 A
5185852 Mayer Feb 1993 A
5194969 DiFrancesco Mar 1993 A
5196922 Yeomans Mar 1993 A
5198661 Anderson et al. Mar 1993 A
5200818 Neta et al. Apr 1993 A
5206868 Deacon Apr 1993 A
5214757 Mauney et al. May 1993 A
5222205 Larson et al. Jun 1993 A
5226109 Dawson et al. Jul 1993 A
5227863 Bilbrey et al. Jul 1993 A
5229593 Cato Jul 1993 A
5230039 Grossman et al. Jul 1993 A
5231388 Stoltz Jul 1993 A
5239625 Bogart et al. Aug 1993 A
5241659 Parulski et al. Aug 1993 A
5242306 Fisher Sep 1993 A
5243448 Banbury Sep 1993 A
5251160 Rockwood et al. Oct 1993 A
5252068 Gryder Oct 1993 A
5255274 Wysocki et al. Oct 1993 A
5266930 Ichikawa et al. Nov 1993 A
5267045 Stroomer Nov 1993 A
5272473 Thompson et al. Dec 1993 A
5276849 Patel Jan 1994 A
5285397 Heier et al. Feb 1994 A
5291317 Newswanger Mar 1994 A
5293233 Billing et al. Mar 1994 A
5297156 Deacon Mar 1994 A
5300942 Dolgoff Apr 1994 A
5301062 Takahashi et al. Apr 1994 A
5311360 Bloom et al. May 1994 A
5315699 Imai et al. May 1994 A
5317576 Leonberger et al. May 1994 A
5317689 Nack et al. May 1994 A
5319744 Kelly et al. Jun 1994 A
5320353 Moore Jun 1994 A
5320534 Thomas Jun 1994 A
5325133 Adachi Jun 1994 A
5325485 Hochmuth et al. Jun 1994 A
5326266 Fisher et al. Jul 1994 A
5329323 Biles Jul 1994 A
5333021 Mitsutake et al. Jul 1994 A
5333245 Vecchione Jul 1994 A
5341460 Tam Aug 1994 A
5345280 Kimura et al. Sep 1994 A
5347433 Sedlmayr Sep 1994 A
5347620 Zimmer Sep 1994 A
5348477 Welch et al. Sep 1994 A
5353390 Harrington Oct 1994 A
5357579 Buchner et al. Oct 1994 A
5359526 Whittington et al. Oct 1994 A
5359704 Rossignac et al. Oct 1994 A
5360010 Applegate et al. Nov 1994 A
5361386 Watkins et al. Nov 1994 A
5363220 Kuwayama et al. Nov 1994 A
5363475 Baker et al. Nov 1994 A
5363476 Kurashige et al. Nov 1994 A
5367585 Ghezzo et al. Nov 1994 A
5367615 Economy et al. Nov 1994 A
5369450 Haseltine et al. Nov 1994 A
5369735 Thier et al. Nov 1994 A
5369739 Akeley Nov 1994 A
5377320 Abi-Ezzi et al. Dec 1994 A
5379371 Usami et al. Jan 1995 A
5380995 Udd et al. Jan 1995 A
5381338 Wysocki et al. Jan 1995 A
5381519 Brown et al. Jan 1995 A
5384719 Baker et al. Jan 1995 A
5388206 Poulton et al. Feb 1995 A
5394414 Kozlovsky et al. Feb 1995 A
5394515 Lentz et al. Feb 1995 A
5394516 Winser Feb 1995 A
5396349 Roberts et al. Mar 1995 A
5398083 Tsujihara et al. Mar 1995 A
5408249 Wharton et al. Apr 1995 A
5408606 Eckart Apr 1995 A
5410371 Lambert Apr 1995 A
5412796 Olive May 1995 A
5422986 Neely Jun 1995 A
5430888 Witek et al. Jul 1995 A
5432863 Benati et al. Jul 1995 A
5444839 Silverbrook et al. Aug 1995 A
5451765 Gerber Sep 1995 A
5459610 Bloom et al. Oct 1995 A
5459835 Trevett Oct 1995 A
5465121 Blalock et al. Nov 1995 A
5465368 Davidson et al. Nov 1995 A
5471545 Negami et al. Nov 1995 A
5471567 Soderberg et al. Nov 1995 A
5473373 Hwung et al. Dec 1995 A
5473391 Usui Dec 1995 A
5479597 Fellous Dec 1995 A
5480305 Montag et al. Jan 1996 A
5487665 Lechner et al. Jan 1996 A
5488687 Rich Jan 1996 A
5489920 Kaasila Feb 1996 A
5490238 Watkins Feb 1996 A
5490240 Foran et al. Feb 1996 A
5493439 Engle Feb 1996 A
5493629 Stange Feb 1996 A
5495563 Winser Feb 1996 A
5499194 Prestidge et al. Mar 1996 A
5500747 Tanide et al. Mar 1996 A
5500761 Goossen et al. Mar 1996 A
5502482 Graham Mar 1996 A
5502782 Smith Mar 1996 A
5504496 Tanaka et al. Apr 1996 A
5506949 Perrin Apr 1996 A
5519518 Watanabe et al. May 1996 A
5535374 Olive Jul 1996 A
5536085 Li et al. Jul 1996 A
5537159 Suematsu et al. Jul 1996 A
5539577 Si et al. Jul 1996 A
5541769 Ansley et al. Jul 1996 A
5544306 Deering et al. Aug 1996 A
5544340 Doi et al. Aug 1996 A
5550960 Shirman et al. Aug 1996 A
5551283 Manaka et al. Sep 1996 A
5557297 Sharp et al. Sep 1996 A
5557733 Hicok et al. Sep 1996 A
5559952 Fujimoto Sep 1996 A
5559954 Sakoda et al. Sep 1996 A
5561745 Jackson et al. Oct 1996 A
5566370 Young Oct 1996 A
5572229 Fisher Nov 1996 A
5574847 Eckart et al. Nov 1996 A
5579456 Cosman Nov 1996 A
5584696 Walker et al. Dec 1996 A
5586291 Lasker et al. Dec 1996 A
5590254 Lippincott et al. Dec 1996 A
5594854 Baldwin et al. Jan 1997 A
5598517 Watkins Jan 1997 A
5604849 Artwick et al. Feb 1997 A
5610665 Berman et al. Mar 1997 A
5612710 Christensen et al. Mar 1997 A
5614961 Gibeau et al. Mar 1997 A
5625768 Dye Apr 1997 A
5627605 Kim May 1997 A
5629801 Staker et al. May 1997 A
5630037 Schindler May 1997 A
5633750 Nogiwa et al. May 1997 A
5638208 Walker Jun 1997 A
5648860 Ooi et al. Jul 1997 A
5650814 Florent et al. Jul 1997 A
5651104 Cosman Jul 1997 A
5657077 DeAngelis et al. Aug 1997 A
5658060 Dove Aug 1997 A
5659490 Imamura Aug 1997 A
5659671 Tannenbaum et al. Aug 1997 A
5661592 Bornstein et al. Aug 1997 A
5661593 Engle Aug 1997 A
5665942 Williams et al. Sep 1997 A
5677783 Bloom et al. Oct 1997 A
5684939 Foran et al. Nov 1997 A
5684943 Abraham et al. Nov 1997 A
5689437 Nakagawa Nov 1997 A
5691999 Ball et al. Nov 1997 A
5694180 Deter et al. Dec 1997 A
5696892 Redmann et al. Dec 1997 A
5696947 Johns et al. Dec 1997 A
5699497 Erdahl et al. Dec 1997 A
5703604 McCutchen Dec 1997 A
5706061 Marshall et al. Jan 1998 A
5715021 Gibeau et al. Feb 1998 A
5719951 Shackleton et al. Feb 1998 A
5724561 Tarolli et al. Mar 1998 A
5726785 Chawki et al. Mar 1998 A
5734386 Cosman Mar 1998 A
5734521 Fukudome et al. Mar 1998 A
5739819 Bar-Nahum Apr 1998 A
5740190 Moulton Apr 1998 A
5742749 Foran et al. Apr 1998 A
5748264 Hegg May 1998 A
5748867 Cosman et al. May 1998 A
5761709 Kranich Jun 1998 A
5764280 Bloom et al. Jun 1998 A
5764311 Bonde et al. Jun 1998 A
5768443 Michael et al. Jun 1998 A
5781666 Ishizawa et al. Jul 1998 A
5793912 Boord et al. Aug 1998 A
5798743 Bloom Aug 1998 A
5808797 Bloom et al. Sep 1998 A
5818456 Cosman et al. Oct 1998 A
5818998 Harris et al. Oct 1998 A
5821944 Watkins Oct 1998 A
5825363 Anderson Oct 1998 A
5825538 Walker Oct 1998 A
5835256 Huibers Nov 1998 A
5837996 Keydar Nov 1998 A
5838328 Roller Nov 1998 A
5838484 Goossen Nov 1998 A
5841443 Einkauf Nov 1998 A
5841447 Drews Nov 1998 A
5841579 Bloom et al. Nov 1998 A
5850225 Cosman Dec 1998 A
5854631 Akeley et al. Dec 1998 A
5854865 Goldberg Dec 1998 A
5860721 Bowron et al. Jan 1999 A
5864342 Kajiya et al. Jan 1999 A
5867166 Myhrvold et al. Feb 1999 A
5867301 Engle Feb 1999 A
5870097 Snyder et al. Feb 1999 A
5870098 Gardiner Feb 1999 A
5874967 West et al. Feb 1999 A
5889529 Jones et al. Mar 1999 A
5900881 Ikedo May 1999 A
5903272 Otto May 1999 A
5905504 Barkans et al. May 1999 A
5908300 Walker et al. Jun 1999 A
5909225 Schinnerer et al. Jun 1999 A
5912670 Lipscomb et al. Jun 1999 A
5912740 Zare et al. Jun 1999 A
5917495 Doi et al. Jun 1999 A
5920361 Gibeau et al. Jul 1999 A
5923333 Stroyan Jul 1999 A
5930740 Mathisen Jul 1999 A
5943060 Cosman et al. Aug 1999 A
5946129 Xu et al. Aug 1999 A
5963788 Barron et al. Oct 1999 A
5969699 Balram et al. Oct 1999 A
5969721 Chen et al. Oct 1999 A
5969726 Rentschler et al. Oct 1999 A
5974059 Dawson Oct 1999 A
5977977 Kajiya et al. Nov 1999 A
5980044 Cannon et al. Nov 1999 A
5982553 Bloom et al. Nov 1999 A
5987200 Fleming et al. Nov 1999 A
5988814 Rohlfing et al. Nov 1999 A
5990935 Rohlfing Nov 1999 A
5999549 Freitag et al. Dec 1999 A
6002454 Kajiwara et al. Dec 1999 A
6002505 Kraenert et al. Dec 1999 A
6005580 Donovan Dec 1999 A
6005611 Gullichsen et al. Dec 1999 A
6014144 Nelson et al. Jan 2000 A
6014163 Houskeeper Jan 2000 A
6021141 Nam et al. Feb 2000 A
6031541 Lipscomb et al. Feb 2000 A
6034739 Rohlfing et al. Mar 2000 A
6038057 Brazas, Jr. et al. Mar 2000 A
6042238 Blackham et al. Mar 2000 A
6052125 Gardiner et al. Apr 2000 A
6052485 Nelson et al. Apr 2000 A
6057909 Yahav et al. May 2000 A
6064392 Rohner May 2000 A
6064393 Lengyel et al. May 2000 A
6069903 Zanger et al. May 2000 A
6072500 Foran et al. Jun 2000 A
6072544 Gleim et al. Jun 2000 A
6078333 Wittig et al. Jun 2000 A
6084610 Ozaki et al. Jul 2000 A
6094226 Ke et al. Jul 2000 A
6094267 Levenson et al. Jul 2000 A
6094298 Luo et al. Jul 2000 A
6100906 Asaro et al. Aug 2000 A
6101036 Bloom Aug 2000 A
6108054 Heizmann et al. Aug 2000 A
6111616 Chauvin et al. Aug 2000 A
6122413 Jiang et al. Sep 2000 A
6124647 Marcus et al. Sep 2000 A
6124808 Budnovitch Sep 2000 A
6124922 Sentoku Sep 2000 A
6124989 Oode et al. Sep 2000 A
6126288 Hewlett Oct 2000 A
6128019 Crocker, III et al. Oct 2000 A
6128021 van der Meulen et al. Oct 2000 A
6130770 Bloom Oct 2000 A
6134339 Luo Oct 2000 A
6137565 Ecke et al. Oct 2000 A
6137932 Kim et al. Oct 2000 A
6141013 Nelson et al. Oct 2000 A
6141025 Oka et al. Oct 2000 A
6141034 McCutchen Oct 2000 A
6144481 Kowarz et al. Nov 2000 A
6147690 Cosman Nov 2000 A
6147695 Bowen et al. Nov 2000 A
6147789 Gelbart Nov 2000 A
6154259 Hargis et al. Nov 2000 A
6175579 Sandford et al. Jan 2001 B1
6184888 Yuasa et al. Feb 2001 B1
6184891 Blinn Feb 2001 B1
6184926 Khosravi et al. Feb 2001 B1
6188427 Anderson et al. Feb 2001 B1
6188712 Jiang et al. Feb 2001 B1
6191827 Segman et al. Feb 2001 B1
6195099 Gardiner Feb 2001 B1
6195484 Brennan, III et al. Feb 2001 B1
6195609 Pilley et al. Feb 2001 B1
6204859 Jouppi et al. Mar 2001 B1
6204955 Chao et al. Mar 2001 B1
6215579 Bloom et al. Apr 2001 B1
6219015 Bloom et al. Apr 2001 B1
6222937 Cohen et al. Apr 2001 B1
6229650 Reznichenko et al. May 2001 B1
6229827 Fernald et al. May 2001 B1
6233025 Wallenstein May 2001 B1
6236408 Watkins May 2001 B1
6240220 Pan et al. May 2001 B1
6262739 Migdal et al. Jul 2001 B1
6262810 Bloomer Jul 2001 B1
6263002 Hsu et al. Jul 2001 B1
6266068 Kang et al. Jul 2001 B1
6268861 Sanz-Pastor et al. Jul 2001 B1
6282012 Kowarz et al. Aug 2001 B1
6282220 Floyd Aug 2001 B1
6285407 Yasuki et al. Sep 2001 B1
6285446 Farhadiroushan Sep 2001 B1
6292165 Lin et al. Sep 2001 B1
6292268 Hirota et al. Sep 2001 B1
6292310 Chao Sep 2001 B1
6297899 Romanovsky Oct 2001 B1
6298066 Wettroth et al. Oct 2001 B1
6301370 Steffens et al. Oct 2001 B1
6304245 Groenenboom Oct 2001 B1
6307558 Mao Oct 2001 B1
6307663 Kowarz Oct 2001 B1
6308144 Bronfeld et al. Oct 2001 B1
6320688 Westbrook et al. Nov 2001 B1
6323984 Trisnadi Nov 2001 B1
6333792 Kimura Dec 2001 B1
6333803 Kurotori et al. Dec 2001 B1
6335765 Daly et al. Jan 2002 B1
6335941 Grubb et al. Jan 2002 B1
6340806 Smart et al. Jan 2002 B1
6356683 Hu et al. Mar 2002 B1
6360042 Long Mar 2002 B1
6361173 Vlahos et al. Mar 2002 B1
6362817 Powers et al. Mar 2002 B1
6362818 Gardiner et al. Mar 2002 B1
6363089 Fernald et al. Mar 2002 B1
6366721 Hu et al. Apr 2002 B1
6369936 Moulin Apr 2002 B1
6370312 Wagoner et al. Apr 2002 B1
6374011 Wagoner et al. Apr 2002 B1
6374015 Lin Apr 2002 B1
6375366 Kato et al. Apr 2002 B1
6381072 Burger Apr 2002 B1
6381385 Watley et al. Apr 2002 B1
6384828 Arbeiter et al. May 2002 B1
6388241 Ang May 2002 B1
6393036 Kato May 2002 B1
6393181 Bulman et al. May 2002 B1
6396994 Philipson et al. May 2002 B1
6404425 Cosman Jun 2002 B1
6407736 Regan Jun 2002 B1
6411425 Kowarz et al. Jun 2002 B1
6421636 Cooper et al. Jul 2002 B1
6424343 Deering et al. Jul 2002 B1
6429876 Morein Aug 2002 B1
6429877 Stroyan Aug 2002 B1
6433823 Nakamura et al. Aug 2002 B1
6433838 Chen Aug 2002 B1
6433840 Poppleton Aug 2002 B1
6437789 Tidwell et al. Aug 2002 B1
6445362 Tegreene Sep 2002 B1
6445433 Levola Sep 2002 B1
6449071 Farhan et al. Sep 2002 B1
6449293 Pedersen et al. Sep 2002 B1
6452667 Fernald et al. Sep 2002 B1
6456288 Brockway et al. Sep 2002 B1
6466206 Deering Oct 2002 B1
6466224 Nagata et al. Oct 2002 B1
6470036 Bailey et al. Oct 2002 B1
6473090 Mayer Oct 2002 B1
6476848 Kowarz et al. Nov 2002 B2
6480513 Kapany et al. Nov 2002 B1
6480634 Corrigan Nov 2002 B1
6490931 Fernald et al. Dec 2002 B1
6496160 Tanner et al. Dec 2002 B1
6507706 Brazas et al. Jan 2003 B1
6510272 Wiegand Jan 2003 B1
6511182 Agostinelli et al. Jan 2003 B1
RE37993 Zhang Feb 2003 E
6519388 Fernald et al. Feb 2003 B1
6522809 Takabayashi et al. Feb 2003 B1
6525740 Cosman Feb 2003 B1
6529310 Huibers et al. Mar 2003 B1
6529531 Everage et al. Mar 2003 B1
6534248 Jain et al. Mar 2003 B2
6538656 Cheung et al. Mar 2003 B1
6549196 Taguchi et al. Apr 2003 B1
6554431 Binsted et al. Apr 2003 B1
6556627 Kitamura et al. Apr 2003 B2
6563968 Davis et al. May 2003 B2
6574352 Skolmoski Jun 2003 B1
6575581 Tsurushima Jun 2003 B2
6577429 Kurtz et al. Jun 2003 B1
6580430 Hollis et al. Jun 2003 B1
6591020 Klassen Jul 2003 B1
6594043 Bloom et al. Jul 2003 B1
6597363 Duluk, Jr. et al. Jul 2003 B1
6598979 Yoneno Jul 2003 B2
6600460 Mays, Jr. Jul 2003 B1
6600830 Lin et al. Jul 2003 B1
6600854 Anderegg et al. Jul 2003 B2
6603482 Tidwell Aug 2003 B1
6643299 Lin Nov 2003 B1
6646645 Simmonds et al. Nov 2003 B2
6650326 Huber et al. Nov 2003 B1
6671293 Kopp et al. Dec 2003 B2
6678085 Kowarz et al. Jan 2004 B2
6690655 Miner et al. Feb 2004 B1
6692129 Gross et al. Feb 2004 B2
6711187 Tanner et al. Mar 2004 B2
6727918 Nason Apr 2004 B1
6738105 Hannah et al. May 2004 B1
6741384 Martin et al. May 2004 B1
6747649 Sanz-Pastor et al. Jun 2004 B1
6747781 Trisnadi Jun 2004 B2
6751001 Tanner et al. Jun 2004 B1
6760036 Tidwell Jul 2004 B2
6763042 Williams et al. Jul 2004 B2
6773142 Rekow Aug 2004 B2
6776045 Fernald et al. Aug 2004 B2
6782205 Trisnadi et al. Aug 2004 B2
6788304 Hart et al. Sep 2004 B1
6788307 Coleman et al. Sep 2004 B2
6789903 Parker et al. Sep 2004 B2
6791562 Cosman et al. Sep 2004 B2
6793350 Raskar et al. Sep 2004 B1
6798418 Sartori et al. Sep 2004 B1
6799850 Hong et al. Oct 2004 B2
6801205 Gardiner et al. Oct 2004 B2
6809731 Muffler et al. Oct 2004 B2
6811267 Allen et al. Nov 2004 B1
6816169 Cosman Nov 2004 B2
6831648 Mukherjee et al. Dec 2004 B2
6840627 Olbrich Jan 2005 B2
6842298 Shafer et al. Jan 2005 B1
6856449 Winkler et al. Feb 2005 B2
6868212 DeWitte et al. Mar 2005 B2
6871958 Streid et al. Mar 2005 B2
6897878 Cosman et al. May 2005 B2
6943803 Cosman et al. Sep 2005 B1
6956582 Tidwell Oct 2005 B2
6956878 Trisnadi Oct 2005 B1
6971576 Tsikos et al. Dec 2005 B2
6984039 Agostinelli Jan 2006 B2
6985663 Catchmark et al. Jan 2006 B2
7012669 Streid et al. Mar 2006 B2
7030883 Thompson Apr 2006 B2
7038735 Coleman et al. May 2006 B2
7043102 Okamoto et al. May 2006 B2
7053911 Cosman May 2006 B2
7053912 Cosman May 2006 B2
7053913 Cosman May 2006 B2
7054051 Bloom May 2006 B1
7091980 Tidwell Aug 2006 B2
7095423 Cosman et al. Aug 2006 B2
7110153 Sakai Sep 2006 B2
7110624 Williams et al. Sep 2006 B2
7111943 Agostinelli et al. Sep 2006 B2
7113320 Tanner Sep 2006 B2
7133583 Marceau et al. Nov 2006 B2
7169630 Moriwaka Jan 2007 B2
7193765 Christensen et al. Mar 2007 B2
7193766 Bloom Mar 2007 B2
7197200 Marceau et al. Mar 2007 B2
7210786 Tamura et al. May 2007 B2
7215840 Marceau et al. May 2007 B2
7237916 Mitomori Jul 2007 B2
7257519 Cosman Aug 2007 B2
7267442 Childers et al. Sep 2007 B2
7277216 Bloom Oct 2007 B2
7286277 Bloom et al. Oct 2007 B2
7317464 Willis Jan 2008 B2
7327909 Marceau et al. Feb 2008 B2
7334902 Streid et al. Feb 2008 B2
7354157 Takeda et al. Apr 2008 B2
7364309 Sugawara et al. Apr 2008 B2
7400449 Christensen et al. Jul 2008 B2
7420177 Williams et al. Sep 2008 B2
7594965 Tanaka Sep 2009 B2
20010002124 Mamiya et al. May 2001 A1
20010010536 Katzir et al. Aug 2001 A1
20010027456 Lancaster et al. Oct 2001 A1
20010047251 Kemp Nov 2001 A1
20020005862 Deering Jan 2002 A1
20020021462 Delfyett et al. Feb 2002 A1
20020030769 Bae Mar 2002 A1
20020042674 Mochizuki et al. Apr 2002 A1
20020067467 Dorval et al. Jun 2002 A1
20020071453 Lin Jun 2002 A1
20020075202 Fergason Jun 2002 A1
20020101647 Moulin Aug 2002 A1
20020136121 Salmonsen et al. Sep 2002 A1
20020145615 Moore Oct 2002 A1
20020145806 Amm Oct 2002 A1
20020146248 Herman et al. Oct 2002 A1
20020154860 Fernald et al. Oct 2002 A1
20020176134 Vohra Nov 2002 A1
20020196414 Manni et al. Dec 2002 A1
20030035190 Brown et al. Feb 2003 A1
20030038807 Demos et al. Feb 2003 A1
20030039443 Catchmark et al. Feb 2003 A1
20030048275 Ciolac Mar 2003 A1
20030081303 Sandstrom et al. May 2003 A1
20030086647 Willner et al. May 2003 A1
20030142319 Ronnekleiv et al. Jul 2003 A1
20030160780 Lefebvre et al. Aug 2003 A1
20030174312 Leblanc Sep 2003 A1
20030214633 Roddy et al. Nov 2003 A1
20030235304 Evans et al. Dec 2003 A1
20040017518 Stern et al. Jan 2004 A1
20040017608 Lantz Jan 2004 A1
20040085283 Wang May 2004 A1
20040136074 Ford et al. Jul 2004 A1
20040165154 Kobori et al. Aug 2004 A1
20040179007 Bower et al. Sep 2004 A1
20040183954 Hannah et al. Sep 2004 A1
20040184013 Raskar et al. Sep 2004 A1
20040196660 Usami Oct 2004 A1
20040207618 Williams et al. Oct 2004 A1
20050018309 McGuire, Jr. et al. Jan 2005 A1
20050024722 Agostinelli et al. Feb 2005 A1
20050047134 Mueller et al. Mar 2005 A1
20050093854 Kennedy et al. May 2005 A1
20050243389 Kihara Nov 2005 A1
20060039051 Baba et al. Feb 2006 A1
20060114544 Bloom et al. Jun 2006 A1
20060176912 Anikitchev Aug 2006 A1
20060221429 Christensen et al. Oct 2006 A1
20060238851 Bloom Oct 2006 A1
20060255243 Kobayashi et al. Nov 2006 A1
20070183473 Bicknell et al. Aug 2007 A1
20080037125 Takamiya Feb 2008 A1
20080218837 Yang et al. Sep 2008 A1
20090033875 King et al. Feb 2009 A1
Foreign Referenced Citations (51)
Number Date Country
2 325 028 Dec 1974 DE
197 21 416 Jan 1999 DE
0 155 858 Sep 1985 EP
0 306 308 Mar 1989 EP
0 319 165 Jul 1989 EP
0 417 039 Mar 1991 EP
0 480 570 Apr 1992 EP
0 488 326 Jun 1992 EP
0 489 594 Jun 1992 EP
0 528 646 Feb 1993 EP
0 530 760 Mar 1993 EP
0 550 189 Jul 1993 EP
0 610 665 Aug 1994 EP
0 621 548 Oct 1994 EP
0 627 644 Dec 1994 EP
0 627 850 Dec 1994 EP
0 643 314 Mar 1995 EP
0 654 777 May 1995 EP
0 658 868 Jun 1995 EP
0 689 078 Dec 1995 EP
0 801 319 Oct 1997 EP
0 880 282 Nov 1998 EP
1 365 584 Nov 2003 EP
2 118 365 Oct 1983 GB
2 144 608 Mar 1985 GB
2 179 147 Feb 1987 GB
2 245 806 Jan 1992 GB
2 251 770 Jul 1992 GB
2 251 773 Jul 1992 GB
2 266 385 Oct 1993 GB
2 293 079 Mar 1996 GB
63-305323 Dec 1988 JP
2-219092 Aug 1990 JP
2000-305481 Nov 2000 JP
8701571 Mar 1987 WO
9212506 Jul 1992 WO
9302269 Feb 1993 WO
9309472 May 1993 WO
9318428 Sep 1993 WO
9511473 Apr 1995 WO
9527267 Oct 1995 WO
9641217 Dec 1996 WO
9641224 Dec 1996 WO
9726569 Jul 1997 WO
9815127 Apr 1998 WO
0146248 Jun 2001 WO
0157581 Aug 2001 WO
0212925 Feb 2002 WO
0223824 Mar 2002 WO
0231575 Apr 2002 WO
03001281 Jan 2003 WO
Non-Patent Literature Citations (118)
Entry
Apte, “Grating Light Valves for High-Resolution Displays,” Ph.D. Dissertation—Stanford University, 1994 (abstract only).
Ellis, “Lo-cost Bimorph Mirrors in Adaptive Optics,” Ph.D. Thesis, Imperial College of Science, Technology and Medicine—University of London, 1999.
Halevi, “Bimorph piezoelectric flexible mirror: graphical solution and comparison with experiment,” J. Opt. Soc. Am., Jan. 1983, pp. 110-113, vol. 73, No. 1.
Kudryashov et al., “Adaptive Optics for High Power Laser ZBeam Control,” Springer Proceedings in Physics, 2005, pp. 237-248, vol. 102.
Safronov, “Bimorph adaptive optics: elements, technology and design principles,” SPIE, 1996, pp. 494-504, vol. 2774.
Solgaard, “Integrated Semiconductor Light Modulators for Fiber-Optic and Display Applications,” Ph.D. Dissertation submitted to the Deparatment of Electrical Engineering and the Committee on Graduate Studies of Stanford University, Feb. 1992.
Steinhaus et al., “Bimorph piezoelectric flexible mirror,” J. Opt. Soc. Am., Mar. 1979, pp. 478-481, vol. 69, No. 3.
Tseng et al., “Development of an Aspherical Bimorph PZT Mirror Bender with Thin Film Resistor Electrode,” Advanced Photo Source, Argonne National Laboratory, Sep. 2002, pp. 271-278.
Vinevich et al., “Cooled and uncooled single-channel deformable mirrors for industrial laser systems,” Quantum Electronics, 1998, pp. 366-369, vol. 28, No. 4.
Abrash, “The Quake Graphics Engine,” CGDC Quake Talk taken from Computer Game Developers Conference on Apr. 2, 1996. http://gamers.org/dEngine/quake/papers/mikeab-cgdc.html.
Akeley, “RealityEngine Graphics,” Computer Graphics Proceedings, Annual Conference Series, 1993.
Allen, J. et al., “An Interactive Learning Environment for VLSI Design,” Proceedings of the IEEE, Jan. 2000, pp. 96-106, vol. 88, No. 1.
Allen, W. et al., “47.4: Invited Paper: Wobulation: Doubling the Addressed Resolution of Projection Displays,” SID 05 Digest, 2005, pp. 1514-1517.
Amm, et al., “5.2: Grating Light Valve™ Technology: Update and Novel Applications,” Presented at Society for Information Display Symposium, May 19, 1998, Anaheim, California.
Apgar et al., “A Display System for the Stellar™ Graphics Supercomputer Model GS1000™,” Computer Graphics, Aug. 1988, pp. 255-262, vol. 22, No. 4.
Baer, Computer Systems Architecture, 1980, Computer Science Press, Inc., Rockville, Maryland.
Barad et al., “Real-Time Procedural Texturing Techniques Using MMX,” Gamasutra, May 1, 1998, http://www. gamasutra.com/features/19980501/mmxtexturing—01.htm.
Bass, “4K GLV Calibration,” E&S Company, Jan. 8, 2008.
Becker et al., “Smooth Transitions between Bump Rendering Algorithms,” Computer Graphics Proceedings, 1993, pp. 183-189.
Bishop et al., “Frameless Rendering: Double Buffering Considered Harmful,” Computer Graphics Proceedings, Annual Conference Series, 1994.
Blinn, “Simulation of Wrinkled Surfaces,” Siggraph '78 Proceedings, 1978, pp. 286-292.
Blinn, “A Trip Down the Graphics Pipeline: Subpixelic Particles,” IEEE Computer Graphics & Applications, Sep./Oct. 1991, pp. 86-90, vol. 11, No. 5.
Blinn et al., “Texture and Reflection in Computer Generated Images,” Communications of the ACM, Oct. 1976, pp. 542-547, vol. 19, No. 10.
Boyd et al., “Parametric Interaction of Focused Gaussian Light Beams,” Journal of Applied Physics, Jul. 1968, pp. 3597-3639vol. 39, No. 8.
Brazas et al., “High-Resolution Laser-Projection Display System Using a Grating Electromechanical System (GEMS),” MOEMS Display and Imaging Systems II, Proceedings of SPIE, 2004, pp. 65-75vol. 5348.
Bresenham, “Algorithm for computer control of a digital plotter,” IBM Systems Journal, 1965, pp. 25-30, vol. 4, No. 1.
Carlson, “An Algorithm and Data Structure for 3D Object Synthesis Using Surface Patch Intersections,” Computer Graphics, Jul. 1982, pp. 255-263, vol. 16, No. 3.
Carpenter, “The A-buffer, an Antialiased Hidden Surface Method,” Computer Graphics, Jul. 1984, pp. 103-108, vol. 18, No. 3.
Carter, “Re: Re seams and creaseAngle (long),” posted on the GeoVRML.org website Feb. 2, 2000, http://www.ai.sri.com/geovrml/archive/msg00560.html.
Catmull, “An Analytic Visible Surface Algorithm for Independent Pixel Processing,” Computer Graphics, Jul. 1984, pp. 109-115, vol. 18, No. 3.
Chasen, Geometric Principles and Procedures for Computer Graphic Applications, 1978, pp. 11-123, Upper Saddle River, New Jersey.
Choy et al., “Single Pass Algorithm for the Generation of Chain-Coded Contours and Contours Inclusion Relationship,” Communications, Computers and Signal Processing—IEEE Pac Rim '93, 1993, pp. 256-259.
Clark et al., “Photographic Texture and CIG: Modeling Strategies for Production Data Bases,” 9th VITSC Proceedings, Nov. 30-Dec. 2, 1987, pp. 274-283.
Corrigan et al., “Grating Light Valve™ Technology for Projection Displays,” Presented at the International Display Workshop—Kobe, Japan, Dec. 9, 1998.
Crow, “Shadow Algorithms for Computer Graphics,” Siggraph '77, Jul. 20-22, 1977, San Jose, California, pp. 242, 248.
Deering et al., “FBRAM: A new Form of Memory Optimized for 3D Graphics,” Computer Graphics Proceedings, Annual Conference Series, 1994.
Drever et al., “Laser Phase and Frequency Stabilization Using an Optical Resonator,” Applied Physics B: Photophysics and Laser Chemistry, 1983, pp. 97-105, vol. 31.
Duchaineau et al., “ROAMing Terrain: Real-time Optimally Adapting Meshes,” Los Alamos National Laboratory and Lawrence Livermore National Laboratory, 1997.
Duff, “Compositing 3-D Rendered Images,” Siggraph '85, Jul. 22-26, 1985, San Francisco, California, pp. 41-44.
Faux et al., Computational Geometry for Design and Manufacture, 1979, Ellis Horwood, Chicester, United Kingdom.
Feiner et al., “Dial: A Diagrammatic Animation Language,” IEEE Computer Graphics & Applications, Sep. 1982, pp. 43-54, vol. 2, No. 7.
Fiume et al., “A Parallel Scan Conversion Algorithm with Anti-Aliasing for a General-Purpose Ultracomputer,” Computer Graphics, Jul. 1983, pp. 141-150, vol. 17, No. 3.
Foley et al., Computer Graphics: Principles and Practice, 2nd ed., 1990, Addison-Wesley Publishing Co., Inc., Menlo Park, California.
Foley et al., Fundamentals of Interactive Computer Graphics, 1982, Addison-Wesley Publishing Co., Inc., Menlo Park, California.
Fox et al., “Development of Computer-Generated Imagery for a Low-Cost Real-Time Terrain Imaging System,” IEEE 1986 National Aerospace and Electronic Conference, May 19-23, 1986, pp. 986-991.
Gambotto, “Combining Image Analysis and Thermal Models for Infrared Scene Simulations,” Image Processing Proceedings, ICIP-94, IEEE International Conference, 1994, vol. 1, pp. 710-714.
Gardiner, “A Method for Rendering Shadows,” E&S Company, Sep. 25, 1996.
Gardiner, “Shadows in Harmony,” E&S Company, Sep. 20, 1996.
Gardner, “Simulation of Natural Scenes Using Textured Quadric Surfaces,” Computer Graphics, Jul. 1984, pp. 11-20, vol. 18, No. 3.
Gardner, “Visual Simulation of Clouds,” Siggraph '85, Jul. 22-26, 1985, San Francisco, California, pp. 297-303.
Giloi, Interactive Computer Graphics: Data Structures, Algorithms, Languages, 1978, Prentice-Hall, Inc., Englewood Cliffs, New Jersey.
Glaskowsky, “Intel Displays 740 Graphics Chip: Auburn Sets New Standard for Quality—But Not Speed,” Microprocessor Report, Feb. 16, 1998, pp. 5-9, vol. 12, No. 2.
Goshtasby, “Registration of Images with Geometric Distortions,” IEEE Transactions on Geoscience and Remote Sensing, Jan. 1988, pp. 60-64, vol. 26, No. 1.
Great Britain Health & Safety Executive, The Radiation Safety of Lasers Used for Display Purposes, Oct. 1996.
Gupta et al., “Filtering Edges for Gray-Scale Displays,” Computer Graphics, Aug. 1981, pp. 1-5, vol. 15, No. 3.
Gupta et al., “A VLSI Architecture for Updating Raster-Scan Displays,” Computer Graphics, Aug. 1981, pp. 71-78, vol. 15, No. 3.
Hearn et al., Computer Graphics, 2nd ed., 1994, pp. 143-183.
Heckbert, “Survey of Texture Mapping,” IEEE Computer Graphics and Applications, Nov. 1986, pp. 56-67.
Heckbert, “Texture Mapping Polygons in Perspective,” New York Institute of Technology, Computer Graphics Lab, Technical Memo No. 13, Apr. 28, 1983.
Heidrich et al., “Applications of Pixel Textures in Visualization and Realistic Image Synthesis,” Symposium on INteractive 3D Graphics, 1990, pp. 127-135, Atlanta, Georgia.
Holten-Lund, Design for Scalability in 3D Computer Graphics Architectures, Ph.D. thesis, Computer Science sand Technology Informatics and Mathematical Modelling, Technical University of Denmark, Jul. 2001.
INTEL740 Graphics Accelerator Datasheet, Apr. 1998.
INTEL470 Graphics Accelerator Datasheet, Architectural Overview, at least as early as Apr. 30, 1998.
Jacob, “Eye Tracking in Advanced Interface Design,” ACM, 1995.
Kelley et al., “Hardware Accelerated Rendering of CSG and Transparency,” SIGGRAPH '94, in Computer Graphics Proceedings, Annual Conference Series, 1994, pp. 177-184.
Klassen, “Modeling the Effect of the Atmosphere on Light,” ACM Transactions on Graphics, Jul. 1987, pp. 215-237, vol. 6, No. 3.
Kleiss, “Tradeoffs Among Types of Scene Detail for Simulating Low-Altitude Flight,” University of Dayton Research Institute, Aug. 1, 1992, pp. 1141-1146.
Lewis, “Algorithms for Solid Noise Synthesis,” SIGGRAPH '89, Computer Graphics, Jul. 1989, pp. 263-270, vol. 23, No. 3.
Lindstrom et al., “Real-Time, Continuous Level of Detail Rendering of Height Fields,” SIGGRAPH '96, Aug. 1996.
McCarty et al., “A Virtual Cockpit for a Distributed Interactive Simulation,” IEEE Computer Graphics & Applications, Jan. 1994, pp. 49-54.
Microsoft Flight Simulator 2004, Aug. 9, 2000. http://www.microsoft.com/games/flightsimulator/fs2000—devdesk.sdk.asp.
Miller et al., “Illumination and Reflection Maps: Simulated Objects in Simulated and Real Environments,” SIGGRAPH '84, Course Notes for Advances Computer Graphics Animation, Jul. 23, 1984.
Mitchell, “Spectrally Optimal Sampling for Distribution Ray Tracing,” SIGGRAPH '91, Computer Graphics, Jul. 1991, pp. 157-165, vol. 25, No. 4.
Mitsubishi Electronic Device Group, “Overview of 3D-RAM and Its Functional Blocks,” 1995.
Montrym et al., “InfiniteReality: A Real-Time Graphics System,” Computer Graphics Proceedings, Annual Conference Series, 1997.
Mooradian et al., “High Power Extended Vertical Cavity Surface Emitting Diode Lasers and Arrays and Their Applications,” Micro-Optics Conference, Tokyo, Nov. 2, 2005.
Musgrave et al., “The Synthesis and Rendering of Eroded Fractal Terrains,” SIGGRAPH '89, Computer Graphics, Jul. 1989, pp. 41-50, vol. 23, No. 3.
Nakamae et al., “Compositing 3D Images with Antialiasing and Various Shading Effects,” IEEE Computer Graphics & Applications, Mar. 1989, pp. 21-29, vol. 9, No. 2.
Newman et al., Principles of Interactive Computer Graphics, 2nd ed., 1979, McGraw-Hill Book Company, San Francisco, California.
Niven, “Trends in Laser Light Sources for Projection Display,” Novalux International Display Workshop, Session LAD2-2, Dec. 2006.
Oshima et al., “An Animation Design Tool Utilizing Texture,” International Workshop on Industrial Applications of Machine Intelligence and Vision, Tokyo, Apr. 10-12, 1989, pp. 337-342.
Parke, “Simulation and Expected Performance Analysis of Multiple Processor Z-Buffer Systems,” Computer Graphics, 1980, pp. 48-56.
Peachey, “Solid Texturing of Complex Surfaces,” SIGGRAPH '85, 1985, pp. 279-286, vol. 19, No. 3.
Peercy et al., “Efficient Bump Mapping Hardware,” Computer Graphics Proceedings, 1997.
Perlin, “An Image Synthesizer,” SIGGRAPH '85, 1985, pp. 287-296, vol. 19, No. 3.
Pineda, “A Parallel Algorithm for Polygon Rasterization,” SIGGRAPH '88, Aug. 1988, pp. 17-20, vol. 22, No. 4.
Porter et al., “Compositing Digital Images,” SIGGRAPH '84, Computer Graphics, Jul. 1984, pp. 253-259, vol. 18, No. 3.
Poulton et al., “Breaking the Frame-Buffer Bottleneck with Logic-Enhanced Memories,” IEEE Computer Graphics & Applications, Nov. 1992, pp. 65-74.
Reeves et al., “Rendering Antialiased Shadows with Depth Maps,” SIGGRAPH '87, Computer Graphics, Jul. 1987, pp. 283-291, vol. 21, No. 4.
Regan et al., “Priority Rendering with a Virtual Reality Address Recalculation Pipeline,” Computer Graphics Proceedings, Annual Conference Series, 1994.
Rhoades et al., “Real-Time Procedural Textures,” ACM, Jun. 1992, pp. 95-100, 225.
Rockwood et al., “Blending Surfaces in Solid Modeling,” Geometric Modeling: Algorithms and New Trends, 1987, pp. 367-383, Society for Industrial and Applied Mathematics, Philadelphia, Pennsylvania.
Röttger et al., “Real-Time Generation of Continuous Levels of Detail for Height Fields,” WSCG '98, 1998.
Saha et al., “Web-based Distributed VLSI Design,” IEEE, 1997, pp. 449-454.
Salzman et al., “VR's Frames of Reference: A Visualization Technique for Mastering Abstract Multidimensional Information,” CHI 99 Papers, May 1999, pp. 489-495.
Sandejas, Silicon Microfabrication of Grating Light Valves, Doctor of Philosophy Dissertation, Stanford University, Jul. 1995.
Scarlatos, “A Refined Triangulation Hierarchy for Multiple Levels of Terrain Detail,” presented at the Image V Conference, Phoenix, Arizona, Jun. 19-22, 1990, pp. 114-122.
Schilling, “A New Simple and Efficient Antialiasing with Subpixel Masks,” SIGGRAPH '91, Computer Graphics, Jul. 1991, pp. 133-141, vol. 25, No. 4.
Schumacker, “A New Visual System Architecture,” Proceedings of the Second Interservices/Industry Training Equipment Conference, Nov. 18-20, 1990, Salt Lake City, Utah.
Segal et al., “Fast Shadows and Lighting Effects Using Texture Mapping,” SIGGRAPH '92, Computer Graphics, Jul. 1992, pp. 249-252, vol. 26, No. 2.
Sick AG, S3000 Safety Laser Scanner Operating Instructions, Aug. 25, 2005.
Silicon Light Machines, “White Paper: Calculating Response Characteristics for the ‘Janis’ GLV Module, Revision 2.0,” Oct. 1999.
Stevens et al., “The National Simulation Laboratory: The Unifying Tool for Air Traffic Control System Development,” Proceedings of the 1991 Winter Simulation Conference, 1991, pp. 741-746.
Stone, High-Performance Computer Architecture, 1987, pp. 278-330, Addison-Wesley Publishing Company, Menlo Park, California.
Tanner et al., “The Clipmap: A Virtual Mipmap,” Silicon Graphics Computer Systems; Proceedings of the 25th Annual Conference on Computer Graphics and Interactive Techniques, Jul. 1998.
Tanriverdi et al., “Interacting with Eye Movements in Virtual Environments,” CHI Letters, Apr. 2000, pp. 265-272, vol. 2, No. 1.
Texas Instruments, DLP® 3-D HDTV Technology, 2007.
Torborg et al., “Talisman: Commodity Realtime 3D Graphics for the PC,” Computer Graphics Proceedings, Annual Conference Series, 1996, pp. 353-363.
Trisnadi, “Hadamard speckle contrast reduction,” Optics Letters, 2004, vol. 29, pp. 11-13.
Trisnadi et al., “Overview and applications of Grating Light Valve™ based optical write engines for high-speed digital imaging”, proceedings of conference “MOEMS Display and Imaging SYstems II,” Jan. 2004, vol. 5328, 13 pages.
Whitton, “Memory Design for Raster Graphics Displays,” IEEE Computer Graphics & Applications, Mar. 1984, pp. 48-65.
Williams, “Casting Curved Shadows on Curved Surfaces,” Computer Graphics Lab, New York Institute of Technology, 1978, pp. 270-274.
Williams, “Pyramidal Parametrics,” Computer Graphics, Jul. 1983, pp. 1-11, vol. 17, No. 3.
Willis et al., “A Method for Continuous Adaptive Terrain,” Presented at the 1996 Image Conference, Jun. 23-28, 1996.
Woo et al., “A Survey of Shadow Algorithms,” IEEE Computer Graphics & Applications, Nov. 1990, pp. 13-32, vol. 10, No. 6.
Wu et al., “A Differential Method for Simultaneous Estimation of Rotation, Change of Scale and Translation,” Signal Processing: Image Communication, 1990, pp. 69-80, vol. 2, No. 1.
Youbing et al., “A Fast Algorithm for Large Scale Terrain Walkthrough,” CAD/Graphics, Aug. 22-24, 2001, 6 pages.
Sollberger et al., “Frequency Stabilization of Semiconductor Lasers for Applications in Coherent Communication Systems,” Journal of Lightwave Technology, Apr. 1987, pp. 485-491, vol. LT-5, No. 4.
Provisional Applications (1)
Number Date Country
61060744 Jun 2008 US