HUD system and method with dynamic light exclusion

Information

  • Patent Grant
  • 9679367
  • Patent Number
    9,679,367
  • Date Filed
    Thursday, April 24, 2014
    10 years ago
  • Date Issued
    Tuesday, June 13, 2017
    7 years ago
Abstract
A system and method can mitigate or prevent contrast ratio issues due to bright light (e.g., light saturation) in a head up display (HUD). The head up display can include a waveguide combiner (or non-waveguide combiner) and a mitigator disposed to prevent or reduce bright light from being provided through the combiner. The bright light can be direct lunar light, direct solar light, or solar/lunar reflections. The mitigator dynamically selects and reduces bright background light which results in an increase of contrast ratio.
Description
CROSS REFERENCE TO RELATED APPLICATIONS

The present application is related to: U.S. patent application Ser. No. 13/892,057, filed on May 10, 2013, by Stratton et al., and U.S. patent application Ser. No. 13/864,991, filed on Apr. 17, 2013 by Tiana et al., incorporated herein by reference herein in their entireties and assigned to the assignee of the present application.


BACKGROUND OF THE INVENTION

The present specification relates to displays. More particularly, the present specification relates to head up displays (HUDs).


Conventional HUDs utilize an image source and a combiner to provide an image (e.g., symbology, enhanced vision, etc.) to a user. The user (e.g., a pilot) can view the image on the combiner as well as an environmental scene through the combiner. Substrate guided HUDs have been proposed which use waveguide combiners to preserve eye box size while reducing lens size. U.S. Pat. No. 4,309,070 issued to St. Leger Searle and U.S. Pat. No. 4,711,512 issued to Upatnieks disclose waveguide HUDs.


HUDs can be susceptible to a loss of contrast ratio in certain environmental conditions. The loss of contrast ratio can be due to bright light in the environment often associated with direct lunar light, direct solar light, lunar reflection or solar reflection. The bright light can strike the combiner or other optical components and dominate the background and/or image or portions thereof. In substrate guided HUDs, the loss of contrast ratio can also be due to interference when the diffractive gratings in substrate waveguides are exposed to bright light.


The susceptibility to loss of contrast ratio is increased in glass cockpit environments, such as, those associated with tactical aircraft. For example, the bright light (e.g., from the sun, bright clouds, the moon and reflections) can directly strike the combiner or other optical components through the windshield, glass canopy, windows, etc. and overcome or saturate the background and/or image or portions thereof. The saturation can occur at specific points or areas on the combiner and does not necessarily affect the entire display. Conventional uses of tinted lens and visors can reduce background saturation, but do not adequately solve lower contrast ratio issues.


Accordingly, there is a need for a HUD that is less susceptible to lower contrast ratios caused by non-uniform bright light from the environment. Further still, there is a need for a HUD having a waveguide combiner that is less susceptible to solar and lunar specular reflection. There is also a need for a system and method of eliminating or mitigating image quality degradation due to bright light in the environment. Further still, there is a need for a combiner which can mitigate the effects of refractive saturation on image quality of a waveguide display. Further still, there is a need for an inexpensive, lightweight solution for preventing or reducing solar or lunar light saturation of the background scene.


SUMMARY OF THE INVENTION

An exemplary embodiment relates to a head up display (HUD) for use in an environment and for providing light from an image source. The head up display includes a sensor array, a combiner, and a dynamic gradient. The combiner can be configured to allow viewing of an outside scene and information from the image source. The dynamic gradient is disposed to selectably reduce or prevent transmission through the combiner of bright light from the environment observed by the sensor array.


Another exemplary embodiment relates to a method of providing information to a pilot in an environment. The method includes providing light associated with the information to a combiner, sensing a location of a bright light from the environment on the combiner, and attenuating the bright light reaching the combiner at the location.


Still another exemplary embodiment relates to a head up display. The head up display includes a saturation mitigator and a waveguide combiner including an input grating and an output grating. The saturation mitigator is disposed to prevent bright light from outside the waveguide combiner from being provided through the output grating and reaching an eye box. The saturation mitigator is selectably opaque of at least one area associated with a location of the bright light.


A further embodiment relates to a HUD. The HUD can include an image source, a sensor array, a combiner, and a dynamic gradient. The combiner is configured to allow viewing of an outside scene and information from the image source. The dynamic gradient is disposed to selectably reduce or prevent transmission of one or more bright lights sensed by the sensor array.





BRIEF DESCRIPTION OF THE DRAWINGS

Exemplary embodiments are hereafter described with reference to the accompanying drawings, wherein like numerals denote like elements; and:



FIG. 1 is a general schematic block diagram of a head up display (HUD) system including reflective or holographic combiner with a bright light mitigator in accordance with an exemplary embodiment;



FIG. 2 is a general schematic block diagram of a HUD system including a waveguide combiner with a bright light mitigator in accordance with another exemplary embodiment;



FIG. 3 is a general schematic block diagram of the waveguide combiner and bright light mitigator for the head up display system illustrated in FIG. 2 showing bright light striking the waveguide combiner in accordance with another exemplary embodiment;



FIG. 4 is side view schematic drawing of yet another embodiment of the HUD system illustrated in FIG. 2 in accordance with another exemplary embodiment; and



FIG. 5 is a flow diagram showing operation of the HUD systems illustrated in FIGS. 1 and 2, according to another exemplary embodiment.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Before describing in detail the particular improved system and method, it should be observed that the invention includes, but is not limited to, a novel structural combination of optical components and not in the particular detailed configurations thereof. Accordingly, the structure, methods, functions, control and arrangement of components have been illustrated in the drawings by readily understandable block representations and schematic drawings, in order not to obscure the disclosure with structural details which will be readily apparent to those skilled in the art, having the benefit of the description herein. Further, the invention is not limited to the particular embodiments depicted in the exemplary diagrams, but should be construed in accordance with the language in the claims.


With reference to FIG. 1, a head up display (HUD) system 10 can be utilized in various applications, including but not limited to aviation, medical, naval, targeting, ground based, military, etc. HUD system 10 advantageously is configured to reduce loss of contrast ratio due to bright light in the environment in one embodiment. System 10 includes a control or processing circuit 47, a saturation mitigator or dynamic gradient 92, a bright light sensor 96, an image source 20, optics 30, and a combiner 40 in one embodiment. System 10 can also include an optional eye sensor 45 and an INS 99 or other navigational sensor in one embodiment.


In operation, system 10 provides images from image source 20 to a pilot or other operator so that the pilot can simultaneously view the images and a real world scene. The images can include graphic and/or text information (e.g., flight path vector, target icons, etc.) related to avionic information in one embodiment. In addition, the images can include synthetic or enhanced vision images. In one embodiment, collimated light representing the image from image source 20 is projected onto combiner 40 by optics 30 so that the pilot can view the image conformally on the real world scene through combiner 40. In one embodiment, combiner 40 is preferably transparent for viewing the real world scene through main surfaces or sides 84 and 88.


System 10 advantageously reduces low contrast ratio issues related to bright light in the environment by sensing the location of the bright light in the environment with respect to eye position and attenuating the bright light using gradient 92 to reduce its effect on the image provided on combiner 40 in one embodiment. In one embodiment, processing circuit 47 dynamically controls areas on gradient 92 so that the areas associated with the location of the sensed bright light are more opaque (e.g., less transmissive in one embodiment) than the remainder of gradient 92. The processing circuit 47 can use gaze information or pupil location data from sensor 45 to determine the appropriate areas to make more opaque.


Combiner 40 can be a flat or curved, semi-reflective or holographic combiner. Combiner 40 can be disposed in a fixed location in the cockpit or can be a worn component. Image source 20 can be any device for providing an image including but not limited to a CRT display, an LED display, an active matrix liquid crystal display (LCD), etc. Image source 20 can be located in front of or behind combiner 40 with respect to the windshield in one embodiment.


Collimating optics 30 can be a projector disposed between combiner 40 and image source 20 in one embodiment. Collimating optics 30 can be a single optical component, such as a lens, or include multiple optical components. In one embodiment, collimating optics 30 are configured as a catadioptric collimator. In one embodiment, optics 30 together with an optically powered combiner 40 cooperate to provide the collimated light to the user. Collimating optics 30 can be any optical component or configuration of optical components that provide light (preferably collimated light) from image source 20 to combiner 40. Collimating optics 30 can be integrated with or spaced apart from image source 20 and/or combiner 40. Optics 30 can be an overhead projector.


Sensor 45 can be any device for determining head position or eye position. In one embodiment, sensor 45 can utilize infrared or other light to detect the area where the pupil of an eye of the user is looking. Sensor 45 can be part of a head tracking system in one embodiment. Sensor 45 can be an inward facing eye tracker in one embodiment. Sensor 45 can be a pupil sensor that provides accurate eye box position information in one embodiment. Sensor 45 can provide a gaze information signal or data to processing circuit 45.


Sensor 96 can be any device for determining a location of bright light in the environment. In one embodiment, sensor 96 can be a two or three axis photovoltaic sensor. In one embodiment, sensor 96 can be a visible light camera. The location of the bright light can be determined in two dimensions by sensor 96. Sensor 96 can provide a bright light location signal or data to processing circuit 47. The signal or data can provide an indication of the location of one or more bright lights in the environment in one embodiment. Sensor 96 can include an integrated circuit array of photovoltaic sensors to sense a relative position of the brightest points of light in the field-of-view of system 10 in one embodiment. Sensor 96 can include a processing circuit to filter, condition and/or process the data for processing circuit 47.


INS 99 can be any device for providing navigation information. In one embodiment, INS 99 can provide velocity and acceleration information associated with system 10 (e.g., associated with the aircraft to processing circuit 47). INS 99 can also provide positional information (e.g., heading, attitude, altitude, longitude, latitude, pitch, roll, etc.) to processing circuit 47 in one embodiment. Alternatively, other sensors or computers can provide navigation information to processing circuit 47.


Processing circuit 47 can be a computer (e.g., part of a HUD computer, flight computer, display computer, navigation computer, display computer, etc.) in one embodiment. Circuit 47 can be a standalone circuit or be integrated with other line replaceable units or computers. Processing circuit 47 can control gradient 92 in response to signals or data from one or more of sensor 45, sensor 96, and INS 99. In one embodiment, processing circuit 47 executes software to control gradient 92 to attenuate bright light virtually displayed on combiner 40.


The software executed on processing circuit 47 can utilize the data from sensor 45 and sensor array 96 to perform a process (e.g., process 700 described in FIG. 5) to track and locate bright lights and control gradient 92 in accordance with the tracking and locating of the bright lights in one embodiment. Tracking software (e.g., ray tracing software) executed by processing circuit 47 determines the perceived location of the bright light on combiner 40 with respect to eye position. In addition, the tracking software can use velocity and positional data from INS 99 to predict future perceived locations of the bright light and control gradient 92 accordingly in one embodiment.


Software executed by processing circuit 47 can triangulate known positions of the sun or moon, the eye, and combiner 40 to provide map data. The map data can indicate bright light position relative to combiner 40. The bright light position can be a solar, lunar, or bright light position or related specular position as perceived by the pilot in one embodiment. Processing circuit 47 can provide a control signal or control data to dynamic gradient 92 so that dynamic gradient 92 will block or attenuate the area associated with the bright light striking combiner 40. In one embodiment, an angular area is blocked (e.g., an angular area of 0.5 degrees).


Gradient 92 can be any device for selectably blocking or attenuating bright light in response to a control signal or control data from circuit 47 in one embodiment. According to one embodiment, gradient 92 is a dynamic gradient for attenuating bright light from the environment and preventing such light from affecting the view of information on combiner 40. In one embodiment, gradient 92 can be a layer or layers that can attenuate bright light from the environment provided through combiner 40. Gradient 92 can be provided between combiner 40 and a windshield or other entry area for environmental bright light.


In one embodiment, gradient 92 is a liquid crystal or OLED film having pixels that can be set to levels of opaqueness. In one embodiment, gradient 92 can be an active matrix LCD or film. In another embodiment, the dynamic gradient is a dynamic neutral density (ND) filter. A neutral density filter can selectably have areas with a rating of 0 (no reduction of intensity) and 1 (a reduction of intensity by a factor of 10), with a rating of 0 and 2, 0 and 3, etc. Alternatively, a neutral filter can have areas of 1 or 2, 1 or 3, etc. In one embodiment, gradient 92 is a layer or medium that includes dynamically controlled polarized layers. Gradient 92 can be disposed directly on combiner 40 or spaced apart from combiner 40.


Gradient 92 (e.g. dynamic filter) filter can have areas which are selectable for providing attenuation so that only an exclusion area associated with the area where bright light would strike combiner 40 is opaque. The area is preferably in a shape corresponding to the shape of the light and is approximately a percentage level larger to prevent corona effects (e.g., 10% larger). Accordingly, bright outside light is not allowed to pass and yet projected symbology/images are unaffected on combiner 40.


The term bright light refers to a light having sufficient intensity to adversely affect the image quality (e.g., contrast ratio) provided by system 10. For example, the reduction of solar intensity will greatly increase contrast ratio and allow the pilot to continue to use the entire field-of-view of the HUD without loss of capability in daylight and moonlight operating conditions in one embodiment.


Given the high velocities in flight, a target on combiner 40 will not likely hide in a particular location due to specular point or solar blockage for a very long time. Targets will most likely pass through the areas allowing the pilot to see the target as it passes through in one embodiment. Accordingly, matching the area of the exclusion (e.g., opaque) area to the size of the bright light provides advantages.


With reference to FIG. 2, a head up display (HUD) system 100, similar to system 10, can be utilized in various applications, including but not limited to aviation, medical, naval, targeting, ground based, military, etc. In one embodiment, HUD system 100 is configured for use in smaller cockpit environments and yet provides an appropriate field of view and eye box for avionic applications. In one embodiment, HUD system 100 is configured for larger cockpit environments or tactical cockpit environments.


HUD system 100 includes INS 199 (similar to INS 99), gradient 192 (similar to gradient 92), collimating optics 130, eye sensor 145 (similar to sensor 45), a sensor 196 (similar to sensor 96), a processing circuit 147 (similar to circuit 47), an image source 120 and a substrate waveguide combiner 140 in one embodiment. Image source 120 can be any device for providing an image including but not limited to a CRT display, an LED display, an active matrix liquid crystal display (LCD), a light emitting diode, laser illuminator, etc. In one embodiment, image source 120 is a micro LCD assembly or liquid crystal on silicon (LCOS) display and can provide linearly polarized light. Image source 120 can include a laser or LED backlight.


In operation, HUD system 110 provides images from image source 120 to a pilot or other operator so that the pilot can simultaneously view the images and a real world scene. The images can include graphic and/or text information (e.g., flight path vector, target icons, etc.) related to avionic information in one embodiment. In addition, the images can include synthetic or enhanced vision images. In one embodiment, collimated light representing the image from image source 120 is provided on substrate waveguide combiner 140 so that the pilot can view the image conformally on the real world scene through substrate waveguide combiner 140. Similar to system 10 discussed with reference to FIG. 1, system 100 includes dynamic gradient 192, sensor 196, eye sensor 195, processing circuit 147, and INS 199 which can advantageously mitigate the effects of bright light from the environment striking combiner 140. INS 99, gradient 192, eye sensor 145, a sensor 196, and a processing circuit 147 can advantageously be used to mitigate contrast ratio reduction due to environmental bright light (e.g., as described below with reference to FIG. 5) according to one embodiment.


Collimating optics 130 can be disposed between substrate waveguide combiner 140 and image source 120. Collimating optics 130 can be a single optical component, such as a lens, or include multiple optical components. In one embodiment, collimating optics 130 are configured as a catadioptric collimator. Collimating optics 130 can be any optical component or configuration of optical components that provide light (preferably collimated light) from image source 120 to substrate waveguide combiner 140. Collimating optics 130 can be integrated with or spaced apart from image source 120 and/or substrate waveguide combiner 140.


Substrate waveguide combiner 140 can be a single glass plate or can be made from two or more fixed glass plates. Substrate waveguide combiner 140 can have a variety of shapes including generally rectangular, oval, circular, tear drop-shaped, hexagonal, rectangular with rounded corners, square-shaped, etc. Combiner 140 can include a coupler 142 and a coupler 144.


Couplers 142 and 144 can be input and output coupling gratings that provide excellent image quality and acceptable brightness in one embodiment. Couplers 142 and 144 are implemented as any type of diffractive element (e.g., dichromated gratings, holographic, Bragg gratings, or blazed surface relief gratings) in one embodiment. Couplers 142 and 144 can be implemented according to a number of techniques including but not limited to mechanical reproduction, holographic formation, embossing, casting (e.g., into a polymer resin), or lithography.


In operation, substrate waveguide combiner 140 advantageously receives light from image source 120 provided through collimating optics 130 at an input 172 and provides light to a user at its output 174. Image source 120 provides information using a single color of light (e.g., green light approximately between 500 and 550 nanometers (nm)) in one embodiment. Light provided to substrate waveguide 40 is preferably linearly S-polarized or P-polarized and collimated in one embodiment. Alternatively, other polarization, multiple colors, or other colors at different wavelengths can be utilized without departing from the scope of the invention. Optics 130 can have an output disposed directly adjacent or attached to coupler 142 in one embodiment.


Substrate waveguide combiner 140 performs two operations in one preferred embodiment. First, substrate waveguide combiner 140 is disposed to provide a medium for transporting light by total internal reflection from input 172 to output 174. Light is reflected multiple times off of opposing main sides 184 and 188 of combiner 140 as it travels from input 172 to output 174. Second, substrate waveguide combiner 140 operates as a combiner allowing the user to view the light from image source 120 at output 174 and light from the real world scene through sides 184 and 188.


Light from collimating optics 130 first strikes coupler 142 at input 172 on side 184 of substrate waveguide combiner 140. Coupler 142 diffracts light toward the length of combiner 140 so that it travels by total internal reflection to output 174 on side 184. At output 174, coupler 144 diffracts the light toward the user and out of the substrate waveguide combiner 140.


Couplers 142 and 144 are disposed on respective opposing sides 184 and 188 of substrate waveguide 140 in one embodiment. Couplers 142 and 144 can also be formed on the same side 184 of or within waveguide combiner 140 in one alternative embodiment. Couplers 142 and 144 can be disposed respectively on sides 188 and 184 of waveguide combiner 140, within waveguide combiner 140 or both on the same side 188 of waveguide combiner 140 in other alternative embodiments.


Couplers 142 and 144 are preferably disposed in respective areas that are rectangular in shape and have the same width as each other in one embodiment. Alternatively, couplers 142 and 44 can have different widths. Coupler 144 has a greater height than coupler 142 in one embodiment. Without the operation of processing circuit 47 and gradient 192, combiner 140 may be susceptible to refractive saturation through gradients due to coupler 144 diffracting light from the environment into the eye box.


Gradient 192 can be provided as a layer over a side 184 of combiner 140 in one embodiment. Gradient 92 can be provided over an area associated with coupler 144 in one embodiment. System parameters and design criteria can affect the placement of gradient 192.


Systems 10 and 100 can utilize partial and dynamic exclusion of saturated areas according to processes similar to machine vision systems (e.g., earth observing satellites, helio dynamic observing satellites, etc.) Advantageously, systems 10 and 100 can also take into account the placement of the human vision to provide more precise placement of saturation areas.


With reference to FIG. 3, substrate waveguide combiner 140 is shown wherein solar light from sun 102 in the environment is virtually provided through combiner 140 into an eye box 104 at a virtual location 108. Processing circuit 147 utilizing data from eye sensor 45 and sensor 196 can advantageously attenuate light at the area associated with virtual image 108 of sun 102.


With reference to FIGS. 1, 2, 3, and 5, processing circuit 47 and 147 can perform process 700 (FIG. 5) to attenuate bright light through respective combiners 40 and 140. Process 700 is described below with respect to system 100. Process 700 can be performed by system 10 or 600 without departing from exemplary embodiments disclosed herein.


At an operation 702, processing circuit 147 receives sensor data from sensor 196. The data from sensor 196 indicates the position of the sun 102 (or other bright light) with respect to the position of combiner 140 (or the aircraft on which system 100 is employed).


Processing circuit 147 can sense the brightest portions in the field of view of sensor 196 and send telemetry data to tracking software associated with circuit 147. The brightest portions can be determined by comparing to a threshold. The threshold can be a fixed value or related to an average value (e.g., more than 20%, 40%, or 60% higher than the average).


At an operation 704, processing circuit 147 receives eye tracking data associated with eye sensor 145. Eye sensor 147 senses the position of the pupil and provides data to software 147 so that tracking can occur.


At an operation 706, processing circuit 147 uses the sensor telemetry data from sensor 196 to determine where the bright light source (e.g., the sun, moon, bright cloud, etc.) is with respect to the aircraft. Processing circuit 147 can utilize data from INS 199 or other sensors to receive velocity and acceleration data (velocity and acceleration deltas). Triangulations with known positions of the sun, the eye and combiner 140 are utilized to provide a map relative to combiner 140 with the sun position (and other specular positions as necessary) as perceived by the pilot.


Processing circuit 147 can use a Kalman filter to translate into pixel space on the combiner 140 in one embodiment. In one embodiment, gradient 192 has a 1000×1000 pixels or more. Alternatively, a lower number of pixels can be utilized. Processing circuit 147 provides signals or data to gradient 192 to block or otherwise attenuate the area associated with the sun 102 at an operation 708.


With reference to FIG. 4, HUD system 600 includes a dynamic gradient 692 provided on a waveguide combiner 640. In one embodiment, gradient 692 is similar to gradient 192. Gradient 692 can be spaced apart from a surface 684. In one embodiment, gradient 692 can be physically separated and parallel to surface 684. System 600 can include the components similar to circuit 147, sensors 145 and 196 and INS 99 of system 100 for controlling gradient 692 to mitigate bright light saturation.


HUD system 600 is comprised of a projector 602. Projector 602 is provided beneath a top surface glare shield 610 in one embodiment. Substrate waveguide combiner 640 can be disposed through a glare shield 610 such that the input end of substrate waveguide 640 is beneath glare shield.


Projector 602 can include image source 620 and collimating optics 630. Image source 620 can include a backlight 612 which can be an LED backlight in one embodiment. Image source 620 can also include display 614 which can be an active matrix LCD, although other types of image sources 620 are available according to alternative embodiments.


In one embodiment, projector 602 is an LCOS based system including a beam splitter 652, an LED or laser illuminator 650, and an LCOS display in the position of display 614. An image shaping lens 654 is provided between illuminator 650 and beam splitter 652. Beam splitter 652 reflects polarized (e.g., s-polarized) light and provides the light to display 614 which selectively rotates the polarization to p-polarized light in one embodiment. The p-polarized light is provided through beam splitter 652 and collimating optics 630.


Collimating optics 630 includes a lens 615, a mirror 616 and a lens 618 in one embodiment. Lens 615, mirror 616 and lens 618 are configured to provide collimated light to couplers 642 and 644. Optics 630 can be embodied in various forms including more or fewer components. HUD system 600 can have a bent waveguide combiner configuration in one embodiment.


It is understood that while the detailed drawings, specific examples, material types, thicknesses, dimensions, and particular values given provide a preferred exemplary embodiment of the present invention, the preferred exemplary embodiment is for the purpose of illustration only. The method and apparatus of the invention is not limited to the precise details and conditions disclosed. For example, although specific types of optical component, dimensions and angles are mentioned, other components, dimensions and angles can be utilized. Various changes may be made to the details disclosed without departing from the spirit of the invention which is defined by the following claims.

Claims
  • 1. A head up display for use in a cockpit environment for providing light from an image source, the head up display comprising: a sensor array;a combiner configured to allow viewing of an outside scene and information from the image source, wherein the combiner comprises a waveguide have an input and an output, the light associated with the information traveling from the input to the output by total internal reflection within the waveguide, wherein bright light from the cockpit environment is virtually provided through the waveguide into an eye box at a virtual location;a processing circuit configured to determine the virtual location in response to data from the sensor array and provide a control signal; anda dynamic gradient disposed to selectably reduce or prevent transmission through the combiner of the bright light from the cockpit environment observed by the sensor array in response to the control signal, the dynamic gradient utilizing at least one selectably opaque area to reduce or prevent transmission of the bright light through the combiner, the at least one selectably opaque area corresponding to the virtual location of the bright light on the combiner, and wherein the at least one selectably opaque area is shaped to a correspond to a shape of the bright location at the virtual location on the combiner.
  • 2. The head up display of claim 1, wherein the dynamic gradient comprises a dynamic filter.
  • 3. The head up display of claim 1, wherein the dynamic gradient is disposed on a side of the combiner closer to a windshield than another side of the combiner.
  • 4. The head up display of claim 1, wherein the combiner comprises at least one diffractive element coupler.
  • 5. The head up display of claim 4, wherein the dynamic gradient covers an area associated with the diffractive element coupler.
  • 6. The head up display of claim 1, wherein the dynamic gradient creates an exclusion area on a pixel-by-pixel basis.
  • 7. The head up display of claim 1, further comprising an eye sensor.
  • 8. The head up display of claim 7, wherein the processing circuit is configured to receive aircraft position, velocity and acceleration parameters and correlate a first position of a user eye, and a second position of a bright light on the sensor array and uses the first and second position to select a location to reduce or prevent transmission of the bright light on the dynamic gradient.
  • 9. The head up display of claim 1, wherein the dynamic gradient comprises a liquid crystal material, organic light emitting diode material, or dynamic neutral density filter.
  • 10. The head up display of claim 1, wherein the processing circuit is configured to determine a future position of the bright light in response to the portion and velocity parameters and control the dynamic gradient accordingly.
  • 11. A method of providing information to a pilot in an aviation environment, the method comprising: providing light associated with the information to a combiner, wherein the combiner comprises a waveguide have an input and an output, the light associated with the information traveling from the input to the output by total internal reflection within the waveguide, wherein bright light in the cockpit environment is provided through the waveguide into an eye box at a virtual location;sensing a location of the bright light from the environment and determining the virtual location in response to data from a sensor array andattenuating the bright light at the virtual location by utilizing at least one selectably opaque area wherein the at least one selectably opaque area is shaped to correspond to a shape of the bright light at the virtual location.
  • 12. The method of claim 11, further comprising executing software to perform real time computational analysis to determine the location of both a source of the bright light and the position of the eye.
  • 13. The method of claim 11, wherein the determining uses an eye location, a velocity parameter of the aircraft, and a position parameter of the aircraft.
  • 14. The method of claim 13, wherein the determining use, the velocity parameter and the position parameter to determine a future location of the bright light in an environment.
  • 15. The method of claim 14, wherein the location of the bright light in the environment is triangulated to the location on the combiner using data from an eye sensor.
  • 16. The method of claim 11, wherein the attenuating uses an LCD or a dynamic filter.
  • 17. A head up display, comprising: a waveguide combiner including an input grating and an output grating, wherein light associated with information travels from the input grating from the output grating by total internal reflection within the waveguide combiner, wherein bright light in an environment of the cockpit is virtually provided through the waveguide combiner into an eye box at a virtual location; anda saturation mitigator disposed to prevent bright light from outside the waveguide combiner from being provided through the output grating and reaching the eye box, the saturation mitigator being selectably opaque in at least one area, the at least one area corresponding to the virtual location of the bright light, and wherein the at least one area is shaped to correspond to a shape of the bright light at the location.
  • 18. The head up display of claim 17, further comprising a bright light sensor.
  • 19. The head up display of claim 17, further comprising an eye sensor.
  • 20. The head up display of claim 17, further comprising a processor for locating the bright light and correlating the location of the bright light to the area.
US Referenced Citations (526)
Number Name Date Kind
2141884 Sonnefeld Dec 1938 A
3620601 Waghorn Nov 1971 A
3851303 Muller Nov 1974 A
3885095 Wolfson et al. May 1975 A
3940204 Withrington Feb 1976 A
4082432 Kirschner Apr 1978 A
4099841 Ellis Jul 1978 A
4178074 Heller Dec 1979 A
4218111 Withrington et al. Aug 1980 A
4232943 Rogers Nov 1980 A
4309070 St. Leger Searle Jan 1982 A
4647967 Kirschner et al. Mar 1987 A
4711512 Upatnieks Dec 1987 A
4714320 Banbury Dec 1987 A
4743083 Schimpe May 1988 A
4749256 Bell et al. Jun 1988 A
4775218 Wood et al. Oct 1988 A
4854688 Hayford et al. Aug 1989 A
4928301 Smoot May 1990 A
4946245 Chamberlin et al. Aug 1990 A
5007711 Wood et al. Apr 1991 A
5035734 Honkanen et al. Jul 1991 A
5076664 Migozzi Dec 1991 A
5079416 Filipovich Jan 1992 A
5117285 Nelson et al. May 1992 A
5124821 Antier et al. Jun 1992 A
5148302 Nagano et al. Sep 1992 A
5151958 Honkanen Sep 1992 A
5153751 Ishikawa et al. Oct 1992 A
5159445 Gitlin et al. Oct 1992 A
5160523 Honkanen et al. Nov 1992 A
5183545 Branca et al. Feb 1993 A
5187597 Kato et al. Feb 1993 A
5210624 Matsumoto et al. May 1993 A
5218360 Goetz et al. Jun 1993 A
5243413 Gitlin et al. Sep 1993 A
5289315 Makita et al. Feb 1994 A
5295208 Caulfield et al. Mar 1994 A
5303085 Rallison Apr 1994 A
5317405 Kuriki et al. May 1994 A
5341230 Smith Aug 1994 A
5351151 Levy Sep 1994 A
5359362 Lewis et al. Oct 1994 A
5363220 Kuwayama et al. Nov 1994 A
5369511 Amos Nov 1994 A
5400069 Braun et al. Mar 1995 A
5408346 Trissel et al. Apr 1995 A
5418584 Larson May 1995 A
5438357 McNelley Aug 1995 A
5455693 Wreede et al. Oct 1995 A
5471326 Hall et al. Nov 1995 A
5473222 Thoeny et al. Dec 1995 A
5496621 Makita et al. Mar 1996 A
5500671 Andersson et al. Mar 1996 A
5510913 Hashimoto et al. Apr 1996 A
5515184 Caulfield et al. May 1996 A
5524272 Podowski et al. Jun 1996 A
5532736 Kuriki et al. Jul 1996 A
5537232 Biles Jul 1996 A
5572248 Allen et al. Nov 1996 A
5579026 Tabata Nov 1996 A
5604611 Saburi et al. Feb 1997 A
5606433 Yin et al. Feb 1997 A
5612733 Flohr Mar 1997 A
5612734 Nelson et al. Mar 1997 A
5619254 McNelley Apr 1997 A
5629259 Akada et al. May 1997 A
5631107 Tarumi et al. May 1997 A
5633100 Mickish et al. May 1997 A
5646785 Gilboa et al. Jul 1997 A
5648857 Ando et al. Jul 1997 A
5661577 Jenkins et al. Aug 1997 A
5661603 Hanano et al. Aug 1997 A
5665494 Kawabata et al. Sep 1997 A
5668907 Veligdan Sep 1997 A
5682255 Friesem et al. Oct 1997 A
5694230 Welch Dec 1997 A
5701132 Kollin et al. Dec 1997 A
5706108 Ando et al. Jan 1998 A
5707925 Akada et al. Jan 1998 A
5724189 Ferrante Mar 1998 A
5726782 Kato et al. Mar 1998 A
5727098 Jacobson Mar 1998 A
5729242 Margerum et al. Mar 1998 A
5731060 Hirukawa et al. Mar 1998 A
5731853 Taketomi et al. Mar 1998 A
5742262 Tabata et al. Apr 1998 A
5751452 Tanaka et al. May 1998 A
5760931 Saburi et al. Jun 1998 A
5764414 King et al. Jun 1998 A
5790288 Jager et al. Aug 1998 A
5812608 Valimaki et al. Sep 1998 A
5822127 Chen et al. Oct 1998 A
5841507 Barnes Nov 1998 A
5856842 Tedesco Jan 1999 A
5868951 Schuck et al. Feb 1999 A
5892598 Asakawa et al. Apr 1999 A
5898511 Mizutani et al. Apr 1999 A
5903395 Rallison et al. May 1999 A
5907416 Hegg et al. May 1999 A
5907436 Perry et al. May 1999 A
5917459 Son et al. Jun 1999 A
5926147 Sehm et al. Jul 1999 A
5929946 Sharp et al. Jul 1999 A
5937115 Domash Aug 1999 A
5942157 Sutherland et al. Aug 1999 A
5945893 Plessky et al. Aug 1999 A
5949302 Sarkka Sep 1999 A
5966223 Friesem et al. Oct 1999 A
5985422 Krauter Nov 1999 A
5991087 Rallison Nov 1999 A
5999314 Asakura et al. Dec 1999 A
6042947 Asakura et al. Mar 2000 A
6043585 Plessky et al. Mar 2000 A
6075626 Mizutani et al. Jun 2000 A
6078427 Fontaine et al. Jun 2000 A
6115152 Popovich et al. Sep 2000 A
6127066 Ueda et al. Oct 2000 A
6137630 Tsou et al. Oct 2000 A
6169613 Amitai et al. Jan 2001 B1
6176837 Foxlin Jan 2001 B1
6195206 Yona et al. Feb 2001 B1
6222675 Mall et al. Apr 2001 B1
6222971 Veligdan et al. Apr 2001 B1
6249386 Yona et al. Jun 2001 B1
6259423 Tokito et al. Jul 2001 B1
6259559 Kobayashi et al. Jul 2001 B1
6317083 Johnson et al. Nov 2001 B1
6317227 Mizutani et al. Nov 2001 B1
6321069 Piirainen Nov 2001 B1
6327089 Hosaki et al. Dec 2001 B1
6333819 Svedenkrans Dec 2001 B1
6340540 Ueda et al. Jan 2002 B1
6351333 Araki et al. Feb 2002 B2
6356172 Koivisto et al. Mar 2002 B1
6359730 Tervonen Mar 2002 B2
6359737 Stringfellow Mar 2002 B1
6366378 Tervonen et al. Apr 2002 B1
6392812 Howard May 2002 B1
6409687 Foxlin Jun 2002 B1
6470132 Nousiainen et al. Oct 2002 B1
6486997 Bruzzone et al. Nov 2002 B1
6504518 Kuwayama et al. Jan 2003 B1
6524771 Maeda et al. Feb 2003 B2
6545778 Ono et al. Apr 2003 B2
6550949 Bauer et al. Apr 2003 B1
6557413 Nieminen et al. May 2003 B2
6563648 Gleckman et al. May 2003 B2
6580529 Amitai et al. Jun 2003 B1
6583873 Goncharov et al. Jun 2003 B1
6587619 Kinoshita Jul 2003 B1
6598987 Parikka Jul 2003 B1
6608720 Freeman Aug 2003 B1
6611253 Cohen Aug 2003 B1
6646810 Harter et al. Nov 2003 B2
6661578 Hedrick Dec 2003 B2
6674578 Sugiyama et al. Jan 2004 B2
6686815 Mirshekarl-Syahkal et al. Feb 2004 B1
6721096 Bruzzone et al. Apr 2004 B2
6741189 Gibbons, II et al. May 2004 B1
6744478 Asakura et al. Jun 2004 B1
6748342 Dickhaus Jun 2004 B1
6750941 Satoh et al. Jun 2004 B2
6757105 Niv et al. Jun 2004 B2
6771403 Endo et al. Aug 2004 B1
6776339 Piikivi Aug 2004 B2
6781701 Sweetser et al. Aug 2004 B1
6805490 Levola Oct 2004 B2
6825987 Repetto et al. Nov 2004 B2
6829095 Amitai Dec 2004 B2
6833955 Niv Dec 2004 B2
6836369 Fujikawa et al. Dec 2004 B2
6844212 Bond et al. Jan 2005 B2
6844980 He et al. Jan 2005 B2
6847274 Salmela et al. Jan 2005 B2
6847488 Travis Jan 2005 B2
6853491 Ruhle et al. Feb 2005 B1
6864861 Schehrer et al. Mar 2005 B2
6864927 Cathey Mar 2005 B1
6885483 Takada Apr 2005 B2
6903872 Schrader Jun 2005 B2
6909345 Salmela et al. Jun 2005 B1
6917375 Akada et al. Jul 2005 B2
6922267 Endo et al. Jul 2005 B2
6926429 Barlow et al. Aug 2005 B2
6940361 Jokio et al. Sep 2005 B1
6950173 Sutherland et al. Sep 2005 B1
6950227 Schrader Sep 2005 B2
6951393 Koide Oct 2005 B2
6952312 Weber et al. Oct 2005 B2
6958662 Salmela et al. Oct 2005 B1
6987908 Bond et al. Jan 2006 B2
7003187 Frick et al. Feb 2006 B2
7018744 Otaki et al. Mar 2006 B2
7021777 Amitai Apr 2006 B2
7026892 Kajiya Apr 2006 B2
7027671 Huck et al. Apr 2006 B2
7034748 Kajiya Apr 2006 B2
7053735 Salmela et al. May 2006 B2
7058434 Wang et al. Jun 2006 B2
7095562 Peng et al. Aug 2006 B1
7101048 Travis Sep 2006 B2
7110184 Yona et al. Sep 2006 B1
7123418 Weber et al. Oct 2006 B2
7126418 Hunton et al. Oct 2006 B2
7126583 Breed Oct 2006 B1
7132200 Ueda et al. Nov 2006 B1
7149385 Parikka et al. Dec 2006 B2
7151246 Fein et al. Dec 2006 B2
7158095 Jenson et al. Jan 2007 B2
7181105 Teramura et al. Feb 2007 B2
7181108 Levola Feb 2007 B2
7184615 Levola Feb 2007 B2
7190849 Katase Mar 2007 B2
7199934 Yamasaki Apr 2007 B2
7205960 David Apr 2007 B2
7205964 Yokoyama et al. Apr 2007 B1
7206107 Levola Apr 2007 B2
7230767 Walck et al. Jun 2007 B2
7242527 Spitzer et al. Jul 2007 B2
7248128 Mattila et al. Jul 2007 B2
7259906 Islam Aug 2007 B1
7268946 Wang Sep 2007 B2
7285903 Cull et al. Oct 2007 B2
7289069 Ranta Oct 2007 B2
7299983 Piikivi Nov 2007 B2
7313291 Okhotnikov et al. Dec 2007 B2
7319573 Nishiyama Jan 2008 B2
7320534 Sugikawa et al. Jan 2008 B2
7323275 Otaki et al. Jan 2008 B2
7336271 Ozeki et al. Feb 2008 B2
7339737 Urey et al. Mar 2008 B2
7339742 Amitai et al. Mar 2008 B2
7375870 Schorpp May 2008 B2
7391573 Amitai Jun 2008 B2
7394865 Borran et al. Jul 2008 B2
7395181 Foxlin Jul 2008 B2
7397606 Peng et al. Jul 2008 B1
7401920 Kranz et al. Jul 2008 B1
7404644 Evans et al. Jul 2008 B2
7410286 Travis Aug 2008 B2
7411637 Weiss Aug 2008 B2
7415173 Kassamakov et al. Aug 2008 B2
7418170 Mukawa et al. Aug 2008 B2
7433116 Islam Oct 2008 B1
7436568 Kuykendall, Jr. Oct 2008 B1
7454103 Parriaux Nov 2008 B2
7457040 Amitai Nov 2008 B2
7466994 Pihlaja et al. Dec 2008 B2
7479354 Ueda et al. Jan 2009 B2
7480215 Makela et al. Jan 2009 B2
7482996 Larson et al. Jan 2009 B2
7483604 Levola Jan 2009 B2
7492512 Niv et al. Feb 2009 B2
7496293 Shamir et al. Feb 2009 B2
7500104 Goland Mar 2009 B2
7528385 Volodin et al. May 2009 B2
7545429 Travis Jun 2009 B2
7550234 Otaki et al. Jun 2009 B2
7567372 Schorpp Jul 2009 B2
7570429 Maliah et al. Aug 2009 B2
7572555 Takizawa et al. Aug 2009 B2
7573640 Nivon et al. Aug 2009 B2
7576916 Amitai Aug 2009 B2
7577326 Amitai Aug 2009 B2
7579119 Ueda et al. Aug 2009 B2
7588863 Takizawa et al. Sep 2009 B2
7589900 Powell Sep 2009 B1
7589901 DeJong et al. Sep 2009 B2
7592988 Katase Sep 2009 B2
7593575 Houle et al. Sep 2009 B2
7597447 Larson et al. Oct 2009 B2
7599012 Nakamura et al. Oct 2009 B2
7600893 Laino et al. Oct 2009 B2
7602552 Blumenfeld Oct 2009 B1
7616270 Hirabayashi et al. Nov 2009 B2
7618750 Ueda et al. Nov 2009 B2
7629086 Otaki et al. Dec 2009 B2
7639911 Lee et al. Dec 2009 B2
7643214 Amitai Jan 2010 B2
7660047 Travis et al. Feb 2010 B1
7672055 Amitai Mar 2010 B2
7710654 Ashkenazi et al. May 2010 B2
7724441 Amitai May 2010 B2
7724442 Amitai May 2010 B2
7724443 Amitai May 2010 B2
7733572 Brown et al. Jun 2010 B1
7747113 Mukawa et al. Jun 2010 B2
7751122 Amitai Jul 2010 B2
7764413 Levola Jul 2010 B2
7777819 Simmonds Aug 2010 B2
7778305 Parriaux et al. Aug 2010 B2
7778508 Hirayama Aug 2010 B2
7847235 Krupkin et al. Dec 2010 B2
7864427 Korenaga et al. Jan 2011 B2
7865080 Hecker et al. Jan 2011 B2
7872804 Moon et al. Jan 2011 B2
7884985 Amitai et al. Feb 2011 B2
7887186 Watanabe Feb 2011 B2
7903921 Ostergard Mar 2011 B2
7907342 Simmonds et al. Mar 2011 B2
7920787 Gentner et al. Apr 2011 B2
7944428 Travis May 2011 B2
7969644 Tilleman et al. Jun 2011 B2
7970246 Travis et al. Jun 2011 B2
7976208 Travis Jul 2011 B2
7999982 Endo et al. Aug 2011 B2
8000491 Brodkin et al. Aug 2011 B2
8004765 Amitai Aug 2011 B2
8016475 Travis Sep 2011 B2
8022942 Bathiche et al. Sep 2011 B2
RE42992 David Dec 2011 E
8079713 Ashkenazi Dec 2011 B2
8082222 Rangarajan et al. Dec 2011 B2
8086030 Gordon et al. Dec 2011 B2
8089568 Brown et al. Jan 2012 B1
8107023 Simmonds et al. Jan 2012 B2
8107780 Simmonds Jan 2012 B2
8132948 Owen et al. Mar 2012 B2
8132976 Odell et al. Mar 2012 B2
8136690 Fang et al. Mar 2012 B2
8137981 Andrew et al. Mar 2012 B2
8149086 Klein et al. Apr 2012 B2
8152315 Travis et al. Apr 2012 B2
8155489 Saarikko et al. Apr 2012 B2
8160409 Large Apr 2012 B2
8160411 Levola et al. Apr 2012 B2
8186874 Sinbar et al. May 2012 B2
8188925 Dejean May 2012 B2
8189263 Wang et al. May 2012 B1
8189973 Travis et al. May 2012 B2
8199803 Hauske et al. Jun 2012 B2
8213065 Mukawa Jul 2012 B2
8233204 Robbins et al. Jul 2012 B1
8253914 Kajiya et al. Aug 2012 B2
8254031 Levola Aug 2012 B2
8295710 Marcus Oct 2012 B2
8301031 Gentner et al. Oct 2012 B2
8305577 Kivioja et al. Nov 2012 B2
8306423 Gottwald et al. Nov 2012 B2
8314819 Kimmel et al. Nov 2012 B2
8321810 Heintze Nov 2012 B2
8335040 Mukawa et al. Dec 2012 B2
8351744 Travis et al. Jan 2013 B2
8354806 Travis et al. Jan 2013 B2
8355610 Simmonds Jan 2013 B2
8369019 Baker et al. Feb 2013 B2
8384694 Powell et al. Feb 2013 B2
8398242 Yamamoto et al. Mar 2013 B2
8403490 Sugiyama et al. Mar 2013 B2
8422840 Large Apr 2013 B2
8427439 Larsen et al. Apr 2013 B2
8432363 Saarikko et al. Apr 2013 B2
8432372 Butler et al. Apr 2013 B2
8472119 Kelly Jun 2013 B1
8477261 Travis et al. Jul 2013 B2
8491121 Tilleman et al. Jul 2013 B2
8491136 Travis et al. Jul 2013 B2
8493366 Bathiche et al. Jul 2013 B2
8493662 Noui Jul 2013 B2
8508848 Saarikko Aug 2013 B2
8547638 Levola Oct 2013 B2
8578038 Kaikuranta et al. Nov 2013 B2
8581831 Travis Nov 2013 B2
8582206 Travis Nov 2013 B2
8593734 Laakkonen Nov 2013 B2
8611014 Valera et al. Dec 2013 B2
8619062 Powell et al. Dec 2013 B2
8633786 Ermolov et al. Jan 2014 B2
8639072 Popovich et al. Jan 2014 B2
8643691 Rosenfeld et al. Feb 2014 B2
8649099 Schultz et al. Feb 2014 B2
8654420 Simmonds Feb 2014 B2
8670029 McEldowney Mar 2014 B2
8693087 Nowatzyk et al. Apr 2014 B2
8736802 Kajiya et al. May 2014 B2
8736963 Robbins et al. May 2014 B2
8767294 Chen et al. Jul 2014 B2
8810600 Bohn et al. Aug 2014 B2
8814691 Haddick et al. Aug 2014 B2
8830584 Saarikko et al. Sep 2014 B2
8913324 Schrader Dec 2014 B2
8938141 Magnusson Jan 2015 B2
20020012064 Yamaguchi Jan 2002 A1
20020021461 Ono et al. Feb 2002 A1
20020131175 Yagi et al. Sep 2002 A1
20030039442 Bond et al. Feb 2003 A1
20030063042 Friesem et al. Apr 2003 A1
20030149346 Arnone et al. Aug 2003 A1
20030228019 Eichler et al. Dec 2003 A1
20040089842 Sutherland et al. May 2004 A1
20040188617 Devitt et al. Sep 2004 A1
20040208446 Bond et al. Oct 2004 A1
20040208466 Mossberg et al. Oct 2004 A1
20050135747 Greiner et al. Jun 2005 A1
20050136260 Garcia Jun 2005 A1
20050259302 Metz et al. Nov 2005 A9
20050269481 David et al. Dec 2005 A1
20060093793 Miyakawa et al. May 2006 A1
20060114564 Sutherland et al. Jun 2006 A1
20060119916 Sutherland et al. Jun 2006 A1
20060132914 Weiss et al. Jun 2006 A1
20060221448 Nivon et al. Oct 2006 A1
20060228073 Mukawa et al. Oct 2006 A1
20060279662 Kapellner et al. Dec 2006 A1
20070019152 Caputo et al. Jan 2007 A1
20070019297 Stewart et al. Jan 2007 A1
20070041684 Popovich et al. Feb 2007 A1
20070045596 King et al. Mar 2007 A1
20070089625 Grinberg et al. Apr 2007 A1
20070133920 Lee et al. Jun 2007 A1
20070133983 Traff Jun 2007 A1
20070188837 Shimizu et al. Aug 2007 A1
20070211164 Olsen et al. Sep 2007 A1
20080043334 Itzkovitch et al. Feb 2008 A1
20080106775 Amitai et al. May 2008 A1
20080136923 Inbar et al. Jun 2008 A1
20080151379 Amitai Jun 2008 A1
20080186604 Amitai Aug 2008 A1
20080198471 Amitai Aug 2008 A1
20080278812 Amitai Nov 2008 A1
20080285140 Amitai Nov 2008 A1
20080309586 Vitale Dec 2008 A1
20090017424 Yoeli et al. Jan 2009 A1
20090019222 Verma et al. Jan 2009 A1
20090052046 Amitai Feb 2009 A1
20090052047 Amitai Feb 2009 A1
20090067774 Magnusson Mar 2009 A1
20090097122 Niv Apr 2009 A1
20090097127 Amitai Apr 2009 A1
20090121301 Chang May 2009 A1
20090122413 Hoffman et al. May 2009 A1
20090122414 Amitai May 2009 A1
20090128902 Niv et al. May 2009 A1
20090128911 Itzkovitch et al. May 2009 A1
20090153437 Aharoni Jun 2009 A1
20090190222 Simmonds et al. Jul 2009 A1
20090213208 Glatt Aug 2009 A1
20090237804 Amitai et al. Sep 2009 A1
20090303599 Levola Dec 2009 A1
20090316246 Asai et al. Dec 2009 A1
20100039796 Mukawa Feb 2010 A1
20100060551 Sugiyama et al. Mar 2010 A1
20100060990 Wertheim et al. Mar 2010 A1
20100079865 Saarikko et al. Apr 2010 A1
20100092124 Magnusson et al. Apr 2010 A1
20100096562 Klunder et al. Apr 2010 A1
20100103078 Mukawa et al. Apr 2010 A1
20100136319 Imai et al. Jun 2010 A1
20100141555 Rorberg et al. Jun 2010 A1
20100165465 Levola Jul 2010 A1
20100171680 Lapidot et al. Jul 2010 A1
20100177388 Cohen et al. Jul 2010 A1
20100214659 Levola Aug 2010 A1
20100231693 Levola Sep 2010 A1
20100231705 Yahav et al. Sep 2010 A1
20100232003 Baldy et al. Sep 2010 A1
20100246004 Simmonds Sep 2010 A1
20100246993 Rieger et al. Sep 2010 A1
20100265117 Weiss Oct 2010 A1
20100277803 Pockett et al. Nov 2010 A1
20100284085 Laakkonen Nov 2010 A1
20100296163 Saarikko Nov 2010 A1
20100315719 Saarikko et al. Dec 2010 A1
20100321781 Levola et al. Dec 2010 A1
20110019250 Aiki et al. Jan 2011 A1
20110019874 Jarvenpaa et al. Jan 2011 A1
20110026128 Baker et al. Feb 2011 A1
20110026774 Flohr et al. Feb 2011 A1
20110038024 Wang et al. Feb 2011 A1
20110050548 Blumenfeld et al. Mar 2011 A1
20110096401 Levola Apr 2011 A1
20110157707 Tilleman et al. Jun 2011 A1
20110164221 Tilleman et al. Jul 2011 A1
20110211239 Mukawa et al. Sep 2011 A1
20110235179 Simmonds Sep 2011 A1
20110238399 Ophir et al. Sep 2011 A1
20110242349 Izuha et al. Oct 2011 A1
20110242661 Simmonds Oct 2011 A1
20110242670 Simmonds Oct 2011 A1
20110310356 Vallius Dec 2011 A1
20120007979 Schneider et al. Jan 2012 A1
20120033306 Valera et al. Feb 2012 A1
20120044572 Simmonds et al. Feb 2012 A1
20120044573 Simmonds et al. Feb 2012 A1
20120062850 Travis Mar 2012 A1
20120099203 Boubis et al. Apr 2012 A1
20120105634 Meidan et al. May 2012 A1
20120120493 Simmonds et al. May 2012 A1
20120127577 Desserouer May 2012 A1
20120224062 Lacoste et al. Sep 2012 A1
20120235884 Miller et al. Sep 2012 A1
20120235900 Border et al. Sep 2012 A1
20120242661 Takagi et al. Sep 2012 A1
20120280956 Yamamoto et al. Nov 2012 A1
20120300311 Simmonds et al. Nov 2012 A1
20120320460 Levola Dec 2012 A1
20130069850 Mukawa et al. Mar 2013 A1
20130093893 Schofield Apr 2013 A1
20130101253 Popovich et al. Apr 2013 A1
20130138275 Nauman et al. May 2013 A1
20130141937 Katsuta et al. Jun 2013 A1
20130170031 Bohn et al. Jul 2013 A1
20130200710 Robbins Aug 2013 A1
20130249895 Westerinen et al. Sep 2013 A1
20130257848 Westerinen et al. Oct 2013 A1
20130258701 Westerinen et al. Oct 2013 A1
20130314793 Robbins et al. Nov 2013 A1
20130322810 Robbins Dec 2013 A1
20130328948 Kunkel et al. Dec 2013 A1
20140104665 Popovich et al. Apr 2014 A1
20140104685 Bohn et al. Apr 2014 A1
20140140653 Brown et al. May 2014 A1
20140140654 Brown et al. May 2014 A1
20140146394 Tout et al. May 2014 A1
20140152778 Ihlenburg et al. Jun 2014 A1
20140168055 Smith Jun 2014 A1
20140168260 O'Brien et al. Jun 2014 A1
20140168735 Yuan et al. Jun 2014 A1
20140172296 Shtukater Jun 2014 A1
20140176528 Robbins Jun 2014 A1
20140204455 Popovich et al. Jul 2014 A1
20140211322 Bohn et al. Jul 2014 A1
20140218801 Simmonds et al. Aug 2014 A1
20150010265 Popovich et al. Jan 2015 A1
20150289762 Popovich et al. Oct 2015 A1
Foreign Referenced Citations (28)
Number Date Country
101881936 Nov 2010 CN
1020060 03 785 Jul 2007 DE
2 110 701 Oct 2009 EP
2 225 592 Sep 2010 EP
2 381 290 Oct 2011 EP
2 733 517 May 2014 EP
2677463 Dec 1992 FR
2 115 178 Sep 1983 GB
EP0822441 Apr 1998 JP
2004-157245 Jun 2004 JP
WO-9952002 Oct 1999 WO
WO-03081320 Oct 2003 WO
WO-2006002870 Jan 2006 WO
WO-2007130130 Nov 2007 WO
WO-2007130130 Nov 2007 WO
WO-2009013597 Jan 2009 WO
WO-2009077802 Jun 2009 WO
WO-2010067114 Jun 2010 WO
WO-2010067117 Jun 2010 WO
WO-2010125337 Nov 2010 WO
WO-2011012825 Feb 2011 WO
WO-2011051660 May 2011 WO
WO-2011055109 May 2011 WO
WO-2011107831 Sep 2011 WO
WO-2013027006 Feb 2013 WO
WO-2013033274 Mar 2013 WO
WO-2013163347 Oct 2013 WO
WO-2014091200 Jun 2014 WO
Non-Patent Literature Citations (78)
Entry
Final Office Action on U.S. Appl. No. 13/250,940 Dated Oct. 17, 2014, 15 pages.
Office Action on U.S. Appl. No. 13/892,026 Dated Dec. 8, 2014, 19 pages.
Office Action on U.S. Appl. No. 13/892,057 Dated Nov. 28, 2014, 17 pages.
Webster's Third New International Dictionary 433 (1986), 3 pages.
Amendment and Reply for U.S. Appl. No. 12/571,262, mail date Dec. 16, 2011, 7 pages.
Amitai, Y., et al. “Visor-display design based on planar holographic optics,” Applied Optics, vol. 34, No. 8, Mar. 10, 1995, pp. 1352-1356.
Ayras, et al., “Exit pupil expander with a large field of view based on diffractive optics”, Journal of the Society for Information Display, 17/8, 2009, pp. 659-664.
Caputo, R. et al., POLICRYPS Switchable Holographic Grating: A Promising Grating Electro-Optical Pixel for High Resolution Display Application; Journal of Display Technology, vol. 2, No. 1, Mar. 2006, pp. 38-51, 14 pages.
Crawford, “Switchable Bragg Gratings”, Optics & Photonics News, Apr. 2003, pp. 54-59.
Extended European Search Report for EP Application No. 13192383, dated Apr. 2, 2014, 7 pages.
Final Office Action in U.S. Appl. No. 13/864,991, dated Apr. 2, 2015, 16 pages.
Final Office Action on U.S. Appl. No. 13/892,026 Dated Apr. 3, 2015, 17 pages.
International Preliminary Report on Patentability for PCT Application No. PCT/US2013/038070, dated Oct. 28, 2014, 6 pages.
International Search Report and Written Opinion regarding PCT/US2013/038070, mail date Aug. 14, 2013, 14 pages.
Levola, et al., “Replicated slanted gratings with a high refractive index material for in and outcoupling of light” Optics Express, vol. 15, Issue 5, pp. 2067-2074 (2007).
Moffitt, “Head-Mounted Display Image Configurations”, retrieved from the internet at http://www.kirkmoffitt.com/hmd—image—configurations.pdf on Dec. 19, 2014, dated May 2008, 25 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,940 Dated Mar. 18, 2015, 17 pages.
Non-Final Office Action on U.S. Appl. No. 13/432,662 Dated May 27, 2015, 15 pages.
Non-Final Office Action on U.S. Appl. No. 13/844,456 Apr. 1, 2015, XX Pages.
Non-Final Office Action on U.S. Appl. No. 13/869,866 Dated May 28, 2014, 16 pages.
Non-Final Office Action on U.S. Appl. No. 14/044,676 Dated Apr. 9, 2015, 13 pages.
Non-Final Office Action on U.S. Appl. No. 14/225,062 Dated May 21, 2015, 11 pages.
Nordin, G., et al., Journal of the Optical Society of America A., vol. 9, No. 12, Dec. 1992, pp. 2206-2217, 12 pages.
Office Action for U.S. Appl. No. 12/571,262, mail date Sep. 28, 2011, 5 pages.
Office Action for U.S. Appl. No. 13/355,360, mail date Sep. 12, 2013, 7 pages.
Press Release, “USAF Awards SBG Labs an SBIR Contract for Wide Field of View HUD”, SBG Labs—DigiLens, Apr. 2013, 1 page.
Press Release: “Navy awards SGB Labs a contract for HMDs for simulation and training”, Press releases, DigiLens, Oct. 2012, pp. 1-2, retrieved from the internet at http://www.digilens.com/pr10-2012.2.php. 2 pages.
Requirement for Restriction/Election on U.S. Appl. No. 13/844,456 Dated Sep. 12, 2014, 23 pages.
Schechter, et al., “Compact beam expander with linear gratings”, Applied Optics, vol. 41, No. 7, Mar. 1, 2002, pp. 1236-1240.
Urey, “Diffractive exit pupil expander for display applications” Applied Optics, vol. 40, Issue 32, pp. 5840-5851 (2001).
Non-Final Office Action on U.S. Appl. No. 13/250,858 dated Jun. 12, 2015, 20 pages.
Final Office Action on U.S. Appl. No. 13/869,866 Dated Oct. 3, 2014, 17 pages.
Final Office Action on U.S. Appl. No. 14/038,400 Dated Aug. 10, 2015, 32 pages.
First office action received in Chinese patent application No. 201380001530.1, dated Jun. 30, 2015, 9 pages with English translation.
Non-Final Office Action on U.S. Appl. No. 13/869,866 Dated Jul. 22, 2015, 28 pages.
Non-Final Office Action on U.S. Appl. No. 13/892,026 Dates Aug. 6, 2015, 22 pages.
Non-Final Office Action on U.S. Appl. No. 13/892,057 dated Jul. 30, 2015, 29 pages.
Non-Final Office Action on U.S. Appl. No. 14/109,551 Dated Jul. 14, 2015, 32 pages.
Non-Final Office Action on U.S. Appl. No. 14/152,756, mail date Aug. 25, 2015, 39 pages.
Non-Final Office Action on U.S. Appl. No. 14/168,173 Dated Jun. 22, 2015, 14 pages.
Notice of Allowance on U.S. Appl. No. 13/355,360 Dated Apr. 10, 2014, 7 pages.
Office Action, USPTO, U.S. Appl. No. 10/696,507, mailed on Nov. 13, 2008 (CTX-290US), 15 pages.
Irie, Masahiro, Photochromic diarylethenes for photonic devices, Pure and Applied Chemistry, 1996, pp. 1367-1371, vol. 68, No. 7, IUPAC.
Non-Final Office Action on U.S. Appl. No. 13/250,858 Dated Sep. 15, 2014, 16 pages.
Non-Final Office Action on U.S. Appl. No. 13/864,991 Dated Oct. 22, 2014, 16 pages.
Notice of Allowance on U.S. Appl. No. 13/250,970 dated Sep. 16, 2014, 7 pages.
Notice of Allowance on U.S. Appl. No. 13/251,087 Dated Jul. 17, 2014, 8 pages.
Plastic has replaced glass in photochromic lens, www.plastemart.com, 2003, 1 page.
Ayras et al., Exit Pupil Expander with a Large Field of View Based on Diffractive Optics, Journal of the SID, 2009, 6 pages.
Cameron, A., The Application of Holograhpic Optical Waveguide Technology to Q-Sight Family of Helmet Mounted Displays, Proc. of SPIE, vol. 7326, 7326OH-1, 2009, 11 pages.
Notice of Allowance for U.S. Appl. No. 12/700,557, mail date Oct. 22, 2013, 9 pages.
Office Action for U.S. Appl. No. 12/700,557, mail date Aug. 9, 2013, 12 pages.
Office Action for U.S. Appl. No. 12/700,557, mail date Feb. 4, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/250,621, mail date May 21, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/250,858 Dated Feb. 19, 2014, 13 page.
Office Action for U.S. Appl. No. 13/250,858, mail date Oct. 28, 2013, 9 pages.
Office Action for U.S. Appl. No. 13/250,940, mail date Aug. 28, 2013, 15 pages.
Office Action for U.S. Appl. No. 13/250,940, mail date Mar. 12, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/250,970, mail date Jul. 30, 2013, 4 pages.
Office Action for U.S. Appl. No. 13/250,994, mail date Sep. 16, 2013, 11 pages.
Office Action on U.S. Appl. No. 13/250,940 Dated Mar. 25, 2014, 12 pages.
Office Action on U.S. Appl. No. 13/251,087 Dated Mar. 28, 2014, 12 pages.
Wisely, P.L., Head up and head mounted display performance improvements through advanced techniques in the manipulation of light, Proc. of SPIE vol. 7327, 732706-1, 2009, 10 pages.
Final Office Action on U.S. Appl. No. 13/250,858 Dated Feb. 4, 2015, 18 pages.
Final Office Action on U.S. Appl. No. 13/892,057 Dated Mar. 5, 2015, 21 pages.
Non-Final Office Action on U.S. Appl. No. 14/038,400 Dated Feb. 5, 2015, 18 pages.
Final Office Action issued in U.S. Appl. No. 13/864,991 dated Jun. 27, 2016.
Non-Final Office Action issued in U.S. Appl. No. 13/892,057, dated May 16, 2016.
Final Office Action on U.S. Appl. No. 13/892,057 dated Nov. 20, 2015, 30 pages.
Non-Final Office Action on U.S. Appl. No. 13/844,456, dated Jan. 15, 2016, 16 Pages.
Non-Final Office Action on U.S. Appl. No. 13/864,991 dated Nov. 30, 2015, 18 pages.
Non-Final Office Action on U.S. Appl. No. 14/044,676, dated Jan. 20, 2016, 21 pages.
Notice of Allowance on U.S. Appl. No. 13/432,662, dated Feb. 18, 2016, 10 pages.
Non-Final Office Action on U.S. Appl. No. 13/892,026 dated Mar. 22, 2016, 16 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,858, dated Mar. 18, 2016, 20 pages.
Non-final Office Action on U.S. Appl. No. 13/250,858, dated Nov. 14, 2016, 18 pages.
Notice of Allowance on U.S. Appl. No. 13/864,991, dated Feb. 2, 2017, 10 pages.
Notice of Allowance on U.S. Appl. No. 13/892,057, dated Nov. 8, 2016, 10 pages.