Radar beam sharpening system and method

Information

  • Patent Grant
  • 10705201
  • Patent Number
    10,705,201
  • Date Filed
    Monday, August 31, 2015
    9 years ago
  • Date Issued
    Tuesday, July 7, 2020
    4 years ago
Abstract
An enhanced vision method or a weather radar system can be used with an aircraft and includes an antenna and a control circuit. The control circuit is configured to provide radar beams via the antenna toward external surroundings and is configured to receive radar returns. The control circuit is configured to process a collection of radar measurements from the radar returns, wherein each of the radar measurements is associated with a location determined using an antenna position, an antenna attitude, a beam sharpening angle, and a range. The radar measurements are processed to determine power density per grid cell associated with the power and location of the radar measurements, and the power density per grid cell is used to provide an image associated with the power and location of the radar measurements.
Description
CROSS-REFERENCE TO RELATED PATENT APPLICATIONS

The present application is related to U.S. Pat. No. 8,773,301 filed May 17, 2012 by Woodell et al., U.S. Pat. No. 9,939,526 filed Nov. 7, 2014 by Jinkins et al., U.S. Pat. No. 9,733,349 filed Sep. 10, 2014 by Wood et al., U.S. Pat. No. 9,384,586 filed Jun. 10, 2014 by McCusker et al., U.S. Pat. No. 9,024,805 filed Sep. 26, 2012 by Jinkins et al., U.S. Pat. No. 8,643,533 filed Sep. 28, 2010 by Woodell et al., U.S. Pat. No. 9,562,788 filed Sep. 30, 2011 by Jinkins et al., U.S. Pat. No. 8,977,491 filed Sep. 28, 2008 by McCusker et al., U.S. Pat. No. 7,889,117 filed Jul. 2, 2008 by Woodell et al., U.S. Pat. No. 8,077,078 filed Jul. 25, 2008 by Woodell et al., U.S. Pat. No. 8,896,480 filed Sep. 28, 2011 by Wilson et al., U.S. Pat. No. 8,515,600 filed Sep. 6, 2007 by McCusker, U.S. Pat. No. 8,755,954 filed Sep. 22, 2011 by McCusker et al., U.S. Pat. No. 8,917,191 filed Oct. 31, 2008 by Tiana et al., U.S. Pat. No. 9,354,633 filed Oct. 31, 2008 by McCusker et al., and U.S. Pat. No. 8,077,078 filed Jul. 25, 2008 by Woodell et al., all of which are herein incorporated by reference in their entireties and assigned to the assignee of the present application.


BACKGROUND

Sensor systems are used by aircraft. For example, an aircraft uses an enhanced vision system (EVS) or enhanced flight visions system (EFVS) to provide imagery to an aircraft crew. An EVS uses either a passive or an active sensing system to acquire data used to generate imagery of the runway environment. A conventional passive sensor, such as a forward looking infrared (FLIR) camera or visible light spectrum camera, receives electromagnetic energy from the environment and outputs data that may be used by the system to generate video images from the point of view of the camera. The camera is installed in an appropriate position, such as in the nose of an aircraft, so that the pilot may be presented with an appropriately scaled and positioned video image on the display. However, while passive sensors provide relatively high quality video imagery, the passive sensors are often be unable to identify required visual references in certain low visibility conditions such as heavy fog.


Active sensing systems, such as millimeter wavelength (MMW) radar systems (e.g., 94 GHz), transmit electromagnetic energy into the environment and receive return electromagnetic energy reflected from the environment. The active sensing system is generally installed in an appropriate position, such as in the nose of an aircraft. Active sensing systems are expensive and require space on-board the aircraft that might already be required for other types of equipment. In addition, millimeter wavelength radar systems require expensive radome technology. Additionally, both FLIR cameras and millimeter wavelength radar systems may have limited range in certain low visibility conditions such as heavy fog, rain, or other precipitation.


Thus, there is a need for real time or near real time sensing systems for and methods of providing enhanced vision at longer ranges and in inclement weather. Further, there is a need for real time or near real time sensing systems for and methods of providing enhanced vision imagery that are less expensive and do not require additional space on the aircraft. There is also a need for display systems for and methods of providing images of the external scene using radar data from a weather radar system. There is still a further need for systems for and methods of providing images of the runway environment derived from weather radar data where such images enable operation below certain specified altitudes during instrument approaches. Further still, there is a need for systems and methods that achieve higher resolution imaging using X-band and C-band radar data.


SUMMARY

In one aspect, embodiments of the inventive concepts disclosed herein are directed to an image processing system for enhanced vision including a radar system, a processor and memory coupled to the processor. The memory contains program instructions that, when executed, cause the processor to instruct the radar system to provide radar beams and receive radar returns with improved angular and/or range resolution for deriving image data of the external scene topography.


In a further aspect, embodiments of the inventive concepts disclosed herein are directed to an enhanced vision system including a weather radar system and a display. The weather radar system is configured to generate image data representative of a runway environment associated with radar returns received by the weather radar system. The radar returns are in an X-band or a C-band, and the weather radar system is configured to process a collection of radar measurements from the radar returns. Each of the radar measurements is associated with a location determined from using an antenna position, an antenna attitude, a beam sharpening angle, and a range. The radar measurements are processed to determine power density per grid cell associated with the power and location of the radar measurements, and the power density per grid cell is used to provide an image associated with the power and location of the radar measurements. The display is in communication with the weather radar system and configured to display an image associated with the image data.


In a further aspect, embodiments of the inventive concepts disclosed herein are directed to a weather radar system for an aircraft. The aircraft includes aircraft sensors. The weather radar system includes an antenna and a control circuit configured to provide radar beams via the antenna toward external surroundings and configured to receive radar returns. The control circuit is configured to process a collection of radar measurements from the radar returns. Each of the radar measurements is associated with a location determined using an antenna position which is determined from aircraft sensors, an antenna attitude which is determined from the aircraft attitude sensed by the aircraft sensors combined with antenna boresight relative to the aircraft attitude, a beam sharpening angle, and a range. The radar measurements are processed to determine power density per grid cell associated with the power and location of the radar measurements. The power density per grid cell is used to provide an image associated with the power and location of the radar measurements.


In a further aspect, embodiments of the inventive concepts disclosed herein are directed to a method of providing a real time sensor image on an electronic display. The method includes determining an antenna position, determining an antenna attitude, receiving radar returns from an X-band or C-band airborne weather radar system, and determining a location of each of a plurality of radar measurements associated with the radar returns using an antenna position, an antenna attitude, a beam sharpening angle, and a range. The method also includes determining a power density per grid cell associated with the power and location of the radar measurements. The power density per grid cell is used to provide a real time sensor image on an electronic display.





BRIEF DESCRIPTION OF THE DRAWINGS

Implementations of the inventive concepts disclosed herein may be better understood when consideration is given to the following detailed description thereof. Such description makes reference to the annexed drawings, which are not necessarily to scale, and in which some features may be exaggerated and some features may be omitted or may be represented schematically in the interest of clarity. Like reference numerals in the figures may represent and refer to the same or similar element, feature, or function. In the drawings:



FIG. 1 is a schematic illustration of an aircraft control center or cockpit according to some exemplary embodiments;



FIG. 2 is a schematic general block diagram of a display system for displaying an image derived from radar data, according to some exemplary embodiments;



FIG. 3 is a flow diagram showing an operational flow used by the display system illustrated in FIG. 2 to determine the measurement locations and accumulate returned measurements from the target environment, according to a further exemplary embodiment of the inventive concepts disclosed herein;



FIG. 4 is an illustration of an image derived from radar data provided by the display system illustrated in FIG. 2, wherein individualized measurements are represented by dot representations, according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 5 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 4, wherein individualized measurements are represented by dot representations, according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 6 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein the individualized measurements represented by the dot representations are provided over Cartesian grid cells, according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 7 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein the individualized measurements are represented by power spatial density for the Cartesian grid cells according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 8 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein the individualized measurements are represented in a revised gray scale for the power spatial density for the Cartesian grid cells, according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 9 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein individualized measurements are represented in filtered power spatial density for the Cartesian grid cells according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 10 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein the individualized measurements represented by the dot representations are provided over polar grid cells, according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 11 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein the individualized measurements are represented in a revised gray scale for the power spatial density for the polar grid cells, according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 12 is an illustration of an image derived from the radar data associated with the image illustrated in FIG. 5, wherein the individualized measurements are represented in filtered power spatial density for the polar grid cells, according to yet another exemplary embodiment of the inventive concepts disclosed herein; and



FIG. 13 is a diagram of an operational flow for the generation of image data for the display system illustrated in FIG. 2, in accordance with some embodiments of the inventive concepts disclosed herein.





DETAILED DESCRIPTION

Before describing in detail the inventive concepts disclosed herein, it should be observed that the inventive concepts disclosed herein include, but are not limited to, a novel structural combination of data/signal processing components, sensors, and communications circuits, and not in the particular detailed configurations thereof. Accordingly, the structure, methods, functions, control and arrangement of components, software, and circuits have, for the most part, been illustrated in the drawings by readily understandable block representations and schematic diagrams, in order not to obscure the disclosure with structural details which will be readily apparent to those skilled in the art, having the benefit of the description herein. Further, the inventive concepts disclosed herein are not limited to the particular embodiments depicted in the exemplary diagrams, but should be construed in accordance with the language in the claims.


According to various exemplary embodiments, an EVS, a sensor system, or a display system can use radar sensing to provide imagery for a pilot or co-pilot of an aircraft display, such as, imagery for a head down display (HDD) or head up display (HUD). For example, the display system may include or use a weather radar system to display an image based upon radar return data. In some embodiments, a Doppler weather radar system may be configured to have enhanced resolution (e.g., angular resolution and/or range resolution). Reflectivity of radar returns from runway structures in an airport terminal or runway environment, such as, an approach lighting system, a threshold lighting system, and or a runway edge lighting system, can be sensed. As will be appreciated, using a weather radar system configured according to the various exemplary embodiments provides greater range than millimeter wavelength radar sensing systems or passive FLIR or visible light camera systems in low visibility conditions, such as, in heavy fog, rain, and snow, given the weather radar system's superior ability to penetrate heavy fog, rain, and snow.


In some embodiments, a weather radar system achieves accurate measurement location using an accurate beam sharpening angle. In some embodiments, the weather radar system creates a two dimensional or three dimensional grid containing power spatial density (PSD) parameters from a collection of radar measurements, each having an individual location and power associated with it. The location of the individual measurement is based on the radar antenna position in some embodiments. From the radar antenna position, the location of the individual measurement is found using the attitude of the antenna, the beam sharpening angle and the range, all specific to that individual measurement in some embodiments. The individual measurements are processed to determine the power spatial density per grid cell associated with the power and location of the individual measurements. The power spatial density per grid cell can be used to generate an image representing the target environment sensed by the radar scan.


Using the weather radar system configured according to the various exemplary embodiments also provides EVS imagery having sufficient accuracy in low visibility conditions (given that many of the visual references required under Title 14 of the Code of Federal Regulations, part 91, such as, approach lighting systems, threshold lighting systems, runway edge lighting systems, and other runway structures, are structures that exhibit high radar reflectivity) in some embodiments. The imagery may allow lower landing minima (e.g., 100 feet or less) in some embodiments. In some embodiments, the lack of radar returns from the runway surface combined with runway structures and lights can provide a suitable image for runway identification by the pilot.


The display system includes a radar processing circuit in communication with the radar system and configured to generate high resolution radar image data of the power spatial density for display in some embodiments. The image data is processed to provide a two-dimensional aircraft situation display (e.g., vertical profile display or plan view display) or three dimensional or perspective aircraft situation display representative of the three dimensional positions of runway structures in an airport terminal or runway environment based on the radar returns as described in U.S. patent application Ser. Nos. 14/301,199, 14/482,681 and 14/536,3300 incorporated herein by reference in their entireties in some embodiments. For example, the radar processing circuit can be embodied as a processor and a non-transitory memory containing program instructions that, when executed, cause the processor to instruct the radar system to provide radar beams and receive radar returns via the antenna and generate image data from the radar returns.


According to certain exemplary embodiments, a radar system, such as, a weather radar system, can be used to sense features of a runway environment. Utilizing the radar cross section associated with runway lighting structures advantageously allows sensing to be achieved whether at day or night, regardless of whether runway lights are on or are off in some embodiments. In one embodiment, the regular, periodic, equal spacing nature of visual aids, such as, approach lighting system, runway edge lights, taxi way lights, and center line lights, can be identified from the image generated from the radar data. In certain embodiments, the systems and methods can be utilized as extension to a combined vision system (CVS).


Referring to FIG. 1, a vision system 10 is provided in an aircraft having an aircraft control center 11 or cockpit. The aircraft control center 11 includes flight displays 20 embodied as head down displays (HDDs). The aircraft control center 11 can also include a combiner 21 association with a head up display (HUD) system. In some embodiments, the combiner 21 is provided as part of a wearable HUD. Conformal images are provided on the combiner 21 in some embodiments.


The flight displays 20 and the combiner 21 can be used to provide information to the flight crew, thereby increasing visual range and enhancing decision-making abilities. In an exemplary embodiment, the flight displays 20 and the combiner 21 can include a weather display, a joint display, a weather radar map and a terrain display. Further, the flight displays 20 may include images from a synthetic vision system (SVS) or an enhanced vision system (EVS) (e.g., an EFVS). For example, the flight displays 20 can include a display configured to display a perspective image of terrain and/or weather information. Other views of terrain and/or weather information may also be provided (e.g., plan view, horizontal view, vertical view, or combinations thereof). Additionally, the flight displays 20 can be implemented using any of a variety of display technologies, including CRT, LCD, organic LED, dot matrix display, and others.


According to some embodiments, the vision system 10 is configured to provide an image based upon radar data to at least one of the displays 20 or the combiner 21. In FIG. 1, the image on the combiner 21 includes a runway 23 or features 29 associated with the runway 23 as viewed from the aircraft (e.g., during approach and/or landing). In some embodiments, at least one of the displays 20 or the combiner 21 displays a merged image of terrain derived from two or more of enhanced vision data, radar data, and SVS data. Advantageously, real time radar data can be provided to provide a real time, all weather detection of the runway features 29 associated with the runway 23 in one embodiment. Advantageously, the radar data allows the runway 23 and its orientation and an extended centerline 27 to be viewed by one or more pilots in challenging weather conditions in some embodiments.


Referring to FIG. 2, the vision system 10 uses radar data from a radar system 102 to provide an image to any of the displays 20 (FIG. 1) via a HDD computer 103 or the combiner 21 (FIG. 1) via a HUD computer 104 in some embodiments. The vision system 10 is in communication with or includes the radar system 102 and can include or be in communication with an SVS and/or an EVS in some embodiments. The EVS and the SVS are optional in some embodiments.


The vision system 10 is in communication with aircraft sensors 130. The aircraft sensors 130 are associated with navigation equipment and flight control devices in some embodiments and can include an inertial navigation system, flight computer, barometric altimeter, radar altimeter, global positioning system (GPS) receiver, air speed sensors, heading sensors, etc.) The aircraft sensors 130 include an aircraft position sensor 134 and an aircraft attitude sensor 136.


The aircraft positon sensor 134 is any device or system for providing an electronic signal or data representing aircraft positon. In some embodiments, the aircraft position includes a position parameter (e.g., latitude and longitude) and an aircraft altitude parameter. The aircraft position sensor 134 is a GPS receiver or inertial navigation system in some embodiments. The aircraft altitude can be provided by a GPS receiver, an altitude sensor (e.g., barometric sensor), an inertial navigation system, or other device. In some embodiments, the aircraft positon also includes velocity and acceleration parameters associated with aircraft positon (e.g., for predicting future positon of the aircraft).


The aircraft attitude sensor 136 is any device for providing an electronic signal or data representing attitude of the aircraft. The aircraft attitude includes a heading parameter, a pitch parameter, a yaw parameter, and a roll parameter in some embodiments. In some embodiments, the aircraft positon sensor 134 and the aircraft attitude sensor 138 are parts of a single positioning or navigation sensor system, such as a GPS receiver or inertial navigation system. In some embodiments, the aircraft attitude also includes velocity and acceleration parameters associated with attitude (e.g., for predicting future attitude position of the aircraft).


The radar system 102 receives data from the aircraft sensors 130 and provides image data for display by vision system 10. The radar system 102 is a weather radar system generally located inside the nose of the aircraft, inside a cockpit of the aircraft, on the top of the aircraft or on the tail of the aircraft in some embodiments. The radar system 102 includes a radar antenna 120, a processor 125, a radar data storage unit 180, and an image frame memory 128. The radar system 102 can be a weather radar system, such as, a Multiscan™ radar system from Rockwell Collins, Inc. configured as described herein. The radar system 102 can utilize a split, half or sub-aperture or other technique for obtaining radar data associated with external surroundings in some embodiments. The radar system 102 can use the split or sub-aperture techniques of the radar systems described in U.S. application Ser. Nos. 13/627,788, 12/892,563, 13/250,798, 12/236,464, and 12/167,200 and U.S. Pat. No. 8,077,078, incorporated herein by reference and assigned to the assignee of the present application. The type of the radar system 102 and data gathering techniques are not discussed in the specification in a limiting fashion.


The processor 125 uses radar data stored in the radar data storage unit 180 and the data from the aircraft sensors 130 to store image data in the image frame memory 128. The processor 125 includes a measurement location determiner 154, an accumulator 155, an antenna scan angle determiner 182, a beam sharpening angle determiner 184, a range determiner 186, and a power determiner 188. The accumulator 155, the measurement location determiner 154, the radar data storage unit 180, the antenna scan angle determiner 182, the beam sharpening angle determiner 184, the range determiner 186, and the power determiner 188 are software modules, circuits, or combinations thereof in some embodiments.


The measurement location determiner 154 uses data from the aircraft sensors 130 and from the radar data storage unit 180, the antenna scan angle determiner 182, the beam sharpening angle determiner 184, and the range determiner 186 to identify a location for individualized measurements associated with returns received at the radar antenna 120 in some embodiments. The data associated with the radar returns are stored in the radar data storage unit 180 in some embodiments. The measurement location determiner 154 uses the position of the radar antenna 120, the position and attitude of the aircraft, the antenna pointing angle, the beam sharpening angle, and the range to determine the location of individualized measurements in some embodiments. Each individualized measurement is accumulated in the accumulator 155 by location in some embodiments. The power determiner 188 is tallied for a return power of the individualized measurements by the accumulator 155 for each return on a location basis in some embodiments.


The processor 125 causes the radar antenna 120 to provide radar signals or beams and to receive radar returns (e.g., weather radar returns data). The processor 125 is an electronic processor that processes the radar returns and provides the radar data to the radar data storage unit 180. The radar signals and radar returns are in the X-band or C-band in some embodiments.


The radar system 102 provides the radar data (e.g., weather radar return data) to the storage unit 180 in one embodiment. The radar data can be processed and filtered for various weather sensing functions as well as measurement location functions. In some embodiments, the processor 125 provides an image frame or image data for storage in the image frame memory 128 using the accumulated individualized returns in accumulator 155. In some embodiments, the accumulator 155 determines power spatial density by grid cell to create the image frame.


The image frame memory 128 is a video or graphic electronic memory in some embodiments. In some embodiments, the image frame memory 128 is a regular electronic memory. The accumulator 155 is an electronic memory, processor circuitry, or combination thereof in some embodiments. In some embodiments, accumulator 155 is part of a digital signal processor or the processor 125 or is a software module executing on the digital signal processor or the processor 125. The accumulator 155 is embodied as a register in some embodiments. The data storage unit 180 is an electronic memory in some embodiments.


The radar data associated with the external surroundings can represent sensed targets and the location of the sensed targets. Targets include terrain, man-made features, objects, runways, etc. Improved angular resolution and range resolution techniques discussed in U.S. patent application Ser. No. 14/536,330 filed Nov. 7, 2014 by Jinkins et al. incorporated herein by reference in its entirety allows the location of the targets to be more accurately determined and represented in image data in some embodiments. The radar system 102 can utilize clutter suppression and Doppler filtering to improve performance in some embodiments.


In some embodiments, the radar system 102 provides data representing a 120 degree field of view in accordance with a weather radar sweep. The sweep can be limited during approach to be a 30 degree sweep in certain embodiments. The sweep is directed toward the surface of the Earth so that returns are obtained which allow runway environment features to be sensed. Various types of sweeps, sweep patterns, and sweep speeds can be utilized without departing from the scope of the invention.


The radar system 102 embodied as a weather radar allows existing avionic equipment to be used as a real-time sensor for providing a radar-derived enhanced image of the external scene topography to the pilot in some embodiments. The image or representation generated by the radar system 102 is provided on the displays 20 (FIG. 1) or the combiner 21 can function as an EVS to provide situation awareness to the pilot in some embodiments. In other embodiments, the image or representation generated by the radar system 102 is provided on the displays 20 or the combiner 21 can function as an EFVS to allow lower landing minima.


The radar system 102 advantageously provides increased range resolution in some embodiments. The radar system 102 also provides increased angle resolution in some embodiments. The increased resolution in range and angle allows a higher resolution for measurement location by the measurement location determiner 154 and hence higher image resolution to be provided on the displays 20 and the combiner 21 in some embodiments.


According to some embodiments, the radar system 102 can use a beam sharpening method to achieve increased angular resolution. In some embodiments, the radar system 102 can utilize techniques such as beam sharpening (e.g., horizontal beam sharpening) and de-convolution of the beam point spread function for improved angular resolution. In some embodiments, the radar system 102 can use beam sharpening as a process that improves the antenna-induced poor angular resolution (e.g., due to the beam width). There are many methods that can be used such as: Monopulse Radar, Sub-Aperture Radar or Split-Aperture Radar, etc. Mathematical methods can be utilized to determine a center of the radar echo for identifying runway features. Techniques for beam sharpening and determining beam sharpening angles and scan angles are discussed in U.S. patent application Ser. Nos. 13/627,788, 12/892,563, 13/250,798, 12/236,464, and 12/167,200 and U.S. Pat. No. 8,077,078 incorporated herein by reference in their entireties.


The radar system 102 uses the radar antenna 120 that toggles between transmitting and receiving on the full aperture and transmitting on the full aperture while receiving on the partial aperture in some embodiments. These techniques can be used to accurately estimate at which angle the measurement was located within the radar beam. The received returns can be processed to determine a high resolution estimate of a measurement angle relative to the boresight of the antenna beam. According to some embodiments, the returns can be processed using a complex conjugate multiplication method to determine the measurement angle. The processing can be related to sequential lobing processing but is executed in the phase domain as opposed to the common amplitude domain in some embodiments.


In some embodiments, the radar system 102 uses sequential lobing techniques where two antennas that are close to the same place may be used, going back and forth between the two antennas. An amplitude signature or phase signature that varies between the two halves of the antennas may be used to obtain data about target position for sensed targets (e.g., an object such as other aircraft, terrain, or towers). Sequential lobing generally does not use phase comparisons with moving targets due to Doppler-induced phase changes that contaminate the phase center measurement. However, using a complex conjugate multiply method allows the Doppler-induced phase changes to be removed by cancellation. Therefore, a change in phase center between multiple different sub-apertures may be determined and used to determine the angle to measurement of the target.


In some embodiments, the effective waveform bandwidth of the radar system 102 is increased to increase range resolution. The radar system 102 can use stepped-frequency compression in some embodiments. To provide higher range resolution, the radar system 102 provides ultra-wideband radar (UWB) pulses (e.g., extremely narrow pulses with high power), or provides intra pulse compression (frequency of phase modulation of the transmitted pulse) in some embodiments. Frequency coding techniques including the common linear frequency modulation (LFM) or chirp method, and discrete coded segments within the pulse can be utilized in some embodiments. Phase coding techniques including binary phase codes as well as various polyphase codes can be utilized in some embodiments. To provide higher range resolution, the radar system 102 provides interpulse pulse compression or stepped frequency compression (e.g., successive pulses with discrete increasing frequency steps) in some embodiments. In some embodiments, stepped frequency compression advantageously achieves high effective bandwidth with narrow instantaneous bandwidth. The receive bandwidth is smaller, has lower noise bandwidth, and a higher signal to noise ratio in some embodiments. Analog-to-digital sampling rates are lower (vs. pulse-compression) in some embodiments. In addition, the stepped frequency compression also has a smaller peak power (e.g., when compared to impulse), provides flexible transmit frequency control, can “hop” over restricted or undesired transmit frequencies, enables adaptive/cognitive frequency use, and rejects later received clutter from earlier transmit pulses in some embodiments. Further, the stepped frequency compression provides returns from clutter in ambiguous ranges that have frequencies that are different from returns from targets and rejects ambiguous clutter returns in the receiver IF filter of the radar system 102 in some embodiments. Stepped frequency compression generally does not achieve range resolution with a single pulse, requires transmit, receive and processing of a group of pulses for any one bin, and has more pronounced range-Doppler coupling (e.g., different Doppler shifts for each frequency) in some embodiments.


The processor 125 can be any hardware and/or software processor or processing architecture capable of executing instructions and operating on navigational and radar data. The processor 125 can be capable of determining navigational information such as altitude, heading, bearing, and location based on data from the aircraft sensors 130. The processor 125 can be, or can include one or more microprocessors, an application specific integrated circuit (ASIC), a circuit containing one or more processing components, a group of distributed processing components, circuitry for supporting a microprocessor, or other hardware configured for processing.


With reference to FIG. 3, a flow 300 can be performed by the vision system 10 in some embodiments. The processor 125 or other computing platform can execute flow 300 to provide an image in response to aircraft sensor parameters and radar data. The aircraft sensor parameters and radar data are processed to provide individualized measurements at locations which have a reflected power. The individualized measurements are further processed on a location basis to determine the power spatial density which is used to provide the image data for an enhanced target image 340. The flow 300 is used to create a two dimensional image (e.g., the enhanced radar image 340) of the power spatial density from a collection of radar measurements that each have an individual location and power associated to that measurement in some embodiments. The individualized measurements are located and accumulated to determine a power spatial density on a grid basis to create the image (e.g., the enhanced radar image 340).


The radar system 102 uses an aircraft position parameter 302 from the aircraft sensors 130 to perform a projection operation 304 to project a local aircraft position and provide a projected aircraft position parameter 306 in some embodiments. The aircraft positon parameter 302 includes positon (e.g., altitude, latitude and longitude) in some embodiments. The projected aircraft positon parameter includes position parameters but projected to a local axis system in some embodiments.


The radar system 102 uses an aircraft attitude parameter 310 from the aircraft sensors 130 to perform a combination operation 314 with an antenna scan angle parameter 312 from the radar system 102. The combination operation 314 provides an antenna pointing angle parameter 316 associated of the radar antenna 120 in some embodiments. The attitude parameter 310 includes heading in some embodiments. In some other embodiments, the attitude parameter 310 could also include yaw, pitch, and roll.


The radar system 102 uses a beam sharpening angle parameter 320, the antenna pointing angle parameter 316, the range parameter 322 and the projected aircraft positon parameter 306 to provide a projected measurement location parameter 324 in a geometry calculation operation 305 in some embodiments. The projected measurement location parameter 324 is a location of an individualized measurement associated with a radar return in some embodiments. In some embodiments, the location of the individual measurement is determined using the position of the radar antenna 120 (e.g., the antenna pointing angle parameter 312), the heading of the aircraft, the position of the aircraft (e.g., the projected position parameter 306), the range parameter 322, and the beam sharpening angle parameter 320, all specific to the individualized measurement.


The heading from the aircraft attitude sensor 310 is the direction of the longitudinal axis of the aircraft with respect to North in some embodiments. The antenna scan angle parameter 312 represents the direction of the boresight of the radar antenna 120 with respect to the heading of the aircraft in some embodiments. The beam sharpening angle parameter 320 represents the radar sensed direction of the returned reflection of the individual measurement with respect to the boresight of the radar antenna 120 in some embodiments. The range parameter 322 is the ground range (e.g., the radar sensed distance between the radar antenna 120 and the returned reflection of the individual measurement projected onto a flat ground surface) in some embodiments. Ground range can be calculated using the Pythagorean theorem by taking a slant range that is sensed by the radar system corrected with the difference in altitude between the radar antenna 120 and returned reflection of the individual measurement in some embodiments. The altitude of the radar antenna 120 can be obtained from the aircraft sensors 130 in some embodiments. There are multiple ways to obtain the altitude of the returned reflection of the individual measurement in some embodiments including using the radar antenna altitude, antenna tilt angle, vertical beam sharpening angle and slant range similar to the horizontal calculation discussed above in some embodiments. The antenna tilt angle is the direction of the vertical boresight of the radar antenna 120 with respect to the horizon in some embodiments. The vertical beam sharpening angle is the radar sensed vertical direction of the returned reflection of the individual measurement with respect to the vertical boresight of the radar antenna 120 in some embodiments. The altitude of the reflection can also be assumed to be coming from an altitude that can be approximated be the surveyed runway threshold altitude, since the observed approach lights are normally located level with the runway threshold in some embodiments. The parameters and their representations discussed in this paragraph are exemplary only.


An accumulation operation 328 accumulates return measurements based on the locations of the individualized measurements in the accumulator 155. The accumulation includes power parameter 326 in some embodiments. The power can be sensed as a decibel (dB) value of electromagnetic energy received at the radar antenna 120. All individualized measurements have a location and reflected power value.


The data accumulated in the accumulation operation 328 is used to provide an enhanced target image 340 for provision in the radar image frame 128 in some embodiments. The enhanced target image 340 is provided on the display 20 (FIG. 1) via the HDD computer 103 (FIG. 2) and the combiner 21 via the HUD computer 104 in some embodiments. The data accumulated in the accumulation operation 328 represents spatial power density on a grid basis in some embodiments.


With reference to FIG. 4, an image 400 of the external scene topography derived from the radar data associated with the returns or measurements includes features 402 associated with runway approach lights of a runway 412. An absence of return energy is associated with the runway 412 in some embodiments.


A portion 410 of the image 400 shows features 414, 416, and 418 associated with runway lights. In some embodiments, the individualized measurements are represented by dots in the image 410. The grey level of each dot represents the reflected power level of the individualized measurements. The darkness and size of features 414, 416, and 418 indicates presence of a large number of strong individualized measurements, such as measurements associated with reflections from approach lights.


With reference to FIG. 5, an image 500 similar to the portion 410, is derived from the radar data associated with the returns or measurements. The image 500 includes features 502, 504, and 506 associated with features 414, 416, and 418 (FIG. 4).


With reference to FIG. 6, an image 600 similar to the image 500 derived from the radar data associated with the radar returns or measurements is provided over a Cartesian grid with grid cells 601. The image 600 includes features 602, 604, and 606 associated with the features 502, 504, and 506 (FIG. 5). The grid cells 601 can be sized to represent a particular area or volume. For example, the grid cells 601 can represent a 10 square foot area or a 10 cubic foot volume. Various dimensions can be used for the grid cells 601. The grid cells 601 can be other shapes and be part of a non-Cartesian grid in some embodiments


The features 602, 604, and 606 fill a relatively large portion of the grid cells 601 associated with their locations. All of the individual measurements can be accumulated per each of the grid cells 601. Within one grid cell 601, the powers of the individual measurements are added together. Based on the area of the grid cell 601, the power per area is calculated to provide the power spatial density on a grid basis in some embodiments.


With reference to FIG. 7, an image 700 similar to the image 600 derived from the radar data associated with the radar returns or measurements shows power spatial density in each of grid cells 701 which are similar to grid cells 601. The power spatial density in each grid cell 701 is represented by a grey level. The image 700 includes features 702, 704, and 706 associated with the features 602, 604, and 606 (FIG. 6). The grid cells 710 are white representing very low power or no individual measurements in the grid cells 701 while grid cells 702, 704, and 706 associated with the features 602, 604, and 606 are black representing the highest power spatial density level or levels. The features 702, 704, and 706 appear as one or more grid cells 701 that are darker than neighboring grid cells 701. In some embodiments, visualization of high power spatial density areas is of main interest so the intensity range for grey levels for the power spatial density are adjusted such the lower power spatial density levels are white and gray levels start at higher power spatial density levels.


With reference to FIG. 8, an image 800 similar to the image 700 derived from the radar data associated with the radar returns or measurements shows power spatial density in each of grid cells 801 which are similar to grid cells 701. The power spatial density in each grid cell 801 is represented by a grey level. The image 800 includes features 802, 804, and 806 associated with the features 702, 704, and 706 (FIG. 7). The intensity range for grey levels for the power spatial density is adjusted using filtering or other image darkness adjustment techniques such that the lower power spatial density levels are white and gray levels start at higher power spatial density levels. For example, power spatial density thresholds for each grey level can be adjusted to emphasize higher power spatial density cells 801 over medium to lower power spatial density grid cells 801 in some embodiments. A threshold table or map can be used to store such thresholds. The threshold table can be adjusted via a user interface in some embodiments. Filtering can be used to reduce fluctuations in the power spatial density in some embodiments.


With reference to FIG. 9, an image 900 similar to the image 800 derived from the radar data associated with the radar returns or measurements shows power spatial density over grid cells 901 which are similar to grid cells 801. The image 900 is a result of spatial filtering of the image 800 to reduce fluctuations in the power spatial density in some embodiments. The image 900 appears less noisy in some embodiments. Digital spatial filtering can be utilized. The image 900 is an effective visualization of power spatial density despite spatial filtering.


With reference to FIG. 10, an image 1000 similar to the image 500 derived from the radar data associated with the storage unit 180 is provided over a polar grid with grid cells 1001. The grid cells 1001 can be sized to represent a particular area associated with features 1002, 1004, and 1006 which are associated with the features 414, 416, and 418 (FIG. 4). For example, the grid cells 1001 can represent a particular area or a volume. Various dimensions can be used for the grid cells 1001. The grid cells 1001 can be other shapes and be part of a non-polar grid in some embodiments


With reference to FIG. 11, an image 1100 similar to the image 1000 derived from the radar data associated with the radar returns or measurements shows power spatial density in each of grid cells 1101 which are similar to grid cells 1001. The power spatial density in each grid cell 1101 is represented by a grey level. The image 1100 includes features 1102, 1104, and 1106 associated with features 1002, 1004, and 1006. The features 1102, 1104, and 1106 appear as one or more grid cells 1101 that are darker than neighboring gird cells 1101. The intensity range for grey levels for the power spatial density are adjusted using filtering or other image darkness adjustment techniques such that the lower power spatial density levels are white and gray levels start at higher power spatial density levels.


With reference to FIG. 12, an image 1200 similar to the image 1100 derived from the radar data associated with the radar returns or measurements shows power spatial density in each of grid cells 1201 which are similar to grid cells 1101. The power spatial density in each grid cell 1201 is represented by a grey level. The image 1200 includes features 1202, 1204, and 1206 associated with features 1102, 1104, and 1106 (FIG. 11). Spatial filtering can be used to reduce fluctuations in the power spatial density in some embodiments. Such filtering can make the image 1200 appear less noisy in some embodiments.


With reference to FIG. 13, a flow 1300 can be performed by the vision system 10 in some embodiments. The processor 125 or other computing platform can execute flow 1300 to provide an image in response to aircraft sensor parameters and radar data. At an operation 1302 radar beams are provided by the radar antenna. At an operation 1304, radar returns are received by the radar antenna 120. At an operation 1306, the radar returns are processed to provide radar data. The radar data is used to determine the power and location of the individual radar measurements. The locations of the measurements are determined using the radar data and parameters from the aircraft sensors 130 (e.g., according to flow 300) in some embodiments. At an operation 1308, the powers of the localized measurements are accumulated over a grid cell. At an operation 1310, power spectral density data is calculated for the grid cells associated with the powers of the localized measurements. At an operation 1312, an image is displayed using image data generated from the power spatial density data at an operation 1310. The image is of a grid-based the power spatial density in some embodiments.


The scope of this disclosure should be determined by the claims, their legal equivalents and the fact that it fully encompasses other embodiments which may become apparent to those skilled in the art. All structural, electrical and functional equivalents to the elements of the above-described disclosure that are known to those of ordinary skill in the art are expressly incorporated herein by reference and are intended to be encompassed by the present claims. A reference to an element in the singular is not intended to mean one and only one, unless explicitly so stated, but rather it should be construed to mean at least one. No claim element herein is to be construed under the provisions of 35 U.S.C. § 112, sixth paragraph, unless the element is expressly recited using the phrase “means for.” Furthermore, no element, component or method step in the present disclosure is intended to be dedicated to the public, regardless of whether the element, component or method step is explicitly recited in the claims.


Embodiments of the inventive concepts disclosed herein have been described with reference to drawings. The drawings illustrate certain details of specific embodiments that implement the systems and methods and programs of the present disclosure. However, describing the embodiments with drawings should not be construed as imposing any limitations that may be present in the drawings. The present disclosure contemplates methods, systems and program products on any machine-readable media for accomplishing its operations. Embodiments of the inventive concepts disclosed herein may be implemented using an existing computer processor, or by a special purpose computer processor incorporated for this or another purpose or by a hardwired system.


Embodiments in the inventive concepts disclosed herein have been described in the general context of method steps which may be implemented in one embodiment by a program product including machine-executable instructions, such as program code, for example in the form of program modules executed by machines in networked environments. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Machine-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represent examples of corresponding acts for implementing the functions described in such steps.


It should be noted that although the diagrams herein may show a specific order and composition of method steps, it is understood that the order of these steps may differ from what is depicted. For example, two or more steps may be performed concurrently or with partial concurrence. Also, some method steps that are performed as discrete steps may be combined, steps being performed as a combined step may be separated into discrete steps, the sequence of certain processes may be reversed or otherwise varied, and the nature or number of discrete processes may be altered or varied. The order or sequence of any element or apparatus may be varied or substituted according to alternative embodiments. Accordingly, all such modifications are intended to be included within the scope of the present disclosure.


The foregoing description of embodiments has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the subject matter to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of the subject matter disclosed herein. The embodiments were chosen and described in order to explain the principals of the disclosed subject matter and its practical application to enable one skilled in the art to utilize the disclosed subject matter in various embodiments and with various modifications as are suited to the particular use contemplated. Other substitutions, modifications, changes and omissions may be made in the design, operating conditions and arrangement of the embodiments without departing from the scope of the presently disclosed subject matter.

Claims
  • 1. A weather radar system for an aircraft, the aircraft including aircraft sensors, the weather radar system comprising: an antenna; anda control circuit coupled with the antenna and configured to provide radar beams via the antenna toward external surroundings and configured to receive radar returns, wherein the control circuit is configured to process a collection of radar measurements from the radar returns, wherein each of the radar measurements is associated with a location determined using an antenna position, an antenna attitude, a beam sharpening angle, and a range, wherein power and the location of the radar measurements are processed to determine power density per grid cell using the location for each of the radar measurements, and wherein the power density per grid cell is used to provide an image associated with the power and the location of the radar measurements, wherein the beam sharpening angle provides a higher resolution value for angle to target than the antenna position and the antenna attitude due to antenna-induced poor angular resolution, wherein the radar beams and the radar returns are provided by the antenna, the antenna being part of the weather radar system, wherein the power density per grid cell is determined for each cell of a grid, wherein each cell is associated with a number of the locations associated with the radar measurements, wherein the power density per grid cell is determined for each cell using a volume or area of each cell and an accumulation of the radar measurements for each location of the locations in each grid cell.
  • 2. The weather radar system of claim 1, wherein the antenna position is determined using latitude, longitude and altitude sensed by the aircraft sensors.
  • 3. The weather radar system of claim 1, wherein the antenna attitude is determined using a direction of a longitudinal axis of the aircraft sensed by the aircraft sensors and a direction of a boresight of the antenna with respect to the direction of the longitudinal axis of the aircraft.
  • 4. The weather radar system of claim 1, wherein the antenna attitude is determined using an attitude sensor associated with the antenna.
  • 5. The weather radar system of claim 1, wherein the beam sharpening angle is a radar sensed direction of an individual radar measurement with respect to the antenna attitude.
  • 6. The weather radar system of claim 5, wherein the beam sharpening angle is generated using a monopulse technique, a sub-aperture radar technique, deconvolution of a beam point spread function, or any combination thereof.
  • 7. The weather radar system of claim 1, wherein the range is a radar sensed distance between the antenna and a returned reflection of an individual measurement.
  • 8. The weather radar system of claim 1, wherein the power density per grid cell is in a form of a two dimensional or three dimensional grid.
  • 9. A method of providing a real time sensor image on an electronic display using an X-band or C-band airborne weather radar system, the method comprising: determining an antenna position,determining an antenna attitude;receiving radar returns from an X-band or C-band airborne weather radar system;determining a location of each of a plurality of radar measurements associated with the radar returns using the antenna position, the antenna attitude, a beam sharpening angle, and a range; determining a power associated with the radar measurements;determining a power density per grid cell associated with the power and the location of the radar measurements, wherein the power density per grid cell is determined for each cell of a grid, wherein each cell is associated with an area or a volume greater than the location of each of the radar measurements such that a plurality of locations are located in each cell, wherein the power density per grid cell is determined for each cell using the volume or area of each cell and an accumulation of the radar measurements for each location of the locations in each grid cell; andusing the power density per grid cell to provide a real time sensor image on an electronic display, wherein the beam sharpening angle is determined by a beam sharpening process that improves antenna-induced poor angular resolution related to beam width.
  • 10. The method of claim 9, wherein the X-band or C-band airborne weather radar system comprises a switched aperture, sequential lobing or monopulse weather radar system, and wherein the real time sensor image is a three dimensional or two dimensional image.
  • 11. The method of claim 9, further comprising: providing radar beams associated with the radar returns, the radar beams being provided using beam sharpening techniques.
  • 12. The method of claim 11, wherein the beam sharpening techniques comprise a sub-aperture or split aperture technique.
  • 13. The method of claim 9, wherein the radar returns are provided using stepped frequency compression.
  • 14. An enhanced vision system, comprising: a weather radar system configured to generate image data representative of a runway environment associated with radar returns received by the weather radar system, the radar returns being in an X-band or a C-band, wherein the weather radar system is configured to process a collection of radar measurements from the radar returns, wherein each of the radar measurements is associated with a location determined using an antenna position, an antenna attitude, a beam sharpening angle, and a range, wherein power and the location of the radar measurements are processed to determine power density per grid cell associated with the power and the location of the radar measurements, wherein the power density per grid cell is used to provide an image, wherein the power density per grid cell is determined for each cell of a grid, wherein each cell is associated with an area or a volume greater than the location of each of the radar measurements, wherein the power density per grid cell is determined for each cell using the volume or area of each cell and an accumulation of the radar measurements for each location of the locations in each grid cell, wherein the beam sharpening angle provides a higher resolution value for angle to target than the antenna attitude due to beam width; anda display in communication with the weather radar system and configured to display the image associated with the image data.
  • 15. The enhanced vision system, of claim 14, wherein the location is determined using a slant range, wherein the slant range is a radar sensed distance between an antenna and a radar measurement associated with the radar returns.
  • 16. The enhanced vision system, of claim 14, wherein the antenna attitude is a direction of a boresight of an antenna.
  • 17. The enhanced vision system, of claim 14, wherein the beam sharpening angle is a radar sensed direction of a returned reflection of an individual measurement with respect to a boresight of an antenna.
  • 18. The enhanced vision system of claim 14, wherein the beam sharpening angle is generated using a monopulse technique, a sub-aperture radar technique, deconvolution of a beam point spread function, or any combination thereof.
  • 19. The enhanced vision system of claim 14, wherein the location is determined using a ground range, wherein the ground range is a radar sensed distance between an antenna and a radar measurement projected onto a flat ground surface.
  • 20. The enhanced vision system of claim 14, wherein the power density per grid cell is in a form of a two dimensional or three dimensional grid.
US Referenced Citations (540)
Number Name Date Kind
2416155 Chubb Feb 1947 A
2849184 Arden et al. Aug 1958 A
2929059 Parker Mar 1960 A
2930035 Altekruse Mar 1960 A
2948892 White Aug 1960 A
2965894 Sweeney Dec 1960 A
2994966 Senitsky et al. Aug 1961 A
3031660 Young Apr 1962 A
3049702 Schreitmueller Aug 1962 A
3064252 Varela Nov 1962 A
3070795 Chambers Dec 1962 A
3071766 Fenn Jan 1963 A
3072903 Meyer Jan 1963 A
3089801 Tierney et al. May 1963 A
3107351 Milam Oct 1963 A
3113310 Standing Dec 1963 A
3129425 Sanner Apr 1964 A
3153234 Begeman et al. Oct 1964 A
3175215 Blasberg et al. Mar 1965 A
3212088 Alexander et al. Oct 1965 A
3221328 Walter Nov 1965 A
3241141 Wall Mar 1966 A
3274593 Varela et al. Sep 1966 A
3325807 Burns et al. Jun 1967 A
3334344 Colby, Jr. Aug 1967 A
3339199 Pichafroy Aug 1967 A
3373423 Levy Mar 1968 A
3397397 Barney Aug 1968 A
3448450 Alfandari et al. Jun 1969 A
3618090 Garrison Nov 1971 A
3680094 Bayle et al. Jul 1972 A
3716855 Asam Feb 1973 A
3739380 Burdic et al. Jun 1973 A
3781878 Kirkpatrick Dec 1973 A
3810175 Bell May 1974 A
3815132 Case et al. Jun 1974 A
3816718 Hall et al. Jun 1974 A
3851758 Makhijani et al. Dec 1974 A
3866222 Young Feb 1975 A
3885237 Kirkpatrick May 1975 A
3956749 Magorian May 1976 A
4024537 Hart May 1977 A
4058701 Gruber et al. Nov 1977 A
4058710 Altmann Nov 1977 A
4063218 Basov et al. Dec 1977 A
4235951 Swarovski Nov 1980 A
4277845 Smith et al. Jul 1981 A
4359732 Martin Nov 1982 A
4405986 Gray Sep 1983 A
4435707 Clark Mar 1984 A
4481519 Margerum Nov 1984 A
4532515 Cantrell et al. Jul 1985 A
4594676 Breiholz et al. Jun 1986 A
4595925 Hansen Jun 1986 A
4598292 Devino Jul 1986 A
4628318 Alitz Dec 1986 A
4646244 Bateman et al. Feb 1987 A
4649388 Atlas Mar 1987 A
4654665 Kiuchi et al. Mar 1987 A
4685149 Smith et al. Aug 1987 A
4698635 Hilton Oct 1987 A
4760396 Barney et al. Jul 1988 A
4828382 Vermilion May 1989 A
4843398 Houston et al. Jun 1989 A
4912477 Lory et al. Mar 1990 A
4914436 Bateman et al. Apr 1990 A
4924401 Bice et al. May 1990 A
4939513 Paterson et al. Jul 1990 A
4951059 Taylor, Jr. Aug 1990 A
4953972 Zuk Sep 1990 A
4965573 Gallagher et al. Oct 1990 A
4987419 Salkeld Jan 1991 A
5045855 Moreira Sep 1991 A
5047779 Hager Sep 1991 A
5047781 Bleakney Sep 1991 A
5049886 Seitz et al. Sep 1991 A
5132693 Werp Jul 1992 A
5166688 Moreira Nov 1992 A
5173703 Mangiapane et al. Dec 1992 A
5175554 Mangiapane et al. Dec 1992 A
5198819 Susnjara Mar 1993 A
5202690 Frederick Apr 1993 A
5247303 Cornelius et al. Sep 1993 A
5273553 Hoshi et al. Dec 1993 A
5311183 Mathews et al. May 1994 A
5329391 Miyamoto et al. Jul 1994 A
5332998 Avignon et al. Jul 1994 A
5345241 Huddle Sep 1994 A
5365356 McFadden Nov 1994 A
5442364 Lee et al. Aug 1995 A
5469167 Polge Nov 1995 A
5481270 Urkowitz Jan 1996 A
5539409 Mathews et al. Jul 1996 A
5559515 Alimena et al. Sep 1996 A
5559518 Didomizio Sep 1996 A
5566840 Waldner et al. Oct 1996 A
5592178 Chang et al. Jan 1997 A
5678303 Wichmann Oct 1997 A
5736957 Raney Apr 1998 A
5820080 Eschenbach Oct 1998 A
5828332 Frederick Oct 1998 A
5831570 Ammar et al. Nov 1998 A
5839080 Muller et al. Nov 1998 A
5847673 DeBell Dec 1998 A
5867119 Corrubia et al. Feb 1999 A
5894286 Morand et al. Apr 1999 A
5900833 Sunlin May 1999 A
5918517 Malapert et al. Jul 1999 A
5920276 Frederick Jul 1999 A
5923279 Bamler et al. Jul 1999 A
5936575 Azzarelli et al. Aug 1999 A
5942062 Hassall et al. Aug 1999 A
5945926 Ammar et al. Aug 1999 A
5950512 Fields Sep 1999 A
5959762 Bandettini et al. Sep 1999 A
5978715 Briffe et al. Nov 1999 A
6002347 Daly et al. Dec 1999 A
6023240 Sutton Feb 2000 A
6061016 Lupinski et al. May 2000 A
6061022 Menegozzi et al. May 2000 A
6064942 Johnson et al. May 2000 A
6075484 Daniel et al. Jun 2000 A
6092009 Glover Jul 2000 A
6112141 Briffe et al. Aug 2000 A
6112570 Hruschak Sep 2000 A
6122570 Muller et al. Sep 2000 A
6127944 Daly et al. Oct 2000 A
6128066 Yokozeki Oct 2000 A
6128553 Gordon et al. Oct 2000 A
6138060 Conner et al. Oct 2000 A
6150901 Auken Nov 2000 A
6154151 McElreath et al. Nov 2000 A
6154169 Kuntman Nov 2000 A
6157339 Sato et al. Dec 2000 A
6157891 Lin Dec 2000 A
6163021 Mickelson Dec 2000 A
6166661 Anderson et al. Dec 2000 A
6169770 Henely Jan 2001 B1
6178391 Anderson et al. Jan 2001 B1
6184816 Zheng et al. Feb 2001 B1
6188330 Glover Feb 2001 B1
6194980 Thon Feb 2001 B1
6199008 Aratow et al. Mar 2001 B1
6201494 Kronfeld Mar 2001 B1
6204806 Hoech Mar 2001 B1
6205400 Lin Mar 2001 B1
6208284 Woodell et al. Mar 2001 B1
6219592 Muller et al. Apr 2001 B1
6233522 Morici May 2001 B1
6236351 Conner et al. May 2001 B1
6259400 Higgins et al. Jul 2001 B1
6266114 Skarohlid Jul 2001 B1
6278799 Hoffman Aug 2001 B1
6281832 McElreath Aug 2001 B1
6285298 Gordon Sep 2001 B1
6285337 West et al. Sep 2001 B1
6285926 Weiler et al. Sep 2001 B1
6289277 Feyereisen et al. Sep 2001 B1
6311108 Ammar et al. Oct 2001 B1
6317468 Meyer Nov 2001 B1
6317690 Gia Nov 2001 B1
6317872 Gee et al. Nov 2001 B1
6340946 Wolfson et al. Jan 2002 B1
6345127 Mitchell Feb 2002 B1
6359585 Bechman et al. Mar 2002 B1
6366013 Leenders et al. Apr 2002 B1
6373418 Abbey Apr 2002 B1
6374286 Gee et al. Apr 2002 B1
6377202 Kropfli et al. Apr 2002 B1
6377892 Johnson et al. Apr 2002 B1
6388607 Woodell May 2002 B1
6388608 Woodell et al. May 2002 B1
6388724 Campbell et al. May 2002 B1
6389354 Hicks et al. May 2002 B1
6401038 Gia Jun 2002 B2
6411890 Zimmerman Jun 2002 B1
6421000 McDowell Jul 2002 B1
6421603 Pratt et al. Jul 2002 B1
6424288 Woodell Jul 2002 B1
6426717 Maloratsky Jul 2002 B1
6426720 Ross et al. Jul 2002 B1
6427122 Lin Jul 2002 B1
6430480 Ammar Aug 2002 B1
6441773 Kelly et al. Aug 2002 B1
6445310 Bateman et al. Sep 2002 B1
6448922 Kelly Sep 2002 B1
6452511 Kelly et al. Sep 2002 B1
6456236 Hauck et al. Sep 2002 B1
6456238 Posey Sep 2002 B1
6462703 Hedrick Oct 2002 B2
6473026 Ali-Mehenni et al. Oct 2002 B1
6473037 Vail et al. Oct 2002 B2
6473240 Dehmlow Oct 2002 B1
6481482 Shimotomai Nov 2002 B1
6492934 Hwang et al. Dec 2002 B1
6501424 Haendel et al. Dec 2002 B1
6512476 Woodell Jan 2003 B1
6512527 Barber et al. Jan 2003 B1
6516272 Lin Feb 2003 B2
6516283 McCall et al. Feb 2003 B2
6520056 Nemeth et al. Feb 2003 B1
6525674 Kelly et al. Feb 2003 B1
6531669 Miller et al. Mar 2003 B1
6549161 Woodell Apr 2003 B1
6567728 Kelly et al. May 2003 B1
6574030 Mosier Jun 2003 B1
6577947 Kronfeld et al. Jun 2003 B1
6590528 Dewulf Jul 2003 B1
6591171 Ammar et al. Jul 2003 B1
6593875 Bergin et al. Jul 2003 B2
6600443 Landt Jul 2003 B2
6603425 Woodell Aug 2003 B1
6614057 Silvernail et al. Sep 2003 B2
6650275 Kelly et al. Nov 2003 B1
6650291 West et al. Nov 2003 B1
6653947 Dwyer et al. Nov 2003 B2
6667710 Cornell et al. Dec 2003 B2
6681668 Smirle Jan 2004 B1
6690298 Barber et al. Feb 2004 B1
6690299 Suiter Feb 2004 B1
6690317 Szeto et al. Feb 2004 B2
6697008 Sternowski Feb 2004 B1
6697012 Lodwig et al. Feb 2004 B2
6710663 Berquist Mar 2004 B1
6714186 Mosier et al. Mar 2004 B1
6724344 Stockmaster et al. Apr 2004 B1
6731236 Hager et al. May 2004 B1
6738011 Evans May 2004 B1
6739929 Furukawa et al. May 2004 B2
6741203 Woodell May 2004 B1
6741208 West et al. May 2004 B1
6744382 Lapis et al. Jun 2004 B1
6744408 Stockmaster Jun 2004 B1
6757624 Hwang et al. Jun 2004 B1
6760155 Murayama et al. Jul 2004 B2
6771626 Golubiewski et al. Aug 2004 B1
6782392 Weinberger et al. Aug 2004 B1
6799095 Owen et al. Sep 2004 B1
6803245 Auch et al. Oct 2004 B2
6804614 McGraw et al. Oct 2004 B1
6806846 West Oct 2004 B1
6807538 Weinberger et al. Oct 2004 B1
6813777 Weinberger et al. Nov 2004 B1
6819983 McGraw Nov 2004 B1
6822617 Mather et al. Nov 2004 B1
6825804 Doty Nov 2004 B1
6832538 Hwang Dec 2004 B1
6839017 Dillman Jan 2005 B1
6842288 Liu et al. Jan 2005 B1
6850185 Woodell Feb 2005 B1
6862323 Loper Mar 2005 B1
6862501 He Mar 2005 B2
6865452 Burdon Mar 2005 B2
6879280 Bull et al. Apr 2005 B1
6879886 Wilkins et al. Apr 2005 B2
6882302 Woodell et al. Apr 2005 B1
6908202 Graf et al. Jun 2005 B2
6917396 Hiraishi et al. Jul 2005 B2
6918134 Sherlock et al. Jul 2005 B1
6933885 Stockmaster et al. Aug 2005 B1
6938258 Weinberger et al. Aug 2005 B1
6950062 Mather et al. Sep 2005 B1
6959057 Tuohino Oct 2005 B1
6972727 West et al. Dec 2005 B1
6977608 Anderson et al. Dec 2005 B1
6984545 Grigg et al. Jan 2006 B2
6990022 Morikawa et al. Jan 2006 B2
6992614 Joyce Jan 2006 B1
6995726 West et al. Feb 2006 B1
6998648 Silvernail Feb 2006 B2
6998908 Sternowski Feb 2006 B1
6999022 Vesel et al. Feb 2006 B1
6999027 Stockmaster Feb 2006 B1
7002546 Stuppi et al. Feb 2006 B1
7010398 Wilkins et al. Mar 2006 B2
7023375 Klausing et al. Apr 2006 B2
7026956 Wenger et al. Apr 2006 B1
7028304 Weinberger et al. Apr 2006 B1
7030945 Umemoto et al. Apr 2006 B2
7034753 Elsallal et al. Apr 2006 B1
7042387 Ridenour et al. May 2006 B2
7053796 Barber May 2006 B1
7057549 Block Jun 2006 B2
7064680 Reynolds et al. Jun 2006 B2
7069120 Koenck et al. Jun 2006 B1
7089092 Wood et al. Aug 2006 B1
7092645 Sternowski Aug 2006 B1
7098913 Etherington et al. Aug 2006 B1
7109912 Paramore et al. Sep 2006 B1
7109913 Paramore et al. Sep 2006 B1
7123260 Brust Oct 2006 B2
7129885 Woodell et al. Oct 2006 B1
7145501 Manfred et al. Dec 2006 B1
7148816 Carrico Dec 2006 B1
7151507 Herting Dec 2006 B1
7158072 Venkatachalam et al. Jan 2007 B1
7161525 Finley et al. Jan 2007 B1
7170446 West et al. Jan 2007 B1
7170959 Abbey Jan 2007 B1
7180476 Guell et al. Feb 2007 B1
7191406 Barber et al. Mar 2007 B1
7196329 Wood et al. Mar 2007 B1
7205933 Snodgrass Apr 2007 B1
7209070 Gilliland et al. Apr 2007 B2
7212216 He et al. May 2007 B2
7218268 Vandenberg May 2007 B2
7219011 Barber May 2007 B1
7236125 Tullsson Jun 2007 B2
7242343 Woodell Jul 2007 B1
7242345 Raestad et al. Jul 2007 B2
7250903 McDowell Jul 2007 B1
7265710 Deagro Sep 2007 B2
7269657 Alexander et al. Sep 2007 B1
7272472 McElreath Sep 2007 B1
7273403 Yokota et al. Sep 2007 B2
7280068 Lee et al. Oct 2007 B2
7289058 Shima Oct 2007 B2
7292178 Woodell et al. Nov 2007 B1
7292180 Schober Nov 2007 B2
7295150 Burlet et al. Nov 2007 B2
7295901 Little et al. Nov 2007 B1
7301496 Honda et al. Nov 2007 B2
7307576 Koenigs Dec 2007 B1
7307577 Kronfeld et al. Dec 2007 B1
7307583 Woodell et al. Dec 2007 B1
7312725 Berson et al. Dec 2007 B2
7312743 Ridenour et al. Dec 2007 B2
7317427 Pauplis et al. Jan 2008 B2
7321332 Focke et al. Jan 2008 B2
7337043 Bull Feb 2008 B2
7349154 Aiura et al. Mar 2008 B2
7352292 Alter et al. Apr 2008 B2
7361240 Kim Apr 2008 B2
7372394 Woodell et al. May 2008 B1
7373223 Murphy May 2008 B2
7375678 Feyereisen et al. May 2008 B2
7379014 Woodell et al. May 2008 B1
7379796 Walsdorf et al. May 2008 B2
7381110 Sampica et al. Jun 2008 B1
7417578 Woodell et al. Aug 2008 B1
7417579 Woodell Aug 2008 B1
7423578 Tietjen Sep 2008 B1
7446697 Burlet et al. Nov 2008 B2
7446938 Miyatake et al. Nov 2008 B2
7452258 Marzen et al. Nov 2008 B1
7474262 Alland Jan 2009 B2
7479920 Niv Jan 2009 B2
7486220 Kronfeld et al. Feb 2009 B1
7486291 Berson et al. Feb 2009 B2
7492304 Woodell et al. Feb 2009 B1
7492305 Woodell et al. Feb 2009 B1
7515087 Woodell et al. Apr 2009 B1
7515088 Woodell et al. Apr 2009 B1
7525448 Wilson et al. Apr 2009 B1
7528765 Woodell et al. May 2009 B1
7528915 Choi et al. May 2009 B2
7541970 Godfrey et al. Jun 2009 B1
7541971 Woodell et al. Jun 2009 B1
7551451 Kim et al. Jun 2009 B2
7557735 Woodell et al. Jul 2009 B1
7566254 Sampica et al. Jul 2009 B2
7570177 Reynolds et al. Aug 2009 B2
7576680 Woodell Aug 2009 B1
7603209 Dwyer et al. Oct 2009 B2
7609200 Woodell et al. Oct 2009 B1
7612706 Honda et al. Nov 2009 B2
7616150 Woodell Nov 2009 B1
7633428 McCusker et al. Dec 2009 B1
7633430 Wichgers et al. Dec 2009 B1
7633584 Umemoto et al. Dec 2009 B2
7639175 Woodell Dec 2009 B1
7664601 Daly, Jr. Feb 2010 B2
7675461 McCusker et al. Mar 2010 B1
7693621 Chamas Apr 2010 B1
7696921 Finley et al. Apr 2010 B1
7714767 Kronfeld et al. May 2010 B1
7733264 Woodell et al. Jun 2010 B1
7783427 Woodell et al. Aug 2010 B1
7783429 Walden et al. Aug 2010 B2
7791529 Filias et al. Sep 2010 B2
7808422 Woodell et al. Oct 2010 B1
7814676 Sampica et al. Oct 2010 B2
7843380 Woodell Nov 2010 B1
7859448 Woodell et al. Dec 2010 B1
7859449 Woodell et al. Dec 2010 B1
7864103 Weber et al. Jan 2011 B2
7868811 Woodell et al. Jan 2011 B1
7872594 Vesel Jan 2011 B1
7889117 Woodell et al. Feb 2011 B1
7889118 Finley et al. Feb 2011 B1
7927440 Matsuhira Apr 2011 B2
7929086 Toyama et al. Apr 2011 B2
7965223 McCusker Jun 2011 B1
7965225 Dickerson et al. Jun 2011 B1
8035547 Flanigan et al. Oct 2011 B1
8038498 Miyauchi et al. Oct 2011 B2
8045098 Kitagawa et al. Oct 2011 B2
8059025 D'Addio Nov 2011 B2
8068050 Christianson Nov 2011 B2
8068984 Smith et al. Nov 2011 B2
8072368 Woodell Dec 2011 B1
8077078 Woodell et al. Dec 2011 B1
8102487 Kitagawa et al. Jan 2012 B2
8118075 Sampica et al. Feb 2012 B2
8137498 Sampica et al. Mar 2012 B2
8140223 Whitehead et al. Mar 2012 B2
8159464 Gribble et al. Apr 2012 B1
8232917 Scherzinger et al. Jul 2012 B2
8296065 Haynie et al. Oct 2012 B2
8373580 Bunch et al. Feb 2013 B2
8410975 Bell et al. Apr 2013 B1
8477062 Kanellis Jul 2013 B1
8486535 Nemeth et al. Jul 2013 B1
8493241 He Jul 2013 B2
8515600 McCusker Aug 2013 B1
8540002 Sampica et al. Sep 2013 B2
8558731 Woodell Oct 2013 B1
8576112 Garrec et al. Nov 2013 B2
8583315 Whitehead et al. Nov 2013 B2
8594879 Roberge et al. Nov 2013 B2
8603288 Sampica et al. Dec 2013 B2
8634993 McClure et al. Jan 2014 B2
8639416 Jones et al. Jan 2014 B2
8643533 Woodell et al. Feb 2014 B1
8691043 Sampica et al. Apr 2014 B2
8717226 Bon et al. May 2014 B2
8773301 Woodell Jul 2014 B1
8896480 Wilson et al. Nov 2014 B1
8909471 Jinkins et al. Dec 2014 B1
8917191 Tiana et al. Dec 2014 B1
8936057 Sampica et al. Jan 2015 B2
9354633 McCusker et al. May 2016 B1
9500746 Miles Nov 2016 B2
9529081 Whelan Dec 2016 B2
9613269 Kilty Apr 2017 B2
10037124 Khatwa Jul 2018 B2
20010023390 Gia Sep 2001 A1
20010050372 Krijn et al. Dec 2001 A1
20010053648 Furukawa et al. Dec 2001 A1
20020039070 Ververs et al. Apr 2002 A1
20020111717 Scherzinger et al. Aug 2002 A1
20020116125 Lin Aug 2002 A1
20020116126 Lin Aug 2002 A1
20020158256 Yamada et al. Oct 2002 A1
20020179229 Chuzles Dec 2002 A1
20020185600 Kerr Dec 2002 A1
20020187284 Kinoshita et al. Dec 2002 A1
20030021491 Brust Jan 2003 A1
20030038916 Nakano et al. Feb 2003 A1
20030043315 Umemoto et al. Mar 2003 A1
20030071828 Wilkins et al. Apr 2003 A1
20030089214 Fukuta et al. May 2003 A1
20030093187 Walker May 2003 A1
20030102999 Bergin et al. Jun 2003 A1
20030156238 Hiraishi et al. Aug 2003 A1
20030160718 Nagasaku Aug 2003 A1
20030174396 Murayama et al. Sep 2003 A1
20030180528 Flosenzier et al. Sep 2003 A1
20030189606 Moon et al. Oct 2003 A1
20030195672 He Oct 2003 A1
20030216859 Martell et al. Nov 2003 A1
20030222887 Wilkins et al. Dec 2003 A1
20040044445 Burdon Mar 2004 A1
20040059473 He Mar 2004 A1
20040066645 Graf et al. Apr 2004 A1
20040072575 Young et al. Apr 2004 A1
20040083038 He Apr 2004 A1
20040160341 Feyereisen et al. Aug 2004 A1
20040160364 Regev Aug 2004 A1
20040181318 Redmond et al. Sep 2004 A1
20040264549 Hoole Dec 2004 A1
20050004748 Pinto et al. Jan 2005 A1
20050052451 Servantie Mar 2005 A1
20050126679 Kim Jun 2005 A1
20050136625 Henseler et al. Jun 2005 A1
20050150289 Osborne Jul 2005 A1
20050174350 Ridenour et al. Aug 2005 A1
20050200502 Reusser et al. Sep 2005 A1
20050230563 Corcoran, III Oct 2005 A1
20060004497 Bull Jan 2006 A1
20060097895 Reynolds et al. May 2006 A1
20060098452 Choi et al. May 2006 A1
20060164284 Pauplis et al. Jul 2006 A1
20060207967 Bocko et al. Sep 2006 A1
20060215265 Miyatake et al. Sep 2006 A1
20060227012 He Oct 2006 A1
20060244636 Rye et al. Nov 2006 A1
20060245171 Kim et al. Nov 2006 A1
20060290253 Yeo et al. Dec 2006 A1
20060290531 Reynolds et al. Dec 2006 A1
20070001897 Alland Jan 2007 A1
20070002078 He et al. Jan 2007 A1
20070008214 Wasiewicz Jan 2007 A1
20070013575 Lee et al. Jan 2007 A1
20070018887 Feyereisen et al. Jan 2007 A1
20070032951 Tanenhaus et al. Feb 2007 A1
20070060063 Wright et al. Mar 2007 A1
20070146364 Aspen Jun 2007 A1
20070171094 Alter et al. Jul 2007 A1
20070176794 Feyereisen et al. Aug 2007 A1
20070179684 He Aug 2007 A1
20070228586 Merrill et al. Oct 2007 A1
20070247350 Ryan Oct 2007 A1
20070279253 Priest Dec 2007 A1
20070297736 Sherman et al. Dec 2007 A1
20080018524 Christianson Jan 2008 A1
20080051947 Kemp Feb 2008 A1
20080074308 Becker et al. Mar 2008 A1
20080111731 Hubbard et al. May 2008 A1
20080145610 Muller et al. Jun 2008 A1
20080180351 He Jul 2008 A1
20080305721 Ohashi et al. Dec 2008 A1
20090040070 Alter et al. Feb 2009 A1
20090040772 Laney Feb 2009 A1
20090046229 Umemoto et al. Feb 2009 A1
20090148682 Higuchi Jun 2009 A1
20090152391 McWhirk Jun 2009 A1
20090153783 Umemoto Jun 2009 A1
20090164067 Whitehead et al. Jun 2009 A1
20090207048 He et al. Aug 2009 A1
20090279030 Toyama et al. Nov 2009 A1
20090279175 Laney et al. Nov 2009 A1
20100033499 Gannon et al. Feb 2010 A1
20100103353 Yamada Apr 2010 A1
20100297406 Schaffer et al. Nov 2010 A1
20100312428 Roberge et al. Dec 2010 A1
20100312461 Haynie et al. Dec 2010 A1
20110054729 Whitehead et al. Mar 2011 A1
20110075070 Kitagawa et al. Mar 2011 A1
20110141405 Kitagawa et al. Jun 2011 A1
20110165361 Sherman et al. Jul 2011 A1
20110184594 Manfred et al. Jul 2011 A1
20110282580 Mohan Nov 2011 A1
20120053831 Halder Mar 2012 A1
20120150426 Conway Jun 2012 A1
20120174445 Jones et al. Jul 2012 A1
20120215410 McClure et al. Aug 2012 A1
20130041529 He et al. Feb 2013 A1
20160011334 Khatwa Jan 2016 A1
20160131739 Jinkins et al. May 2016 A1
Foreign Referenced Citations (23)
Number Date Country
196 49 838 Apr 1998 DE
0 556 351 Jun 1995 EP
0 962 752 Dec 1999 EP
0 814 744 Jun 1959 GB
01-210328 Aug 1989 JP
05-200880 Aug 1993 JP
05-293895 Nov 1993 JP
06-051484 Feb 1994 JP
H08-220547 Aug 1996 JP
09-057779 Mar 1997 JP
10-156853 Jun 1998 JP
10-244589 Sep 1998 JP
2000-141388 May 2000 JP
2004-233590 Aug 2004 JP
2004-354645 Dec 2004 JP
2006-218658 Aug 2006 JP
2006-334912 Dec 2006 JP
2006-348208 Dec 2006 JP
2007-206559 Aug 2007 JP
2007-302398 Nov 2007 JP
2008-238607 Jan 2008 JP
WO-9305634 Mar 1993 WO
WO-2011089474 Jul 2011 WO
Non-Patent Literature Citations (143)
Entry
Non-Final Office Action on U.S. Appl. No. 13/250,798 dated Feb. 26, 2016, 9 pages.
Notice of Allowance on U.S. Appl. No. 12/263,282 dated Jan. 29, 2016, 8 pages.
Notice of Allowance on U.S. Appl. No. 14/301,199 dated Mar. 1, 2016, 11 pages.
U.S. Appl. No. 11/851,323, filed Sep. 6, 2007, McCusker.
U.S. Appl. No. 11/863,219, filed Sep. 27, 2007, Woodell.
U.S. Appl. No. 11/863,221, filed Sep. 27, 2007, Woodell.
U.S. Appl. No. 11/899,801, filed Sep. 6, 2007, Woodell et al.
U.S. Appl. No. 11/900,002, filed Sep. 6, 2007, Woodell et al.
U.S. Appl. No. 12/167,200, filed Jul. 2, 2008, Woodell et al.
U.S. Appl. No. 12/167,203, filed Jul. 2, 2008, Woodell.
U.S. Appl. No. 12/167,208, filed Jul. 2, 2008, Dickerson et al.
U.S. Appl. No. 12/180,293, filed Jul. 25, 2008, Woodell et al.
U.S. Appl. No. 12/236,464, filed Sep. 23, 2008, Rockwell Collins.
U.S. Appl. No. 12/786,169, filed May 24, 2010, Nemeth et al.
U.S. Appl. No. 13/224,992, filed Sep. 2, 2011, Hufnagel et al.
U.S. Appl. No. 13/250,307, filed Sep. 30, 2011, Jinkins.
U.S. Appl. No. 13/250,798, filed Sep. 30, 2011, Jinkins.
U.S. Appl. No. 13/627,788, filed Sep. 26, 2012, Rockwell Collins.
U.S. Appl. No. 13/857,955, filed Apr. 5, 2013, Barber et al.
U.S. Appl. No. 13/250,798, filed Sep. 30, 2011, Rockwell Collins.
U.S. Appl. No. 14/301,199, filed Jun. 10, 2014, Rockwell Collins.
U.S. Appl. No. 14/482,681, filed Sep. 10, 2014, Rockwell Collins.
“MountainScope™ on a TabletPC,” PCAvionics™, printed from website www.pcavionics.com on Aug. 28, 2007, 1 page.
TAWS Class A and Class B, Terrain Awareness and Warning Systems, Universal® Avionics Systems Corporation, Sep. 2007, 6 pages.
“TAWS Terrain Awareness and Warning System,” Universal® Avionics, printed from website www.uasc.com on Aug. 28, 2007, 2 pages.
Adams, Charlotte, “Synthetic Vision: Picturing the Future,” Avionics magazine, Oct. 1, 2006, printed from website www.aviationtoday.com, 4 pages.
Adams, Charlotte, “Synthetic Vision: Picturing the Future,” Avionics magazine, Solutions for Global Airspace Electronics, Oct. 2006, cover and pp. 22-29.
Airports Authority of India, Chapter 7: Visual Aids for Navigation—Lights, available prior to Jan. 1, 2005, retrieved from the internet at: http://www.aai.aero/aai_employees/chapter_7.pdf on Sep. 26, 2014, 33 pages.
Blue Mountain Avionics' Products, printed from website www.bluemountainavionics.com on Aug. 28, 2007, 4 pages.
Brailovsky et al., REVS122: A Radar-Based Enhanced Vision System for Degraded Visual Environments, Proc. of SPIE vol. 9087 908708-1, retrieved from the internet at http://proceedings.spiedigitallibrary.org on Jun. 25, 2014, 13 pages.
Carter, S. P., D. D. Blankenship, M. E. Peters, D. A. Young, J. W. Holt, and D. L. Morse (2007), Radar-based subglacial lake classification in Antarctica, Geochem. Geophys. Geosyst., 8, 003016, doi:10.1029/2006GC001408, 20 pages.
Federal Aviation Administration, Advisory Circular AC 90-106, “Enhanced Flight Vision Systems”, initiated by AFS-400, dated Jun. 2, 2010, 55 pages.
Federal Aviation Administration, Aeronautical Information Manual (AIM) Basic Flight Information and ATC Procedures, dated Jul. 24, 2014, 2 pages.
Final Office Action on U.S. Appl. No. 13/250,798 dated Sep. 4, 2014, 22 pages.
Final Office Action on U.S. Appl. No. 13/867,556 dated Jul. 3, 2014, 11 pages.
Final Office Action on U.S. Appl. No. 13/250,307 dated Jun. 11, 2014, 8 pages.
Final Office Action on U.S. Appl. No. 12/236,464, dated Dec. 5, 2011, 15 pages.
Final Office Action on U.S. Appl. No. 13/250,798 dated Aug. 7, 2015, 21 pages.
Fountain, J.R., Digital Terrain Systems, Airborne Navigation Systems Workshop (Digest No. 1997/169), IEE Colloquium, pp. 4/1-4/6, Feb. 21, 1997.
G2000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=97668 on Jun. 28, 2011, 2 pages.
G3000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=66916 on Jun. 28, 2011, 2 pages.
G5000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=90821&ra=true on Apr. 20, 2011, 2 pages.
Honeywell, RDR-4B Forward looking windshear detection / weather radar system user's manual with radar operating guidelines, Rev. 6, Jul. 2003, 106 pages.
Johnson, A., et al., Vision Guided Landing of an Autonomous Helicopter in Hazardous Terrain, Robotics and Automation, 2005. ICRA 2005. Proceedings of the 2005 IEEE International Conference, pp. 3966-3971, Apr. 18-22, 2005.
Kuntman, D., Airborne system to address leading cause of injuries in non-fatal airline accidents, ICAO Journal, Mar. 2000, 4 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798 dated Mar. 18, 2015, 21 pages.
Non-Final Office Action on U.S. Appl. No. 14/301,199 dated Sep. 9, 2015, 18 pages.
Notice of Allowance for U.S. Appl. No. 11/863,215, dated Oct. 13, 2009, 8 pages.
Notice of Allowance for U.S. Appl. No. 11/863,219, dated Jun. 23, 2009, 7 pages.
Notice of Allowance for U.S. Appl. No. 11/863,221, dated Aug. 2, 2010, 9 pages.
Notice of Allowance for U.S. Appl. No. 11/899,801, dated Aug. 19, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 11/900,002, dated Sep. 14, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 12/009,372, dated Oct. 13, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/009,373, dated Jun. 16, 2010, 4 pages.
Notice of Allowance for U.S. Appl. No. 12/009,472, dated Sep. 5, 2013, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/167,200, dated Oct. 28, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 12/167,203, dated Jun. 21, 2013, 7 pages.
Notice of Allowance for U.S. Appl. No. 12/167,208, dated Mar. 21, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/180,293, dated Aug. 4, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/786,169, dated Mar. 28, 2013, 6 pages.
Notice of Allowance for U.S. Appl. No. 13/538,957, dated Oct. 3, 2013, 13 pages.
Notice of Allowance on U.S. Appl. No. 13/241,051 dated Aug. 28, 2014, 9 pages.
Notice of Allowance on U.S. Appl. No. 13/247,742 dated Jul. 30, 2014, 9 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Dec. 15, 2010, 13 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Aug. 6, 2009, 23 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Jul. 5, 2012, 23 pages.
Office Action for U.S. Appl. No. 11/863,215, dated May 27, 2009, 5 pages.
Office Action for U.S. Appl. No. 11/863,215, dated Nov. 12, 2008, 8 pages.
Office Action for U.S. Appl. No. 11/863,219, dated Dec. 12, 2008, 7 pages.
Office Action for U.S. Appl. No. 11/863,221, dated Dec. 18, 2009, 5 pages.
Office Action for U.S. Appl. No. 11/863,221, dated Dec. 8, 2008, 8 pages.
Office Action for U.S. Appl, No. 11/863,221, dated May 26, 2009, 5 pages.
Office Action for U.S. Appl. No. 12/009,372, dated Dec. 20, 2010, 10 pages.
Office Action for U.S. Appl. No. 12/009,372, dated Jun. 13, 2011, 9 pages.
Office Action for U.S. Appl. No. 12/009,373, dated Dec. 30, 2009, 14 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Apr. 16, 2012, 16 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Jan. 14, 2011, 14 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Mar. 20, 2013, 15 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Nov. 3, 2011, 15 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Nov. 9, 2012, 15 pages.
Office Action for U.S. Appl. No. 12/167,200, dated Jul. 21, 2010, 6 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Aug. 26, 2010, 9 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Jul. 20, 2011, 6 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Mar. 7, 2013, 5 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Oct. 31, 2011, 5 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Sep. 21, 2012, 6 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Dec. 30, 2009, 10 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Feb. 7, 2011, 8 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Jun. 3, 2010, 11 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Oct. 19, 2010, 8 pages.
Office Action for U.S. Appl. No. 12/180,293, dated Jan. 4, 2011, 5 pages.
Office Action for U.S. Appl. No. 12/180,293, dated Jul. 28, 2010, 8 pages.
Office Action for U.S. Appl. No. 12/263,282, dated Jan. 5, 2012, 10 pages.
Office Action for U.S. Appl. No. 12/786,169, dated Jan. 18, 2013, 14 pages.
Office Action for U.S. Appl. No. 12/892,563, dated Feb. 19, 2013, 12 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Feb. 15, 2012, 8 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Jul. 10, 2012, 4 pages.
Office Action for U.S. Appl. No. 12/976,871, dated May 6, 2013, 5 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Nov. 21, 2012, 5 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Oct. 9, 2013, 5 pages.
Office Action for U.S. Appl. No. 13/183,314, dated Aug. 14, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/183,314, dated Mar. 28, 2013, 12 pages.
Office Action for U.S. Appl. No. 13/224,992, dated Feb. 28, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/250,307, dated Nov. 5, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/474,559, dated Aug. 28, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/474,559, dated Dec. 28, 2012, 8 pages.
Office Action for U.S. Appl. No. 13/538,957, dated Apr. 4, 2013, 19 pages.
Office Action for U.S. Appl. No. 13/538,957, dated Oct. 5, 2012, 18 pages.
Office Action for U.S. Appl. No. 13/743,182, dated Apr. 8, 2013, 10 pages.
Office Action for U.S. Appl. No. 12/786,169, dated Jul. 20, 2012, 8 pages.
Office Action in Japanese Patent Application 2015-116688, dated Aug. 25, 2015, 4 pages.
Office Action in Japanese Patent Application 2015-116716, dated Aug. 25, 2015, 3 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Jul. 12, 2013, 17 pages.
Office Action on U.S. Appl. No. 11/787,460, dated Mar. 19, 2010, 16 pages.
Office Action on U.S. Appl. No. 11/787,460, dated Sep. 16, 2009, 15 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Feb. 11, 2014, 21 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Jun. 22, 2011, 14 pages.
Office Action on U.S. Appl. No. 12/892,563, dated May 7, 2013, 6 pages.
Office Action on U.S. Appl. No. 12/892,563, dated Oct. 10, 2012, 12 pages.
Office Action on U.S. Appl. No. 13/241,051 dated Feb. 27, 2014, 21 pages.
Office Action on U.S. Appl. No. 13/247,742 dated Dec. 3, 2013, 11 pages.
Office Action on U.S. Appl. No. 13/250,798 dated Apr. 23, 2014, 15 pages.
Office Action on U.S. Appl. No. 13/627,788, dated Jul. 28, 2014, 10 pages.
Office Action on U.S. Appl. No. 13/867,556 dated Feb. 7, 2014, 11 pages.
Office Action U.S. Appl. No. 11/787,460, dated Aug. 31, 2010, 18 pages.
Office Action with English Translation received in Korean Patent Application 10-2010-7017278, dated Aug. 26, 2015, 5 pages.
Pictures of DELPHINS, printed from website www.tunnel-in-the-sky.tudelft.nl on Aug. 28, 2007, 4 pages.
REVS Product Information Sheet, Sierra Nevada Corporation, dated May 7, 2014, 2 pages.
Skolnik, Introduction to Radar Systems, McGraw Hill Book Company, 2001, 3 pages.
Skolnik, Radar Handbook (McGraw Hill Book Company), 1990, 23 pages.
Synthetic Vision System, en.wikipedia.org/wiki/Synthetic_vision_system, retrieved Feb. 28, 2013, 4 pages.
Technical Standard Order, TSO-C115b, Airborne Area Navigation Equipment Using Multi-Sensor Inputs, Department of Transportation, Federal Aviation Administration, Sep. 30, 1994, 11 pages.
U.S. Office Action on U.S. Appl. No. 11/900,002 dated Jun. 8, 2010, 7 pages.
U.S. Office Action on U.S. Appl. No. 13/247,742 dated Apr. 16, 2014, 15 pages.
Vadlamani, A., et al., Improving the detection capability of spatial failure modes using downward-looking sensors in terrain database integrity monitors, Digital Avionics Systems Conference, 2003. DASC-03. The 22nd, vol. 2, pp. 9C.5-91-12 vol. 2, Oct. 12-16, 2003.
Van Kasteren, Joost, “Tunnel-in-the-Sky, Synthetic vision simplifies the pilot's job and enhances safety,” printed from website www.delftoutlook.tudelft.nl on Aug. 28, 2007, 13 pages.
Walker, GD-Itronix Dynavue Technology, The Ultimate Outdoor-Readable Touch-Screen Display, Rugged PC Review, 4 pages.
Wang et al., A Simple Based on DSP Antenna Controller of Weather Radar, 2001 CIE International Conference, 4 pages.
First Office Action on Korean Patent Application No. 10-2016-7013740, dated Sep. 19, 2016, 7 pages.
McGray et al., Air Operators, Airlines, Manufacturers and Interested Industry Stakeholders & Aero Chart Forum-Utilizing EFVS technology and incorporating it into Faa NextGen, Federal Aviation Administration, Apr. 23 & 30, 2014, 34 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798, dated Sep. 9, 2016, 6 pages.
Non-Final Office Action on U.S. Appl. No. 14/536,330 dated Jul. 13, 2016, 12 pages.
Notice of Allowance on U.S. Appl. No. 13/250,798, dated Sep. 28, 2016, 10 pages.