System for and method of image processing for low visibility landing applications

Information

  • Patent Grant
  • 10928510
  • Patent Number
    10,928,510
  • Date Filed
    Monday, June 19, 2017
    7 years ago
  • Date Issued
    Tuesday, February 23, 2021
    3 years ago
Abstract
An apparatus is for use with an aircraft radar system having a radar antenna. The apparatus includes processing electronics configured to cause the radar antenna to emit radar pulses having a pulse width less than 6 microseconds and configured to receive radar data associated with signals associated with the radar antenna. The radar data is processed to identify centroids associated with indications of the runway lights for a runway in the radar data. A best fit analysis of the centroids is used to identify a runway centerline associated with the runway.
Description
BACKGROUND

Embodiments of the inventive concepts disclosed herein relate generally to aircraft sensing and/or display systems. Embodiments of the inventive concepts disclosed herein more specifically relate to an apparatus or method for the detection and/or display of runways and/or runway features using radar.


Displays are used in head down display (HDD) systems, head up display (HUD) systems including but not limited to fixed combiner displays and wearable displays such as helmet mounted display (HMD) and head worn display (HWD) systems. In aircraft applications, HUD and HDD systems advantageously display information from aircraft systems and sensors in a graphical and alphanumeric format. The display information can include an enhanced vision image from a camera or other imaging sensor (such as a visible light imaging sensor, infrared imaging sensor, millimeter wave radar imager, or combinations thereof) and/or a synthetic vision image from a synthetic vision computer in certain applications. The enhanced vision image can be merged with a synthetic vision image to provide a single image to the pilot. The image can further contain or be augmented by alphanumeric and iconic symbology that highlights or enhances the image content.


Enhanced flight vision systems (EFVSs) are often utilized to view the runway during the approach or landing phases. Infrared cameras and light sensors used by conventional enhanced vision systems can have a limited ability to penetrate certain challenging weather conditions, such as CAT III fog and heavy precipitation. For example, infrared-based systems can have difficulty detecting and displaying runways in low visibility conditions (e.g., under 300 feet runway visual range (RVR)).


EFVSs have been proposed which use radio frequency (RF) signals such as millimeter wave (mmW) radar and/or weather radar signals (X-band, K-band, etc.). RF signals are better able to penetrate challenging weather conditions. However, conventional use of RF signals often deliver images of lower resolution when compared to infrared-based cameras and light sensors. Various weather radar systems capable of providing radar data for an image of a runway are described in the applications incorporated herein by reference.


FAA-certified EFVS can allow pilots landing under instrument flight rules to operate below certain specified altitudes during instrument approaches even when the airport environment is not visible to the unaided eye. Conventional SVS cannot provide enhanced flight visibility, especially the capability to show a real world sensed image of the runway during an impending landing. The use of an integrity monitor for an SVS may allow for higher design assurance levels which could lead to the use of monitored SVS for lower landing minimum credit (e.g., would allow a pilot with a monitored SVS display system to land where a non-monitored SVS pilot would not be otherwise allowed to land due to the current low visibility or RVR restrictions). Accordingly, there have been proposals to provide a monitor for an SVS system based upon various instruments. The use of additional equipment to provide an integrity monitor for the SVS can add to the cost and weight of the aircraft.


SUMMARY

In one aspect, the inventive concepts disclosed herein are directed to an apparatus for use with an aircraft radar system having a radar antenna. The apparatus includes processing electronics configured to cause the radar antenna to emit radar pulses having a pulse width less than 6 microseconds and configured to receive radar data associated with signals associated with the radar antenna. The radar data is processed to identify centroids associated with indications of runway lights or runway infrastructure (e.g., light stanchions) for a runway in the radar data. A best fit analysis of the centroids is used to identify a runway centerline associated with the runway.


In another aspect, the inventive concepts disclosed herein are directed to a method of using radar return data from a radar system on an aircraft. The method includes receiving the radar return data from the radar system, and processing image data associated with the radar return data to identify a runway centerline using centroids associated with groups of returns in the image data. The method also includes providing image or symbolic data including a representation of the runway centerline.


In yet another aspect, the inventive concepts disclosed herein are directed to a weather radar system. The weather radar system includes means for receiving weather radar data. The weather radar system also includes means for identifying runway lights from the weather radar data and providing a runway centerline depiction on an electronic display.





BRIEF DESCRIPTION OF THE DRAWINGS

Implementations of the inventive concepts disclosed herein may be better understood when consideration is given to the following detailed description thereof. Such description makes reference to the included drawings, which are not necessarily to scale, and in which some features may be exaggerated and some features may be omitted or may be represented schematically in the interest of clarity. Like reference numerals in the drawings may represent and refer to the same or similar element, feature, or function. In the drawings:



FIG. 1 is a schematic illustration of an aircraft control center or cockpit according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 2 is a schematic general block diagram of a display system for displaying an image or symbols derived from radar data according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 3 is a flow diagram showing an exemplary process used by the system illustrated in FIG. 2 according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 4 is an illustration of image data generated from radar data in an environment including runway lights using the display system illustrated in FIG. 2 according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 5 is an illustration of the image data illustrated in FIG. 4 processed to assist the identification of the runway lights according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 6 is an illustration of a Hough transform of the image data illustrated in FIG. 5 according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 7 is an illustration of the image data illustrated in FIG. 5 including an extended centerline drawn through centroids of radar return blobs according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 8 is an illustration of image data generated from radar data in an environment including with oil rigs using the display system illustrated in FIG. 2 according to exemplary aspects of the inventive concepts disclosed herein;



FIG. 9 is an illustration of an image generated from the image data illustrated in FIG. 8 identifying locations of the oil rigs according to exemplary aspects of the inventive concepts disclosed herein; and



FIG. 10 is an illustration of an image generated from radar data in an environment including an oil platform with a helipad using the display system illustrated in FIG. 2 according to exemplary aspects of the inventive concepts disclosed herein





DETAILED DESCRIPTION OF THE EXEMPLARY EMBODIMENTS

Before explaining at least one embodiment of the inventive concepts disclosed herein in detail, it is to be understood that the inventive concepts are not limited in their application to the details of construction and the arrangement of the components or steps or methodologies set forth in the following description or illustrated in the drawings. In the following detailed description of embodiments of the instant inventive concepts, numerous specific details are set forth in order to provide a more thorough understanding of the inventive concepts. However, it will be apparent to one of ordinary skill in the art having the benefit of the instant disclosure that the inventive concepts disclosed herein may be practiced without these specific details. In other instances, well-known features may not be described in detail to avoid unnecessarily complicating the instant disclosure. The inventive concepts disclosed herein are capable of other embodiments or of being practiced or carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein is for the purpose of description and should not be regarded as limiting.


Unless expressly stated to the contrary, “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), or both A and B are true (or present).


In addition, use of the “a” or “an” are employed to describe elements and components of embodiments of the instant inventive concepts. This is done merely for convenience and to give a general sense of the inventive concepts, and “a” and “an” are intended to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise.


Finally, as used herein any reference to “one embodiment” or “some embodiments” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the inventive concepts disclosed herein. The appearances of the phrase “in some embodiments” in various places in the specification are not necessarily all referring to the same embodiment, and embodiments of the inventive concepts disclosed may include one or more of the features expressly described or inherently present herein, or any combination or sub-combination of two or more such features, along with any other features which may not necessarily be expressly described or inherently present in the instant disclosure.


Embodiments of the inventive concepts disclosed herein are directed to a radar system, such as, a weather radar system, that can be used to detect features of a runway environment. In one embodiment, features having a high radar cross-section (RCS) are utilized to detect features associated with the runway. For example, metal stanchions that support the Approach Light System (ALS) and parabolic reflectors of the light bulbs whether on or off can be identified in radar data associated with scans of the runway. In another embodiment, metal content of embedded runway centerlines can be identified in the radar data. Utilizing the high radar cross section associated with metal content of runway lighting advantageously allows detection to be achieved whether at day or night, regardless of whether runway lights are on or are off.


In some embodiments, radar data from a legacy weather radar system can be used to provide a situational awareness sensor that can provide alignment confirmation in all weather conditions in the final phases of flight. In some embodiments, the systems and methods can be utilized as an extension to focal plane array-based (FPA-based) EFVS or as extended centerline generators using sensing rather than other navigation. The extended centerlines are provided on a display of a combined vision system (CVS) and is used as an alignment confirmation tool for an SVS solution, thereby providing higher system integrity for the SVS, or as an enabler for a lower minimum minima for low visibility operations.


The systems and methods detect and display a runway centerline from weather radar data (e.g., using a HUD including runway symbology or icons derived from radar data) from legacy weather radar hardware already available on existing aircraft in some embodiments. In some embodiments, the weather radar system is configured to provide short pulse lengths (e.g., less than 6 microseconds, less than 3 microseconds, etc.). The ALS installations consist of 7-12 light stanchions separated by 100′ in range in some embodiments.


With adequate range resolution, as well as the location of the runway threshold, the presence of the ALS or runway centerline lights is detectable in some embodiments. Without improvements in the azimuth resolution of legacy weather radar hardware, the returned detections of the 5-light and 15-light ALS barrettes are substantially spread in azimuth, and centroiding techniques are used to provide an image of the runway centerline in some embodiments. An individual group of returns or blobs has a centroid that may not be overly accurate in positioning the blob itself, but statistical regression of the location of each centroid in each group or blob leads to a best fit extended runway centerline with the specificity required to deliver the desired accuracy for lateral alignment monitoring in some embodiments.


In some embodiments, the systems and methods take advantage of regular, periodic, equal spacing nature of visual aids such as ALS lights and runway centerline lights. The strict definition associated with the spacing of such visual aids which by design are very specific to air field environments can be used to uniquely identify runways on approach in one embodiment. For example, the spacing and location (the spatial frequency) of the features can be specified with very high accuracy according to Federal Aviation Administration (FAA) and International Civil Aviation Organization (ICAO) standards, rules and specifications. In one embodiment, Fourier transforms or related techniques, such as, discrete cosine transform (DCT) functions, etc., can be used to detect or assist detection of the periodic pattern.


Systems and methods can be provided which detect characteristics of runway features using runway lights disposed in various patterns. According one example, runway centerline lights are spaced at 50-foot intervals. See “Aeronautical Information Manual Official Guide to Basic Flight Information and ATC Procedures, Federal Aviation Administration,” Jul. 24, 2014. According to another example, runway touchdown zone lights are provided for runways intended for precision approach Category II or III operations. Runway touchdown zone lights extend from the runway threshold for a distance of 900 m in a pattern formed by pairs of barrettes symmetrically located about the runway centerline. Each barrette consists of three light units at 1.5 m apart. The first pair of barrettes are located at 60 m from the threshold, and subsequent barrettes are spaced longitudinally at 60 m apart according to one example. According to another example, runway center lights are placed at longitudinal intervals of 60 m. According to another example, lights for a precision approach Category II and III lighting system include row of lights on the extended centerline of the runway extending, wherever possible, over a distance of 900 m from the runway threshold. In addition, the system can have two side rows of lights, extending 270 m from the threshold, and 5 crossbars, at 150 m, 300 m, 450 m, 600 m and 750 m from the threshold. In another example, a precision landing system includes the lights forming the centerline placed at longitudinal intervals of 30 m with the innermost light located 30 m from the threshold. See, CHAPTER 7: VISUAL AIDS FOR NAVIGATION-LIGHTS, aai.aero/aai_employees/chapter_7.pdf. Embodiments are not limited to any of the specific examples given above or any particular standards, specifications, rules, or governing bodies.


Referring to FIG. 1, an illustration of a display system 10 for use in an aircraft control center 11 or cockpit is shown, according to some embodiments. The aircraft control center 11 includes flight displays 20 embodied as head down displays (HDDs). Aircraft control center 11 can also include a combiner 21 association with a head up display (HUD) system. In some embodiments, the combiner 21 is provided as part of a wearable HUD. Conformal images are provided on the combiner 21 for viewing by the pilot or flight crew.


The flight displays 20 and the combiner 21 can be used to provide information to the flight crew, thereby increasing visual range and enhancing decision-making abilities. In some embodiments, the flight displays 20 and combiner 21 can include a weather display, a joint display, a weather radar map, a terrain display, an SVS image or an EFVS image. For example, the flight displays 20 and the combiner 21 can include a display configured to display a three dimensional perspective image of terrain and/or weather information or other view of terrain and/or weather information (e.g., plan view, horizontal view, vertical view, or combinations thereof).


According to an exemplary embodiment, at least one of the flight displays 20 or combiner 21 is configured to provide an image of a runway 22 or features associated with the runway 22 (e.g., during approach and landing). In some embodiments, a symbol or icon for the runway 22 can be provided on the displays 20 or combiner 21. A set of runway features, such as, ALS lights 23 or runway centerline lights 24, can be indicated on the flight displays 20 or combiner 21. In some embodiments, an extended runway centerline 27 is provided on the displays 20 or combiner 21. The placement of extended runway centerline 27 is determined at least in part from the detection of runway features (e.g., the ALS lights 23 and runway centerline lights 24) in image data associated with radar data. In some embodiment, at least one of the displays 20 or combiner 21 displays a merged image of terrain derived from two or more of enhanced vision data, radar data, and SVS data.


The flight displays 20 and/or combiner 21 can be configured to provide an indication to a flight crew as to whether the extended runway centerline 27 (or the ALS lights 23 and runway centerline lights 24) associated with the radar data and/or SVS data displayed on the flight displays 20 or combiner 21 are correct or incorrect. In some embodiments, such an indication notifies the crew if the integrity of the SVS is sufficient for lower authorized decision heights and minimum descent altitudes in low visibility conditions.


Referring to FIG. 2, the display system 10 can be utilized for providing an image to any of the flight displays 20 or combiner 21. The display system 10 includes a processor 30 (including an extended centerline generator 32, an image merge module 36, an image merge control/configuration module 38, a filter 39, and a runway database 40), an HDD computer 42, a credit monitor 43, a HUD computer 44, a memory 52, and a memory 53. The display system 10 includes or is in communication with a navigation unit 55, a radar system 102 an SVS 111, and an EVS 112. In some embodiments, the processor 30, and computers 42 and 44 are implemented as a single processing/computing resource providing multiple applications.


The EVS 112 and SVS 111 are optional in some embodiments. The memory 53 is for storing enhanced vision frame from the EVS 112, and the memory 52 is for storing enhanced vision frame from the SVS 111. According to some embodiments, the SVS 111 can be any electronic system or device for providing a computer generated image of the external scene topography. According to some embodiments, the EVS 112 can be any electronic system or device for providing a sensed image of the external scene topography. The EVS 112 can be an infrared camera in some embodiments.


The filter 39, extended centerline generator 32, image merge module 36, image merge control/configuration module 38, and credit monitor 43 can be embodied as software modules operating on a computing platform or the processor 30 and can be stored on a non-transitory medium. The processor 30 can be part of or integrated with the radar system 102, SVS 111, EVS 112, HDD computer 42, or HUD computer 44 in some embodiments. In one embodiment, the processor 30 is an independent computing platform.


The radar system 102 is a weather radar system generally located inside the nose of the aircraft, inside a cockpit of the aircraft, on the top of the aircraft or on the tail of the aircraft. The radar system 102 includes a radar antenna 122 and a processor 125. The radar system 102 can be a legacy weather radar system, such as, a Multiscan™ radar system from Rockwell Collins, Inc. The type of radar system 102 and data gathering techniques are not discussed in the specification in a limiting fashion.


The processor 125 includes a radar data storage unit 128 in some embodiments. The processor 125 receives radar returns (e.g., weather radar return data) from the radar antenna 122, processes the radar returns and stores radar return data in the radar data storage unit 128. The data stored in the radar data storage unit 128 includes an image frame representing the data from a radar scan of the Earth's surface (e.g., near the runway 22) at an approach phase of flight in some embodiments. The processor 125 converts the radar return data to the image frame in some embodiments.


The radar system 102 provides the radar data (e.g., weather radar return data) or the image data in the radar data storage unit 128 to the filter 39 in some embodiments. Alternatively, the processor 30 can build the image frame based upon radar return data (filtered or unfiltered) from the radar system 102. Similarly, the SVS 111 can provide data or an image frame for the SVS image. Alternatively, the display system 10 can provide the data or image frame to the memory 52 in response to data from the SVS 111. The EVS 112 similarly provides data or an image frame to the memory 53.


In some embodiments, the radar system 102 provides data representing a 120 degree field of view in accordance with a weather radar sweep which takes approximately five seconds to complete in one embodiment. The sweep can be limited during approach to be a 30 degree sweep which requires five seconds before new data is available for display in certain embodiments. The sweep is directed toward the surface of the Earth so that returns are obtained which allow runway features to be detected. The sweep can provide pulses having a width of less than 6 microseconds (e.g., less than 3-8 microseconds) in some embodiments. Various types of sweeps, scans and timings of sweeps and scans can be utilized without departing from the scope of the invention. In some embodiments, the radar system 102 is embodied as a weather radar, thereby allowing the display system 10 to use existing avionic equipment to achieve a precise runway centerline detector as described below. Other radar systems may or may not operate according to the above specifications and yet utilize aspects for the inventive concepts disclosed herein.


The display system 10 combines or fuses images from the SVS 111 or EVS 112 and the representations of runway features (e.g., the ALS lights 23 and runway centerline lights 24) derived from radar data from the radar system 102 to provide an overall image provided to the pilot according to one embodiment. The images are fused in a format that integrates features of the SVS 111, EVS 112 and the representation of the runway features from radar return data associated with the radar system 102 in some embodiments. In some embodiments, the representations of runway features are fused with HUD symbology for display on the combiner 21.


The SVS 111 and the extended centerline generator 32 receive aircraft position data from an aircraft data source, such as, the navigation unit 55. The navigation unit 55 can include any system or sensor (or combination thereof) that provides navigation data or aircraft flight parameters. For example, navigation unit 55 can have numerous sub-systems. Sub-systems which provide aircraft position data and flight parameter data could include, but are not limited to, an inertial navigation system (INS), a global navigation satellite system (e.g., global positioning system (GPS)), air data sensors, compasses, and a flight management computer (FMC).


The extended centerline generator 32 utilizes the processed or filtered data from the filter 39 or unfiltered data stored in the radar data storage unit 128 to identify particular runway features. In some embodiments, the extended centerline generator 32 or the processor 125 processes the radar data or the filtered radar data to provide image data where intensity is representative of radar return energy. The data can be processed so that only higher energy returns are shown in the image data (e.g., radar returns having an energy below a threshold are eliminated). The image data is processed to determine the location of groups of radar returns or blobs. The groups of radar returns or blobs are identified in the image data by identifying higher intensity pixels in close proximity (e.g., a group of pixels with neighboring pixels at a higher intensity level). Such groups are associated with radar returns that come from objects with high radar reflectivity, such as, the ALS lights 23 and runway centerline lights 24 or infrastructure thereof. The objects or runway infrastructure can be passive runway reflectors, lights, metal stanchions, or other man-made radar reflecting objects in some embodiments. In some embodiments, the radar data includes discernible returns from stanchions as opposed to other light hardware. The centroids of the blobs are determined and a line through the centroids of the blobs is displayed as the extended runway centerline 27 (FIG. 1). In some embodiments, at least five to seven centroids of blobs are utilized.


In some embodiments, the extended centerline generator 32 uses navigation parameters, phase of flight, and information the runway database to qualify the blobs as potential reflections from the ALS lights 23 and runway centerline lights 24. In some embodiments, the extended centerline generator 32 begins processing data in response to an approach phase of flight and identifies blobs that are in the expected proximity and alignment based upon the runway positon data in the runway database 40 and the positon and orientation of the aircraft.


In some embodiments, the filter 39 processes the radar data for identification of runway features. For example, the filter 39 can be embodied as a Fourier transform, a discrete cosine transform (DCT), a spatial frequency filter, a pattern matching filter, for detecting periodic patterns associated with features of the runway in the radar return data stored in the radar data storage unit 128. A Fourier transform of the radar return data in the radar data storage unit 128 has extremely high values at the spatial frequency associated with light barrette installations and/or other high radar cross-section features and provides strong confirmation of alignment to a runway or taxi way as identified by an extended centerline generator 32. Proper design of the filter 39 can ensure detection of these spatial frequencies as they happen in a rotation-invariant and scale-invariant fashion. The threshold for radar returns can be set relatively high to eliminate returns that are not from high radar-cross section objects in some embodiments.


In some embodiments, the Fourier transform can be performed on the radar data in the radar data storage unit 128 in the plan position indicator (PPI) domain. In some embodiments, the radar data in the PPI domain is filtered before perspective transformation into C-scope domain. Alternatively, the transform can be performed on the C-scope transformation of the radar data taking into account perspective compression with elevation. In some embodiments, the transform can be set to the expected spatial frequencies of runway light features. Alternatively, the PPI domain radar data can be used directly for comparison to a PPI-like SVS solution, thereby altogether obviating the need for processing the radar data into a C-scope domain.


The runway features (e.g., the ALS lights 23 and runway centerline lights 24) can be identified by comparing patterns with reference patterns stored in the runway database 40 for such features in some embodiments. The database 40 can also store expected spatial frequencies for the filter 39. The particular patterns can be chosen utilizing a location parameter from the navigation unit 55. For example, the runway light pattern for particular airports and approaches can be stored in database 40.


The extended centerline generator 32 can utilize roll, pitch, and speed of the aircraft to adjust the patterns before comparing to stored patterns. The reference patterns can be stored as mathematical representations, in a look-up table, as a video frame, etc. In one embodiment, longitudinal patterns, lateral patterns or both lateral and longitudinal patterns can be detected. All runway approach lighting systems in typical modern installations fall into few possible spacing categories, so that a detection of a runway may not use an explicit database of runway features in some embodiments.


The extended centerline generator 32 can provide data for icons or symbols (e.g., the extended runway centerline 27) for merger or provision on the flight displays 20 or combiner 21 via the HDD computer 42 or HUD computer 44. The symbols can include dashes, circles, or other indications of the presence of runway lights in some embodiments.


In some embodiments, data for runway feature symbols are provided directly to HDD computer 42 or HUD computer 44 for providing indications on displays 20 or combiner 21. In another embodiment, the data for runway feature symbols can be provided to image merge function module 26. Image merge function module 26 receives an EVS frame from memory 53 or an SVS frame from memory 52 and merges the data for runway feature symbols to appropriately display an EVS image or an SVS image with the runway feature symbols.


The processor 30 executes a fusion processing algorithm in the module 26 for fusing the frames from the memory 52, memory 53, and extended centerline generator 32 provided as video signals. This fusion process may include special formatting (positioning, sizing, cropping, etc.) of specific features or the entire image from a specific image source based on other sensor inputs or aircraft. After the combined or fused image has been completed, the entire image is offset and scaled to fit appropriately within the total HUD field-of-view and conformally overlay the outside scene, which is viewed through combiner 21 of the HUD. In addition, the overall fused image contrast is standardized with the brightness/contrast to support the brightness/contrast controls of the HUD.


Precise localization of the exact spacing of runway light locations in actual physical distance units (e.g., 100 feet, 200 feet) can allow exact and very reliable confirmation of ALS light and runway centerline light detection via the filter 39 in one embodiment. In one embodiment, localization of the pattern of runway lights allows accurate measurement in the longitudinal location of the aircraft with respect to the runway. Accurate longitudinal location is particularly advantageous for pilots because judging longitudinal location from perspective imagery such as imagery from focal-plane based enhanced flight vision systems or C-scope transform can be difficult.


In some embodiments, the extended runway centerline 27 can be extracted using a very precise line fit using the filter 39 or extended centerline generator 32. The extended runway centerline 27 can even be extracted in the presence of some lateral direction uncertainty. A Hough transform or statistical regression can be utilized to reduce lateral uncertainty.


The credit monitor 43 can receive data from the SVS 111, EVS 112, and radar system 102 to determine position accuracy for SVS 111. The credit monitor 43 can also receive signals from various aircraft sensors including integrated landing system (ILS), radio altimeters, an inertial navigation system (INS) and/or other sensors. The credit monitor 43 can provide an indication of the result of an integrity check for display via HDD computer 42 and HUD computer 44.


The extended centerline generator 32 can provide data associated with the centroiding of blobs associated with the location of the ALS lights 23 and runway centerline lights 24 (or runway infrastructure (e.g., runway reflectors, metal stanchions, or other man-made radar reflecting objects)) to confirm alignment for the SVS solution of SVS 111. In some embodiments, the image data of the extended runway centerline from the extended centerline generator 32 is compared to the image data from the SVS 111. A comparison of the extended runway centerlines in each image provides an indication of alignment. In some embodiments, both extended centerlines are displayed so the pilot can view any misalignment. The confirmed alignment derived from radar data can effectively be used in the credit monitor 43 for the HUD computer 44, thereby allowing the use of lower landing minima in SVS 111 without EVS according to one embodiment.


Various matching algorithms can be utilized to determine if the solutions provided by the SVS and extended centerline generator 32 are accurate. If the compare function indicates that there is not a match, the credit monitor 43 indicates that lower minimum approaches are not enabled as there may be a database error or error with system 102 or SVS 111. The indication can be provided on the combiner 21 associated with the HUD computer 44 or on the flight displays 20. Applicants believe that use of the credit monitor 43 and the display system 10 may assist in a reduction of minimum from a 200 foot decision height to a 100 foot for SVS 111, or lower, due to the high design assurance and integrity brought about by the comparison of the information from multiple independent subsystems.


The image merge control/configuration module 38 can provide format adjustments to data. The SVS 111 and radar system 102 can have their own specific interface type and format. Also, each display of the flight displays 20 and combiner 21 may require specific formatting. A standard format can be a format used in HUD processing functions. Module 38 can be implemented in hardware, software, or combinations thereof.


With reference to FIG. 3, a flow 300 can be performed by the display system 10 for runway identification and display of runway data. The flow 300 is performed when in the vicinity of the airport or upon approach in some embodiments.


At an operation 301, the radar system 102 captures radar data in a radar sweep. The radar sweep is performed with pulses having smaller widths for greater range resolution in some embodiments. At an operation 304, the radar data is processed to provide image data containing groups of returns or blobs having higher return energy. As discussed above, filtering or related techniques by the filter 39 can be utilized to identify pre-determined patterns or accentuate patterns associated with the ALS lights 23 and runway centerline lights 24. In some embodiments, operation 304 uses runway data to assist the identification of the blobs. In some embodiments, the radar data or image data is skeletonized to remove lower energy returns (e.g., returns below a threshold level are removed).


At an operation 306, centroiding is utilized to determine a center of each of the blobs. At an operation 308, the extended runway centerline 27 is determined using a line through the centroids of the blobs. Identification of an individual blob's centroid may not be overly accurate in positioning the blob itself, but statistical regression of the location of the centroid of each of the blobs leads to a best fit extended runway centerline with the specificity required to deliver the desired accuracy for lateral alignment monitoring in some embodiments.


At an operation 310, the extended runway centerline 27 (FIG. 1) is displayed via a display computer such as the HDD computer 42 or HUD computer 44 (FIG. 2). In some embodiments, only the extended runway centerline 27 is displayed to provide a cleaner image (e.g., blobs and other images associated with the radar data are not displayed). In some embodiments, any or all of five display layers are provided and displayed (e.g., a radar data layer, an SVS data layer, an EVS layer, a HUD layer, and an extended centerline layer).


At an operation 312, the credit monitor 43 can use the determined locations associated with the runway features (e.g., the extended centerline) to perform an integrity check associated with the SVS data in memory 52. After operation 310, flow 300 can return to the operation 301.


The processor 30 can be any hardware and/or software processor or processing architecture capable of executing instructions and operating on navigational and radar data. The processor 30 can be capable of determining navigational information such as altitude, heading, bearing, and location based on data from aircraft sensors. Applicants note that flow 300 and runway identification can be performed in various equipment on the aircraft including in a HUD computer, a display processor, radar system 102, a navigation system, SVS 111, etc. in accordance with an exemplary embodiment.


With reference to FIG. 4, a representation 400 of radar data (e.g., as image data) associated with the radar data storage unit 128 includes a set of blobs 402 associated with the ALS lights 23 and runway centerline lights 24 and additional blobs 404. With reference to FIG. 5, the radar data associated with representation 400 is processed (e.g., skeletonized to provide a cleaner image as shown in representation 500. The processing can be performed by the filter 39 and removes returns that are not in groups having energy return and proximity to neighboring return characteristics associated with reflections from the ALS lights 23 and runway centerline lights 24 in some embodiments.


With reference to FIG. 6, a representation 600 of a Hough transform for the image data associated with representation 500 is provided on an X-axis 602 indicating the angle θ and a Y axis 604 representing a radius p. The Hough transform is a process for feature extraction such as identification of lines in an image. The peak group at the angle θ of zero degrees and the radius p at a maximum is an indication of points for the extended centerline and are represented by rectangles 610.


With reference to FIG. 7, the Hough transform is used to provide an extended runway centerline 702 for the image data associated with representation 500. The extended runway centerline 702 can be extrapolated using a best fit technique.


With reference to FIG. 8, a representation 800 of radar data (e.g., as image data) associated with the radar data storage unit 128 includes a set of blobs 804 associated with oil rigs or platforms. With reference to FIG. 9, the radar data associated with representation 800 is processed to determine locations 902, 904, 906908, 910, 912, and 194. The platforms can be identified by their relative locations as compared to a database. The data is processed using centroiding to determine the location of each platform. The processing can include skeletonization and use of a Hough transform. The target platform 916 can be identified with an outline graphic in some embodiments.


With reference to FIG. 10, the display system 10 can be used to locate a helipad 1002 on an oil platform using image data 1000. The oil platform includes structures such as cranes, buildings etc. associated with blobs 1004. Centroiding of the blobs 1004 enables the helipad 1002 location to be confirmed. The height of the centroids and expected relative locations can be used to identify the helipad 1002. Approaches to helipads are discussed in U.S. patent application Ser. No. 15/222,923 incorporated herein by reference.


It is to be understood that embodiments of the methods according to the inventive concepts disclosed herein may include one or more of the steps described herein. Further, such steps may be carried out in any desired order and two or more of the steps may be carried out simultaneously with one another. Two or more of the steps disclosed herein may be combined in a single step, and in some embodiments, one or more of the steps may be carried out as two or more sub-steps. Further, other steps or sub-steps may be carried out in addition to, or as substitutes to one or more of the steps disclosed herein.


From the above description, it is clear that the inventive concepts disclosed herein are well adapted to carry out the objects and to attain the advantages mentioned herein as well as those inherent in the inventive concepts disclosed herein. While presently preferred embodiments of the inventive concepts disclosed herein have been described for purposes of this disclosure, it will be understood that numerous changes may be made which will readily suggest themselves to those skilled in the art and which are accomplished within the broad scope and coverage of the inventive concepts disclosed and claimed herein.

Claims
  • 1. An apparatus for use with a weather aircraft radar system having a radar antenna, the apparatus comprising: processing electronics configured to cause the radar antenna to emit radar pulses having a pulse width less than 6 microseconds and configured to receive radar data associated with signals associated with the radar antenna, wherein the radar data is processed to identify a plurality of centroids associated with indications of a plurality of runway lights or infrastructure in the radar data, wherein (1) a best fit analysis of the centroids and (2) a detection of a periodic pattern of the plurality of runway lights or infrastructure is used to identify a runway centerline;wherein the detection of the periodic pattern of the plurality of runway lights is performed using a fixed spatial frequency of at least one of the group selected from ALS lights and runway centerline lights.
  • 2. The apparatus of claim 1, wherein the processing electronics utilize statistical regression to identify the runway centerline.
  • 3. The apparatus of claim 2, wherein the radar antenna is a weather radar antenna and the radar data is from a weather radar system without an enhanced lateral resolution capability.
  • 4. The apparatus of claim 1, further comprising a display configured to provide an image, the image including a representation of the runway centerline detected by the processing electronics.
  • 5. The apparatus of claim 4, wherein the image is conformally provided onto a combiner of a head up display system.
  • 6. The apparatus of claim 1, wherein the runway lights comprise runway centerline lights, approach lighting system lights, or stanchions.
  • 7. The apparatus of claim 1, wherein the centroids comprise at least four centroids.
  • 8. The apparatus of claim 7, wherein the centroids comprise at least seven centroids.
  • 9. The apparatus of claim 8, wherein the runway lights are approach lighting system lights spaced at a distance of 200 feet or 100 feet.
  • 10. A method of using radar return data from a weather radar system on an aircraft, the method comprising: receiving the radar return data using the radar system;using an electronic processor, processing image data associated with the radar return data to identify a plurality of centroids associated with indications of a plurality of runway lights;identifying a runway centerline by (1) a best fit analysis of the centroids and (2) a detection of a periodic pattern of the plurality of runway lights,wherein the detection of the periodic pattern of the plurality of runway lights is performed using a filter selected from the group consisting of a Fourier transform, a discrete cosine transform (DCT), a spatial frequency, or pattern matching; andproviding image or symbolic data including a representation of the runway centerline on an aircraft electronic display.
  • 11. The method of claim 10, further comprising: correlating the runway centerline with a synthetic vision system centerline.
  • 12. The method of claim 11, further comprising: using the correlation to enable a lower minima approach.
  • 13. The method of claim 11, further comprising: using the correlation to indicate an error.
  • 14. The method of claim 13, further comprising: displaying an image associated with the image data or the symbolic data on an electronic display.
  • 15. The method of claim 14, wherein the electronic display is a head up display and the identifying step is performed in a head up display computer, the weather radar system, a synthetic vision system, or a display computer.
  • 16. The method of claim 12, wherein the identifying uses a filter to identify a pattern associated with the runway lights and the pattern comprises the groups of returns.
  • 17. The method of claim 16, further comprising: using statistical regression or a Hough transform to detect the runway centerline in the image data.
  • 18. A weather radar system, the weather radar system comprising: means for receiving weather radar data; andmeans for identifying runway lights or runway light infrastructure from the weather radar data and providing a runway centerline depiction on an electronic display using a plurality of centroids associated with indications of the runway lights or the infrastructure in the weather radar data, wherein the runway centerline depiction is performed using (1) a best fit analysis of the centroids and (2) a detection of a periodic pattern of the plurality of runway lights,wherein the detection of the periodic pattern of the plurality of runway lights is performed using a fixed spatial frequency of at least one of the group selected from ALS lights and runway centerline lights.
  • 19. The weather radar system claim 18, wherein the weather radar data is generated without using beam sharpening.
  • 20. The weather radar system of claim 18, wherein the centroids are in a pattern in the weather radar data and a Hough transform using the centroids is used to provide the runway centerline depiction.
CROSS-REFERENCE TO RELATED PATENT APPLICATIONS

The present application is a continuation-in-part of U.S. patent application Ser. No. 14/482,681, filed Sep. 10, 2014 issued as U.S. Pat. No. 9,733,349 and is related to U.S. patent application Ser. No. 15/222,923, filed on Jul. 28, 2016 issued as U.S. Pat. No. 10,353,068, U.S. patent application Ser. No. 13/247,742 filed on Sep. 28, 2011 issued as U.S. Pat. No. 8,896,480, U.S. patent application Ser. No. 11/851,323 filed on Sep. 6, 2007 issued as U.S. Pat. No. 8,515,600, U.S. patent application Ser. No. 11/904,491 filed on Sep. 26, 2007 issued as U.S. Pat. No. 8,755,954, U.S. patent application Ser. No. 13/241,051, filed on Sep. 22, 2011 issued as U.S. Pat. No. 8,515,600, U.S. patent application Ser. No. 12/263,282, filed on Oct. 31, 2008 issued as U.S. Pat. No. 9,354,633 and U.S. patent application Ser. No. 12/180,293 filed Jul. 25, 2008 issued as U.S. Pat. No. 8,077,078, all of which are herein incorporated by reference in their entirety and assigned to the assignee of the present application.

US Referenced Citations (565)
Number Name Date Kind
2416155 Chubb Feb 1947 A
2849184 Arden et al. Aug 1958 A
2929059 Parker Mar 1960 A
2930035 Altekruse Mar 1960 A
2948892 White Aug 1960 A
2965894 Sweeney Dec 1960 A
2994966 Senitsky et al. Aug 1961 A
3031660 Young Apr 1962 A
3049702 Schreitmueller Aug 1962 A
3064252 Varela Nov 1962 A
3070795 Chambers Dec 1962 A
3071766 Fenn Jan 1963 A
3072903 Meyer Jan 1963 A
3089801 Tierney et al. May 1963 A
3107351 Milam Oct 1963 A
3113310 Standing Dec 1963 A
3129425 Sanner Apr 1964 A
3153234 Begeman et al. Oct 1964 A
3175215 Blasberg et al. Mar 1965 A
3212088 Alexander et al. Oct 1965 A
3221328 Walter Nov 1965 A
3241141 Wall Mar 1966 A
3274593 Varela et al. Sep 1966 A
3325807 Burns et al. Jun 1967 A
3334344 Colby, Jr. Aug 1967 A
3339199 Pichafroy Aug 1967 A
3373423 Levy Mar 1968 A
3396391 Anderson Aug 1968 A
3397397 Barney Aug 1968 A
3448450 Alfandari et al. Jun 1969 A
3618090 Garrison Nov 1971 A
3680094 Bayle et al. Jul 1972 A
3716855 Asam Feb 1973 A
3739380 Burdic et al. Jun 1973 A
3781878 Kirkpatrick Dec 1973 A
3810175 Bell May 1974 A
3815132 Case et al. Jun 1974 A
3816718 Hall et al. Jun 1974 A
3851758 Makhijani et al. Dec 1974 A
3866222 Young Feb 1975 A
3885237 Kirkpatrick May 1975 A
3956749 Magorian May 1976 A
4024537 Hart May 1977 A
4058701 Gruber et al. Nov 1977 A
4058710 Altmann Nov 1977 A
4063218 Basov et al. Dec 1977 A
4103300 Gendreu et al. Jul 1978 A
4235951 Swarovski Nov 1980 A
4277845 Smith et al. Jul 1981 A
4405986 Gray Sep 1983 A
4435707 Clark Mar 1984 A
4481519 Margerum Nov 1984 A
4509048 Jain Apr 1985 A
4532515 Cantrell et al. Jul 1985 A
4594676 Breiholz et al. Jun 1986 A
4595925 Hansen Jun 1986 A
4598292 Devino Jul 1986 A
4628318 Alitz Dec 1986 A
4646244 Bateman et al. Feb 1987 A
4649388 Atlas Mar 1987 A
4654665 Kiuchi et al. Mar 1987 A
4685149 Smith et al. Aug 1987 A
4723124 Boles Feb 1988 A
4760396 Barney et al. Jul 1988 A
4828382 Vermilion May 1989 A
4843398 Houston et al. Jun 1989 A
4912477 Lory et al. Mar 1990 A
4914436 Bateman et al. Apr 1990 A
4924401 Bice et al. May 1990 A
4939513 Paterson et al. Jul 1990 A
4951059 Taylor, Jr. Aug 1990 A
4953972 Zuk Sep 1990 A
4965573 Gallagher et al. Oct 1990 A
4987419 Salkeld Jan 1991 A
5045855 Moreira Sep 1991 A
5047779 Hager Sep 1991 A
5047781 Bleakney Sep 1991 A
5049886 Seitz et al. Sep 1991 A
5053778 Imhoff Oct 1991 A
5166688 Moreira Nov 1992 A
5173703 Mangiapane et al. Dec 1992 A
5175554 Mangiapane et al. Dec 1992 A
5198819 Susnjara Mar 1993 A
5202690 Frederick Apr 1993 A
5247303 Cornelius et al. Sep 1993 A
5273553 Hoshi et al. Dec 1993 A
5311183 Mathews et al. May 1994 A
5329391 Miyamoto et al. Jul 1994 A
5332998 Avignon et al. Jul 1994 A
5345241 Huddle Sep 1994 A
5365356 McFadden Nov 1994 A
5383457 Cohen Jan 1995 A
5442364 Lee et al. Aug 1995 A
5530440 Danzer et al. Jun 1996 A
5539409 Mathews et al. Jul 1996 A
5559515 Alimena et al. Sep 1996 A
5559518 Didomizio Sep 1996 A
5566840 Waldner et al. Oct 1996 A
5592178 Chang et al. Jan 1997 A
5678303 Wichmann Oct 1997 A
5736957 Raney Apr 1998 A
5820080 Eschenbach Oct 1998 A
5828332 Frederick Oct 1998 A
5831570 Ammar Nov 1998 A
5839080 Muller et al. Nov 1998 A
5867119 Corrubia et al. Feb 1999 A
5894286 Morand et al. Apr 1999 A
5918517 Malapert et al. Jul 1999 A
5920276 Frederick Jul 1999 A
5923279 Bamler et al. Jul 1999 A
5936575 Azzarelli et al. Aug 1999 A
5942062 Hassall et al. Aug 1999 A
5945926 Ammar et al. Aug 1999 A
5950512 Fields Sep 1999 A
5959762 Bandettini et al. Sep 1999 A
5978715 Briffe et al. Nov 1999 A
6002347 Daly et al. Dec 1999 A
6023240 Sutton Feb 2000 A
6061016 Lupinski et al. May 2000 A
6061022 Menegozzi et al. May 2000 A
6064942 Johnson et al. May 2000 A
6075484 Daniel et al. Jun 2000 A
6092009 Glover Jul 2000 A
6112141 Briffe et al. Aug 2000 A
6112570 Hruschak Sep 2000 A
6122570 Muller et al. Sep 2000 A
6127944 Daly et al. Oct 2000 A
6128066 Yokozeki Oct 2000 A
6128553 Gordon et al. Oct 2000 A
6138060 Conner et al. Oct 2000 A
6150901 Auken Nov 2000 A
6154151 McElreath et al. Nov 2000 A
6154169 Kuntman Nov 2000 A
6157339 Sato et al. Dec 2000 A
6157891 Lin Dec 2000 A
6163021 Mickelson Dec 2000 A
6166661 Anderson et al. Dec 2000 A
6169770 Henely Jan 2001 B1
6178391 Anderson et al. Jan 2001 B1
6184816 Zheng et al. Feb 2001 B1
6188330 Glover Feb 2001 B1
6194980 Thon Feb 2001 B1
6199008 Aratow et al. Mar 2001 B1
6201494 Kronfeld Mar 2001 B1
6204806 Hoech Mar 2001 B1
6205400 Lin Mar 2001 B1
6208284 Woodell et al. Mar 2001 B1
6219592 Muller et al. Apr 2001 B1
6233522 Morici May 2001 B1
6236351 Conner et al. May 2001 B1
6259400 Higgins et al. Jul 2001 B1
6266114 Skarohlid Jul 2001 B1
6278799 Hoffman Aug 2001 B1
6281832 McElreath Aug 2001 B1
6285298 Gordon Sep 2001 B1
6285337 West et al. Sep 2001 B1
6285926 Weiler et al. Sep 2001 B1
6289277 Feyereisen et al. Sep 2001 B1
6311108 Ammar et al. Oct 2001 B1
6317468 Meyer Nov 2001 B1
6317690 Gia Nov 2001 B1
6317872 Gee et al. Nov 2001 B1
6340946 Wolfson et al. Jan 2002 B1
6345127 Mitchell Feb 2002 B1
6359585 Bechman et al. Mar 2002 B1
6366013 Leenders et al. Apr 2002 B1
6373418 Abbey Apr 2002 B1
6374286 Gee et al. Apr 2002 B1
6377202 Kropfli et al. Apr 2002 B1
6377892 Johnson et al. Apr 2002 B1
6388607 Woodell May 2002 B1
6388608 Woodell et al. May 2002 B1
6388724 Campbell et al. May 2002 B1
6389354 Hicks et al. May 2002 B1
6401038 Gia Jun 2002 B2
6411890 Zimmerman Jun 2002 B1
6421000 McDowell Jul 2002 B1
6421603 Pratt et al. Jul 2002 B1
6424288 Woodell Jul 2002 B1
6426717 Maloratsky Jul 2002 B1
6426720 Ross et al. Jul 2002 B1
6427122 Lin Jul 2002 B1
6441773 Kelly et al. Aug 2002 B1
6445310 Bateman et al. Sep 2002 B1
6448922 Kelly Sep 2002 B1
6452511 Kelly et al. Sep 2002 B1
6456236 Hauck et al. Sep 2002 B1
6456238 Posey Sep 2002 B1
6462703 Hedrick Oct 2002 B2
6473026 Ali-Mehenni et al. Oct 2002 B1
6473037 Vail et al. Oct 2002 B2
6473240 Dehmlow Oct 2002 B1
6481482 Shimotomai Nov 2002 B1
6492934 Hwang et al. Dec 2002 B1
6501424 Haendel et al. Dec 2002 B1
6512476 Woodell Jan 2003 B1
6512527 Barber et al. Jan 2003 B1
6516272 Lin Feb 2003 B2
6516283 McCall et al. Feb 2003 B2
6520056 Nemeth et al. Feb 2003 B1
6525674 Kelly et al. Feb 2003 B1
6531669 Miller et al. Mar 2003 B1
6549161 Woodell Apr 2003 B1
6567728 Kelly et al. May 2003 B1
6574030 Mosier Jun 2003 B1
6577947 Kronfeld et al. Jun 2003 B1
6590528 Dewulf Jul 2003 B1
6591171 Ammar et al. Jul 2003 B1
6593875 Bergin et al. Jul 2003 B2
6600443 Landt Jul 2003 B2
6603425 Woodell Aug 2003 B1
6614057 Silvernail et al. Sep 2003 B2
6650275 Kelly et al. Nov 2003 B1
6650291 West et al. Nov 2003 B1
6653947 Dwyer et al. Nov 2003 B2
6667710 Cornell et al. Dec 2003 B2
6681668 Smirle Jan 2004 B1
6690298 Barber et al. Feb 2004 B1
6690299 Suiter Feb 2004 B1
6690317 Szeto et al. Feb 2004 B2
6697008 Sternowski Feb 2004 B1
6697012 Lodwig et al. Feb 2004 B2
6710663 Berquist Mar 2004 B1
6714186 Mosier et al. Mar 2004 B1
6720890 Ezroni et al. Apr 2004 B1
6724344 Stockmaster et al. Apr 2004 B1
6731236 Hager et al. May 2004 B1
6738011 Evans May 2004 B1
6739929 Furukawa et al. May 2004 B2
6741203 Woodell May 2004 B1
6741208 West et al. May 2004 B1
6744382 Lapis et al. Jun 2004 B1
6744408 Stockmaster Jun 2004 B1
6757624 Hwang et al. Jun 2004 B1
6760155 Murayama et al. Jul 2004 B2
6771626 Golubiewski et al. Aug 2004 B1
6782392 Weinberger et al. Aug 2004 B1
6799095 Owen et al. Sep 2004 B1
6803245 Auch et al. Oct 2004 B2
6804614 McGraw et al. Oct 2004 B1
6806846 West Oct 2004 B1
6807538 Weinberger et al. Oct 2004 B1
6813777 Weinberger et al. Nov 2004 B1
6819983 McGraw Nov 2004 B1
6822617 Mather et al. Nov 2004 B1
6825804 Doty Nov 2004 B1
6832538 Hwang Dec 2004 B1
6839017 Dillman Jan 2005 B1
6842288 Liu et al. Jan 2005 B1
6850185 Woodell Feb 2005 B1
6862323 Loper Mar 2005 B1
6862501 He Mar 2005 B2
6865452 Burdon Mar 2005 B2
6879280 Bull et al. Apr 2005 B1
6879886 Wilkins et al. Apr 2005 B2
6882302 Woodell et al. Apr 2005 B1
6908202 Graf et al. Jun 2005 B2
6917396 Hiraishi et al. Jul 2005 B2
6918134 Sherlock et al. Jul 2005 B1
6933885 Stockmaster et al. Aug 2005 B1
6938258 Weinberger et al. Aug 2005 B1
6950062 Mather et al. Sep 2005 B1
6959057 Tuohino Oct 2005 B1
6972727 West et al. Dec 2005 B1
6977608 Anderson et al. Dec 2005 B1
6984545 Grigg et al. Jan 2006 B2
6990022 Morikawa et al. Jan 2006 B2
6992614 Joyce Jan 2006 B1
6995726 West et al. Feb 2006 B1
6998648 Silvernail Feb 2006 B2
6998908 Sternowski Feb 2006 B1
6999022 Vesel et al. Feb 2006 B1
6999027 Stockmaster Feb 2006 B1
7002546 Stuppi et al. Feb 2006 B1
7010398 Wilkins et al. Mar 2006 B2
7023375 Klausing et al. Apr 2006 B2
7026956 Wenger et al. Apr 2006 B1
7028304 Weinberger et al. Apr 2006 B1
7030945 Umemoto et al. Apr 2006 B2
7034753 Elsallal et al. Apr 2006 B1
7042387 Ridenour et al. May 2006 B2
7053796 Barber May 2006 B1
7057549 Block Jun 2006 B2
7064680 Reynolds et al. Jun 2006 B2
7069120 Koenck et al. Jun 2006 B1
7089092 Wood et al. Aug 2006 B1
7092645 Sternowski Aug 2006 B1
7098913 Etherington et al. Aug 2006 B1
7109912 Paramore et al. Sep 2006 B1
7109913 Paramore et al. Sep 2006 B1
7123260 Brust Oct 2006 B2
7129885 Woodell et al. Oct 2006 B1
7145501 Manfred et al. Dec 2006 B1
7148816 Carrico Dec 2006 B1
7151507 Herting Dec 2006 B1
7158072 Venkatachalam et al. Jan 2007 B1
7161525 Finley et al. Jan 2007 B1
7170446 West et al. Jan 2007 B1
7170959 Abbey Jan 2007 B1
7180476 Guell et al. Feb 2007 B1
7191406 Barber et al. Mar 2007 B1
7196329 Wood et al. Mar 2007 B1
7205933 Snodgrass Apr 2007 B1
7209070 Gilliland et al. Apr 2007 B2
7212216 He et al. May 2007 B2
7218268 Vandenberg May 2007 B2
7219011 Barber May 2007 B1
7242343 Woodell Jul 2007 B1
7242345 Raestad et al. Jul 2007 B2
7250903 McDowell Jul 2007 B1
7265710 Deagro Sep 2007 B2
7269657 Alexander et al. Sep 2007 B1
7272472 McElreath Sep 2007 B1
7273403 Yokota et al. Sep 2007 B2
7280068 Lee et al. Oct 2007 B2
7289058 Shima Oct 2007 B2
7292178 Woodell et al. Nov 2007 B1
7292180 Schober Nov 2007 B2
7295150 Burlet et al. Nov 2007 B2
7295901 Little et al. Nov 2007 B1
7301496 Honda et al. Nov 2007 B2
7307576 Koenigs Dec 2007 B1
7307577 Kronfeld et al. Dec 2007 B1
7307583 Woodell et al. Dec 2007 B1
7312725 Berson et al. Dec 2007 B2
7312743 Ridenour et al. Dec 2007 B2
7317427 Pauplis et al. Jan 2008 B2
7321332 Focke et al. Jan 2008 B2
7337043 Bull Feb 2008 B2
7349154 Aiura et al. Mar 2008 B2
7352292 Alter et al. Apr 2008 B2
7361240 Kim Apr 2008 B2
7372394 Woodell et al. May 2008 B1
7373223 Murphy May 2008 B2
7375678 Feyereisen et al. May 2008 B2
7379014 Woodell et al. May 2008 B1
7379796 Walsdorf et al. May 2008 B2
7381110 Sampica et al. Jun 2008 B1
7417578 Woodell et al. Aug 2008 B1
7417579 Woodell Aug 2008 B1
7423578 Tietjen Sep 2008 B1
7446697 Burlet et al. Nov 2008 B2
7446938 Miyatake et al. Nov 2008 B2
7452258 Marzen et al. Nov 2008 B1
7474262 Alland Jan 2009 B2
7479920 Niv Jan 2009 B2
7486220 Kronfeld et al. Feb 2009 B1
7486291 Berson et al. Feb 2009 B2
7492304 Woodell et al. Feb 2009 B1
7492305 Woodell et al. Feb 2009 B1
7515069 Dorneich et al. Apr 2009 B2
7515087 Woodell et al. Apr 2009 B1
7515088 Woodell et al. Apr 2009 B1
7525448 Wilson et al. Apr 2009 B1
7528765 Woodell et al. May 2009 B1
7528915 Choi et al. May 2009 B2
7541970 Godfrey et al. Jun 2009 B1
7541971 Woodell et al. Jun 2009 B1
7551451 Kim et al. Jun 2009 B2
7557735 Woodell et al. Jul 2009 B1
7566254 Sampica et al. Jul 2009 B2
7570177 Reynolds et al. Aug 2009 B2
7576680 Woodell Aug 2009 B1
7579978 Finley Aug 2009 B1
7603209 Dwyer et al. Oct 2009 B2
7609200 Woodell et al. Oct 2009 B1
7612706 Honda et al. Nov 2009 B2
7616150 Woodell Nov 2009 B1
7633428 McCusker et al. Dec 2009 B1
7633430 Wichgers et al. Dec 2009 B1
7633584 Umemoto et al. Dec 2009 B2
7639175 Woodell Dec 2009 B1
7664601 Daly, Jr. Feb 2010 B2
7675461 McCusker et al. Mar 2010 B1
7693621 Chamas Apr 2010 B1
7696921 Finley et al. Apr 2010 B1
7714767 Kronfeld et al. May 2010 B1
7733264 Woodell et al. Jun 2010 B1
7783427 Woodell et al. Aug 2010 B1
7783429 Walden et al. Aug 2010 B2
7791529 Filias et al. Sep 2010 B2
7808422 Woodell et al. Oct 2010 B1
7814676 Sampica et al. Oct 2010 B2
7843380 Woodell Nov 2010 B1
7859448 Woodell et al. Dec 2010 B1
7859449 Woodell et al. Dec 2010 B1
7864103 Weber et al. Jan 2011 B2
7868811 Woodell et al. Jan 2011 B1
7872594 Vesel Jan 2011 B1
7889117 Woodell et al. Feb 2011 B1
7889118 Finley et al. Feb 2011 B1
7925117 Hamza Apr 2011 B2
7927440 Matsuhira Apr 2011 B2
7929086 Toyama et al. Apr 2011 B2
7965223 McCusker Jun 2011 B1
7965225 Dickerson et al. Jun 2011 B1
8035547 Flanigan et al. Oct 2011 B1
8038498 Miyauchi et al. Oct 2011 B2
8045098 Kitagawa et al. Oct 2011 B2
8059025 D'Addio Nov 2011 B2
8068984 Smith et al. Nov 2011 B2
8072368 Woodell Dec 2011 B1
8077078 Woodell et al. Dec 2011 B1
8077081 Bateman et al. Dec 2011 B2
8102487 Kitagawa et al. Jan 2012 B2
8118075 Sampica et al. Feb 2012 B2
8137498 Sampica et al. Mar 2012 B2
8140223 Whitehead et al. Mar 2012 B2
8159464 Gribble et al. Apr 2012 B1
8232917 Scherzinger et al. Jul 2012 B2
8289202 Christianson Oct 2012 B1
8296065 Haynie et al. Oct 2012 B2
8330650 Goldman Dec 2012 B2
8373580 Bunch et al. Feb 2013 B2
8410975 Bell et al. Apr 2013 B1
8477062 Kanellis Jul 2013 B1
8486535 Nemeth et al. Jul 2013 B1
8493241 He Jul 2013 B2
8515600 McCusker Aug 2013 B1
8540002 Sampica et al. Sep 2013 B2
8558731 Woodell Oct 2013 B1
8576112 Garrec et al. Nov 2013 B2
8583315 Whitehead et al. Nov 2013 B2
8594879 Roberge et al. Nov 2013 B2
8603288 Sampica et al. Dec 2013 B2
8634993 McClure et al. Jan 2014 B2
8639416 Jones et al. Jan 2014 B2
8643533 Woodell et al. Feb 2014 B1
8691043 Sampica et al. Apr 2014 B2
8717226 Bon et al. May 2014 B2
8755954 McCusker et al. Jun 2014 B1
8773301 Woodell Jul 2014 B1
8847794 Buratto et al. Sep 2014 B2
8896480 Wilson et al. Nov 2014 B1
8909471 Jinkins et al. Dec 2014 B1
8917191 Tiana et al. Dec 2014 B1
8936057 Sampica et al. Jan 2015 B2
8976042 Chiew Mar 2015 B1
8977491 McCusker et al. Mar 2015 B1
9024805 Jinkins et al. May 2015 B1
9174746 Bell Nov 2015 B1
9354633 McCusker et al. May 2016 B1
20010023390 Gia Sep 2001 A1
20010050372 Krijn et al. Dec 2001 A1
20010053648 Furukawa et al. Dec 2001 A1
20020039070 Ververs et al. Apr 2002 A1
20020111717 Scherzinger et al. Aug 2002 A1
20020116125 Lin Aug 2002 A1
20020116126 Lin Aug 2002 A1
20020158256 Yamada et al. Oct 2002 A1
20020179229 Chuzles Dec 2002 A1
20020185600 Kerr Dec 2002 A1
20020187284 Kinoshita et al. Dec 2002 A1
20030021491 Brust Jan 2003 A1
20030038916 Nakano et al. Feb 2003 A1
20030043315 Umemoto et al. Mar 2003 A1
20030071828 Wilkins et al. Apr 2003 A1
20030089214 Fukuta et al. May 2003 A1
20030093187 Walker May 2003 A1
20030102999 Bergin et al. Jun 2003 A1
20030156238 Hiraishi et al. Aug 2003 A1
20030160718 Nagasaku Aug 2003 A1
20030174396 Murayama et al. Sep 2003 A1
20030180528 Flosenzier et al. Sep 2003 A1
20030189606 Moon et al. Oct 2003 A1
20030195672 He Oct 2003 A1
20030216859 Martell et al. Nov 2003 A1
20030222887 Wilkins et al. Dec 2003 A1
20040044445 Burdon Mar 2004 A1
20040059473 He Mar 2004 A1
20040066645 Graf et al. Apr 2004 A1
20040072575 Young et al. Apr 2004 A1
20040083038 He Apr 2004 A1
20040130073 Yamazaki et al. Jul 2004 A1
20040145499 Schmidt et al. Jul 2004 A1
20040160341 Feyereisen et al. Aug 2004 A1
20040160364 Regev Aug 2004 A1
20040181318 Redmond et al. Sep 2004 A1
20040264549 Hoole Dec 2004 A1
20050004748 Pinto et al. Jan 2005 A1
20050052451 Servantie Mar 2005 A1
20050073455 Chow et al. Apr 2005 A1
20050126679 Kim Jun 2005 A1
20050136625 Henseler et al. Jun 2005 A1
20050150289 Osborne Jul 2005 A1
20050174350 Ridenour et al. Aug 2005 A1
20050200502 Reusser et al. Sep 2005 A1
20050225461 Tsai et al. Oct 2005 A1
20050225481 Bonthron Oct 2005 A1
20050230563 Corcoran, III Oct 2005 A1
20060004497 Bull Jan 2006 A1
20060097895 Reynolds et al. May 2006 A1
20060098452 Choi et al. May 2006 A1
20060164284 Pauplis et al. Jul 2006 A1
20060207967 Bocko et al. Sep 2006 A1
20060215265 Miyatake et al. Sep 2006 A1
20060227012 He Oct 2006 A1
20060244636 Rye et al. Nov 2006 A1
20060245171 Kim et al. Nov 2006 A1
20060290253 Yeo et al. Dec 2006 A1
20060290531 Reynolds et al. Dec 2006 A1
20070001897 Alland Jan 2007 A1
20070002078 He et al. Jan 2007 A1
20070008214 Wasiewicz Jan 2007 A1
20070013575 Lee et al. Jan 2007 A1
20070018887 Feyereisen et al. Jan 2007 A1
20070032951 Tanenhaus et al. Feb 2007 A1
20070060063 Wright et al. Mar 2007 A1
20070146364 Aspen Jun 2007 A1
20070171094 Alter et al. Jul 2007 A1
20070176794 Feyereisen et al. Aug 2007 A1
20070179684 He Aug 2007 A1
20070228586 Merrill et al. Oct 2007 A1
20070247350 Ryan Oct 2007 A1
20070279253 Priest Dec 2007 A1
20070297736 Sherman et al. Dec 2007 A1
20080018524 Christianson Jan 2008 A1
20080051947 Kemp Feb 2008 A1
20080065286 Han Mar 2008 A1
20080074308 Becker et al. Mar 2008 A1
20080111731 Hubbard et al. May 2008 A1
20080145610 Muller et al. Jun 2008 A1
20080180351 He Jul 2008 A1
20080305721 Ohashi et al. Dec 2008 A1
20090021397 Wipf et al. Jan 2009 A1
20090040070 Alter et al. Feb 2009 A1
20090040772 Laney Feb 2009 A1
20090046229 Umemoto et al. Feb 2009 A1
20090126872 Sampica et al. May 2009 A1
20090148682 Higuchi Jun 2009 A1
20090152391 McWhirk Jun 2009 A1
20090153783 Umemoto Jun 2009 A1
20090164067 Whitehead et al. Jun 2009 A1
20090207048 He et al. Aug 2009 A1
20090219189 Bateman Sep 2009 A1
20090279030 Toyama et al. Nov 2009 A1
20090279175 Laney et al. Nov 2009 A1
20090324009 Schulz Dec 2009 A1
20100033499 Gannon et al. Feb 2010 A1
20100103353 Yamada Apr 2010 A1
20100297406 Schaffer et al. Nov 2010 A1
20100312428 Roberge et al. Dec 2010 A1
20100312461 Haynie et al. Dec 2010 A1
20110037616 Leutelt et al. Feb 2011 A1
20110054729 Whitehead et al. Mar 2011 A1
20110075070 Kitagawa et al. Mar 2011 A1
20110141405 Kitagawa et al. Jun 2011 A1
20110165361 Sherman et al. Jul 2011 A1
20110184594 Manfred et al. Jul 2011 A1
20110273325 Goldman Nov 2011 A1
20110282580 Mohan Nov 2011 A1
20110304479 Chen et al. Dec 2011 A1
20120053831 Halder Mar 2012 A1
20120133546 Reiter May 2012 A1
20120150426 Conway Jun 2012 A1
20120174445 Jones et al. Jul 2012 A1
20120176497 Shadmi Jul 2012 A1
20120215410 McClure et al. Aug 2012 A1
20130041529 He et al. Feb 2013 A1
20130188049 Koukol Jul 2013 A1
20130234884 Bunch et al. Sep 2013 A1
20130285847 Ward Oct 2013 A1
20140009324 Ranney et al. Jan 2014 A1
20150211883 He Jul 2015 A1
20160131739 Jinkins et al. May 2016 A1
Foreign Referenced Citations (29)
Number Date Country
102269809 Dec 2011 CN
104898833 Sep 2015 CN
19649838 Apr 1998 DE
19949737 Apr 2001 DE
0 556 351 Jun 1995 EP
0 962 752 Dec 1999 EP
1 092 821 Apr 2001 EP
0 814 744 Jun 1959 GB
1 092 821 Nov 1967 GB
01-210328 Aug 1989 JP
05-200880 Aug 1993 JP
05-293895 Nov 1993 JP
06-051484 Feb 1994 JP
H08-220547 Aug 1996 JP
09-057779 Mar 1997 JP
10-156853 Jun 1998 JP
10-244589 Sep 1998 JP
2000-141388 May 2000 JP
2004-233590 Aug 2004 JP
2004-354645 Dec 2004 JP
2006-218658 Aug 2006 JP
2006-334912 Dec 2006 JP
2006-348208 Dec 2006 JP
2007-206559 Aug 2007 JP
2007-302398 Nov 2007 JP
2008-238607 Jan 2008 JP
WO-9305634 Mar 1993 WO
WO-2009133102 Nov 2009 WO
WO-2011089474 Jul 2011 WO
Non-Patent Literature Citations (154)
Entry
Notice of Allowance for U.S. Appl. No. 15/222,923 dated Feb. 2, 2019. 7 pages
Corrected Notice of Allowability on U.S. Appl. No. 13/250,798, dated Jan. 12, 2017, 2 pages.
Non-Final Office Action for U.S. Appl. No. 15/166,191 dated Apr. 16, 2018. 13 pages.
Non-Final Office Action for U.S. Appl. No. 15/222,923 dated Aug. 31, 2018. 11 pages.
U.S. Appl. No. 11/851,323, filed Sep. 6, 2007, McCusker.
U.S. Appl. No. 11/863,219, filed Sep. 27, 2007, Woodell.
U.S. Appl. No. 11/863,221, filed Sep. 27, 2007, Woodell.
U.S. Appl. No. 11/899,801, filed Sep. 6, 2007, Woodell et al.
U.S. Appl. No. 11/900,002, filed Sep. 6, 2007, Woodell et al.
U.S. Appl. No. 12/167,200, filed Jul. 2, 2008, Woodell et al.
U.S. Appl. No. 12/167,203, filed Jul. 2, 2008, Woodell.
U.S. Appl. No. 12/167,208, filed Jul. 2, 2008, Dickerson et al.
U.S. Appl. No. 12/180,293, filed Jul. 25, 2008, Woodell et al.
U.S. Appl. No. 14/841,558, filed Aug. 31, 2015, Rockwell Collins, Inc.
U.S. Appl. No. 15/222,923, filed Jul. 28, 2016, Rockwell Collins, Inc.
“MountainScope™ on a TabletPC,” PCAvionics™, printed from website www.pcavionics.com on Aug. 28, 2007, 1 page.
“PCAvionics: Makers of MountainScope™, A New Dimension in Situational Awareness,” PCAvionics™, printed from website www.pcavionics.com on Aug. 28, 2007, 1 page.
TAWS Class A and Class B, Terrain Awareness and Warning Systems, Universal® Avionics Systems Corporation, Sep. 2007, 6 pages.
“TAWS Terrain Awareness and Warning System,” Universal® Avionics, printed from website www.uasc.com on Aug. 28, 2007, 2 pages.
Adams, Charlotte, “Synthetic Vision: Picturing the Future,” Avionics magazine, Solutions for Global Airspace Electronics, Oct. 2006, cover and pp. 22-29.
Advisory Action for U.S. Appl. No. 12/009,472, dated Feb. 25, 2013, 3 pages.
Advisory Action for U.S. Appl. No. 13/538,957, dated Jun. 14, 2013, 6 pages.
Airports Authority of India, Chapter 7: Visual Aids for Navigation-Lights, available prior to Jan. 1, 2005, retrieved from the internet at: http://www.aai.aero/aai_employees/chapter_7.pdf on Sep. 26, 2014, 33 pages.
Blue Mountain Avionics' Products, printed from website www.bluemountainavionics.com on Aug. 28, 2007, 4 pages.
Brailovsky et al., REVS122: A Radar-Based Enhanced Vision System for Degraded Visual Environments, Proc. of SPIE vol. 9087 908708-1, retrieved from the internet at http://proceedings.spiedigitallibrary.org on Jun. 25, 2014, 13 pages.
Carter, S. P., D. D. Blankenship, M. E. Peters, D. A. Young, J. W. Holt, and D. L. Morse (2007), Radar-based subglacial lake classification in Antarctica, Geochem. Geophys. Geosyst., 8, 003016, doi:10.1029/2006GC001408, 20 pages.
English Translation of Japanese Notice of Reasons for Rejection in Japanese Application No. 2016-001165, dated Apr. 25, 2017, 1 page.
Federal Aviation Administration, Advisory Circular AC 90-106, “Enhanced Flight Vision Systems”, initiated by AFS-400, dated Jun. 2, 2010, 55 pages.
Federal Aviation Administration, Aeronautical Information Manual (AIM) Basic Flight Information and ATC Procedures, dated Jul. 24, 2014, 2 pages.
Final Office Action on U.S. Appl. No. 13/250,798 dated Sep. 4, 2014, 22 pages.
Final Office Action on U.S. Appl. No. 13/867,556 dated Jul. 3, 2014, 11 pages.
Final Office Action on U.S. Appl. No. 14/536,330, dated Jan. 23, 2017, 14 pages.
Final Office Action on U.S. Appl. No. 13/250,307 dated Jun. 11, 2014, 8 pages.
Final Office Action on U.S. Appl. No. 12/236,464, dated Dec. 5, 2011, 15 pages.
Final Office Action on U.S. Appl. No. 13/250,798 dated Aug. 7, 2015, 21 pages.
First Office Action on Korean Patent Application No. 10-2016-7013740, dated Sep. 19, 2016, 7 pages.
First Office Action with English Translation of Chinese Application No. 201510005057.5, dated Apr. 25, 2017, 8 pages.
Fountain, J.R., Digital Terrain Systems, Airborne Navigation Systems Workshop (Digest No. 1997/169), IEE Colloquium, pp. 4/1-4/6, Feb. 21, 1997.
G2000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=97668 on Jun. 28, 2011, 2 pages.
G3000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=66916 on Jun. 28, 2011, 2 pages.
G5000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=90821&ra=true on Apr. 20, 2011, 2 pages.
Honeywell, RDR-4B Forward looking windshear detection / weather radar system user's manual with radar operating guidelines, Rev. 6, Jul. 2003, 106 pages.
Johnson, A., et al., Vision Guided Landing of an Autonomous Helicopter in Hazardous Terrain, Robotics and Automation, 2005. ICRA 2005. Proceedings of the 2005 IEEE International Conference, pp. 3966-3971, Apr. 18-22, 2005.
Kuntman, D., Airborne system to address leading cause of injuries in non-fatal airline accidents, ICAO Journal, Mar. 2000, 4 pages.
McGray et al., Air Operators, Airlines, Manufacturers and Interested Industry Stakeholders & Aero Chart Forum-Utilizing EFVS technology and incorporating it into FAA NextGen, Federal Aviation Administration, Apr. 23 & 30, 2014, 34 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798 dated Feb. 26, 2016, 9 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798 dated Mar. 18, 2015, 21 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798, dated Sep. 9, 2016, 6 pages.
Non-Final Office Action on U.S. Appl. No. 14/270,587, dated May 8, 2017, 16 pages.
Non-Final Office Action on U.S. Appl. No. 14/301,199 dated Sep. 9, 2015, 18 pages.
Non-Final Office Action on U.S. Appl. No. 14/482,681, dated Dec. 20, 2016, 9 pages.
Non-Final Office Action on U.S. Appl. No. 14/536,330 dated Jul. 13, 2016, 12 pages.
Notice of Allowability on U.S. Appl. No. 12/263,282 dated Apr. 13, 2016, 2 pages.
Notice of Allowance for U.S. Appl. No. 11/863,215, dated Oct. 13, 2009, 8 pages.
Notice of Allowance for U.S. Appl. No. 11/863,219, dated Jun. 23, 2009, 7 pages.
Notice of Allowance for U.S. Appl. No. 11/863,221, dated Aug. 2, 2010, 9 pages.
Notice of Allowance for U.S. Appl. No. 11/899,801, dated Aug. 19, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 11/900,002, dated Sep. 14, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 12/009,372, dated Oct. 13, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/009,373, dated Jun. 16, 2010, 4 pages.
Notice of Allowance for U.S. Appl. No. 12/009,472, dated Sep. 5, 2013, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/167,200, dated Oct. 28, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 12/167,203, dated Jun. 21, 2013, 7 pages.
Notice of Allowance for U.S. Appl. No. 12/167,208, dated Mar. 21, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/180,293, dated Aug. 4, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/786,169, dated Mar. 28, 2013, 6 pages.
Notice of Allowance for U.S. Appl. No. 13/538,957, dated Oct. 3, 2013, 13 pages.
Notice of Allowance on U.S. Appl. No. 12/263,282 dated Jan. 29, 2016, 8 pages.
Notice of Allowance on U.S. Appl. No. 13/241,051 dated Aug. 28, 2014, 9 pages.
Notice of Allowance on U.S. Appl. No. 13/247,742 dated Jul. 30, 2014, 9 pages.
Notice of Allowance on U.S. Appl. No. 13/250,798, dated Sep. 28, 2016, 10 pages.
Notice of Allowance on U.S. Appl. No. 14/301,199 dated Mar. 1, 2016, 11 pages.
Notice of Allowance on U.S. Appl. No. 14/482,681, dated Mar. 28, 2017, 5 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Aug. 6, 2009, 23 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Dec. 15, 2010, 13 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Jul. 5, 2012, 23 pages.
Office Action for U.S. Appl. No. 11/863,215, dated May 27, 2009, 5 pages.
Office Action for U.S. Appl. No. 11/863,215, dated Nov. 12, 2008, 8 pages.
Office Action for U.S. Appl. No. 11/863,219, dated Dec. 12, 2008, 7 pages.
Office Action for U.S. Appl. No. 11/863,221, dated Dec. 18, 2009, 5 pages.
Office Action for U.S. Appl. No. 11/863,221, dated Dec. 8, 2008, 8 pages.
Office Action for U.S. Appl. No. 11/863,221, dated May 26, 2009, 5 pages.
Office Action for U.S. Appl. No. 12/009,372, dated Dec. 20, 2010, 10 pages.
Office Action for U.S. Appl. No. 12/009,372, dated Jun. 13, 2011, 9 pages.
Office Action for U.S. Appl. No. 12/009,373, dated Dec. 30, 2009, 14 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Apr. 16, 2012, 16 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Jan. 14, 2011, 14 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Mar. 20, 2013, 15 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Nov. 3, 2011, 15 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Nov. 9, 2012, 15 pages.
Office Action for U.S. Appl. No. 12/167,200, dated Jul. 21, 2010, 6 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Aug. 26, 2010, 9 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Jul. 20, 2011, 6 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Mar. 7, 2013, 5 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Oct. 31, 2011, 5 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Sep. 21, 2012, 6 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Dec. 30, 2009, 10 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Feb. 7, 2011, 8 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Jun. 3, 2010, 11 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Oct. 19, 2010, 8 pages.
Office Action for U.S. Appl. No. 12/180,293, dated Jan. 4, 2011, 5 pages.
Office Action for U.S. Appl. No. 12/180,293, dated Jul. 28, 2010, 8 pages.
Office Action for U.S. Appl. No. 12/263,282, dated Jan. 5, 2012, 10 pages.
Office Action for U.S. Appl. No. 12/786,169, dated Jan. 18, 2013, 14 pages.
Office Action for U.S. Appl. No. 12/892,563, dated Feb. 19, 2013, 12 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Feb. 15, 2012, 8 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Jul. 10, 2012, 4 pages.
Office Action for U.S. Appl. No. 12/976,871, dated May 6, 2013, 5 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Nov. 21, 2012, 5 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Oct. 9, 2013, 5 pages.
Office Action for U.S. Appl. No. 13/183,314, dated Aug. 14, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/183,314, dated Mar. 28, 2013, 12 pages.
Office Action for U.S. Appl. No. 13/224,992, dated Feb. 28, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/250,307, dated Nov. 5, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/474,559, dated Aug. 28, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/474,559, dated Dec. 28, 2012, 8 pages.
Office Action for U.S. Appl. No. 13/538,957, dated Apr. 4, 2013, 19 pages.
Office Action for U.S. Appl. No. 13/538,957, dated Oct. 5, 2012, 18 pages.
Office Action for U.S. Appl. No. 13/743,182, dated Apr. 8, 2013, 10 pages.
Office Action for U.S. Appl. No. 12/786,169, dated Jul. 20, 2012, 8 pages.
Office Action in Japanese Patent Application 2015-116688, dated Aug. 25, 2015, 4 pages.
Office Action in Japanese Patent Application 2015-116716, dated Aug. 25, 2015, 3 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Jul. 12, 2013, 17 pages.
Office Action on U.S. Appl. No. 11/787,460, dated Mar. 19, 2010, 16 pages.
Office Action on U.S. Appl. No. 11/787,460, dated Sep. 16, 2009, 15 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Feb. 11, 2014, 21 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Jun. 22, 2011, 14 pages.
Office Action on U.S. Appl. No. 12/892,563, dated May 7, 2013, 6 pages.
Office Action on U.S. Appl. No. 12/892,563, dated Oct. 10, 2012, 12 pages.
Office Action on U.S. Appl. No. 13/241,051 dated Feb. 27, 2014, 21 pages.
Office Action on U.S. Appl. No. 13/247,742 dated Dec. 3, 2013, 11 pages.
Office Action on U.S. Appl. No. 13/250,798 dated Apr. 23, 2014, 15 pages.
Office Action on U.S. Appl. No. 13/627,788, dated Jul. 28, 2014, 10 pages.
Office Action on U.S. Appl. No. 13/867,556 dated Feb. 7, 2014, 11 pages.
Office Action U.S. Appl. No. 11/787,460, dated Aug. 31, 2010, 18 pages.
Office Action with English Translation received in Korean Patent Application 10-2010-7017278, dated Aug. 26, 2015, 5 pages.
Pictures of DELPHINS, printed from website www.tunnel-in-the-sky.tudelft.nl on Aug. 28, 2007, 4 pages.
REVS Product Information Sheet, Sierra Nevada Corporation, dated May 7, 2014, 2 pages.
SKOLNIK, Introduction to Radar Systems, McGraw Hill Book Company, 2001, 3 pages.
SKOLNIK, Radar Handbook (McGraw Hill Book Company), 1990, 23 pages.
Synthetic Vision System, en.wikipedia.org/wiki/Synthetic_vision_system, retrieved Feb. 28, 2013, 4 pages.
Technical Standard Order, TSO-C115b, Airborne Area Navigation Equipment Using Multi-Sensor Inputs, Department of Transportation, Federal Aviation Administration, Sep. 30, 1994, 11 pages.
U.S. Office Action on U.S. Appl. No. 11/900,002 dated Jun. 8, 2010, 7 pages.
U.S. Office Action on U.S. Appl. No. 13/247,742 dated Apr. 16, 2014, 15 pages.
U.S. Office Action on U.S. Appl. No. 14/536,330, dated Aug. 11, 2017.
Vadlamani, A., et al., Improving the detection capability of spatial failure modes using downward-looking sensors in terrain database integrity monitors, Digital Avionics Systems Conference, 2003. DASC-03. The 22nd, vol. 2, pp. 9C.5-91-12 vol. 2, Oct. 12-16, 2003.
Van Kasteren, Joost, “Tunnel-in-the-Sky, Synthetic vision simplifies the pilot's job and enhances safety,” printed from website www.delftoutlook.tudelft.nl on Aug. 28, 2007, 13 pages.
Walker, GD-Itronix Dynavue Technology, The Ultimate Outdoor-Readable Touch-Screen Display, Rugged PC Review, 4 pages.
Wang et al., a Simple Based on DSP Antenna Controller of Weather Radar, 2001 CIE International Conference, 4 pages.
Notice of Allowance for U.S. Appl. No. 14/482,681 dated May 1, 2017. 2 pages.
Notice of Allowance for U.S. Appl. No. 14/536,330, dated Dec. 13, 2017. 7 pages.
Notice of Allowance for U.S. Appl. No. 15/166,191 dated Oct. 18, 2108. 8 pages.
Diani et al., “Ground clutter model for airborne MPRF radars in look-down search mode”, IDD Proc.-Radar, Sonar Navig., vol. 143, No. 2. Apr. 1996. 8 pages.
First Office Acton for CN Patent Application No. 201510604462.9 dated Jul. 25, 2019.
Continuation in Parts (1)
Number Date Country
Parent 14482681 Sep 2014 US
Child 15627268 US