Weather radar enabled offshore operation system and method

Information

  • Patent Grant
  • 10353068
  • Patent Number
    10,353,068
  • Date Filed
    Thursday, July 28, 2016
    8 years ago
  • Date Issued
    Tuesday, July 16, 2019
    5 years ago
Abstract
A weather radar system can be used as an airborne sensor for providing an image on an electronic display during low visibility offshore IFR operations (e.g. for a helicopter approach to an offshore platform, such as, a petroleum rig or other structure). The weather radar sensed image is representative of the external surroundings of the maritime environment associated with radar returns received by the weather radar system. Beam sharpening technology produces higher angular resolution of the sensed objects in the radar image which reduces the interpreted azimuth errors from the sensed radar image. Accordingly, beam sharpening technology advantageously allows object isolation of closely clustered offshore platforms and nearby objects. With operational credit provided to these capabilities, the minimal distance for obtaining visual reference with the target platform could be reduced, increasing the success rate of completing offshore operations in low visibility IFR conditions.
Description
CROSS-REFERENCE TO RELATED PATENT APPLICATIONS

The present application is related to U.S. application Ser. No. 15/166,191, filed by Jinkins et al. on May 26, 206, U.S. patent application Ser. No. 14/841,558 filed by Jinkins et al. on Aug. 31, 2015, U.S. Pat. No. 8,773,301, U.S. patent application Ser. No. 14/536,330 filed Nov. 7, 2014 by Jinkins et al., now U.S. Pat. No. 9,934,256, U.S. patent application Ser. No. 14/482,681 filed Sep. 10, 2014 by Wood et al., now U.S. Pat. No. 9,733,349, U.S. patent application Ser. No. 14/301,199 filed on Jun. 10, 2014 by McCusker et al, now U.S. Pat. No. 9,384,586, U.S. patent application Ser. No. 13/627,788 filed on Sep. 26, 2012 by Jinkins et al., U.S. patent application Ser. No. 12/892,563 filed on Sep. 28, 2010 by Woodell et al., now U.S. Pat. No. 8,643,533, U.S. patent application Ser. No. 13/250,798 filed on Sep. 30, 2011 by Jinkins et al., now U.S. Pat. No. 9,562,788, U.S. patent application Ser. No. 12/236,464 filed on Sep. 23, 2008 by McCusker et al, now U.S. Pat. No. 8,977,491, U.S. patent application Ser. No. 12/167,200 filed on Jul. 2, 2008 by Woodell et al., U.S. patent application Ser. No. 12/180,293 filed on Jul. 25, 2008 by Woodell et al., now U.S. Pat. No. 8,077,078, U.S. patent application Ser. No. 13/247,742 filed on Sep. 28, 2011 by Wilson et al., now U.S. Pat. No. 8,896,480, U.S. patent application Ser. No. 11/851,323 filed on Sep. 6, 2007 by McCusker, now U.S. Pat. No. 8,515,600, U.S. patent application Ser. No. 11/900,002 on Sep. 26, 2007, now U.S. Pat. No. 7,859,449, U.S. patent application Ser. No. 13/241,051 filed on Sep. 22, 2011 by Tiana et al., now U.S. Pat. No. 8,917,191, U.S. patent application Ser. No. 12/263,282 filed on Oct. 31, 2008 by McCusker et al., U.S. Pat. Nos. 8,896,480 and 8,077,078, all of which are herein incorporated by reference in their entireties and assigned to the assignee of the present application.


BACKGROUND

Aircraft are utilized to deliver equipment and personal to a variety of locations. Helicopters often land at onshore and offshore helipads which are designated landing areas for helicopters providing a clearly marked hard surface. The direct surroundings of a helipad are not always clear of obstacles and that is one of the reasons why visual reference is required for any operations close to the helipad. The Federal Aviation Administration (FAA) regulations allow part of the approach operation to be performed in low visibility conditions, but at some distance from the helipad visual reference must be obtained.


In one aircraft application, helicopters flying instrument flight rules (IFR) during offshore operations (e.g., approaching an offshore platform such as an oil drilling rig, an oil pumping platform, or a ship) must avoid fixed and moving obstacles and see the intended offshore platform for landing. IFR offshore operations often occur in low visibility conditions, such as, heavy fog associated with maritime environments (e.g., maritime fog). Offshore operations can occur in environments with moving ships having heights up to 400 feet. The missed operation point (MOP) for offshore operations is generally 0.6-0.7 nautical miles (NM) from an offshore helipad or platform.


According to a Delta 30° Offshore Standard Instrument Operation (OSIO), a helicopter or rotary wing aircraft approaching an offshore platform is flown into the wind and at a distance of 1.1 NM from the landing site, a delta 30° turn is made either to the right or left depending on the clear area determined by the helicopter operator based on the image provided by the onboard weather radar. When visual reference with the landing site is established before reaching the MOP at 0.6 NM, the pilot proceeds visually to the landing area, otherwise a missed operation must be executed at the MOP.


According to a Parallel Offset OSIO, a helicopter approaching an offshore helipad or platform is flown into the wind and at a distance of 5 NM from the landing site, either a left or right turn is made based on the clear area determined by the helicopter operator based on the image provided by the onboard weather radar. An offset path of 0.5 NM to the left or right of the landing site is established before reaching the decision point altitude at no less than 2 NM from the landing site. The helicopter is flown along the offset path to a 0.7 NM MOP. When visual reference with the landing site is established before reaching the MOP at 0.7 NM from the landing site, the pilot proceeds visually to the landing area, otherwise a missed operation must be executed at MOP. Offshore helicopter approach operations are discussed in the FAA's Advisory Circular AC 90-80B “Approval of Offshore Helicopter Approaches” (Apr. 19, 1992) and “Airborne Radar Approach FAA/NASA Gulf of Mexico Helicopter Flight Test Program,” Department of Transportation, (Report No. AFO-507-78-2, January 1980). The new nomenclature of the draft Advisory Circular AC 90-80C “Approval of Offshore Standard Instrument Operations, Airborne Radar Operations, and Helicopter En Route Descent Areas” has been adopted, instead of using the nomenclature of the current AC 90-80B.


SUMMARY

In one aspect, embodiments of the inventive concepts disclosed herein are directed to a radar system, a processor, and memory coupled to the processor. The memory contains program instructions that, when executed, cause the processor to instruct the radar system to provide radar beams and receive radar returns with improved angular and/or range resolution for deriving image data of the external scene topography during offshore operations. The radar system derives the image data using intensity and locations of individualized radar returns determined as a relative location to the radar system.


In a further aspect, embodiments of the inventive concepts disclosed herein are directed to a system for use in a helicopter. The system includes a weather radar system and a display. The weather radar system is configured to process the radar returns to determine a location and an intensity of the radar returns and generate image data representative of an offshore environment during offshore operations. The radar returns are in an X-band or a C-band. The display is in communication with the weather radar system and is configured to display an image associated with the image data.


In a further aspect, embodiments of the inventive concepts disclosed herein are directed to a weather radar system for use in a helicopter. The weather radar system includes an antenna and a control circuit coupled with the antenna. The control circuit is configured to provide radar beams via the antenna toward external surroundings, to receive radar returns, process the radar returns to determine the intensity and location associated with the radar returns, and provide data derived from the radar returns for provision of a radar image. The location is determined with high resolution angle and range.


In a further aspect, embodiments of the inventive concepts disclosed herein are directed to a method of providing radar image data for provision of a radar image on an electronic display during offshore operations to avoid obstacles. The method includes determining an antenna attitude, receiving radar returns from at least one of an X-band or C-band weather radar system, and determining a location associated with each of a number of radar returns using the antenna attitude, a beam sharpening angle, and a range. The method also includes determining an intensity associated with each of the radar returns, and providing the radar image data to the electronic display during offshore operations. The radar image data is derived from the intensity and location of the radar returns.





BRIEF DESCRIPTION OF THE DRAWINGS

Implementations of the inventive concepts disclosed herein may be better understood when consideration is given to the following detailed description thereof. Such description makes reference to the annexed drawings, which are not necessarily to scale, and in which some features may be exaggerated and some features may be omitted or may be represented schematically in the interest of clarity. Like reference numerals in the figures may represent and refer to the same or similar element, feature, or function. In the drawings:



FIG. 1A is a side view schematic illustration of a helicopter equipped with weather radar approaching an offshore platform according to an exemplary embodiment of the inventive concepts disclosed herein;



FIG. 1B is a side view and top view schematic illustration of a helicopter equipped with weather radar approaching an offshore platform according to a Parallel Offset OSIO in accordance with an exemplary embodiment of the inventive concepts disclosed herein;



FIG. 1C is a side view and top view schematic illustration of a helicopter equipped with weather radar approaching an offshore platform according to a Delta 30° OSIO in accordance with an exemplary embodiment of the inventive concepts disclosed herein;



FIG. 2 is a schematic general block diagram of a display system for the helicopter illustrated in FIG. 1A, the display system is for providing an image derived from radar data according to another embodiment of the inventive concepts disclosed herein;



FIG. 3A is an illustration of a top-down radar image of an offshore environment derived from radar data provided by a conventional radar display system for the helicopter illustrated in FIG. 1;



FIG. 3B is an illustration of the offshore environment associated with the top-down image of FIG. 3A;



FIG. 4A is an illustration of a top-down radar image of the offshore environment illustrated in FIG. 1A provided by the display system illustrated in FIG. 2 according to yet another exemplary embodiment of the inventive concepts disclosed herein;



FIG. 4B is an illustration of the offshore environment associated with the top-down image of FIG. 4A; and



FIG. 5 is a flow diagram showing operations for the generation of image data for the display system illustrated in FIG. 2 in accordance with some embodiments of the inventive concepts disclosed herein.





DETAILED DESCRIPTION

Before describing in detail the inventive concepts disclosed herein, it should be observed that the inventive concepts disclosed herein include, but are not limited to, a novel structural combination of data/signal processing components, sensors, and/or communications circuits, and not in the particular detailed configurations thereof. Accordingly, the structure, methods, functions, control and arrangement of components, software, and circuits have, for the most part, been illustrated in the drawings by readily understandable block representations and schematic diagrams, in order not to obscure the disclosure with structural details which will be readily apparent to those skilled in the art, having the benefit of the description herein. Further, the inventive concepts disclosed herein are not limited to the particular embodiments depicted in the exemplary diagrams, but should be construed in accordance with the language in the claims.


The weather radar system's superior ability to penetrate heavy fog allows the radar system to detect offshore platforms and other objects associated with offshore environments (e.g., maritime environments).


In some embodiments, a weather radar system employs radar beam sharpening technology to significantly increase the azimuthal accuracy, thereby providing sufficient indication of lateral separation to gain operational credit for less lateral clearance with the maritime obstacles in offshore environments. Without beam sharpening technology, conventional radar systems tend to spread the visualization of the sensed obstacles (e.g., the offshore platform) in azimuth with an angular width that is proportional to the width of the antenna beam. Beam sharpening technology produces higher angular resolution radar estimation of the sensed obstacles which reduces the interpreted azimuth errors from the visualization on the radar display in some embodiments. Accordingly, beam sharpening technology advantageously allows obstruction isolation of closely clustered offshore platforms and nearby obstructions in some embodiments. Beam sharpening is performed using several techniques, including but not limited to mono-pulse technique, a sub-aperture radar technique, a deconvolution of a beam point spread function, or any combination thereof in some embodiments.


In some embodiments, enhancing the range resolution provides the capability to identify the target platform by sensing detailed platform shapes and orientation. Correct identification of the intended target platform reduces the possibility of pilots landing on a non-intended platform that is closely spaced in a cluster of platforms including the intended target platform. The range resolution is enhanced by using techniques including but not limited to ultra-wideband pulsing, stepped-frequency compression, pulse compression, or other compression techniques or combinations thereof in some embodiments.


In some embodiments, a weather radar system is used as an imaging sensor to provide a visualization of the surrounding environment during offshore operations. The visualization can be presented to the pilot on a display (e.g., head up display (HUD) or head down display (HDD)), thereby allowing the pilot to see the surrounding environment in low visibility conditions, such as maritime fog. In some embodiments, a Doppler weather radar system is configured to have enhanced resolution (e.g., angular resolution and/or range resolution) for providing an image derived from radar reflections from offshore structures (e.g., petroleum rigs, petroleum platforms, and/or ships) in some embodiments. In some embodiments, the weather radar system operates in a ground mapping mode enhancing the resulting radar image using beam sharpening with optional increased range resolution. In some embodiments, the intensity, is mapped to position using angle and range without using absolute determined position such as provided by a global navigation satellite system (GNSS).


In some embodiments, a weather radar system achieves accurate location determination using an accurate beam sharpening angle. In some embodiments, the weather radar system creates a two dimensional or three dimensional grid containing the spatial density of the intensity of the radar returns. Each radar return has an associated location (e.g., location relative to the radar system) that is based on the radar antenna position in some embodiments. From the radar antenna position, the location of the radar returns is found using the attitude of the antenna, the beam sharpening angle and the range (e.g., specific to that individual radar return) in some embodiments. In some embodiments the radar returns are processed to determine the spatial density of the intensity associated with the radar returns. The spatial density of the intensity can be used to generate an image representing the offshore environment sensed by the radar. A reference image (e.g., the coordinate of the landing site) can be combined with the radar image to provide additional information of the maritime environment that can help the pilot navigate the maritime environment, but the provided reference image is depending on a database and on an absolute positioning and attitude sources which integrity cannot be guaranteed. The image generated by the radar can be generated using only relative positioning and attitude sources. Combining the two independently created images confirms the alignment and colocation of the radar image and the reference image in order to validate the integrity of the positioning and attitude sources.


The radar return data is processed to provide a two-dimensional helicopter situation display (e.g., a plan position indicator (PPI) providing range and azimuth information or plan view display (e.g., top-down display)) or a three dimensional perspective exocentric or perspective egocentric display representative of the offshore environment based on the radar returns as described in U.S. patent application Ser. Nos. 14/841,558, 14/301,199, now U.S. Pat. No. 9,384,586, Ser. No. 14/482,681, now U.S. Pat. No. 9,733,349, and Ser. No. 14/536,330, now U.S. Pat. No. 9,939,526, incorporated herein by reference in their entireties in some embodiments. For example, the radar processing circuit can be embodied as a processor and a non-transitory memory containing program instructions that, when executed, cause the processor to instruct the radar system to provide radar beams and receive radar returns via the antenna and generate image data from the radar returns.


Referring to FIG. 1A, an aircraft, such as a helicopter 20, approaches an offshore platform 24 in a maritime environment 25. The offshore platform 24 includes a petroleum rig 26 and a landing site 28. Current FAA regulations prohibit the helicopter 20 from descending below an altitude of 500 feet at location 21 not less than 2 NM from the landing site 28 unless the path of the helicopter 20 is laterally clear of all obstructions by at least 0.5 NM. Obstructions such as ships as tall as 400 feet can be in the path to the landing site 28 on the offshore platform 24. The helicopter 20 is required to sense and avoid all obstructions (other offshore platforms, ships, etc.) using an on-board weather radar, such as a radar system 30 (shown in more detail in FIG. 2). In some embodiments, operational credit is provided to the capabilities of the radar system 30, which reduces the minimal distance for obtaining visual reference with the target platform, resulting in higher success rates of completing offshore operations in low visibility IFR conditions. With reference to FIGS. 1B-C, the helicopter 20 performs an approach according to the Parallel Offset OSIO or the Delta 30° OSIO in some embodiments.


In some embodiments, the helicopter 20 using the radar system 30 descends to an altitude 50 feet above the altitude of the landing site 28 (e.g., to an altitude of 200 feet) at a distance of at least 0.6 NM (e.g., at a location 31 between 2.0 NM and 0.6 NM) from the landing site 28. The radar system 30 senses obstacles and the offshore platform 24 at a distance of 2.0 NM through a fog 32. In some embodiments, a missed operation point (MOP) 34 is located at a distance of 0.6-0.7 NM from the landing site. When visual reference with the offshore platform 24 is established before reaching the MOP, the pilot proceeds visually to the landing site 28, otherwise a missed operation must be executed at the MOP. With the radar system 30 employed in the helicopter 20, Applicants believe that a MOP at location 36 closer to the offshore platform 24 can be achieved in some embodiments.


Although offshore environments are discussed herein, the radar system 30 can be utilized with other environments. Although offshore platform applications are discussed herein, the radar system 30 can be utilized with other applications, including but not limited to military applications, rescue applications, policing applications, construction applications, helicopter delivery applications, and shipping applications. In addition, the radar system 30 can be utilized with other types of aircraft than helicopters including unmanned or unmanned aircraft and drones.


With reference to FIG. 2, the helicopter 20 (FIG. 1A) includes a display system 50 which can be used in offshore operations, such as, petroleum rig operations, to view the maritime environment 25 (FIG. 1A). The display system 50 includes display 52 embodied as head down display (HDD) or other flight display. In some embodiments, the display 52 can also be embodied as a head up display (HUD). In some embodiments, the display 52 is provided as a wearable display. The display system 50 can be a weather radar system in communication with a separate display.


The display 52 can be used to provide information to the flight crew, thereby increasing visual range and enhancing decision-making abilities. In some embodiments, the display 52 is configured to display a reference image (e.g., top-down image) of the maritime environment around the helicopter 20. In another embodiment, the display 52 is configured to display a perspective egocentric image or perspective exocentric image. The display 52 can be implemented using any of a variety of display technologies, including CRT, LCD, organic LED, dot matrix display, and others.


Advantageously, real time radar data can be provided to provide a real time, all weather detection of obstacles and the offshore platform 24 in the maritime environment 25 in some embodiments.


Referring to FIG. 2, the display system 50 uses radar data from a radar system 30 to provide an image to display 52. The display system 50 is in communication with or includes the radar system 30 and onboard sensors 105 and can include or be in communication with a reference display system 107 in some embodiments. The reference display system 107 is optional in some embodiments. The radar system 30 includes a radar antenna 120, a processor 125, a radar data storage unit 126, and an image frame memory 128 in some embodiments.


The onboard sensors 105 are associated with navigation equipment and flight control devices in some embodiments and can include one or more of an inertial navigation system (INS), an inertial reference system (IRS), a flight computer, a barometric altimeter, a radar altimeter, a global navigation satellite system (GNSS) receiver, air speed sensors, heading sensors, etc. The onboard sensors 105 include an attitude sensor 106 and a position sensor 108 in some embodiments.


The position sensor 108 is any device or system for providing an electronic signal or data representing position of the helicopter 20. In some embodiments, the position includes a position parameter (e.g., latitude and longitude) and an altitude parameter. The position sensor 108 is a GNSS receiver or INS in some embodiments. The altitude can be provided by a GNSS receiver, an altitude sensor (e.g., barometric sensor), a radio altimeter, or other device. In some embodiments, the position also includes velocity and acceleration parameters associated with position (e.g., for predicting future position of the helicopter 20).


The attitude sensor 106 is any device for providing an electronic signal or data representing attitude of the helicopter 20. The attitude includes a heading parameter, a pitch parameter, a yaw parameter, and a roll parameter in some embodiments. In some embodiments, the position sensor 108 and the attitude sensor 106 are parts of a single positioning or navigation sensor system, such as a GNSS receiver, INS, or integrated GNSS/INS. In some embodiments, the attitude also includes angular velocity and acceleration parameters associated with attitude (e.g., for predicting future attitude of the helicopter 20). In some embodiments, the radar antenna 120 has an attitude sensor for providing attitude and changes in attitude of the radar antenna 120 with respect to the helicopter 20. The attitude of the radar antenna 120 relative to the attitude of the helicopter can be represented by a scan angle parameter for the radar antenna 120.


The radar system 30 receives data from the onboard sensors 105 and provides radar image data for display by the display system 50. The radar system 30 is a weather radar system generally located inside the nose of the helicopter 20 in some embodiments. The radar system 30 can be a weather radar system, such as, a Multiscan™ radar system from Rockwell Collins, Inc. configured as described herein. The radar system 30 utilizes a split, half or sub-aperture or other technique for obtaining radar data associated with external surroundings in some embodiments. The radar system 30 can use the split or sub-aperture techniques of the radar systems described in U.S. application Ser. Nos. 14/841,558, 13/627,788, 12/892,563, 13/250,798, 12/236,464, and 12/167,200 and U.S. Pat. No. 8,077,078, incorporated herein by reference in their entirety and assigned to the assignee of the present application. The type of the radar system 30 and data gathering techniques are not discussed in the specification in a limiting fashion.


The processor 125 uses radar data stored in the radar data storage unit 126 and the data from the onboard sensors 105 to provide and store radar image data in the image frame memory 128. The radar data can be provided directly from the transmit/receive circuits associated with the radar antenna 120 or be stored in the processor 125 in some embodiments. The processor 125 includes a location determiner 134, an accumulator 136, and an antenna scan angle determiner 138, a beam sharpening angle determiner 140, a range determiner 142, an intensity determiner 144, and a radar image generator 127. The accumulator 136, the location determiner 134, the radar data storage unit 126, the antenna scan angle determiner 138, the beam sharpening angle determiner 140, the range determiner 142, the intensity determiner 144, and the radar image generator 127 are software modules, circuits, or combinations thereof in some embodiments.


The location determiner 134 uses data from the onboard sensors 105 and from the radar data storage unit 126, the antenna scan angle determiner 138, the beam sharpening angle determiner 140, and the range determiner 142 to identify a location associated with returns received at the radar antenna 120 in some embodiments. The locations are relative to the helicopter 20 or the radar system 30 (e.g., the radar antenna 120) and are determined without the use of an absolute position source (e.g., a GNSS receiver) in some embodiments. Data associated with the radar returns are stored in the radar data storage unit 126 in some embodiments. The location determiner 134 uses the position of the radar antenna 120, the attitude of the helicopter 20, the antenna scan angle, the beam sharpening angle, and the range to determine the location of the radar returns in some embodiments. The intensity determiner 144 determines an intensity associated with the radar returns in some embodiments. The intensity of the collection of radar returns is accumulated by location in the accumulator 136 in some embodiments. The accumulated intensity by location is used by the radar image generator 127 to provide the radar image data in some embodiments.


The processor 125 causes the radar antenna 120 to provide radar signals or beams and to receive radar returns (e.g., weather radar return data). The processor 125 is an electronic processor that processes the radar returns and provides the radar data associated with the radar returns to the radar data storage unit 126. The radar signals and radar returns are in the X-band or C-band in some embodiments.


The radar system 30 provides the radar data (e.g., weather radar return data) to the radar data storage unit 126 in one embodiment. The radar data can be processed and filtered for various weather sensing functions as well as location functions. In some embodiments, the radar image generator 127 provides the radar image data (e.g., in image frame format) for storage in the image frame memory 128 using the accumulated returns from the accumulator 136. In some embodiments, the accumulator 136 determines spatial density of the intensity by area or volume to create the radar image data. The spatial density is used by the radar image generator 127 to provide the radar image data indicative of an intensity of reflections from objects at locations associated with that area or volume in some embodiments.


The image frame memory 128 is a video or graphic electronic memory in some embodiments. The radar image data represents a plan position indicator view, a plan view, a perspective exocentric, or perspective egocentric view image in some embodiments. The radar image generator 127 and the accumulator 136 are an electronic memory, processor circuitry, or combination thereof in some embodiments. In some embodiments, the radar image generator 127 and accumulator 136 are part of a digital signal processor or the processor 125 or is a software module executing on the digital signal processor or the processor 125. The accumulator 136 is embodied as a register in some embodiments. The radar data storage unit 126 is an electronic memory in some embodiments.


The radar data associated with the external surroundings can represent sensed objects and the location of the sensed objects. Objects include ships, platforms, etc. Improved angular resolution and range resolution techniques discussed in U.S. patent application Ser. No. 14/536,330 filed Nov. 7, 2014 by Jinkins et al. incorporated herein by reference in its entirety allows the location of the objects to be more accurately determined and represented in the radar image data in some embodiments. The radar system 30 can utilize clutter suppression and Doppler filtering to improve performance in some embodiments.


In some embodiments, the radar system 30 provides radar data representing a 120 degree field of view in accordance with a weather radar sweep. In some embodiments, the sweep is directed toward the path of the helicopter 20. In yet some other embodiments, the sweep is directed towards the location of the landing site. The width and direction of the sweep can be varied in certain embodiments. Various types of sweeps, beam widths, sweep patterns, and sweep speeds can be utilized without departing from the scope of the inventive concepts disclosed herein.


The radar system 30 embodied as a weather radar allows existing avionic equipment to be used as a real-time sensor for providing a radar-derived image of the external scene topography to the pilot in some embodiments. The image or representation generated by the radar system 30 is provided on the display 52 can provide situational awareness to the pilot in some embodiments. In other embodiments, the image or representation generated by the radar system 30 is provided on the display 52 can be part of an offshore operations system approved for use with closer MOPs.


The radar system 30 advantageously provides increased range resolution in some embodiments. The increased resolution in range and angle allows a higher resolution for location determination by the location determiner 134 and hence higher image resolution to be provided on the display 52 in some embodiments.


According to some embodiments, the radar system 30 uses a beam sharpening method to achieve increased angular resolution. In some embodiments, the radar system 30 uses techniques, such as, beam sharpening (e.g., horizontal beam sharpening) and de-convolution of the beam point spread function for improved angular resolution. In some embodiments, the radar system 30 can use beam sharpening as a process that improves the antenna-induced poor angular resolution (e.g., due to the beam width). There are many methods that can be used such as: Monopulse Radar, Sub-Aperture Radar or Split-Aperture Radar, etc. Mathematical methods can be utilized to determine a center of the radar returns from an object like the offshore platform 24, ships and other obstacles. Techniques for beam sharpening and determining beam sharpening angles and directions of radar objects are discussed in U.S. patent application Ser. Nos. 13/627,788, 12/892,563, 13/250,798, 12/236,464, and 12/167,200 and U.S. Pat. No. 8,077,078 incorporated herein by reference in their entireties.


The radar system 30 uses the radar antenna 120 that toggles between transmitting and receiving on the full aperture and transmitting on the full aperture while receiving on the partial aperture in some embodiments. These techniques can be used to accurately estimate at which angle the radar return was located within the radar beam in some embodiments. The received returns are processed to determine a high resolution estimate of a beam sharpening angle that is relative to the boresight of the antenna in some embodiments. According to some embodiments, the returns are processed using a complex conjugate multiplication method to determine the beam sharpening angle determined by the beam sharpening angle determiner 140. The processing can be related to sequential lobing processing but is executed in the phase domain as opposed to the common amplitude domain in some embodiments.


In some embodiments, the radar system 30 uses sequential lobing techniques where two antennas that are close to the same place may be used, going back and forth between the two antennas. An amplitude signature or phase signature that varies between the two halves of the antennas may be used to obtain data associated with the returned radar reflection from sensed objects (e.g., an object such as ships or platforms). Sequential lobing generally does not use phase comparisons with moving objects due to Doppler-induced phase changes that contaminate the phase center measurement. However, using a complex conjugate multiply method allows the Doppler-induced phase changes to be removed by cancellation. Therefore, a change in phase center between multiple different sub-apertures may be determined and used to determine the beam sharpening angle associated with the returned radar reflection from a sensed object.


In some embodiments, the effective waveform bandwidth of the radar system 30 is increased to increase range resolution. To provide higher range resolution, the radar system 30 provides ultra-wideband radar (UWB) pulses (e.g., extremely narrow pulses with high power), or provides intra pulse compression (frequency of phase modulation of the transmitted pulse) in some embodiments. Frequency coding techniques including the common linear frequency modulation (LFM) or chirp method and discrete coded segments within the pulse are utilized in some embodiments. Phase coding techniques including binary phase codes as well as various polyphase codes can be utilized in some embodiments. To provide higher range resolution, the radar system 30 provides interpulse pulse compression or stepped frequency compression (e.g., successive pulses with discrete increasing frequency steps) in some embodiments. In some embodiments, stepped frequency compression advantageously achieves high effective bandwidth with narrow instantaneous bandwidth. The receive bandwidth is smaller, has lower noise bandwidth, and a higher signal to noise ratio in some embodiments. Analog-to-digital sampling rates are lower (vs. pulse-compression) in some embodiments. In addition, the stepped frequency compression also has a smaller peak power (e.g., when compared to impulse), provides flexible transmit frequency control, can “hop” over restricted or undesired transmit frequencies, enables adaptive/cognitive frequency use, and rejects later received clutter from earlier transmit pulses in some embodiments. Further, the stepped frequency compression techniques can provide returns from clutter in ambiguous ranges that have frequencies that are different from returns from objects and rejects ambiguous clutter returns in the receiver IF filter of the radar system 30 in some embodiments. Stepped frequency compression generally does not achieve range resolution with a single pulse, requires transmit, receive and processing of a group of pulses for any one bin, and has more pronounced range-Doppler coupling (e.g., different Doppler shifts for each frequency) in some embodiments.


The processor 125 can be any hardware and/or software processor or processing architecture configured to execute instructions and operate on navigational and radar data. The processor 125 can be capable of determining navigational information such as altitude, heading, bearing, location, and changes thereof based on data from the onboard sensors 105. The processor 125 can be, or can include one or more microprocessors, an application specific integrated circuit (ASIC), a circuit containing one or more processing components, a group of distributed processing components, circuitry for supporting a microprocessor, or other hardware configured for processing.


In some embodiments, the display system 50 includes the reference display system 107 including a reference database 150, a reference image generator 152, a frame combiner 154, a reference image frame memory 156, and a combined image frame memory 162. Reference image frame memory 156 and the combined image frame memory 162 are similar to the image frame memory 128. In some embodiments, the display system 50 is configured to provide a combined image using the radar image data stored in the image frame memory 128 and a reference image stored in the reference image frame memory 156.


The reference image generator 152 is configured to provide an image based upon position as sensed by onboard sensors 105 and reference feature data stored in the reference data base 150. The reference image is generated as a plan position indicator view, a top-down view, a perspective exocentric, or perspective egocentric view in some embodiments. The reference image can include a textual or graphic indication of the location of the offshore platform 24, other platforms, and the landing site 28 in some embodiments. In some embodiments, a user interface can be used to input a selected platform, route, or landing location.


In some embodiments, the reference database 150 is memory unit storing reference feature data indicating locations of objects in the maritime environment referenced to absolute location. In some embodiments, the reference image generator 152 and the frame combiner 154 are software modules (e.g., executed on processor 125), circuits, or combinations thereof in some embodiments. A hardware platform associated with the radar system 30 (e.g., the processor 125), the display computer 103, an SVS, or an EVS can be used to provide the operations of the reference display system 107.


In some embodiments, the frame combiner 154 receives the reference image data from the reference image frame memory 156 or reference image generator 152, receives the radar image data from the image frame memory 128 or the radar image generator 127, and provides combined image data to the combined image frame memory 162. The combined image data can be displayed on the display 52 via display computer 103, respectively. In some embodiments, both the reference image data and the radar image data are generated from the same viewpoint using the same projection method so that both images are in the same location, format and size and are overlaid upon each other to provide the combined image data. In some embodiments, the radar image data is generated only using relative positioning and attitude sources and the reference image data is generated using absolute positioning and attitude sources. In some embodiments, the reference image data has priority over the radar image data so that sensed images associated with the radar image data are not blocking or overwriting the features in the reference image data. Other merging techniques can be utilized by the frame combiner 154.


In some embodiments, a location translation operation is performed by the radar image generator 127 to project the radar measurement data into radar image data. In some embodiments, the radar image generator 127 is using a projection viewpoint reference for generating the radar image data that is using the same location and attitude reference as the projection viewpoint that the reference image generator 152 is using for generating the reference image data to generate a plan position indicator view, a top-down view, a perspective exocentric, or perspective egocentric view.


A flow can be performed by the display system 50 in some embodiments similar to flow 300 discussed in U.S. patent application Ser. No. 14/841,558 and U.S. application Ser. No. 15/166,191. The processor 125, reference display system 107, or other computing platform can execute software to provide an image while the helicopter 20 is involved in offshore operations in response to helicopter sensor parameters and radar data. The helicopter sensor parameters from the onboard sensors 105 and the radar returns received by the radar antenna 120 are processed to determine locations and intensity associated with the radar returns to provide locations which have an intensity of the radar return. In some embodiments, these radar returns with their associated location and intensity are accumulated over an area or volume at an operation 136. The locations are determined as relative locations from the helicopter 20, radar system 30 or radar antenna 120 in some embodiments.


With reference to FIG. 3A, an image 400 of the maritime environment 25 (shown in FIG. 3B) of the radar data associated with the radar returns obtained by a radar system without beam sharpening is displayed. The image 400 is a plan position indicator (PPI) or top-down view of a real world scene 402 including the offshore platform 24 (including the petroleum rig 26 and landing site 28) and a ship 403 shown in FIG. 3B. A representation 404 of the position of the helicopter 20 is shown approximately 1.75 NM from the landing site 28. A region 406 represents radar returns from the offshore platform 24 associated with the offshore platform 24 and has a much wider expanse than the width of the offshore platform 24 in real world dimensions. The metal material and shape associated with the offshore platform 24 results in high intensity returns over a large area in contrast with the water surface which has relatively low intensity returns. The intensity of returns is represented by colors where levels above 24 decibels relative a return of a drop of rain with a diameter of 1 millimeter (mm) (dBZ) are represented by the color white, levels between 24 and 12 dBZ are represented by the color pink, levels between 12 and 4 dBZ are represented by the color red, levels between 4 and −4 dBZ are represented by the color yellow, levels between −4 dBZ and −0 dBZ are represented by the color pink, and minimal or zero intensity levels are represented by the color black. A key 405 shows an exemplary color level representation. The levels discussed above are exemplary only.


The reflected radar return of the real world offshore platform 24 is visualized by the region 406. The region 406 has a lateral expanse of approximately 0.5 NM with the lateral expanse of the core being 0.25 NM. The region 406 has a depth of approximately 0.25 NM with the core being 0.1 NM. The large expanse makes it difficult to estimate the actual location of the offshore platform 24 using the radar image. The image 400 also includes a region 412 associated with an obstacle. The obstacle is a ship 403 in the path between the region 406 and the representation 404 of the position of the helicopter 20. The region 412 has a much broader expanse and depth than the real world dimensions which make it more difficult to estimate a precise location of the obstacle.


With reference to FIG. 4A, an image 500 of the maritime environment 25 (shown in FIG. 4B) derived from the radar data associated with the radar returns obtained by the radar system 30 with beam sharpening is displayed. The image 500 is a plan position indicator (PPI) or top-down view of the maritime environment 25 represented by the image 402 (FIG. 4B). A representation 504 of the position of the helicopter 20 is shown 1.75 NM from a landing site symbol 506. A region 508 represents reflections from the offshore platform 24 and has a narrower expanse than the region 406 (FIG. 3A).


The region 508 has a lateral expanse of less than 0.25 NM with the lateral expanse of the core being less than 0.1 NM. The region 508 has a depth of approximately 0.20 NM with the core being less than 0.1 NM. The smaller expanse of region 508 makes it easier to estimate a more accurate location of the offshore platform 24 using the radar data and allows a pilot to safely get closer to the offshore platform 24 before visually identifying it and confirming its actual location. The current lateral separation of 0.5 NM for the parallel offset OSIO procedure and the Delta 30° OSIO procedure is based on radars without beam sharpening that have a large expanse of the visualized radar reflection. Beam sharpening allows the pilot to safely use less lateral separation, which would result in the MOP also positioned closer to the offshore platform. Even small reductions in the distance of the MOP can result in higher possibilities of visual confirmation and less missed approaches in the fog 32 (FIG. 1A). In addition, larger expanses of region 508 can result in difficulty in distinguishing groups of offshore platforms in a small area as the returns for the group would appear as one large region and increases the possibility of landing at the wrong offshore platform.


The image 500 also includes a region 512 which identifies an obstacle. The obstacle is the ship 403 in the path between the region 508 and the representation 504 of the position of the helicopter 20. The region 512 has a narrower expanse than the region 412 which makes it possible to estimate a more accurate location of the obstacle in order to safely avoid the obstacle. In some embodiments, the enhanced processing performed by the radar system 30 allows a ship to be more readily identified as compared to offshore platforms. The longer, thinner profile of the region 512 indicates that it is more likely a ship traveling toward or away from the helicopter 20 as opposed to an offshore platform which has a larger lateral expanse with respect to radar returns.


In some embodiments, a landing site symbol 506 can be added to the visualization on the displayed radar image. In some embodiments, the visualization of the landing site symbol 506 is based on the absolute coordinates of the landing site and a projection of those absolute coordinates to local display coordinates. For a PPI display those local coordinates are azimuth and range from the position of the helicopter taking into account the attitude of the antenna. For a top-down view those local coordinates could be longitudinal distance and lateral distance from the position of the helicopter taking into account the attitude of the antenna. Colocation of the region 508 and the landing site symbol 506 validates the integrity of the absolute positioning and attitude sources and validates the integrity of the coordinate of the landing site.


With reference to FIG. 5, a flow 700 can be performed by the display system 50 in some embodiments. The processor 125 or other computing platform can execute flow 700 to provide an image in response to sensor parameters and radar data. At an operation 702, radar beams are provided by the radar antenna 120. At an operation 704, radar returns are received by the radar antenna 120. At an operation 706, radar data and sensor data associated with the radar returns are processed to provide the intensity and location of the radar returns. In some embodiments, the radar system 30 provides data representing a 120 degree field of view in accordance with a weather radar sweep. The width and direction of the sweep can be varied in certain embodiments. The sweep is directed toward the surface of the ocean so that returns are obtained which allow a radar image to be created from the surrounding maritime environment. Various types of sweeps, scans and timings of sweeps and scans can be utilized without departing from the scope of the invention.


The locations of the radar returns are determined using the radar data and parameters from the onboard sensors 105 in some embodiments. The location is determined using beam sharpening techniques in some embodiments. At an operation 710, radar image data is generated from the radar returns with their associated intensity and location data. In some embodiments, the radar image data is generated based on azimuth and range from the position of the helicopter. In some embodiments, the radar image data is generated based on a coordinate transformation or projection method. At an operation 712, reference data is provided from a reference database. At an operation 714, reference image data is generated from the reference data. The reference image data is generated based on a transformation using a viewpoint and projection method. The location of the viewpoint is related to the absolute coordinates of the reference database in some embodiments. In some embodiments the viewpoint location and projection method for operation 710 and operation 714 are the same, even though the viewpoint location can be referenced relative or absolute for operations 710 and 714. At an operation 716, the radar image data from operation 710 and reference image data from operation 714 is combined to provide combined image data for display. The combined image includes a feature (e.g., the landing site symbol 506 in FIG. 4) for the landing site 28. At an operation 718 the image data is displayed as an image on a display. In some embodiments operations 712, 714 and 716 are not performed, in which case the image data from operation 710 is displayed on a display in operation 718 without combining it with the reference image data.


The scope of this disclosure should be determined by the claims, their legal equivalents and the fact that it fully encompasses other embodiments which may become apparent to those skilled in the art. All structural, electrical and functional equivalents to the elements of the above-described disclosure that are known to those of ordinary skill in the art are expressly incorporated herein by reference and are intended to be encompassed by the present claims. A reference to an element in the singular is not intended to mean one and only one, unless explicitly so stated, but rather it should be construed to mean at least one. No claim element herein is to be construed under the provisions of 35 U.S.C. § 112, sixth paragraph, unless the element is expressly recited using the phrase “means for.” Furthermore, no element, component or method step in the present disclosure is intended to be dedicated to the public, regardless of whether the element, component or method step is explicitly recited in the claims.


Embodiments of the inventive concepts disclosed herein have been described with reference to drawings. The drawings illustrate certain details of specific embodiments that implement the systems and methods and programs of the present disclosure. However, describing the embodiments with drawings should not be construed as imposing any limitations that may be present in the drawings. The present disclosure contemplates methods, systems and program products on any machine-readable media for accomplishing its operations. Embodiments of the inventive concepts disclosed herein may be implemented using an existing computer processor, or by a special purpose computer processor incorporated for this or another purpose or by a hardwired system.


Embodiments in the inventive concepts disclosed herein have been described in the general context of method steps which may be implemented in one embodiment by a program product including machine-executable instructions, such as program code, for example in the form of program modules executed by machines in networked environments. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Machine-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represent examples of corresponding acts for implementing the functions described in such steps.


It should be noted that although the diagrams herein may show a specific order and composition of method steps, it is understood that the order of these steps may differ from what is depicted. For example, two or more steps may be performed concurrently or with partial concurrence. Also, some method steps that are performed as discrete steps may be combined, steps being performed as a combined step may be separated into discrete steps, the sequence of certain processes may be reversed or otherwise varied, and the nature or number of discrete processes may be altered or varied. The order or sequence of any element or apparatus may be varied or substituted according to alternative embodiments. Accordingly, all such modifications are intended to be included within the scope of the present disclosure.


The foregoing description of embodiments has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the subject matter to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of the subject matter disclosed herein. The embodiments were chosen and described in order to explain the principals of the disclosed subject matter and its practical application to enable one skilled in the art to utilize the disclosed subject matter in various embodiments and with various modifications as are suited to the particular use contemplated. Other substitutions, modifications, changes and omissions may be made in the design, operating conditions and arrangement of the embodiments without departing from the scope of the presently disclosed subject matter.

Claims
  • 1. A weather radar system for use in a helicopter, the weather radar system comprising: an antenna;a control circuit coupled with the antenna and configured to: provide radar beams via the antenna toward external surroundings;receive radar returns;process data associated with the radar returns to determine an intensity associated with the radar returns and a location associated with the radar returns, the location being determined with high resolution angle and high resolution range using at least the following parameters: a beam sharpening angle, and an antenna scan angle; andprocess the radar returns with the associated intensity and location to provide radar image data.
  • 2. The weather radar system of claim 1, wherein the control circuit is further configured to: determine the location associated with each of the radar returns using an antenna attitude, the beam sharpening angle, and a range.
  • 3. The weather radar system of claim 2, wherein motion of the antenna is taken into account in determining the location associated with the radar returns.
  • 4. The weather radar system of claim 3, wherein an attitude of the antenna is determined using an attitude sensor associated with the antenna.
  • 5. The weather radar system of claim 3, wherein the beam sharpening angle is a radar sensed direction of an individual radar return with respect to an attitude of the antenna.
  • 6. The weather radar system of claim 5, wherein the beam sharpening angle is generated using a monopulse technique, a sub-aperture radar technique, deconvolution of a beam point spread function, or any combination thereof.
  • 7. The weather radar system of claim 3, wherein the control circuit is further configured to provide combined data, the combined data being comprised of landing site data and the radar image data.
  • 8. The weather radar system of claim 1, wherein the control circuit is further configured to: determine the location associated with each of the radar returns using ultra-wideband pulsing, stepped-frequency compression, or pulse compression.
  • 9. A method of providing radar image data for provision of a radar image during offshore operations to avoid obstacles, the method comprising: determining an antenna attitude of an antenna;receiving radar returns from at least one of an X-hand and a C-band weather radar system;determining a location of each of the radar returns using the antenna attitude, a beam sharpening angle, and a range;determining an intensity associated with each of the radar returns; andproviding radar image data, the radar image data being derived from the intensity and location associated with the radar returns.
  • 10. The method of claim 9, wherein motion of the antenna is taken into account in determining the location associated with each of the radar returns.
  • 11. The method of claim 9, wherein the beam sharpening angle is determined using at least one of: a monopulse technique, a sub-aperture or split aperture technique, deconvolution of a beam point spread function, or a combination thereof.
  • 12. The method of claim 9, wherein the radar image data is provided to identify a landing site.
  • 13. The method of claim 12, wherein the radar image data is combined with landing site image data.
  • 14. The method of claim 13, further comprising: comparing the radar image data to the landing site image data to confirm colocation of a radar sensed landing site and the landing site associated with the landing site image data.
  • 15. A system for a helicopter, comprising: a weather radar system configured to: process the radar returns to determine an intensity and location of the radar returns to provide image data representative of an offshore environment associated with radar returns received by the weather radar system during offshore operations, the radar returns being in an X-band or a C-band, wherein the location is determined using at least an antenna position and a beam sharpening angle; anda display in communication with the weather radar system and configured to display an image associated with the image data.
  • 16. The system of claim 15, wherein the image is a combined image including a landing site symbol and the image data derived from intensities and locations associated with the radar returns.
  • 17. The system of claim 16, wherein the weather radar system is configured to provide radar beams having pulses using ultra-wideband pulsing, stepped-frequency compression, or pulse compression.
  • 18. The system of claim 15, wherein the radar returns are processed using the beam sharpening angle, the beam sharpening angle being a radar sensed direction of an individual radar return with respect to a boresight of an antenna.
  • 19. The system of claim 18, wherein the beam sharpening angle is generated using a monopulse technique, a sub-aperture radar technique, deconvolution of a beam point spread function, or any combination thereof.
  • 20. The system of claim 15, wherein the image data is compared to landing site reference data to check integrity of helicopter location and orientation or integrity of the landing site reference data.
US Referenced Citations (552)
Number Name Date Kind
2416155 Chubb Feb 1947 A
2849184 Arden et al. Aug 1958 A
2929059 Parker Mar 1960 A
2930035 Altekruse Mar 1960 A
2948892 White Aug 1960 A
2965894 Sweeney Dec 1960 A
2994966 Senitsky et al. Aug 1961 A
3031660 Young Apr 1962 A
3049702 Schreitmueller Aug 1962 A
3064252 Varela Nov 1962 A
3070795 Chambers Dec 1962 A
3071766 Fenn Jan 1963 A
3072903 Meyer Jan 1963 A
3089801 Tierney et al. May 1963 A
3107351 Milam Oct 1963 A
3113310 Standing Dec 1963 A
3129425 Sanner Apr 1964 A
3153234 Begeman et al. Oct 1964 A
3175215 Blasberg et al. Mar 1965 A
3212088 Alexander et al. Oct 1965 A
3221328 Walter Nov 1965 A
3241141 Wall Mar 1966 A
3274593 Varela et al. Sep 1966 A
3325807 Burns et al. Jun 1967 A
3334344 Colby, Jr. Aug 1967 A
3339199 Pichafroy Aug 1967 A
3373423 Levy Mar 1968 A
3396391 Anderson Aug 1968 A
3397397 Barney Aug 1968 A
3448450 Alfandari et al. Jun 1969 A
3618090 Garrison Nov 1971 A
3680094 Bayle et al. Jul 1972 A
3716855 Asam Feb 1973 A
3739380 Burdic et al. Jun 1973 A
3781878 Kirkpatrick Dec 1973 A
3810175 Bell May 1974 A
3815132 Case et al. Jun 1974 A
3816718 Hall et al. Jun 1974 A
3851758 Makhijani et al. Dec 1974 A
3866222 Young Feb 1975 A
3885237 Kirkpatrick May 1975 A
3956749 Magorian May 1976 A
4024537 Hart May 1977 A
4058701 Gruber et al. Nov 1977 A
4058710 Altmann Nov 1977 A
4063218 Basov et al. Dec 1977 A
4103300 Gendreu et al. Jul 1978 A
4235951 Swarovski Nov 1980 A
4277845 Smith et al. Jul 1981 A
4405986 Gray Sep 1983 A
4435707 Clark Mar 1984 A
4481519 Margerum Nov 1984 A
4509048 Jain Apr 1985 A
4532515 Cantrell et al. Jul 1985 A
4594676 Breiholz et al. Jun 1986 A
4595925 Hansen Jun 1986 A
4598292 Devino Jul 1986 A
4628318 Alitz Dec 1986 A
4646244 Bateman et al. Feb 1987 A
4649388 Atlas Mar 1987 A
4654665 Kiuchi et al. Mar 1987 A
4685149 Smith et al. Aug 1987 A
4723124 Boles Feb 1988 A
4760396 Barney et al. Jul 1988 A
4828382 Vermilion May 1989 A
4843398 Houston et al. Jun 1989 A
4912477 Lory et al. Mar 1990 A
4914436 Bateman et al. Apr 1990 A
4924401 Bice et al. May 1990 A
4939513 Paterson et al. Jul 1990 A
4951059 Taylor, Jr. Aug 1990 A
4953972 Zuk Sep 1990 A
4965573 Gallagher et al. Oct 1990 A
4987419 Salkeld Jan 1991 A
5045855 Moreira Sep 1991 A
5047779 Hager Sep 1991 A
5047781 Bleakney Sep 1991 A
5049886 Seitz et al. Sep 1991 A
5053778 Imhoff Oct 1991 A
5166688 Moreira Nov 1992 A
5173703 Mangiapane et al. Dec 1992 A
5175554 Mangiapane et al. Dec 1992 A
5198819 Susnjara Mar 1993 A
5202690 Frederick Apr 1993 A
5247303 Cornelius et al. Sep 1993 A
5273553 Hoshi et al. Dec 1993 A
5311183 Mathews et al. May 1994 A
5329391 Miyamoto et al. Jul 1994 A
5332998 Avignon et al. Jul 1994 A
5345241 Huddle Sep 1994 A
5365356 McFadden Nov 1994 A
5383457 Cohen Jan 1995 A
5442364 Lee et al. Aug 1995 A
5530440 Danzer et al. Jun 1996 A
5539409 Mathews et al. Jul 1996 A
5559515 Alimena et al. Sep 1996 A
5559518 Didomizio Sep 1996 A
5566840 Waldner et al. Oct 1996 A
5592178 Chang et al. Jan 1997 A
5678303 Wichmann Oct 1997 A
5736957 Raney Apr 1998 A
5820080 Eschenbach Oct 1998 A
5828332 Frederick Oct 1998 A
5831570 Ammar et al. Nov 1998 A
5839080 Muller et al. Nov 1998 A
5867119 Corrubia et al. Feb 1999 A
5894286 Morand et al. Apr 1999 A
5918517 Malapert et al. Jul 1999 A
5920276 Frederick Jul 1999 A
5923279 Bamler et al. Jul 1999 A
5936575 Azzarelli et al. Aug 1999 A
5942062 Hassall et al. Aug 1999 A
5945926 Ammar et al. Aug 1999 A
5950512 Fields Sep 1999 A
5959762 Bandettini et al. Sep 1999 A
5978715 Briffe et al. Nov 1999 A
6002347 Daly et al. Dec 1999 A
6023240 Sutton Feb 2000 A
6061016 Lupinski et al. May 2000 A
6061022 Menegozzi et al. May 2000 A
6064942 Johnson et al. May 2000 A
6075484 Daniel et al. Jun 2000 A
6092009 Glover Jul 2000 A
6112141 Briffe et al. Aug 2000 A
6112570 Hruschak Sep 2000 A
6122570 Muller et al. Sep 2000 A
6127944 Daly et al. Oct 2000 A
6128066 Yokozeki Oct 2000 A
6128553 Gordon et al. Oct 2000 A
6138060 Conner et al. Oct 2000 A
6150901 Auken Nov 2000 A
6154151 McElreath et al. Nov 2000 A
6154169 Kuntman Nov 2000 A
6157339 Sato et al. Dec 2000 A
6157891 Lin Dec 2000 A
6163021 Mickelson Dec 2000 A
6166661 Anderson et al. Dec 2000 A
6169770 Henely Jan 2001 B1
6178391 Anderson et al. Jan 2001 B1
6184816 Zheng et al. Feb 2001 B1
6188330 Glover Feb 2001 B1
6194980 Thon Feb 2001 B1
6199008 Aratow et al. Mar 2001 B1
6201494 Kronfeld Mar 2001 B1
6204806 Hoech Mar 2001 B1
6205400 Lin Mar 2001 B1
6208284 Woodell et al. Mar 2001 B1
6219592 Muller et al. Apr 2001 B1
6233522 Morici May 2001 B1
6236351 Conner et al. May 2001 B1
6259400 Higgins et al. Jul 2001 B1
6266114 Skarohlid Jul 2001 B1
6278799 Hoffman Aug 2001 B1
6281832 McElreath Aug 2001 B1
6285298 Gordon Sep 2001 B1
6285337 West et al. Sep 2001 B1
6285926 Weiler et al. Sep 2001 B1
6289277 Feyereisen et al. Sep 2001 B1
6311108 Ammar et al. Oct 2001 B1
6317468 Meyer Nov 2001 B1
6317690 Gia Nov 2001 B1
6317872 Gee et al. Nov 2001 B1
6340946 Wolfson et al. Jan 2002 B1
6345127 Mitchell Feb 2002 B1
6359585 Bechman et al. Mar 2002 B1
6366013 Leenders et al. Apr 2002 B1
6373418 Abbey Apr 2002 B1
6374286 Gee et al. Apr 2002 B1
6377202 Kropfli et al. Apr 2002 B1
6377892 Johnson et al. Apr 2002 B1
6388607 Woodell May 2002 B1
6388608 Woodell et al. May 2002 B1
6388724 Campbell et al. May 2002 B1
6389354 Hicks et al. May 2002 B1
6401038 Gia Jun 2002 B2
6411890 Zimmerman Jun 2002 B1
6421000 McDowell Jul 2002 B1
6421603 Pratt et al. Jul 2002 B1
6424288 Woodell Jul 2002 B1
6426717 Maloratsky Jul 2002 B1
6426720 Ross et al. Jul 2002 B1
6427122 Lin Jul 2002 B1
6441773 Kelly et al. Aug 2002 B1
6445310 Bateman et al. Sep 2002 B1
6448922 Kelly Sep 2002 B1
6452511 Kelly et al. Sep 2002 B1
6456236 Hauck et al. Sep 2002 B1
6456238 Posey Sep 2002 B1
6462703 Hedrick Oct 2002 B2
6473026 Ali-Mehenni et al. Oct 2002 B1
6473037 Vail et al. Oct 2002 B2
6473240 Dehmlow Oct 2002 B1
6481482 Shimotomai Nov 2002 B1
6492934 Hwang et al. Dec 2002 B1
6501424 Haendel et al. Dec 2002 B1
6512476 Woodell Jan 2003 B1
6512527 Barber et al. Jan 2003 B1
6516272 Lin Feb 2003 B2
6516283 McCall et al. Feb 2003 B2
6520056 Nemeth et al. Feb 2003 B1
6525674 Kelly et al. Feb 2003 B1
6531669 Miller et al. Mar 2003 B1
6549161 Woodell Apr 2003 B1
6567728 Kelly et al. May 2003 B1
6574030 Mosier Jun 2003 B1
6577947 Kronfeld et al. Jun 2003 B1
6590528 Dewulf Jul 2003 B1
6591171 Ammar et al. Jul 2003 B1
6593875 Bergin et al. Jul 2003 B2
6600443 Landt Jul 2003 B2
6603425 Woodell Aug 2003 B1
6614057 Silvernail et al. Sep 2003 B2
6650275 Kelly et al. Nov 2003 B1
6650291 West et al. Nov 2003 B1
6653947 Dwyer et al. Nov 2003 B2
6667710 Cornell et al. Dec 2003 B2
6681668 Smirle Jan 2004 B1
6690298 Barber et al. Feb 2004 B1
6690299 Suiter Feb 2004 B1
6690317 Szeto et al. Feb 2004 B2
6697008 Sternowski Feb 2004 B1
6697012 Lodwig et al. Feb 2004 B2
6710663 Berquist Mar 2004 B1
6714186 Mosier et al. Mar 2004 B1
6720890 Ezroni et al. Apr 2004 B1
6724344 Stockmaster et al. Apr 2004 B1
6731236 Hager et al. May 2004 B1
6738011 Evans May 2004 B1
6739929 Furukawa et al. May 2004 B2
6741203 Woodell May 2004 B1
6741208 West et al. May 2004 B1
6744382 Lapis et al. Jun 2004 B1
6744408 Stockmaster Jun 2004 B1
6757624 Hwang et al. Jun 2004 B1
6760155 Murayama et al. Jul 2004 B2
6771626 Golubiewski et al. Aug 2004 B1
6782392 Weinberger et al. Aug 2004 B1
6799095 Owen et al. Sep 2004 B1
6803245 Auch et al. Oct 2004 B2
6804614 McGraw et al. Oct 2004 B1
6806846 West Oct 2004 B1
6807538 Weinberger et al. Oct 2004 B1
6813777 Weinberger et al. Nov 2004 B1
6819983 McGraw Nov 2004 B1
6822617 Mather et al. Nov 2004 B1
6825804 Doty Nov 2004 B1
6832538 Hwang Dec 2004 B1
6839017 Dillman Jan 2005 B1
6842288 Liu et al. Jan 2005 B1
6850185 Woodell Feb 2005 B1
6862323 Loper Mar 2005 B1
6862501 He Mar 2005 B2
6865452 Burdon Mar 2005 B2
6879280 Bull et al. Apr 2005 B1
6879886 Wilkins et al. Apr 2005 B2
6882302 Woodell et al. Apr 2005 B1
6908202 Graf et al. Jun 2005 B2
6917396 Hiraishi et al. Jul 2005 B2
6918134 Sherlock et al. Jul 2005 B1
6933885 Stockmaster et al. Aug 2005 B1
6938258 Weinberger et al. Aug 2005 B1
6950062 Mather et al. Sep 2005 B1
6959057 Tuohino Oct 2005 B1
6972727 West et al. Dec 2005 B1
6977608 Anderson et al. Dec 2005 B1
6984545 Grigg et al. Jan 2006 B2
6990022 Morikawa et al. Jan 2006 B2
6992614 Joyce Jan 2006 B1
6995726 West et al. Feb 2006 B1
6998648 Silvernail Feb 2006 B2
6998908 Sternowski Feb 2006 B1
6999022 Vesel et al. Feb 2006 B1
6999027 Stockmaster Feb 2006 B1
7002546 Stuppi et al. Feb 2006 B1
7010398 Wilkins et al. Mar 2006 B2
7023375 Klausing et al. Apr 2006 B2
7026956 Wenger et al. Apr 2006 B1
7028304 Weinberger et al. Apr 2006 B1
7030945 Umemoto et al. Apr 2006 B2
7034753 Elsallal et al. Apr 2006 B1
7042387 Ridenour et al. May 2006 B2
7053796 Barber May 2006 B1
7057549 Block Jun 2006 B2
7064680 Reynolds et al. Jun 2006 B2
7069120 Koenck et al. Jun 2006 B1
7089092 Wood et al. Aug 2006 B1
7092645 Sternowski Aug 2006 B1
7098913 Etherington et al. Aug 2006 B1
7109912 Paramore et al. Sep 2006 B1
7109913 Paramore et al. Sep 2006 B1
7123260 Brust Oct 2006 B2
7129885 Woodell et al. Oct 2006 B1
7145501 Manfred et al. Dec 2006 B1
7148816 Carrico Dec 2006 B1
7151507 Herting Dec 2006 B1
7158072 Venkatachalam et al. Jan 2007 B1
7161525 Finley et al. Jan 2007 B1
7170446 West et al. Jan 2007 B1
7170959 Abbey Jan 2007 B1
7180476 Guell et al. Feb 2007 B1
7191406 Barber et al. Mar 2007 B1
7196329 Wood et al. Mar 2007 B1
7205933 Snodgrass Apr 2007 B1
7209070 Gilliland et al. Apr 2007 B2
7212216 He et al. May 2007 B2
7218268 Vandenberg May 2007 B2
7219011 Barber May 2007 B1
7242343 Woodell Jul 2007 B1
7242345 Raestad et al. Jul 2007 B2
7250903 McDowell Jul 2007 B1
7265710 Deagro Sep 2007 B2
7269657 Alexander et al. Sep 2007 B1
7272472 McElreath Sep 2007 B1
7273403 Yokota et al. Sep 2007 B2
7280068 Lee et al. Oct 2007 B2
7289058 Shima Oct 2007 B2
7292178 Woodell et al. Nov 2007 B1
7292180 Schober Nov 2007 B2
7295150 Burlet et al. Nov 2007 B2
7295901 Little et al. Nov 2007 B1
7301496 Honda et al. Nov 2007 B2
7307576 Koenigs Dec 2007 B1
7307577 Kronfeld et al. Dec 2007 B1
7307583 Woodell et al. Dec 2007 B1
7312725 Berson et al. Dec 2007 B2
7312743 Ridenour et al. Dec 2007 B2
7317427 Pauplis et al. Jan 2008 B2
7321332 Focke et al. Jan 2008 B2
7337043 Bull Feb 2008 B2
7349154 Aiura et al. Mar 2008 B2
7352292 Alter et al. Apr 2008 B2
7361240 Kim Apr 2008 B2
7372394 Woodell et al. May 2008 B1
7373223 Murphy May 2008 B2
7375678 Feyereisen et al. May 2008 B2
7379014 Woodell et al. May 2008 B1
7379796 Walsdorf et al. May 2008 B2
7381110 Sampica et al. Jun 2008 B1
7417578 Woodell et al. Aug 2008 B1
7417579 Woodell Aug 2008 B1
7423578 Tietjen Sep 2008 B1
7446697 Burlet et al. Nov 2008 B2
7446938 Miyatake et al. Nov 2008 B2
7452258 Marzen et al. Nov 2008 B1
7474262 Alland Jan 2009 B2
7479920 Niv Jan 2009 B2
7486220 Kronfeld et al. Feb 2009 B1
7486291 Berson et al. Feb 2009 B2
7492304 Woodell et al. Feb 2009 B1
7492305 Woodell et al. Feb 2009 B1
7515069 Dorneich et al. Apr 2009 B2
7515087 Woodell et al. Apr 2009 B1
7515088 Woodell et al. Apr 2009 B1
7525448 Wilson et al. Apr 2009 B1
7528765 Woodell et al. May 2009 B1
7528915 Choi et al. May 2009 B2
7541970 Godfrey et al. Jun 2009 B1
7541971 Woodell et al. Jun 2009 B1
7551451 Kim et al. Jun 2009 B2
7557735 Woodell et al. Jul 2009 B1
7566254 Sampica et al. Jul 2009 B2
7570177 Reynolds et al. Aug 2009 B2
7576680 Woodell Aug 2009 B1
7603209 Dwyer et al. Oct 2009 B2
7609200 Woodell et al. Oct 2009 B1
7612706 Honda et al. Nov 2009 B2
7616150 Woodell Nov 2009 B1
7633428 McCusker et al. Dec 2009 B1
7633430 Wichgers et al. Dec 2009 B1
7633584 Umemoto et al. Dec 2009 B2
7639175 Woodell Dec 2009 B1
7664601 Daly, Jr. Feb 2010 B2
7675461 McCusker et al. Mar 2010 B1
7693621 Chamas Apr 2010 B1
7696921 Finley et al. Apr 2010 B1
7714767 Kronfeld et al. May 2010 B1
7733264 Woodell et al. Jun 2010 B1
7783427 Woodell et al. Aug 2010 B1
7783429 Walden et al. Aug 2010 B2
7791529 Filias et al. Sep 2010 B2
7808422 Woodell et al. Oct 2010 B1
7814676 Sampica et al. Oct 2010 B2
7843380 Woodell Nov 2010 B1
7859448 Woodell et al. Dec 2010 B1
7859449 Woodell et al. Dec 2010 B1
7864103 Weber et al. Jan 2011 B2
7868811 Woodell et al. Jan 2011 B1
7872594 Vesel Jan 2011 B1
7889117 Woodell et al. Feb 2011 B1
7889118 Finley et al. Feb 2011 B1
7927440 Matsuhira Apr 2011 B2
7929086 Toyama et al. Apr 2011 B2
7965223 McCusker Jun 2011 B1
7965225 Dickerson et al. Jun 2011 B1
8035547 Flanigan et al. Oct 2011 B1
8038498 Miyauchi et al. Oct 2011 B2
8045098 Kitagawa et al. Oct 2011 B2
8059025 D'Addio Nov 2011 B2
8068984 Smith et al. Nov 2011 B2
8072368 Woodell Dec 2011 B1
8077078 Woodell et al. Dec 2011 B1
8077081 Bateman et al. Dec 2011 B2
8102487 Kitagawa et al. Jan 2012 B2
8118075 Sampica et al. Feb 2012 B2
8137498 Sampica et al. Mar 2012 B2
8140223 Whitehead et al. Mar 2012 B2
8159464 Gribble et al. Apr 2012 B1
8232917 Scherzinger et al. Jul 2012 B2
8289202 Christianson Oct 2012 B1
8296065 Haynie et al. Oct 2012 B2
8373580 Bunch et al. Feb 2013 B2
8410975 Bell et al. Apr 2013 B1
8477062 Kanellis Jul 2013 B1
8486535 Nemeth et al. Jul 2013 B1
8493241 He Jul 2013 B2
8515600 McCusker Aug 2013 B1
8540002 Sampica et al. Sep 2013 B2
8558731 Woodell Oct 2013 B1
8576112 Garrec et al. Nov 2013 B2
8583315 Whitehead et al. Nov 2013 B2
8594879 Roberge et al. Nov 2013 B2
8603288 Sampica et al. Dec 2013 B2
8634993 McClure et al. Jan 2014 B2
8639416 Jones et al. Jan 2014 B2
8643533 Woodell et al. Feb 2014 B1
8691043 Sampica et al. Apr 2014 B2
8717226 Bon et al. May 2014 B2
8755954 McCusker et al. Jun 2014 B1
8773301 Woodell Jul 2014 B1
8847794 Buratto et al. Sep 2014 B2
8896480 Wilson et al. Nov 2014 B1
8909471 Jinkins et al. Dec 2014 B1
8917191 Tiana et al. Dec 2014 B1
8936057 Sampica et al. Jan 2015 B2
8976042 Chiew et al. Mar 2015 B1
9354633 McCusker et al. May 2016 B1
20010023390 Gia Sep 2001 A1
20010050372 Krijn et al. Dec 2001 A1
20010053648 Furukawa et al. Dec 2001 A1
20020039070 Ververs et al. Apr 2002 A1
20020111717 Scherzinger et al. Aug 2002 A1
20020116125 Lin Aug 2002 A1
20020116126 Lin Aug 2002 A1
20020158256 Yamada et al. Oct 2002 A1
20020179229 Chuzles Dec 2002 A1
20020185600 Kerr Dec 2002 A1
20020187284 Kinoshita et al. Dec 2002 A1
20030021491 Brust Jan 2003 A1
20030038916 Nakano et al. Feb 2003 A1
20030043315 Umemoto et al. Mar 2003 A1
20030071828 Wilkins et al. Apr 2003 A1
20030089214 Fukuta et al. May 2003 A1
20030093187 Walker May 2003 A1
20030102999 Bergin et al. Jun 2003 A1
20030156238 Hiraishi et al. Aug 2003 A1
20030160718 Nagasaku Aug 2003 A1
20030174396 Murayama et al. Sep 2003 A1
20030180528 Flosenzier et al. Sep 2003 A1
20030189606 Moon et al. Oct 2003 A1
20030195672 He Oct 2003 A1
20030216859 Martell et al. Nov 2003 A1
20030222887 Wilkins et al. Dec 2003 A1
20040044445 Burdon Mar 2004 A1
20040059473 He Mar 2004 A1
20040066645 Graf et al. Apr 2004 A1
20040072575 Young et al. Apr 2004 A1
20040083038 He Apr 2004 A1
20040145499 Schmidt et al. Jul 2004 A1
20040160341 Feyereisen et al. Aug 2004 A1
20040160364 Regev Aug 2004 A1
20040181318 Redmond et al. Sep 2004 A1
20040264549 Hoole Dec 2004 A1
20050004748 Pinto et al. Jan 2005 A1
20050052451 Servantie Mar 2005 A1
20050073455 Chow Apr 2005 A1
20050126679 Kim Jun 2005 A1
20050136625 Henseler et al. Jun 2005 A1
20050150289 Osborne Jul 2005 A1
20050174350 Ridenour et al. Aug 2005 A1
20050200502 Reusser et al. Sep 2005 A1
20050225481 Bonthron Oct 2005 A1
20050230563 Corcoran, III Oct 2005 A1
20060004497 Bull Jan 2006 A1
20060097895 Reynolds et al. May 2006 A1
20060098452 Choi et al. May 2006 A1
20060164284 Pauplis et al. Jul 2006 A1
20060207967 Bocko et al. Sep 2006 A1
20060215265 Miyatake et al. Sep 2006 A1
20060227012 He Oct 2006 A1
20060244636 Rye et al. Nov 2006 A1
20060245171 Kim et al. Nov 2006 A1
20060290253 Yeo et al. Dec 2006 A1
20060290531 Reynolds et al. Dec 2006 A1
20070001897 Alland Jan 2007 A1
20070002078 He et al. Jan 2007 A1
20070008214 Wasiewicz Jan 2007 A1
20070013575 Lee et al. Jan 2007 A1
20070018887 Feyereisen et al. Jan 2007 A1
20070032951 Tanenhaus et al. Feb 2007 A1
20070060063 Wright et al. Mar 2007 A1
20070146364 Aspen Jun 2007 A1
20070171094 Alter et al. Jul 2007 A1
20070176794 Feyereisen et al. Aug 2007 A1
20070179684 He Aug 2007 A1
20070228586 Merrill et al. Oct 2007 A1
20070247350 Ryan Oct 2007 A1
20070279253 Priest Dec 2007 A1
20070297736 Sherman et al. Dec 2007 A1
20080018524 Christianson Jan 2008 A1
20080051947 Kemp Feb 2008 A1
20080074308 Becker et al. Mar 2008 A1
20080111731 Hubbard et al. May 2008 A1
20080145610 Muller et al. Jun 2008 A1
20080180351 He Jul 2008 A1
20080305721 Ohashi et al. Dec 2008 A1
20090021397 Wipf et al. Jan 2009 A1
20090040070 Alter et al. Feb 2009 A1
20090040772 Laney Feb 2009 A1
20090046229 Umemoto et al. Feb 2009 A1
20090148682 Higuchi Jun 2009 A1
20090152391 McWhirk Jun 2009 A1
20090153783 Umemoto Jun 2009 A1
20090164067 Whitehead et al. Jun 2009 A1
20090207048 He et al. Aug 2009 A1
20090279030 Toyama et al. Nov 2009 A1
20090279175 Laney et al. Nov 2009 A1
20100033499 Gannon et al. Feb 2010 A1
20100103353 Yamada Apr 2010 A1
20100297406 Schaffer et al. Nov 2010 A1
20100312428 Roberge et al. Dec 2010 A1
20100312461 Haynie et al. Dec 2010 A1
20110037616 Leutelt et al. Feb 2011 A1
20110054729 Whitehead et al. Mar 2011 A1
20110075070 Kitagawa et al. Mar 2011 A1
20110141405 Kitagawa et al. Jun 2011 A1
20110165361 Sherman et al. Jul 2011 A1
20110184594 Manfred et al. Jul 2011 A1
20110273325 Goldman Nov 2011 A1
20110282580 Mohan Nov 2011 A1
20110304479 Chen et al. Dec 2011 A1
20120053831 Halder Mar 2012 A1
20120133546 Reiter May 2012 A1
20120150426 Conway Jun 2012 A1
20120174445 Jones et al. Jul 2012 A1
20120176497 Shadmi Jul 2012 A1
20120215410 McClure et al. Aug 2012 A1
20130041529 He et al. Feb 2013 A1
20130234884 Bunch et al. Sep 2013 A1
20130285847 Ward Oct 2013 A1
20140009324 Ranney et al. Jan 2014 A1
20150211883 He Jul 2015 A1
20160131739 Jinkins et al. May 2016 A1
Foreign Referenced Citations (26)
Number Date Country
19649838 Apr 1998 DE
19949737 Apr 2001 DE
0 556 351 Jun 1995 EP
0 962 752 Dec 1999 EP
0 814 744 Jun 1959 GB
1 092 821 Nov 1967 GB
01-210328 Aug 1989 JP
05-200880 Aug 1993 JP
05-293895 Nov 1993 JP
06-051484 Feb 1994 JP
H08-220547 Aug 1996 JP
09-057779 Mar 1997 JP
10-156853 Jun 1998 JP
10-244589 Sep 1998 JP
2000-141388 May 2000 JP
2004-233590 Aug 2004 JP
2004-354645 Dec 2004 JP
2006-218658 Aug 2006 JP
2006-334912 Dec 2006 JP
2006-348208 Dec 2006 JP
2007-206559 Aug 2007 JP
2007-302398 Nov 2007 JP
2008-238607 Jan 2008 JP
WO-9305634 Mar 1993 WO
WO-2009133102 Nov 2009 WO
WO-2011089474 Jul 2011 WO
Non-Patent Literature Citations (154)
Entry
U.S. Appl. No. 15/222,923, filed Jul. 28, 2016, Rockwell Collins, Inc.
English Translation of Japanese Notice of Reasons for Rejection in Japanese Application No. 2016-001165, dated Apr. 25, 2017, 1 page.
Final Office Action on U.S. Appl. No. 14/536,330, dated Jan. 23, 2017, 14 pages.
First Office Action with English Translation of Chinese Application No. 201510005057.5, dated Apr. 25, 2017, 8 pages.
Non-Final Office Action on U.S. Appl. No. 14/270,587, dated May 8, 2017, 16 pages.
Notice of Allowance on U.S. Appl. No. 14/482,681, dated Mar. 28, 2017, 5 pages.
U.S. Office Action on U.S. Appl. No. 14/536,330, dated Aug. 11, 2017.
Corrected Notice of Allowability on U.S. Appl. No. 13/250,798, dated Jan. 12, 2017, 2 pages.
Notice of Allowance for U.S. Appl. No. 14/536,330 dated Dec. 13, 2017. 7 pages.
Non-Final Office Action for U.S. Appl. No. 15/166,191 dated Apr. 16, 2018. 13 pages.
U.S. Appl. No. 11/851,323, filed Sep. 6, 2007, McCusker.
U.S. Appl. No. 11/863,219, filed Sep. 27, 2007, Woodell.
U.S. Appl. No. 11/863,221, filed Sep. 27, 2007, Woodell.
U.S. Appl. No. 11/899,801, filed Sep. 6, 2007, Woodell et al.
U.S. Appl. No. 11/900,002, filed Sep. 6, 2007, Woodell et al.
U.S. Appl. No. 12/167,200, filed Jul. 2, 2008, Woodell et al.
U.S. Appl. No. 12/167,203, filed Jul. 2, 2008, Woodell.
U.S. Appl. No. 12/167,208, filed Jul. 2, 2008, Dickerson et al.
U.S. Appl. No. 12/180,293, filed Jul. 25, 2008, Woodell et al.
U.S. Appl. No. 12/236,464, filed Sep. 23, 2008, Rockwell Collins.
U.S. Appl. No. 12/786,169, filed May 24, 2010, Nemeth et al.
U.S. Appl. No. 13/224,992, filed Sep. 2, 2011, Hufnagel et al.
U.S. Appl. No. 13/250,307, filed Sep. 30, 2011, Jinkins.
U.S. Appl. No. 13/250,798, filed Sep. 30, 2011, Jinkins.
U.S. Appl. No. 13/627,788, filed Sep. 26, 2012, Rockwell Collins.
U.S. Appl. No. 13/857,955, filed Apr. 5, 2013, Barber et al.
U.S. Appl. No. 13/250,798, filed Sep. 30, 2011, Rockwell Collins.
U.S. Appl. No. 14/301,199, filed Jun. 10, 2014, Rockwell Collins.
U.S. Appl. No. 14/482,681, filed Sep. 10, 2014, Rockwell Collins.
U.S. Appl. No. 14/841,558, filed Aug. 31, 2015, Rockwell Collins, Inc.
“MountainScope™ on a TabletPC,” PCAvionics™, printed from website www.pcavionics.com on Aug. 28, 2007, 1 page.
TAWS Class A and Class B, Terrain Awareness and Warning Systems, Universal® Avionics Systems Corporation, Sep. 2007, 6 pages.
“TAWS Terrain Awareness and Warning System,” Universal® Avionics, printed from website www.uasc.com on Aug. 28, 2007, 2 pages.
Adams, Charlotte, “Synthetic Vision: Picturing the Future,” Avionics magazine, Oct. 1, 2006, printed from website www.aviationtoday.com, 4 pages.
Adams, Charlotte, “Synthetic Vision: Picturing the Future,” Avionics magazine, Solutions for Global Airspace Electronics, Oct. 2006, cover and pp. 22-29.
Airports Authority of India, Chapter 7: Visual Aids for Navigation—Lights, available prior to Jan. 1, 2005, retrieved from the internet at: http://www.aai.aero/aai_employees/chapter_7.pdf on Sep. 26, 2014, 33 pages.
Blue Mountain Avionics' Products, printed from website www.bluemountainavionics.com on Aug. 28, 2007, 4 pages.
Brailovsky et al., REVS122: A Radar-Based Enhanced Vision System for Degraded Visual Environments, Proc. of SPIE vol. 9087 908708-1, retrieved from the internet at http://proceedings.spiedigitallibrary.org on Jun. 25, 2014, 13 pages.
Carter, S. P., D. D. Blankenship, M. E. Peters, D. A. Young, J. W. Holt, and D. L. Morse (2007), Radar-based subglacial lake classification in Antarctica, Geochem. Geophys. Geosyst., 8, 003016, doi:10.1029/2006GC001408, 20 pages.
Federal Aviation Administration, Advisory Circular AC 90-106, “Enhanced Flight Vision Systems”, initiated by AFS-400, dated Jun. 2, 2010, 55 pages.
Federal Aviation Administration, Aeronautical Information Manual (AIM) Basic Flight Information and ATC Procedures, dated Jul. 24, 2014, 2 pages.
Final Office Action on U.S. Appl. No. 13/250,798 dated Sep. 4, 2014, 22 pages.
Final Office Action on U.S. Appl. No. 13/867,556 dated Jul. 3, 2014, 11 pages.
Final Office Action on U.S. Appl. No. 13/250,307 dated Jun. 11, 2014, 8 pages.
Final Office Action on U.S. Appl. No. 13/250,798 dated Aug. 7, 2015, 21 pages.
First Office Action on Korean Patent Application No. 10-2016-7013740, dated Sep. 19, 2016, 7 pages.
Fountain, J.R., Digital Terrain Systems, Airborne Navigation Systems Workshop (Digest No. 1997/169), IEE Colloquium, pp. 4/1-4/6, Feb. 21, 1997.
G2000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=97668 on Jun. 28, 2011, 2 pages.
G3000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=66916 on Jun. 28, 2011, 2 pages.
G5000, Garmin, printed from website https://buy.garmin.com/shop/shop.do?cID=153&pID=90821&ra=true on Apr. 20, 2011, 2 pages.
Johnson, A., et al., Vision Guided Landing of an Autonomous Helicopter in Hazardous Terrain, Robotics and Automation, 2005. ICRA 2005. Proceedings of the 2005 IEEE International Conference, pp. 3966-3971, Apr. 18-22, 2005.
Kuntman, D., Airborne system to address leading cause of injuries in non-fatal airline accidents, ICAO Journal, Mar. 2000, 4 pages.
McGray et al., Air Operators, Airlines, Manufacturers and Interested Industry Stakeholders & Aero Chart Forum-Utilizing EFVS technology and incorporating it into FAA NextGen, Federal Aviation Administration, Apr. 23 & 30, 2014, 34 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798 dated Feb. 26, 2016, 9 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798 dated Mar. 18, 2015, 21 pages.
Non-Final Office Action on U.S. Appl. No. 13/250,798, dated Sep. 9, 2016, 6 pages.
Non-Final Office Action on U.S. Appl. No. 14/301,199 dated Sep. 9, 2015, 18 pages.
Non-Final Office Action on U.S. Appl. No. 14/482,681, dated Dec. 20, 2016, 9 pages.
Non-Final Office Action on U.S. Appl. No. 14/536,330 dated Jul. 13, 2016, 12 pages.
Notice of Allowance for U.S. Appl. No. 11/863,215, dated Oct. 13, 2009, 8 pages.
Notice of Allowance for U.S. Appl. No. 11/863,219, dated Jun. 23, 2009, 7 pages.
Notice of Allowance for U.S. Appl. No. 11/863,221, dated Aug. 2, 2010, 9 pages.
Notice of Allowance for U.S. Appl. No. 11/899,801, dated Aug. 19, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 11/900,002, dated Sep. 14, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 12/009,372, dated Oct. 13, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/009,373, dated Jun. 16, 2010, 4 pages.
Notice of Allowance for U.S. Appl. No. 12/009,472, dated Sep. 5, 2013, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/167,200, dated Oct. 28, 2010, 5 pages.
Notice of Allowance for U.S. Appl. No. 12/167,203, dated Jun. 21, 2013, 7 pages.
Notice of Allowance for U.S. Appl. No. 12/167,208, dated Mar. 21, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/180,293, dated Aug. 4, 2011, 8 pages.
Notice of Allowance for U.S. Appl. No. 12/786,169, dated Mar. 28, 2013, 6 pages.
Notice of Allowance for U.S. Appl. No. 13/538,957, dated Oct. 3, 2013, 13 pages.
Notice of Allowance on U.S. Appl. No. 12/263,282 dated Jan. 29, 2016, 8 pages.
Notice of Allowance on U.S. Appl. No. 13/241,051 dated Aug. 28, 2014, 9 pages.
Notice of Allowance on U.S. Appl. No. 13/247,742 dated Jul. 30, 2014, 9 pages.
Notice of Allowance on U.S. Appl. No. 13/250,798, dated Sep. 28, 2016, 10 pages.
Notice of Allowance on U.S. Appl. No. 14/301,199 dated Mar. 1, 2016, 11 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Aug. 6, 2009, 23 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Dec. 15, 2010, 13 pages.
Office Action for U.S. Appl. No. 11/851,323, dated Jul. 5, 2012, 23 pages.
Office Action for U.S. Appl. No. 11/863,215, dated May 27, 2009, 5 pages.
Office Action for U.S. Appl. No. 11/863,215, dated Nov. 12, 2008, 8 pages.
Office Action for U.S. Appl. No. 11/863,219, dated Dec. 12, 2008, 7 pages.
Office Action for U.S. Appl. No. 11/863,221, dated Dec. 18, 2009, 5 pages.
Office Action for U.S. Appl. No. 11/863,221, dated Dec. 8, 2008, 8 pages.
Office Action for U.S. Appl. No. 11/863,221, dated May 26, 2009, 5 pages.
Office Action for U.S. Appl. No. 12/009,372, dated Dec. 20, 2010, 10 pages.
Office Action for U.S. Appl. No. 12/009,372, dated Jun. 13, 2011, 9 pages.
Office Action for U.S. Appl. No. 12/009,373, dated Dec. 30, 2009, 14 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Apr. 16, 2012, 16 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Jan. 14, 2011, 14 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Mar. 20, 2013, 15 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Nov. 3, 2011, 15 pages.
Office Action for U.S. Appl. No. 12/009,472, dated Nov. 9, 2012, 15 pages.
Office Action for U.S. Appl. No. 12/167,200, dated Jul. 21, 2010, 6 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Aug. 26, 2010, 9 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Jul. 20, 2011, 6 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Mar. 7, 2013, 5 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Oct. 31, 2011, 5 pages.
Office Action for U.S. Appl. No. 12/167,203, dated Sep. 21, 2012, 6 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Dec. 30, 2009, 10 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Feb. 7, 2011, 8 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Jun. 3, 2010, 11 pages.
Office Action for U.S. Appl. No. 12/167,208, dated Oct. 19, 2010, 8 pages.
Office Action for U.S. Appl. No. 12/180,293, dated Jan. 4, 2011, 5 pages.
Office Action for U.S. Appl. No. 12/180,293, dated Jul. 28, 2010, 8 pages.
Office Action for U.S. Appl. No. 12/263,282, dated Jan. 5, 2012, 10 pages.
Office Action for U.S. Appl. No. 12/786,169, dated Jan. 18, 2013, 14 pages.
Office Action for U.S. Appl. No. 12/892,563, dated Feb. 19, 2013, 12 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Feb. 15, 2012, 8 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Jul. 10, 2012, 4 pages.
Office Action for U.S. Appl. No. 12/976,871, dated May 6, 2013, 5 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Nov. 21, 2012, 5 pages.
Office Action for U.S. Appl. No. 12/976,871, dated Oct. 9, 2013, 5 pages.
Office Action for U.S. Appl. No. 13/183,314, dated Aug. 14, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/183,314, dated Mar. 28, 2013, 12 pages.
Office Action for U.S. Appl. No. 13/224,992, dated Feb. 28, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/250,307, dated Nov. 5, 2013, 11 pages.
Office Action for U.S. Appl. No. 13/474,559, dated Aug. 28, 2013, 10 pages.
Office Action for U.S. Appl. No. 13/474,559, dated Dec. 28, 2012, 8 pages.
Office Action for U.S. Appl. No. 13/538,957, dated Apr. 4, 2013, 19 pages.
Office Action for U.S. Appl. No. 13/538,957, dated Oct. 5, 2012, 18 pages.
Office Action for U.S. Appl. No. 13/743,182, dated Apr. 8, 2013, 10 pages.
Office Action for U.S. Appl. No. 12/786,169, dated Jul. 20, 2012, 8 pages.
Office Action in Japanese Patent Application 2015-116688, dated Aug. 25, 2015, 4 pages.
Office Action in Japanese Patent Application 2015-116716, dated Aug. 25, 2015, 3 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Jul. 12, 2013, 17 pages.
Office Action on U.S. Appl. No. 11/787,460, dated Mar. 19, 2010, 16 pages.
Office Action on U.S. Appl. No. 11/787,460, dated Sep. 16, 2009, 15 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Feb. 11, 2014, 21 pages.
Office Action on U.S. Appl. No. 12/236,464, dated Jun. 22, 2011, 14 pages.
Office Action on U.S. Appl. No. 12/892,563, dated May 7, 2013, 6 pages.
Office Action on U.S. Appl. No. 12/892,563, dated Oct. 10, 2012, 12 pages.
Office Action on U.S. Appl. No. 13/241,051 dated Feb. 27, 2014, 21 pages.
Office Action on U.S. Appl. No. 13/247,742 dated Dec. 3, 2013, 11 pages.
Office Action on U.S. Appl. No. 13/250,798 dated Apr. 23, 2014, 15 pages.
Office Action on U.S. Appl. No. 13/627,788, dated Jul. 28, 2014, 10 pages.
Office Action on U.S. Appl. No. 13/867,556 dated Feb. 7, 2014, 11 pages.
Office Action U.S. Appl. No. 11/787,460, dated Aug. 31, 2010, 18 pages.
Office Action with English Translation received in Korean Patent Application 10-2010-7017278, dated Aug. 26, 2015, 5 pages.
Pictures of DELPHINS, printed from website www.tunnel-in-the-sky.tudelft.nl on Aug. 28, 2007, 4 pages.
REVS Product Information Sheet, Sierra Nevada Corporation, dated May 7, 2014, 2 pages.
Skolnik, Introduction to Radar Systems, McGraw Hill Book Company, 2001, 3 pages.
Skolnik, Radar Handbook (McGraw Hill Book Company), 1990, 23 pages.
Synthetic Vision System, en.wikipedia.org/wiki/Synthetic_vision_system, retrieved Feb. 28, 2013, 4 pages.
Technical Standard Order, TSO-C115b, Airborne Area Navigation Equipment Using Multi-Sensor Inputs, Department of Transportation, Federal Aviation Administration, Sep. 30, 1994, 11 pages.
U.S. Office Action on U.S. Appl. No. 11/900,002 dated Jun. 8, 2010, 7 pages.
U.S. Office Action on U.S. Appl. No. 13/247,742 dated Apr. 16, 2014, 15 pages.
Vadlamani, A., et al., Improving the detection capability of spatial failure modes using downward-looking sensors in terrain database integrity monitors, Digital Avionics Systems Conference, 2003. DASC-03. The 22nd, vol. 2, pp. 9C.5-91-12 vol. 2, Oct. 12-16, 2003.
Van Kasteren, Joost, “Tunnel-in-the-Sky, Synthetic vision simplifies the pilot's job and enhances safety,” printed from website www.delftoutlook.tudelft.nl on Aug. 28, 2007, 13 pages.
Walker, GD-Itronix Dynavue Technology, The Ultimate Outdoor-Readable Touch-Screen Display, Rugged PC Review, 4 pages.
Wang et al., A Simple Based on DSP Antenna Controller of Weather Radar, 2001 CIE International Conference, 4 pages.
Notice of Allowance for U.S. Appl. No. 15/166,191 dated Oct. 18, 2108. 8 pages.