Weather radar system and method with path attenuation shadowing

Information

  • Patent Grant
  • 10684366
  • Patent Number
    10,684,366
  • Date Filed
    Friday, March 17, 2017
    7 years ago
  • Date Issued
    Tuesday, June 16, 2020
    4 years ago
Abstract
An avionic weather radar system and method can sense a path attenuation condition using radar returns of received via a radar antenna onboard the aircraft. Images of weather can be displayed using an outside source and the radar returns. The images are displayed using information from the outside source when the path attenuation condition is sensed. The images can be displayed using speckled areas, cross hatched areas or other symbols to represent the information from the outside source. The images of the weather can be provided on an avionic display.
Description
BACKGROUND

The present disclosure relates generally to the field of weather display systems. More particularly, the present disclosure relates to a weather display system and method configured to display weather which is obscured by path attenuation.


Weather display systems are often used to alert operators of vehicles, such as aircraft pilots, of weather hazards in the area near the vehicle. Conventional weather display systems are configured to display weather data in two dimensions and often operate according to ARINC 453 and 708 standards. A horizontal plan view provides an overview of weather patterns near an aircraft mapped onto a horizontal plane. Generally the horizontal plan view provides indications of precipitation rates in the vicinity of the aircraft (images of weather). Red, yellow, and green colors are used to symbolize areas of respective precipitation rates, and black color symbolizes areas of very little or no precipitation. Each color is associated with radar reflectivity range which corresponds to a respective precipitation rate range. Red indicates the highest rates of precipitation while green represents the lowest (non-zero) rates of precipitation. Certain displays can also utilize a magenta color to indicate regions of turbulence.


Onboard avionic weather radar systems (e.g., x-band weather radar systems) cannot always penetrate or sense areas beyond intense precipitation or rainfall. Accordingly, cells (e.g. storms) behind a nearby line of cells may be shadowed by the closer cells. This shadowing by the closer cells can deprive the pilot of a view of the complete weather situation through which the aircraft is flying. Conventional avionic weather radar systems such as the WX2100 weather radar manufactured by Rockwell Collins, Inc. can execute a path attenuation compensation (PAC) algorithm that flags any radial where the radar return is too badly attenuated to be corrected. Path attenuation is described in U.S. Pat. No. 5,047,775 incorporated herein by reference


Conventional displays can mark such radials with a yellow bar and/or provide the text “PAC Alert” so that the pilot can be aware that hidden weather may be behind displayed weather cells. However, such systems do not provide an indication of whether or not weather actually exists behind the weather cell or what type of weather exists behind a weather cell.


Thus, there is a need for a weather information display that provides more information than a conventional display. Further, there is a need for a threat depiction system and method that can augment the display of information when weather detection is obscured by path attenuation. Further still, there is a need for an avionic weather system including a path attenuation algorithm that provides an indication of weather beyond the attenuating cell. Yet further, there is a need for a system and method that displays weather despite path attenuation issues. Yet further, there is a need for a weather display that can display weather images using merged data from two different sources.


SUMMARY

An exemplary embodiment relates to a method of displaying images associated with weather near an aircraft. The method includes sensing a path attenuation condition using the radar returns and displaying the images of the weather using an outside source and the radar returns. The images of the weather are displayed using information from the outside source when the path attenuation condition is sensed.


Another exemplary embodiment relates to an avionic weather radar system. The avionic weather radar system includes a processor configured to sense a path attenuation condition using radar returns received via a radar antenna onboard the aircraft. The processor is also configured to cause images of weather to be displayed using outside source and the radar returns. The weather is displayed using information from the outside source when the path attenuation condition is sensed.


According to certain embodiments outside sources can be another weather radar source or data from the onboard weather radar system operating at a different frequency. In other embodiments, the outside source can be a non-radar airborne source, a light detection and r ranging (LIDAR) source, an infrared source (e.g., infrared sensor/camera), etc.).


Another embodiment relates to a method of displaying images associated with weather near an aircraft. The method includes receiving radar returns associated with a radar antenna of the aircraft, and receiving data from a source remote from the aircraft. The method also includes displaying the images associated with the weather using at least a first color, a second color and a third color associated with a respective first precipitation rate or threat range, a respective second precipitation rate or threat range, and a respective third precipitation rate range or threat level. The first precipitation rate range or threat level is less than the second precipitation rate range or threat level and the third precipitation rate range or threat level is more than the second precipitation rate range or threat level. The images associated with the weather include at least a first image provided using the radar returns and a second image provided using the data from the source remote from the aircraft. The second image is provided at a range greater than a range of a path attenuation condition.


Another embodiment relates to a system comprising an electronic processor configured to provide the images of weather on a display according to any of the methods described above. Another embodiment relates to one or more non-transitory computer-readable storage media having instructions stored thereon that are executable by one or more processors to execute one or more of the methods described above.





BRIEF DESCRIPTION OF THE DRAWINGS

The disclosure will become more fully understood from the following detailed description, taken in conjunction with the accompanying figures, wherein like reference numerals refer to like elements, in which:



FIG. 1 is a general block diagram of an exemplary weather radar system that may be used, for example, on an aircraft according to an exemplary embodiment;



FIG. 2 is a more detailed block diagram of the weather radar system of FIG. 1 according to an exemplary embodiment;



FIG. 3 is a flow diagram of an exemplary process for generating images of weather using the system illustrated in FIG. 1 according to an exemplary embodiment;



FIG. 4 is a drawing of a screen including display images of weather on a conventional display;



FIG. 5 is a drawing of a screen including display images of the weather illustrated in FIG. 4 provided by the system illustrated in FIG. 1 including images derived from weather data from an outside source during a path attenuation condition according to an exemplary embodiment;



FIG. 6 is a drawing of a screen including display images of the weather illustrated in FIG. 4 provided by the system illustrated in FIG. 1 including images derived from weather data from an outside source during a path attenuation condition according to another exemplary embodiment;



FIG. 7 is a drawing of a screen including display images of the weather illustrated in FIG. 4 provided by the system illustrated in FIG. 1 including images derived from weather data from an outside source during a path attenuation condition according to another exemplary embodiment; and



FIG. 8 is a drawing of a screen including display images of the weather illustrated in FIG. 4 provided by the system illustrated in FIG. 1 including images derived from weather data from an outside source during a path attenuation condition according to another exemplary embodiment.





DETAILED DESCRIPTION

Before turning to the figures, which illustrate the exemplary embodiments in detail, it should be understood that the application is not limited to the details or methodology set forth in the description or illustrated in the figures. It should also be understood that the terminology is for the purpose of description only and should not be regarded as limiting. As discussed below, the systems and methods can be utilized in a number of display devices for various types of applications or sensing systems.


Referring generally to the figures, systems and methods for providing visual representations of weather in proximity to an aircraft or other vehicle are provided. According to one embodiment, weather images are displayed even when sensing the weather is blocked by path attenuation condition. For example, an embodiment of the system and method described herein can advantageously display blocked cells using a source of weather data or information remote from the aircraft. The blocked cells can advantageously be depicted using display formats compatible with ARINC 453, and ARINC 661 avionic display platforms. The display of blocked cells advantageously gives pilots a more complete view of weather conditions.


In one embodiment, the images are displayable on at least one of navigation (NAV) displays, primary flight displays, electronic flight bag displays, tablets such as iPad® computers manufactured by Apple, Inc. or tablet computers, synthetic vision system displays, head up displays (HUDs), wearable displays, Google glasses, etc. In addition, the on-board sensed weather data and remote weather data can be synchronized before merging the data for display, according to one embodiment.


Referring generally to FIGS. 1 through 3, systems and methods that may be used to generate images of weather are shown and described according to exemplary embodiments. In some embodiments, only weather associated with radar returns received by a radar antenna onboard the aircraft are displayed when a path attenuation condition does not exist. Alternatively, weather data from a remote source can be merged with the weather associated with onboard weather radar return data and the merging algorithm can use a path attenuation factor to weight the data for the merging operation. In another embodiment, the merging operation can choose the remote data at ranges past the path attenuation condition over the data from onboard radar returns, or can choose the remote data at a range past the path attenuation condition on a radial basis over the data from onboard radar returns. According to another embodiment, weather within a threshold range is depicted using data derived from returns received onboard the aircraft, and weather outside of the threshold range is depicted using data from a remote source. Advantageously, systems and methods according to certain embodiments can utilize storm motion vectors from the information from the remote source (e.g., ground-based source) to adjust the position of the weather data from the remote source (e.g., NEXRAD data) to compensate for the delay. The data from the remote source can be differentiated visually from an onboard data as there may be residual error in what is presented in the data from the remote source.


Referring specifically to FIG. 1, a block diagram of a weather radar system 200 that may be used, for example, on a vehicle such as an aircraft is depicted according to an exemplary embodiment. In one embodiment, system 200 may be an avionics multi-scan, system manufactured by Rockwell Collins, such as the RTA-4218 MULTISCAN weather radar system or WXR-2100 weather radar system configured in accordance with the principles described herein. System 200 includes a weather image module 212 and a weather image module 214 in one embodiment. Modules 212 and 214 can be a circuit, module, software program or routine. Weather image module 212 advantageously determines the presence of weather at ranges and locations (and altitudes in certain embodiments) for display to the pilot of an aircraft.


Weather image module 212 may be used to create a spatially correct weather interpretation (e.g., weather images) that may be displayed in the range, altitude, and/or time domains. A combination of horizontal and selected vertical antenna beams may facilitate the collection of radar return data (IQ data) that can be used to compute parameters to construct an accurate weather model. The weather model may be altitude-based, and an estimate of the reflectivity based on radar returns may be obtained for any scan angle, range, and altitude combination. Weather image module 212 may store the altitude-based weather model in a memory 206, and the weather model may be used to extract weather information in a defined manner (e.g., based on user actions). Weather image module 212 can use data from remote source 152 and data from onboard the aircraft associated with radar returns received by radar antenna 104.


Avionics weather radar system 200 includes a weather radar receiver/transmitter 102, weather radar adjustable antenna 104, a processor 208, and a memory 206 (e.g., a multi-scan, multi-tilt angle memory) in one embodiment. System 200 also includes a tilt control 203 for automatically controlling the tilt angle (mechanical or electronic) of the antenna 104 in one embodiment; this auto control may include an additional manual control feature as well.


Memory 206 may be capable of storing in a readily addressable and rapidly retrievable manner at least two data sets resulting from two or more antenna sweeps at different angles. Memory 206 can include any type of machine-readable storage device capable of storing radar returns or associated data for analysis/processing by processor 208. In some embodiments, memory 206 can store parameters of a weather model. The data in memory 206 can represent factors for a mathematical relationship defining reflectivity as a function of altitude in one embodiment. The data in memory 206 can be used by path attenuation module 214 to determine whether a path attenuation condition exists. Although a multi-scan, multi-tilt scanning and data sets are described, it should be understood by one of ordinary skill in the art that a single scan of data may also be used in some embodiments.


Memory 206 can also be a three dimensional storage buffer for storing weather radar parameters according to X, Y and Z coordinates according to one embodiment. The storage of radar data and the form of the weather data stored therein is not disclosed in a limiting fashion. A variety of storage techniques for weather data can be used without departing from the scope of the invention.


Weather data from returns received by antenna 104 and weather data from remote source 152 can be stored in memory 206. The weather data can be based on received horizontal and/or vertical scans and from other aircraft or ground based sources. The weather data from remote source 152 can be received using a communications unit 153, such as a radio or other wireless communication device.


In some embodiments, the data may be stored as a mathematical equation representation of the information. The mathematical equation representation may be a piecewise linear function, piecewise nonlinear function, coefficients of a cubic spline, coefficients of a polynomial function, etc. that represent vertical representations of the weather based on the horizontal scan data and/or horizontal representation of the weather based on the vertical scan data. The function may be an equation based on weather parameters that may be sensor driven, model driven, a merger of sensor and model, etc. Although horizontal scan data is described, alternative embodiments may include X, Y Cartesian coordinates, rho/theta input, latitude and longitude coordinates, altitude, etc. Weather may be estimated for any required point in space with the vertical dimension being the subject of the weather equation.


Display 110 can be part of an avionic multi-function display (MFD) unit in one embodiment. In some embodiments, display 110 may be any of a variety of display types, such as a navigation display, an electronic flight bag, a tablet computing device, a synthetic vision system, a heads up display (HUD), a dedicated weather display, or another type of display system. In some embodiments, display 110 may be a remote display not included within a cockpit of the aircraft, such as a ground-based support display or remote display configured to display information regarding whether near an unmanned aerial vehicle (UAV).


Processor 208 may be included as part of a multi-scan, multi-tilt angle weather radar system and may perform the customary functions performed by a conventional weather radar return processing unit. Processor 208 may also perform several additional operations based upon the additional data and/or instructions provided in memory 206. In general, processor 208 can merge or cross qualify portions, or ranges, of the radar returns of several different antenna sweeps at several different tilt angles, so that a single, relatively clutter-free image may be presented to the pilot based upon the several separate scans. The radar returns may be processed by processor 208 to generate a three-dimensional weather profile of the weather near the aircraft.


System 200 may perform a scanning operation by transmitting at least two beams at different tilt angles. In some embodiments, system 200 may use a global positioning system (GPS), terrain database, or other tool to control the tilt control 203. Data obtained from the radar returns (e.g., reflectivity data) may be stored in memory 206. For example, known ground clutter may be removed from each of the returns using ground clutter suppression techniques.


Referring again to FIG. 1, in some embodiments, processor 208 may be configured to store parametric characteristics of the vertical extent of the weather in memory 206 instead of detecting and storing the entire reflectivity profile in a traditional three-dimensional memory. For example, since the radar knows current altitude and outside air temperature, processor 208 can be configured to estimate the altitude of the local atmospheric freezing layer. Further, processor 208 can estimate the height of the troposphere based on latitude, season, or more precise FMS data. Using this information and radar elevation scan data collected from scanning processes (e.g., normal MULTISCAN two-tilt or three-tilt processes), processor 208 may generate a linear fit of the reflectivity data from the ground to the freezing layer and another linear fit between the freezing layer and top of the troposphere.


Some embodiments may use more complex curve fits. Specific curve fits can be utilized depending on geographical location, time of day, time of year, etc. For example, over Europe, a linear curve fit may suffice to determine vertical reflectivity between freezing layer and troposphere height whereas a second order fit may be required over the equatorial Pacific. A tag can be given to each cell denoting the specific order of the data fit and, correspondingly, the method used to obtain vertical data from the stored parameters. As additional data is collected, perhaps through dedicated vertical scans of convective cells, the functions can be updated or increased in complexity in order to improve their accuracy.


Processor 208 may process weather radar returns to identify or sense the presence of weather in front of or in view of the aircraft. Weather image module 212 may utilize the altitude and the range to the weather to generate a vertical profile associated with the weather. Weather merge module 212 may scan across an array of azimuths to generate a three-dimensional weather profile of the weather near the aircraft, which may be stored for later presentation or immediately presented on display 110. Weather images can be displayed on either view screen 111 or 113 or both screen 111 and 113. Screens 111 and 113 plan view only displays in one embodiment. Although vertical profiles are discussed herein, system 200 includes embodiments that do not use vertical profile processing and/or that do not display vertical profiles.


In some embodiments, additional visual indicators other than the representation of weather are provided on screens 111 or 113. In some embodiments, a range and bearing matrix having range markers indicating distance from a current position of the aircraft and bearing markers indicating azimuths from a current flight path or bearing of the aircraft may be provided and may assist the pilot in cognitive recognition of weather features from the pilot's perspective.


With reference to FIG. 2, processor 208 can provide a mean velocity parameter 260 and a spectral width parameter 262 derived from weather radar returns. Alternatively, other types of velocity parameters can be utilized. In addition, processor 208 can provide a reflectivity parameter 264 and a range parameter 268 to circuit 212. Computed range or range parameter 268 along with scan angle position can be used to plot the position of weather on display 110. Processor 208 can also receive a temperature parameter 66, an azimuth 68, a position 70, a date 72, flight plan 75 and time 74. Alternatively, a separate temperature sensor 66 can be coupled to processor 208. Parameters 260, 262, 264, and 268 can be computed by processor 208 using data stored in memory 206.


Path attenuation module 214 receives weather radar returns associated with antenna 104 and determines whether a path attenuation condition exists. Module 214 can employ an algorithm similar to the algorithms employed in the RTA-4218 MULTISCAN™ radar system to determine whether X-band weather radar returns are experiencing a path attenuation condition. These algorithms generally determine whether returns beyond a range associated with large reflections have been too badly attenuated for compensation. If so, the path attenuation condition warning is provided.


In one embodiment, the path attenuation condition can be sensed when expected returns (e.g., noise returns) fall off rapidly. The path attenuation condition can be sensed on a radial basis. Alternatively, weather data from remote source 152 can be compared to weather data from radar returns received by radar antenna 104 to determine if a path attenuation conditions exists. For example, if there is a cell or storm at a particular range in a particular azimuth range according to remote data and the weather radar returns do not show any storms or cell beyond that particular range in the particular azimuth range, then path attenuation module 214 can determine that a path attenuation condition exists. Module 214 may consider whether a storm is in front of the missing storm from the remote data. Temporal correlation can be necessary in certain embodiments as returns received by antenna 104 generally provide more current data than weather data received via communications unit. In one embodiment, data from source 152 includes storm vector information, and the locations of the cells are adjusted with respect to time using the storm vector information.


If a path attenuation condition exists, weather image module 212 can utilize weather data from remote source 152 and radar returns to provide a screen 111 using weather data from a remote source. If a path condition does not exist, weather image module 214 can provide a screen 113 that provides images only using weather data derived from radar returns received on antenna 104. Screen 111 can include both images derived from weather data from remote source 152 and weather data derived from returns received an antenna 104. The images derived from weather data from remote source 152 and weather data derived from returns can be spatially correlated using time of sensing information and storm vector values. The storm vector values can be from a Nowcast or forecast weather data source such as the Corridor Integrated Weather System (CIWS) or Consolidated Storm Prediction for Aviation (COSPA) weather data source in one embodiment. Along with storm vector information, growth and decay information can be received, which may be used to increase or decrease the size, shape, and intensity of the weather data displayed in accordance with time in one embodiment.


Weather data stored in memory 206 can include weather data from remote source 152 (FIG. 1) and weather data associated with radar returns received an antenna 104. The weather data can include a time of sensing data and storm vector data for temporal and spatial correlation.


In one embodiment, if a path attenuation condition exists, weather image module 212 can utilize weather data from another aircraft or other onboard weather sources and radar returns associated with antenna 104 to provide a screen 111. The weather data can be from another weather radar source or data from the onboard weather radar system operating at a different frequency. The different frequency can be a millimeter frequency, a Ka band, frequency, a W band frequency etc. In other embodiments, the weather data can be from a non-radar airborne source (a LIDAR source, an infrared source, etc.).


With reference to FIGS. 1-3, processor 208 and weather image module 212 and path attenuation module 214 can utilize process 140 to provide images on display 110 in one embodiment. At a step 142, processor 208 receives weather data associated with weather returns at multiple tilt angles from antenna 104 in one embodiment. Alternatively, the radar returns do not have to be at multiple tilt angles in certain embodiments. At a step 144, weather image module 212 determines the presence of, location of, and range of weather from weather data derived from radar returns in one embodiment.


At a step 145, path attenuation module 214 determines if a path attenuation condition exists in one embodiment. At a step 146, if a path attenuation condition exists, weather data from remote source 152 are adjusted for spatial and temporal correlation in accordance with one embodiment. At a step 148, the weather data from remote source 152 and weather data derived from radar returns from antenna 104 are merged in one embodiment. At a step 150, weather images are generated using the merged data in one embodiment.


If a path condition does not exist at step 145, weather images can be generated at step 150 using only return data from radar returns received by radar antenna 104 in one embodiment. In one embodiment, a path attenuation factor can be utilized to select weather data for merging. The weather data from source 152 can be merged with the radar return data to create weather images in one embodiment. For example, if the attenuation factor indicates substantial attenuation at a particular azimuth and range, weather images can be provided using a greater percentage of data form source 152 at that range in one embodiment. If the attenuation factor indicates a complete path attenuation condition, then the data from source 152 is entirely used to provide the image at that range and azimuth in one embodiment. Further, data can be merged in accordance with its temporal standing. More recent data is given preference over older data in one embodiment. In one embodiment, images are provided using data from source 152 at ranges beyond a threshold range, and images are provided using data from radar returns received by antenna 104 at ranges below the threshold range.


Processor 208 and weather image module 212 cause display 110 to display areas associated with weather data from remote source 152 differently than weather images derived from radar returns from antenna 104 in one embodiment. Weather images using data from remote source 152 can be displayed using outlines, cross-hatching, uniformly organized speckles, etc. The outlines, cross-hatching and speckles can have red, yellow, or green color according to standard precipitation rates in one embodiment. In one embodiment, the cross-hatched regions are provided with a background color corresponding to the precipitation rate and a line color for the cross hatching differs from the background color.


The speckled pattern or cross hatched regions can either be bounded (e.g., outlined) or unbounded and can have a variety of shape profiles including rounded or straight edges in one embodiment. The speckles can vary in shape and size to represent age of data in one embodiment. The cross hatching can also vary in thickness to represent age of data in one embodiment. According to another embodiment, boundary color can be the same or different as the speckled colors. According to another embodiment processor 208 and weather image module 212 cause display 110 to display areas associated with weather data from remote source 152 as outlined regions without cross hatching and speckles in one embodiment.


In some embodiments, process 140 may be implemented using instructions stored on a machine-readable medium (e.g., memory 206) and executed on a processor (e.g., processor 208). Processor 208 may be configured to operate system 200 as a pulse Doppler multi-scan, multi-tilt angle radar system or a volumetric radar system in which radar beams are produced at more than one tilt angle. Processor 208 receives radar returns through receiver/transmitter circuit 102.


Processor 208 can receive the radar returns (or signals/data related thereto) directly or through memory 206 at a step 142. Receiver/transmitter circuit 102 can be a single path or can have separate circuits for a receive path and a transmit path. Processor 208 may determine power associated with the radar returns.


With reference to FIG. 4, a conventional screen (or portion thereof) displays weather images 402 on a screen 400. Images 402 have a first color 404, a second color 406 and a third color 408. Colors 404, 406 and 408 symbolize a respective precipitation rainfall rate range and can be red, yellow and green respectively. Weather images 405 and 407 are displayed in a fashion similar to weather images 402.


A PAC alert is provided on screen 400 as a bar 412 as well as a textual message 414 stating the term “PAC ALERT”. Bar 412 indicates the radials through which the path attenuation condition exists and is provided on a range arc 416 with azimuth indications. Accordingly, the pilot is aware that additional storms may exist beyond the range of weather images 402. Since bar 412 is not provided behind weather images 405 which are displayed in a similar fashion to weather images 402, the pilot knows that weather does not exist beyond the range of weather images 405.


With reference to FIG. 5, a screen 500 or portion of a screen is provided by system 200 in one embodiment. Weather images 502 are provided behind weather images 402 (further in range from the aircraft). Weather 502 is generated by weather image module 212 in response to weather data from remote source 152. Weather 502 is cross-hatched and includes colors 404, 406 and 408 as background in one embodiment. The line color for the cross hatching can be white, black or another color. Weather images 405 and weather images 407 on screen 500 are not different from their depiction in screen 400 because the path attenuation condition is not associated with the radials corresponding to images 405 and 407. Advantageously, system 200 allows weather images 502 to be viewed by the pilot despite a path attenuation condition due to the weather associated with images 402. In one embodiment, colors 404, 406 and 408 in FIGS. 6-8 can represent a threat level or range.


With reference to FIG. 6, a screen 600 or portion of a screen is provided by system 200 in one embodiment. Screen 600 includes weather images 402, 405, and 602. Weather images 602 correspond to weather images 502 on screen 500 and are shown in an outline form indicating that weather images 602 are provided from data from remote source 152 according to one embodiment. The outline can be color coded to indicate precipitation rates using red, yellow and green according to one embodiment. Weather images 402, 405 and 407 are displayed as on screens 400 and 500 in FIGS. 4 and 5.


With reference to FIG. 7, a screen 700 or portion of a screen is provided by system 200 in one embodiment. Screen 700 includes weather images 402, 405, and 702. Weather images 702 correspond to images 502 and are shown as speckled regions indicating that weather images 502 are derived from data from remote source 152 according to one embodiment. The speckles can be color coded to indicate precipitation rates using red, yellow and green according to one embodiment. Weather images 402, 405 and 407 are displayed as on screens 400 and 500 as in FIGS. 4 and 5.


Screen 700 can also include a PAC Alert message 704 and/or a bar 706 indicating the radials affected by the path attenuation condition according to one embodiment. Bar 706 can be yellow or another appropriate color. Bar 706 can be provided on the outer most range arc in one embodiment. Message 704 and bar 706 can also be used on screens 500 and 600 according to one embodiment.


With reference to FIG. 8, a screen 800 is provided by system 200 in one embodiment. Screen 800 includes weather images 402, 405, and 802. Screen 800 includes a range marker 828. Marker 828 can be a 100 nautical mile (nm) range arc or marker from the aircraft as represented by symbol 826. Alternatively, marker 828 can indicate other ranges including ranges selectable by the pilot. A numerical indicator can be provided at a left end of marker 828 to indicate the current range. Screen 800 also includes weather images 822 at range greater than the range represented by marker 828 in one embodiment. Marker 828 can provide delineation between weather images 822 and weather images 402, 405, and 802. U.S. Pat. No. 6,441,773 assigned to the assignee of the present application and incorporated herein by reference shows exemplary displays using a delineation marker between types of weather data.


Weather images 802 correspond to images 502 and are configured to indicate that weather images 802 are derived from data from remote source 152 according to one embodiment. Weather images 402, 405 and 407 are displayed as on screens 400 and 500 as in FIGS. 4 and 5.


Weather images 822 are derived from data from source 152. In one embodiment, system 200 uses data from source 152 at greater ranges because weather is best sensed via radar antenna 104 at shorter ranges (e.g., 60 nm, 80 nm, and 100 nm) in one embodiment. In another embodiment, system 200 uses data from source 152 for areas outside the azimuth capabilities of antenna 104. Data from returns received by antenna 104 are updated much more rapidly than outside sources like NEXRAD sources. Weather images 822 can be derived from data from outside source 152 at greater ranges because the relevant weather can be sampled more accurately via antenna 104 as the aircraft comes closer to it as the aircraft travels along flight path 824. The locations of weather images 822 can be adjusted in accordance with time and vector parameters in one embodiment. Screen 800 can also include a PAC Alert message 704 and/or a bar 706 indicating the radials affected by the path attenuation condition on marker 820 according to one embodiment. Weather image module 212 can merge the weather data from source 152 with the radar return data associated with antenna 104 by using a threshold range or greater to choose the weather data from source 152 and choosing the weather data from source 152 when a path attenuation condition exists to create screen 800 in one embodiment. In one embodiment, a zone around marker 828 can define an area where merged data is used to provide weather images. In one embodiment, the merged data includes radar return data and data from source 152, each weighted according to its range from the aircraft where radar return data is more heavily weighted the closer it is to the aircraft. The weather images in the merged data area can be configured similar to images 502, 602, 702, or 802. In one embodiment, weather images 822 can be depicted at a lower brightness, luminance or intensity level or similar to images 502, 602, 702, or 802 to indicate they are derived from data from outside source 152. In one embodiment, the brightness, luminance or intensity level is varied to represent the age of data associated with images 822.


The disclosure is described above with reference to drawings. These drawings illustrate certain details of specific embodiments that implement the systems and methods and programs of the present disclosure. However, describing the disclosure with drawings should not be construed as imposing on the disclosure any limitations that may be present in the drawings. The present disclosure contemplates methods, systems and program products on any machine-readable media for accomplishing its operations. The embodiments of the present disclosure may be implemented using an existing computer processor, or by a special purpose computer processor incorporated for this or another purpose or by a hardwired system. No claim element herein is to be construed under the provisions of 35 U.S.C. § 112, sixth paragraph, unless the element is expressly recited using the phrase “means for.” Furthermore, no element, component or method step in the present disclosure is intended to be dedicated to the public, regardless of whether the element, component or method step is explicitly recited in the claims.


As noted above, embodiments within the scope of the present disclosure include program products comprising machine-readable storage media for carrying or having machine-executable instructions or data structures stored thereon. Such machine-readable storage media can be any available media which can be accessed by a general purpose or special purpose computer or other machine with a processor. By way of example, such machine-readable storage media can comprise RAM, ROM, EPROM, EEPROM, CD ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium (e.g., non-transitory medium) which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor. Combinations of the above are also included within the scope of machine-readable storage media. Machine-executable instructions comprise, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing machine to perform a certain function or group of functions.


Embodiments of the disclosure are described in the general context of method steps which may be implemented in one embodiment by a program product including machine-executable instructions, such as program code, for example, in the form of program modules executed by machines in networked environments. Generally, program modules include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types. Machine-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represent examples of corresponding acts for implementing the functions described in such steps.


Embodiments of the present disclosure may be practiced in a networked environment using logical connections to one or more remote computers having processors. Logical connections may include a local area network (LAN) and a wide area network (WAN) that are presented here by way of example and not limitation. Such networking environments are commonplace in office-wide or enterprise-wide computer networks, intranets and the Internet and may use a wide variety of different communication protocols. Those skilled in the art will appreciate that such network computing environments will typically encompass many types of computer system configurations, including personal computers, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, servers, minicomputers, mainframe computers, and the like. Embodiments of the disclosure may also be practiced in distributed computing environments where tasks are performed by local and remote processing devices that are linked (either by hardwired links, wireless links, or by a combination of hardwired or wireless links) through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.


An exemplary system for implementing the overall system or portions of the disclosure might include a general purpose computing device in the form of a computer, including a processing unit, a system memory, and a system bus that couples various system components including the system memory to the processing unit. The system memory may include read only memory (ROM) and random access memory (RAM) or other non-transitory storage medium. The computer may also include a magnetic hard disk drive for reading from and writing to a magnetic hard disk, a magnetic disk drive for reading from or writing to a removable magnetic disk, and an optical disk drive for reading from or writing to a removable optical disk such as a CD ROM or other optical media. The drives and their associated machine-readable media provide nonvolatile storage of machine-executable instructions, data structures, program modules, and other data for the computer.


It should be noted that although the flowcharts provided herein show a specific order of method steps, it is understood that the order of these steps may differ from what is depicted. Also two or more steps may be performed concurrently or with partial concurrence. Such variation will depend on the software and hardware systems chosen and on designer choice. It is understood that all such variations are within the scope of the disclosure. Likewise, software and web implementations of the present disclosure could be accomplished with standard programming techniques with rule based logic and other logic to accomplish the various database searching steps, correlation steps, comparison steps and decision steps. It should also be noted that the word “component” as used herein and in the claims is intended to encompass implementations using one or more lines of software code, and/or hardware implementations, and/or equipment for receiving manual inputs.


The foregoing description of embodiments of the disclosure have been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the disclosure to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of the disclosure. The embodiments were chosen and described in order to explain the principals of the disclosure and its practical application to enable one skilled in the art to utilize the disclosure in various embodiments and with various modifications as are suited to the particular use contemplated.

Claims
  • 1. A method of displaying images of weather near an aircraft, the aircraft receiving radar returns through a weather radar antenna of the aircraft, the method comprising: sensing a path attenuation condition using the radar returns;wherein the path attenuation condition is sensed by determining when expected returns fall off rapidly; anddisplaying the images of the weather using an outside source and the radar returns in response to sensing the path attenuation condition;wherein the images of the weather associated with the outside source are displayed as a pattern, the pattern based on a characteristic of the weather;wherein components of the pattern have a display characteristic based on a time of sensing the weather by the outside source.
  • 2. The method of claim 1, wherein the images of the weather associated with the outside source are displayed to have an outline.
  • 3. The method of claim 1, further comprising: displaying a line at a range associated with the path attenuation condition, the line configured to alert a viewer of a display.
  • 4. The method of claim 1, wherein the components of the pattern are uniformly distributed.
  • 5. The method of claim 1, wherein the display characteristic comprises at least one of a size, shape, intensity, thickness, intensity, brightness, and luminance.
  • 6. The method of claim 1, wherein the images of the weather displayed using the outside source are positioned on a display based on a motion vector and growth and decay information and the time of sensing the weather.
  • 7. The method of claim 6, wherein the motion vector and growth and decay information and the time of sensing the weather is used to compensate for a delay time associated with receiving information regarding the weather from the outside source.
  • 8. An avionic weather radar system, comprising: a processor composing a path attenuation module and a weather image module, wherein the path attenuation module is configured to determine a path attenuation condition using characteristics of radar returns received via a radar antenna onboard an aircraft in response to a presence of the path attenuation condition and wherein the path attenuation module is configured to cause images of weather to be displayed using a second source and the radar returns, wherein the weather is displayed within a boundary of the path attenuation condition using information from the second source based on the path attenuation condition being sensed, wherein the images of the weather associated with the second source have a display characteristic indicating that the images of the weather associated with the second source are based on the information from the second source, and wherein the display characteristic varies based on a time of sensing the weather.
  • 9. The system of claim 8, further comprising: a display for displaying the images of the weather.
  • 10. The system of claim 8, wherein the weather is displayed using at least a first color, a second color, and a third color associated with a respective first precipitation rate range or threat level, a respective second precipitation rate range or threat level, and a respective third precipitation rate range or threat level, wherein the first precipitation rate range or threat level is less than the second precipitation rate range or threat range and the third precipitation rate range or threat level is more than the second precipitation rate range or threat level.
  • 11. The system of claim 8, wherein the processor further provides a textual message indicating detection of the path attenuation condition.
  • 12. The system of claim 10, wherein the processor further provides an arc across an azimuth range of the path attenuation condition.
  • 13. The system of claim 12, wherein the arc across the azimuth range of the path attenuation condition includes azimuth indications.
  • 14. The system of claim 8, wherein weather images beyond a particular range are displayed using data from the second source, the second source being an off aircraft source.
  • 15. The system of claim 8, wherein the second source comprises another aircraft, a ground-based radar system, a LIDAR source, an infrared source, or a data source storing weather data from second radar returns having a different frequency than the radar returns received via the radar antenna onboard the aircraft.
  • 16. The system of claim 8, wherein the processor is further configured to sense the path attenuation condition by determining an area where expected radar returns rapidly fall off.
  • 17. The system of claim 8, wherein the processor is further configured to sense the path attenuation condition by detecting a storm or cell at a particular azimuth range and determining that radar returns do not indicate a presence of a storm or cell beyond the particular azimuth range.
  • 18. A method of displaying images associated with weather near an aircraft, the method comprising: receiving radar returns associated with a radar antenna of the aircraft;sensing a path attenuation condition, wherein the path attenuation condition is sensed by determining when expected returns fall off rapidly;receiving data from a source remote from the aircraft; anddisplaying the images associated with the weather, wherein the images associated with the weather comprise a first image provided using the radar returns and a second image provided using the data from the source remote from the aircraft, wherein the second image is provided at a range greater than a range of the path attenuation condition, wherein the second image has a display characteristic indicating that the second image is based on the data from the source remote from the aircraft.
  • 19. The method of claim 18, wherein the display characteristic comprises a boundary of the path attenuation condition.
  • 20. The method of claim 18, wherein the data from the source remote from the aircraft is weather data provided from a remote aircraft or from a NEXRAD source.
CROSS-REFERENCE TO RELATED PATENT APPLICATIONS

This application is a continuation of U.S. Pat. No. 9,599,707, filed Jan. 23, 2014, which is incorporated herein by reference in its entirety and for all purposes.

US Referenced Citations (278)
Number Name Date Kind
650275 Reeve May 1900 A
3251057 Buehler et al. May 1966 A
3359557 Fow et al. Dec 1967 A
3404396 Buchler et al. Oct 1968 A
3465339 Marner Sep 1969 A
3491358 Hicks et al. Jan 1970 A
3508259 Andrews Apr 1970 A
3540829 Collinson et al. Nov 1970 A
3567915 Altshuler et al. Mar 1971 A
3646555 Atlas Feb 1972 A
3715748 Hicks Feb 1973 A
3764719 Dell Oct 1973 A
3781530 Britland et al. Dec 1973 A
3781878 Kirkpatrick Dec 1973 A
3803609 Lewis et al. Apr 1974 A
3885237 Kirkpatrick May 1975 A
3943511 Evans et al. Mar 1976 A
3964064 Brandao et al. Jun 1976 A
3968490 Gostin Jul 1976 A
4015257 Fetter Mar 1977 A
4043194 Tanner Aug 1977 A
4179693 Evans et al. Dec 1979 A
4223309 Payne Sep 1980 A
4240108 Levy Dec 1980 A
4283715 Choisnet Aug 1981 A
4283725 Chisholm Aug 1981 A
4318100 Shimizu et al. Mar 1982 A
4346595 Frosch et al. Aug 1982 A
4430654 Kupfer Feb 1984 A
4435707 Clark Mar 1984 A
4459592 Long Jul 1984 A
4533915 Lucchi et al. Aug 1985 A
4555703 Cantrell Nov 1985 A
4600925 Alitz et al. Jul 1986 A
4613938 Hansen et al. Sep 1986 A
4649388 Atlas Mar 1987 A
4658255 Nakamura et al. Apr 1987 A
4684950 Long Aug 1987 A
4742353 D'Addio et al. May 1988 A
4761650 Masuda et al. Aug 1988 A
4835536 Piesinger et al. May 1989 A
RE33152 Atlas Jan 1990 E
4914444 Pifer et al. Apr 1990 A
4928131 Onozawa May 1990 A
4940987 Frederick Jul 1990 A
5036334 Henderson et al. Jul 1991 A
5047775 Alitz Sep 1991 A
5049886 Seitz et al. Sep 1991 A
5057820 Markson et al. Oct 1991 A
5077558 Kuntman Dec 1991 A
5105191 Keedy Apr 1992 A
5159407 Churnside et al. Oct 1992 A
5164731 Borden et al. Nov 1992 A
5173704 Buehler et al. Dec 1992 A
5177487 Taylor et al. Jan 1993 A
5198819 Susnjara Mar 1993 A
5202690 Frederick Apr 1993 A
5208600 Rubin May 1993 A
5221924 Wilson, Jr. Jun 1993 A
5262773 Gordon Nov 1993 A
5291208 Young Mar 1994 A
5296865 Lewis Mar 1994 A
5311183 Mathews et al. May 1994 A
5311184 Kuntman May 1994 A
5331330 Susnjara Jul 1994 A
5396220 Markson et al. Mar 1995 A
5402116 Ashley Mar 1995 A
5469168 Anderson Nov 1995 A
5479173 Yoshioka et al. Dec 1995 A
5485157 Long Jan 1996 A
5517193 Allison et al. May 1996 A
5521603 Young May 1996 A
5534868 Gjessing et al. Jul 1996 A
5568151 Merritt Oct 1996 A
5583972 Miller Dec 1996 A
5592171 Jordan Jan 1997 A
5602543 Prata et al. Feb 1997 A
5615118 Frank Mar 1997 A
5648782 Albo et al. Jul 1997 A
5654700 Prata et al. Aug 1997 A
5657009 Gordon Aug 1997 A
5686919 Jordan et al. Nov 1997 A
5726656 Frankot Mar 1998 A
5757322 Ray et al. May 1998 A
5771020 Markson et al. Jun 1998 A
5828332 Frederick Oct 1998 A
5838239 Stern et al. Nov 1998 A
5839080 Muller et al. Nov 1998 A
5907568 Reitan, Jr. May 1999 A
5920276 Frederick Jul 1999 A
5945926 Ammar et al. Aug 1999 A
5973635 Albo Oct 1999 A
6034760 Rees Mar 2000 A
6043756 Bateman et al. Mar 2000 A
6043757 Patrick Mar 2000 A
6081220 Fujisaka et al. Jun 2000 A
6138060 Conner et al. Oct 2000 A
6154151 McElreath et al. Nov 2000 A
6154169 Kuntman Nov 2000 A
6177873 Cragun Jan 2001 B1
6184816 Zheng et al. Feb 2001 B1
6201494 Kronfeld Mar 2001 B1
6208284 Woodell et al. Mar 2001 B1
6236351 Conner et al. May 2001 B1
6240369 Foust May 2001 B1
6246367 Markson et al. Jun 2001 B1
6281832 McElreath Aug 2001 B1
6289277 Feyereisen et al. Sep 2001 B1
6297772 Lewis Oct 2001 B1
6339747 Daly et al. Jan 2002 B1
6340946 Wolfson et al. Jan 2002 B1
6381538 Robinson et al. Apr 2002 B1
6384830 Baron et al. May 2002 B2
6388607 Woodell May 2002 B1
6388608 Woodell et al. May 2002 B1
RE37725 Yamada Jun 2002 E
6405134 Smith et al. Jun 2002 B1
6424288 Woodell Jul 2002 B1
6441773 Kelly et al. Aug 2002 B1
6456226 Zheng et al. Sep 2002 B1
6480142 Rubin Nov 2002 B1
6496252 Whiteley Dec 2002 B1
6501392 Gremmert et al. Dec 2002 B2
6512476 Woodell Jan 2003 B1
6518914 Peterson et al. Feb 2003 B1
6549161 Woodell Apr 2003 B1
6560538 Schwinn et al. May 2003 B2
6563452 Zheng et al. May 2003 B1
6577947 Kronfeld et al. Jun 2003 B1
6590520 Steele et al. Jul 2003 B1
6597305 Szeto et al. Jul 2003 B2
6603425 Woodell Aug 2003 B1
6606564 Schwinn et al. Aug 2003 B2
6614382 Cannaday et al. Sep 2003 B1
6650275 Kelly et al. Nov 2003 B1
6650972 Robinson et al. Nov 2003 B1
6667710 Cornell et al. Dec 2003 B2
6670908 Wilson et al. Dec 2003 B2
6677886 Lok Jan 2004 B1
6683609 Baron et al. Jan 2004 B1
6690317 Szeto et al. Feb 2004 B2
6703945 Kuntman et al. Mar 2004 B2
6720906 Szeto et al. Apr 2004 B2
6738010 Steele et al. May 2004 B2
6741203 Woodell May 2004 B1
6744382 Lapis et al. Jun 2004 B1
6771207 Lang Aug 2004 B1
6788043 Murphy et al. Sep 2004 B2
6791311 Murphy et al. Sep 2004 B2
6828922 Gremmert et al. Dec 2004 B1
6828923 Anderson Dec 2004 B2
6839018 Szeto et al. Jan 2005 B2
6850185 Woodell Feb 2005 B1
6856908 Devarasetty et al. Feb 2005 B2
6879280 Bull et al. Apr 2005 B1
6882302 Woodell et al. Apr 2005 B1
6917860 Robinson et al. Jul 2005 B1
6977608 Anderson et al. Dec 2005 B1
7030805 Ormesher et al. Apr 2006 B2
7039505 Southard et al. May 2006 B1
7042387 Ridenour et al. May 2006 B2
7082382 Rose et al. Jul 2006 B1
7109912 Paramore et al. Sep 2006 B1
7109913 Paramore et al. Sep 2006 B1
7116266 Vesel et al. Oct 2006 B1
7129885 Woodell et al. Oct 2006 B1
7132974 Christianson Nov 2006 B1
7139664 Kelly et al. Nov 2006 B2
7145503 Abramovich et al. Dec 2006 B2
7161525 Finley et al. Jan 2007 B1
7200491 Rose et al. Apr 2007 B1
7205928 Sweet Apr 2007 B1
7242343 Woodell Jul 2007 B1
7259714 Cataldo Aug 2007 B1
7292178 Woodell et al. Nov 2007 B1
7307576 Koenigs Dec 2007 B1
7307577 Kronfeld et al. Dec 2007 B1
7307583 Woodell et al. Dec 2007 B1
7307586 Peshlov et al. Dec 2007 B2
7307756 Walmsley Dec 2007 B2
7352317 Finley et al. Apr 2008 B1
7352929 Hagen et al. Apr 2008 B2
7365674 Tillotson et al. Apr 2008 B2
7372394 Woodell et al. May 2008 B1
7383131 Wey et al. Jun 2008 B1
7417579 Woodell Aug 2008 B1
7427943 Kronfeld et al. Sep 2008 B1
7436361 Paulsen et al. Oct 2008 B1
7471995 Robinson Dec 2008 B1
7486219 Woodell et al. Feb 2009 B1
7486220 Kronfeld et al. Feb 2009 B1
7492304 Woodell et al. Feb 2009 B1
7492305 Woodell et al. Feb 2009 B1
7515087 Woodell et al. Apr 2009 B1
7515088 Woodell et al. Apr 2009 B1
7528613 Thompson et al. May 2009 B1
7541971 Woodell et al. Jun 2009 B1
7557735 Woodell et al. Jul 2009 B1
7576680 Woodell Aug 2009 B1
7581441 Barny et al. Sep 2009 B2
7598901 Tillotson et al. Oct 2009 B2
7598902 Woodell et al. Oct 2009 B1
7633428 McCusker et al. Dec 2009 B1
7633431 Wey et al. Dec 2009 B1
7664601 Daly, Jr. Feb 2010 B2
7696920 Finley Apr 2010 B1
7696921 Finley et al. Apr 2010 B1
7698058 Chen et al. Apr 2010 B2
7714767 Kronfeld et al. May 2010 B1
7728758 Varadarajan et al. Jun 2010 B2
7733264 Woodell et al. Jun 2010 B1
7859448 Woodell et al. Dec 2010 B1
7868811 Woodell et al. Jan 2011 B1
7869953 Kelly et al. Jan 2011 B1
7917255 Finley Mar 2011 B1
7932853 Woodell et al. Apr 2011 B1
7973698 Woodell et al. Jul 2011 B1
7982658 Kauffman et al. Jul 2011 B2
8022859 Bunch et al. Sep 2011 B2
8054214 Bunch Nov 2011 B2
8072368 Woodell Dec 2011 B1
8081106 Yannone Dec 2011 B2
8089391 Woodell et al. Jan 2012 B1
8098188 Costes et al. Jan 2012 B2
8098189 Woodell et al. Jan 2012 B1
8111186 Bunch et al. Feb 2012 B2
8159369 Koenigs et al. Apr 2012 B1
8217828 Kirk Jul 2012 B2
8228227 Bunch et al. Jul 2012 B2
8314730 Musiak et al. Nov 2012 B1
8332084 Bailey et al. Dec 2012 B1
8902100 Woodell et al. Dec 2014 B1
9019146 Finley et al. Apr 2015 B1
9134418 Kronfeld et al. Sep 2015 B1
9507022 Breiholz et al. Nov 2016 B1
9535158 Breiholz et al. Jan 2017 B1
9599707 Kronfeld Mar 2017 B1
20020039072 Gremmert et al. Apr 2002 A1
20020126039 Dalton et al. Sep 2002 A1
20030001770 Cornell et al. Jan 2003 A1
20030025627 Wilson et al. Feb 2003 A1
20030193411 Price Oct 2003 A1
20040183695 Ruokangas et al. Sep 2004 A1
20040239550 Daly, Jr. Dec 2004 A1
20050049789 Kelly et al. Mar 2005 A1
20050174350 Ridenour Aug 2005 A1
20050222770 McKewon et al. Oct 2005 A1
20060036366 Kelly et al. Feb 2006 A1
20070005249 Dupree et al. Jan 2007 A1
20080040038 Southard et al. Feb 2008 A1
20080158048 Ridenour Jul 2008 A1
20080158049 Southard et al. Jul 2008 A1
20090177343 Bunch et al. Jul 2009 A1
20090219197 Bunch Sep 2009 A1
20100019938 Bunch Jan 2010 A1
20100042275 Kirk Feb 2010 A1
20100194628 Christianson et al. Aug 2010 A1
20100201565 Khatwa Aug 2010 A1
20100245164 Kauffman Sep 2010 A1
20100245165 Kauffman Sep 2010 A1
20100302093 Bunch Dec 2010 A1
20100302094 Bunch Dec 2010 A1
20100328143 Kirk Dec 2010 A1
20100332056 Kirk Dec 2010 A1
20110074624 Bunch Mar 2011 A1
20110148692 Christianson Jun 2011 A1
20110148694 Bunch et al. Jun 2011 A1
20120029786 Calandra et al. Feb 2012 A1
20120086596 Insanic et al. Apr 2012 A1
20120133551 Pujol et al. May 2012 A1
20120139778 Bunch et al. Jun 2012 A1
20130226452 Watts Aug 2013 A1
20130234884 Bunch et al. Sep 2013 A1
20130321442 Van Os et al. Dec 2013 A1
20130345982 Liu et al. Dec 2013 A1
20140156133 Cullinane et al. Jun 2014 A1
20140176362 Sneed Jun 2014 A1
20140362088 Veillette et al. Dec 2014 A1
Foreign Referenced Citations (6)
Number Date Country
1 329 738 Jul 2003 EP
2658617 Aug 1991 FR
WO-9807047 Feb 1998 WO
WO-9822834 May 1998 WO
WO-03005060 Jan 2003 WO
WO-2009137158 Nov 2009 WO
Non-Patent Literature Citations (82)
Entry
U.S. Appl. No. 13/246,769, filed Sep. 27, 2011, Rockwell Collins.
U.S. Appl. No. 13/717,052, filed Dec. 17, 2012, Woodell et al.
U.S. Appl. No. 13/837,538, filed Mar. 15, 2013, Kronfeld et al.
U.S. Appl. No. 13/841,893, filed Mar. 15, 2013, Rockwell Collins, Inc.
U.S. Appl. No. 13/919,406, filed Jun. 17, 2013, Rockwell Collins, Inc.
U.S. Appl. No. 14/086,844, filed Nov. 21, 2013, Rockwell Collins, Inc.
U.S. Appl. No. 14/162,035, filed Jan. 23, 2014, Kevin M. Kronfeld et al.
U.S. Appl. No. 14/206,239, filed Mar. 12, 2014, Rockwell Collins.
U.S. Appl. No. 14/206,651, filed Mar. 12, 2014, Rockwell Collins, Inc.
U.S. Appl. No. 14/207,034, filed Mar. 12, 2014, Rockwell Collins, Inc.
U.S. Appl. No. 14/323,766, filed Jul. 3, 2014, Weichbrod et al.
U.S. Appl. No. 14/465,730, filed Aug. 21, 2014, Breiholz et al.
U.S. Appl. No. 14/465,753, filed Aug. 21, 2014, Arlen E. Breiholz et al.
U.S. Appl. No. 14/608,071, filed Jan. 28, 2015, Breiholz et al.
3-D Weather Hazard and Avoidance System, Honeywell InteVue Brochure dated Nov. 2008, 4 pages.
Boudevillain et al., 2003, Assessment of Vertically Integrated Liquid (VIL) Water Content Radar Measurement, J. Atmos. Oceanic Technol., 20, 807-819.
Bovith et al., Detecting Weather Radar Clutter by Information Fusion with Satellite Images and Numerical Weather Prediction Model Output; Jul. 31-Aug. 4, 2006, 4 pages.
Burnham et al., Thunderstorm Turbulence and Its Relationship to Weather Radar Echoes, J. Aircraft, Sep.-Oct. 1969, 8 pages.
Corridor Integrated Weather System (CIWS), www.ll.mit.edu/mission/aviation/faawxsystems/ciws.html, received on Aug. 19, 2009, 3 pages.
Decision on Appeal for Inter Parties Reexamination Control No. 95/001,860, dated Oct. 17, 2014, 17 pages.
Doviak et al., Doppler Radar and Weather Observations, 1984, 298 pages.
Dupree et al.,FAA Tactical Weather Forecasting in the United States National Airspace, Proceedings from the World Weather Research Symposium on Nowcasting and Very Short Term Forecasts, Toulouse, France, 2005, 29 pages.
Final Office Action on U.S. Appl. No. 12/892,663 dated Mar. 7, 2013, 13 pages.
Final Office Action on U.S. Appl. No. 13/238,606 dated Apr. 1, 2014, 11 pages.
Final Office Action on U.S. Appl. No. 13/238,606 dated Jan. 22, 2015, 6 pages.
Final Office Action on U.S. Appl. No. 13/246,769 dated Sep. 16, 2014, 18 pages.
Goodman et al., LISDAD Lightning Observations during the Feb. 22-23, 1998 Central Florida Tornado Outbreak, http:www.srh.noaa.gov/topics/attach/html/ssd98-37.htm, Jun. 1, 1998, 5 pages.
Greene et al., Vertically Integrated Liquid Water—A New Analysis Tool, Monthly Weather Review, Jul. 1972, 5 pages.
Hodanish, Integration of Lightning Detection Systems in a Modernized National Weather Service Office, http://www.srh.noaa.gov/mlb/hoepub.html, retrieved on Aug. 6, 2007, 5 pages.
Honeywell, RDR-4B Forward Looking Windshear Detection/Weather Radar System User's Manual with Radar Operation Guidelines, Jul. 2003, 106 pages.
Keith, Transport Category Airplane Electronic Display Systems, Jul. 16, 1987, 34 pages.
Klingle-Wilson et al., Description of Corridor Integrated Weather System (CIWS) Weather Products, Aug. 1, 2005, 120 pages.
Kuntman et al, Turbulence Detection and Avoidance System, Flight Safety Foundation 53rd International Air Safety Seminar (IASS), Oct. 29, 2000.
Kuntman, Airborne System to Address Leading Cause of Injuries in Non-Fatal Airline Accidents, ICAO Journal, Mar. 2000.
Kuntman, Satellite Imagery: Predicting Aviation Weather Hazards, ICAO Journal, Mar. 2000, 4 pps.
Lahiff, 2005, Vertically Integrated Liquid Density and Its Associated Hail Size Range Across the Burlington, Vermont County Warning Area, Eastern Regional Technical Attachment, No. 05-01, 20 pages.
Liu, Chuntao et al., Relationships between lightning flash rates and radar reflectivity vertical structures in thunderstorms over the tropics and subtropics, Journal of Geophysical Research, vol. 177, D06212, doi:10.1029/2011JDo17123,2012, American Geophysical Union, Mar. 27, 2012, 19 pages.
Meteorological/KSC/L71557/Lighting Detection and Ranging (LDAR), Jan. 2002, 12 pages.
Nathanson, Fred E., “Radar and Its Composite Environment,” Radar Design Principles, Signal Processing and the Environment, 1969, 5 pages, McGraw-Hill Book Company, New York et al.
Non-Final Office Action on U.S. Appl. No. 13/238,606 dated May 27, 2015, 14 pages.
Non-Final Office Action on U.S. Appl. No. 14/452,235 dated Apr. 23, 2015, 9 pages.
Non-Final Office Action on U.S. Appl. No. 14/681,901 dated Jun. 17, 2015, 21 pages.
Non-Final Office Action on U.S. Appl. No. 12/892,663 dated May 29, 2013, 14 pages.
Non-Final Office Action on U.S. Appl. No. 13/238,606 dated Jul. 8, 2014, 12 pages.
Non-Final Office Action on U.S. Appl. No. 13/238,606 dated Mar. 27, 2015, 21 pages.
Non-Final Office Action on U.S. Appl. No. 13/238,606 dated Sep. 23, 2013, 15 pages.
Non-Final Office Action on U.S. Appl. No. 13/717,052 dated Feb. 11, 2015, 15 pages.
Non-Final Office Action on U.S. Appl. No. 13/717,052 dated Sep. 9, 2014, 8 pages.
Non-Final Office Action on U.S. Appl. No. 13/841,893 dated Jun. 22, 2015, 27 pages.
Non-Final Office Action on U.S. Appl. No. 13/913,100 dated May 4, 2015, 25 pages.
Non-Final Office Action on U.S. Appl. No. 13/919,406 dated Jul. 14, 2015, 23 pages.
Non-Final Office Action on U.S. Appl. No. 14/162,035 dated Jul. 11, 2016, 10 pages.
Non-Final Office Action on U.S. Appl. No. 14/162,035, dated Feb. 4, 2016, 9 pages.
Non-Final Office Action on U.S. Appl. No. 14/323,766, dated Feb. 8, 2017, 21 pages.
Non-Final Office Action on U.S. Appl. No. 14/465,753, dated Apr. 4, 2016, 12 pages.
Non-Final Office Action on U.S. Appl. No. 14/608,071, dated Jan. 23, 2017, 15 pages.
Non-Final Office Action on U.S. Appl. No. 14/086,844, dated Nov. 10, 2015, 17 pages.
Notice of Allowability on U.S. Appl. No. 14/162,035, dated Feb. 17, 2017, 2 pages.
Notice of Allowance on U.S. Appl. No. 12/075,103 dated Aug. 4, 2014, 10 pages.
Notice of Allowance on U.S. Appl. No. 13/246,769 dated Jan. 8, 2015, 10 pages.
Notice of Allowance on U.S. Appl. No. 13/707,438 dated Feb. 25, 2015, 11 pages.
Notice of Allowance on U.S. Appl. No. 14/162,035, dated Nov. 9, 2016, 7 pages.
Notice of Allowance on U.S. Appl. No. 14/465,730, dated Feb. 28, 2017, 5 pages.
Notice of Allowance on U.S. Appl. No. 14/608,071, dated Aug. 31, 2017, 9 pages.
Notice of Allowance on U.S. Appl. No. 14/681,901, dated Dec. 23, 2015, 8 pages.
Office Action for U.S. Appl. No. 12/892,663, dated Oct. 22, 2012, 12 pages.
Office Action for U.S. Appl. No. 13/717,052, dated Aug. 22, 2013, 15 pages.
Office Action on U.S. Appl. No. 12/075,103 dated Jul. 31, 2013, 8 pages.
Office Action on U.S. Appl. No. 13/246,769 dated Apr. 21, 2014, 18 pages.
Pessi et al., On the Relationship Between Lightning and Convective Rainfall Over the Central Pacific Ocean, date unknown, 9 pages.
Robinson et al., En Route Weather Depiction Benefits of the Nexrad Vertically Integrated Liquid Water Product Utilized by the Corridor Integrated Weather System, 10th Conference on Aviation, Range, and Aerospace Meteorology (ARAM), 2002, 4 pages.
Stormscope Lightning Detection Systems, L3 Avionics Systems, retrieved on Jul. 11, 2011, 6 pages.
TOA Technology, printed from website: http://www.toasystems.com/technology.html on Dec. 29, 2010, 2 pages.
Triangulation, from Wikipedia, printed from website: http://en.wikipedia.org/wiki/Triangulation on Dec. 29, 2010, 6 pages.
Waldvogel et al., The Kinetic Energy of Hailfalls. Part I: Hailstone Spectra, Journal of Applied Meteorology, Apr. 1978, 8 pages.
Wilson et al., The Complementary Use of Titan-Derived Radar and Total Lightning Thunderstorm Cells, paper presented on Oct. 16, 2005, 10 pages.
Zipser et al., The Vertical Profile of Radar Reflectivity and Convective Cells: A Strong Indicator of Storm Intensity and Lightning Probability? America Meteorological Society, 1994, 9 pages.
Notice of Allowance for U.S. Appl. No. 14/872,942 dated Jan. 17, 2019.
Notice of Allowance for U.S. Appl. No. 14/323,766 dated Jun. 19, 2017. 7 pages.
Non-Final Office Action for U.S. Appl. No. 14/872,942 dated Oct. 6, 2017. 15 pages.
Corrected Notice of Allowance for U.S. Appl. No. 14/608,071 dated Dec. 15, 2017. 3 pages.
Final Office Action for U.S. Appl. No. 14/872,942 dated Jun. 1, 2018. 17 pages.
Continuations (1)
Number Date Country
Parent 14162035 Jan 2014 US
Child 15462712 US