This disclosure is related to modeling geospatial data. More particularly, the embodiments disclosed herein are directed at enhancing the resolution of geospatial data for location-based applications and analyses.
Radar-derived raster digital surface models (DSMs) provide a critical component for many modern applications, including flood risk analysis, telecommunications, pipeline routing, military, agriculture, and others. Interferometric Synthetic Aperture Radar (IFSAR) technologies have historically been able to produce DSMs with resolutions that range from 30 m up to 5 m depending on the sensor design and the operational parameters.
Noise gets introduced to the DSM when the DSM is processed at the same resolution as the image(s) from which the DSM is derived. The noise reduces the vertical accuracy of the data and can obscure spatial features that would otherwise be detectable. To address this issue, filtering is normally applied. However, filtering typically reduces the noise level at the expense of DSM resolution. This results in the DSM being generated at a lower resolution than the original images, e.g., as much as 4-8 times lower resolution than the image. Accordingly, there is a need for methods to recover the DSM resolution that gets lost due to filtering of the noise.
This disclosure is directed at systems and methods of enhancing or restoring details associated with high resolution images into a filtered DSM. The disclosed methods include mapping the changing gray scale values (intensity) from the images to changes in elevation in the DSM using a regression over a local neighborhood of pixels. Further, the disclosed methods do not rely on information about the radar illumination geometry and are extendable to be able to utilize any types of images (e.g. optical images). Although, the present discussions are couched using examples of IFSAR DSMs and images, the technique is generalizable to include DSMs and images from any source and any resolution scale. Additionally, since the present methods do not attempt to reconstruct a surface normal vector, the performance of the methods in enhancing resolution of the DSMs are similar regardless of terrain slope.
The disclosed method takes a model (e.g., a DSM) as input and improves the spatial content and resolution using an image that includes more features than those included in the model. The method is applicable to DEMs and images generated from any sensor technology, including but not limited to IFSAR. Further, the disclosed methods are not dependent on the resolution scale of the input model. The methods can be applied to models of any resolution scale, using imagery of any resolution scale, when the imagery includes features that are not evident in the model. Additionally, the disclosed embodiments are sensor agnostic. That is, the disclosed methods can be applied on any type of images collected by any type of sensor. Examples of image types can be a thermal image, a multi-spectral image, a hyper-spectral image, an optical image, a medical image, a radar image, a weather image, a fused image from multiple types of sensors, a color image, a gray scale image, or a LiDAR intensity image, or any image spatially referenced to the DSM associated with the disclosed methods.
In some embodiments, shape-from-shading (SFS) methodologies are used to extract higher resolution information from images and translate such information into improved terrain definition in the elevation data. Conventional SFS methodologies are based on advanced knowledge about radar image phenomenology in order to properly utilize the radar geometry. These methods, however, are affected by common radar imaging phenomena like speckle, foreshortening and layover. Such phenomena can introduce undesirable artifacts or changes in the gray values of images. These methodologies are also very specific to the sensor technology being used. For example, shape-from-shading techniques for radar imaging are very different than they are for optical imaging, and both are dependent on sensor characteristics and imaging geometry. Furthermore, shape-from-shading techniques exploit variations in the image gray values to reconstruct a normal vector to the DSM surface at each image location, which although useful for accentuating changes in sloped areas, do not offer significant enhancement in flat terrain.
In some embodiments, the disclosed methods are applied to images contemporaneously when the DSM is generated. In some embodiments, errors between the geo-registration quality of the DSM and the images are avoided or minimized. Avoiding or minimizing the errors results in avoiding undesirable spatial features to be added at incorrect locations in the DSM.
IFSAR systems use two antennae separated by an interferometric baseline (B) to image the earth's surface by transmitting radar pulses toward the terrain. The reflected energy is recorded by both antennae, simultaneously providing the system with two SAR images that include amplitude and phase of the same point on the ground, with the two images being separated only by the phase difference created by the space between the two antennae. In addition, as the aircraft passes over the terrain, global positioning system (GPS) data from both aircraft- and ground-based GPS devices as well as navigation data from an inertial measurement unit (IMU) onboard the aircraft can be collected. This navigation data is processed to provide the precise position of the aircraft.
The phase difference between the antennae for each image point, along with range, baseline, GPS, and navigation data, is used to infer the precise topographic height of the terrain being imaged. This enables the creation of an interferogram (depicting the phase difference) from which the DSMs can be derived. Through additional processing, the disclosed DTM is generated.
The DSM is a topographic model of the earth's surface that can be manipulated using a computer. Surface elevation models play a critical role in applications such as biomass studies, flood analysis, geologic and topographic mapping, environmental hazard assessment, oil and gas, telecommunications, and many other applications. The DSM includes elevation measurements that are laid out on a grid. These measurements are derived from the return signals received by two radar antennae mounted on an aircraft. The signals bounce back from first surface they strike, making the DSM a representation of any object large enough to be resolved, including buildings and roads, as well as vegetation and other natural terrain features.
As technologies advance, the demand for higher resolution DSMs that can meet the specifications of modern applications is rising. In such instances where high accuracy and densely sampled elevation data are desirable objectives, other technologies such as Light Detection and Ranging (LIDAR) and stereo photogrammetry can be employed. However, the costs associated with utilizing these technologies can be prohibitive. The higher cost places a limitation on the extent of data that can practically be acquired. For example, in some situations, the dataset can be limited to a smaller size. When compared against these technologies, in some instances, IFSAR can be a more efficient and economical data collection platform since IFSAR is able to penetrate through cloud, smoke, fog and haze and can collect wider swaths of data by aircraft flying at higher altitudes, yielding greater ground coverage.
Preparing Input
In some embodiments, the disclosed method is based on the raster DSM pixels being coincident with the image pixels. Thus, the first step is that the DSM is resampled so that for every image pixel, there is a corresponding DSM pixel. The resampling can be done using techniques such as bilinear resampling, bicubic resampling, nearest neighbor resampling, natural neighbor resampling, kriging resampling, box average resampling, or box median resampling. In some embodiments, the images for input are in grayscale format with a single intensity value for each pixel. Therefore, if a color optical image is being used, the color image is first converted into a grayscale format.
Isolating Surface Features
Low frequency terrain variations typically have a negative impact on the results because the algorithm maps localized changes in elevation to the image grayscale changes. According to disclosed embodiments, slopes present in the terrain are interpreted as elevation change(s), but are not related to the localized distinguishable grayscale changes in the image. Therefore, these slopes are removed to eliminate this confounding effect and isolate the surface feature elevations.
There are many possible ways to identify the low frequency terrain variations. In some embodiments, low frequency terrain variations are identified by applying a coarse smoothing operation to the DSM. The specific parameters of a smoothing filter (e.g., an averaging filter or a median filter) are selected so that the surface features are removed without over-smoothing the actual terrain. For example, over-smoothing can be prevented by ensuring the filter width is not too large. That is, the filter width is chosen to be large enough to remove the surface features, but no so large that it causes the over smoothing of the terrain. When a smoothing filter is applied to the DSM, the size of the filter is defined typically by the number of raster pixels included in the filter kernel. For example, if a DSM has pixels that are 5 meters wide, and a 5×5 smoothing filter is applied, the filter may have a size of 25 m×25 m. This size of kernel can be effective at smoothing over features that are smaller than 25 m in size. Features larger than this may be smoothed to some degree, but not removed from the DSM. As a side-effect of this process, terrain definition can be reduced to some extent because sharp break and drain lines can be rounded off with a radius proportional to the smoothing filter size. When implementing the smoothing process, the area being operated on is analyzed to select a kernel larger than the largest surface feature that needs to be removed. As an example, if a particular area has buildings that are no larger than 18 m in length and in width, a 4×4 kernel size (20 m by 20 m based on 5 m pixels) may be effective at removing the surface features while preserving as much terrain definition as possible. If an area has buildings that are 47 m in length and in width, a larger filter can be used (10×10 kernel size for example).
Another approach is to use a Fourier domain filter to identify low-order terrain variations from the DSM and remove the high frequency content. Additional low frequency terrain identification algorithms can be used that may be more complex, but produce superior results. In some embodiments, a Digital Terrain Model (DTM) is used.
The DTM is a topographic model of the bare earth that can be manipulated using a computer. Vegetation, buildings, and other cultural features have been digitally removed from the DTM, leaving just the underlying terrain. (A DTM is created by removing vegetation, buildings, and other cultural features from a DSM. This is achieved using the disclosed methods, according to which terrain elevations are derived from measurements of bare ground included in the original radar data as well as by manually reviewing and editing every pixel. One key feature of a DTM is that the DTM infers the terrain characteristics that may be hidden in the DSM.
Regardless of how the low frequency terrain is obtained, the process of isolating the surface features is based on subtraction of the low frequency terrain from the DSM. The result is a difference surface that is typically flat except for noise and surface features. This process is illustrated in
Δ=DSM−T (1)
where:
Δ is the isolated surface features,
DSM denotes the upsampled DSM, and
T is the low frequency terrain.
Adjusting Surface Features
The process operates on the surface features (Δ from Equation (1)) and the grayscale image, iterating on a pixel-by-pixel basis. At each pixel, a correspondence table is constructed that provides a mapping of the difference values (e.g., Δ values) and the image values for all pixels in a neighborhood. The neighborhood size may vary.
Y=mx+b (2)
where:
Y is the surface feature elevation (Δ), (e.g., shown in
x is the image gray value, (e.g., shown in
m is the slope of the best-fit line, and
b is the y-intercept of the best-fit line.
Using the pixel neighborhood provides a set of points that can be used to determine the slope and intercept values for Equation (2). As an example, the 21×21 pixel neighborhood provides 441 points that can be used to determine the slope and intercept values for Equation (2). Upon determining the slope and intercept, the relationship for mapping gray values to elevation adjustment is established for the specific neighborhood of the target pixel. The adjusted elevation for the target pixel can be computed by applying Equation (2) to the gray value of the target pixel (at the center of the kernel).
After all pixels are adjusted, these surface features are added back to the low frequency terrain surface. This produces the output DSM. According to disclosed embodiments, the output DSM is invariant to changes in one or more pixel values in the original image. A linear model described herein is for discussion purposes only. In some embodiments, the fit/regression model can be a non-linear model, e.g., a second order or a third order model. Further, the neighborhood kernel can be of any size.
Example Results
In some embodiments, the first model and the second model can be generated using a digital surface model (DSM), a weather model, a medical imaging/tomographic model, or a three dimensional (3D) digital model.
Results for a sample area are shown in
To demonstrate the flexibility of the disclosed method, the method is applied to an optical image instead of a radar image. For example,
In some embodiments, the disclosed methods are applied using an iterative procedure for refining the model for a pre-specified number of iterations. In such embodiments, the second model is produced during the first iteration. The second model becomes the first model in the second iteration. In this iterative procedure, the first iteration can be based on a model that did not fully capture the feature content that was evident in the image. After applying the disclosed methods, the second model is an improvement to the first model and more accurately captures those features. As the iterative procedure continues, the model values move closer to the real values.
In some embodiments, the disclosed methods can be applied progressively, where resolution and feature content are added in stages. For example, the disclosed methods could be used to enhance a 10 m resolution DSM to 1.25 m either directly in one step, or progressively in stages. The stages could be to first enhance from 10 m to 5 m, then from 5 m to 2.5, and then from 2.5 m to 1.25 m, for example. In some applications, a progressive approach produces better results, such as when there is a large gap between the input model resolution and the output model resolution. The second model becomes the first model in the second step. In this progressive procedure, the first step can be based on a model that did not fully capture the feature content that was evident in the image. After applying the disclosed methods, the second model is an improvement to the first model and more accurately captures those features. As the progressive procedure continues, the resolution moves closer to the desired target resolution which is typically the full resolution of the input image (e.g. 1.25 m in this example).
This document discloses a new method for enhancing DSM resolution and spatial content by using detailed images to guide the resampling procedure. The method presented can be applied to optical or radar image inputs and performs consistently regardless of terrain slope. That is, the disclosed methods equally enhance the resolution of equally two identical features, even if one of those features is on highly sloped terrain, while the other is on flat terrain. In other words, the presence of slope(s) of the terrain has/have no effect on the performance of enhancement of the resolution of the features on the terrain. Additionally, the disclosed methods are easier to implement than traditional shape from shading techniques and can be applied to input DSM sources of any resolution scale.
The results (depicted in
Enhancing Resolutions of Digital Terrain Models (DTMs):
The DTM is a topographic model of the bare earth that can be manipulated using a computer. Vegetation, buildings, and other cultural features are not included in the DTM. Because a DTM primarily includes the underlying terrain, a DTM can be created by removing vegetation, buildings, and other cultural features from a DSM. A DTM can also be generated from classified multi-return LiDAR points.
The process of generating DTMs from DSMs often involves steps that can be destructive to the texture and definition of the terrain. For example, a smoothing process may be used to remove surface texture in vegetated and urban areas prior to shifting elevations down to the true ground level. Smoothing may also be used in transition zones to blend the boundary between previously obstructed areas and the surrounding bare earth terrain. Thus, a DTM infers the terrain characteristics that may be hidden in the DSM.
In another example, multi-return LiDAR points can be used to produce DTMs. In these examples, there may be a sparse availability of ground return points (e.g., a dense forest canopy may not reveal much of the underlying terrain hidden by the canopy). Alternatively, there may be no observable ground points, which occurs where buildings are located in a dense urban location (e.g., downtown New York City). When ground points are sparse, large area interpolation can be used to fill in the terrain elevation voids. However, these interpolated areas turn out to be overly smooth, artificial-looking, and devoid of the details and texture associated with natural terrains.
Embodiments of the present disclosure are directed at enhancing the resolution of DTMs. The resolution of the enhanced DTM is invariant to scale changes in the original image or the noise characteristics of the sensor used to capture the original image. The process for enhancing the resolution of a DTM approximates the noise characteristics of the sensor type used to produce the DTM. Non-limiting examples of sensors can be IFSAR, LiDAR, stereo photogrammetry, stereo SAR, or other suitable types of sensors for capturing images. The process further involves adding real terrain signatures from the images to restore the terrain data to the full definition (e.g., higher resolution) that is observed in the images.
In some embodiments (e.g., as shown in
Some of the embodiments described herein are described in the general context of methods or processes, which may be implemented in one embodiment by a computer program product, embodied in a computer-readable medium, including computer-executable instructions, such as program code, executed by computers in networked environments. A computer-readable medium may include removable and non-removable storage devices including, but not limited to, Read Only Memory (ROM), Random Access Memory (RAM), compact discs (CDs), digital versatile discs (DVD), etc. Therefore, the computer-readable media may include a non-transitory storage media. Generally, program modules may include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Computer- or processor-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps or processes.
Some of the disclosed embodiments may be implemented as devices or modules using hardware circuits, software, or combinations thereof. For example, a hardware circuit implementation may include discrete analog and/or digital components that are, for example, integrated as part of a printed circuit board. Alternatively, or additionally, the disclosed components or modules may be implemented as an Application Specific Integrated Circuit (ASIC) and/or as a Field Programmable Gate Array (FPGA) device. Some implementations may additionally or alternatively include a digital signal processor (DSP) that is a specialized microprocessor with an architecture optimized for the operational needs of digital signal processing associated with the disclosed functionalities of this application. Similarly, the various components or sub-components within each module may be implemented in software, hardware or firmware. The connectivity between the modules and/or components within the modules may be provided using any one of the connectivity methods and media that is known in the art, including, but not limited to, communications over the Internet, wired, or wireless networks using the appropriate protocols.
The foregoing description of embodiments has been presented for purposes of illustration and description. The foregoing description is not intended to be exhaustive or to limit embodiments of the present invention(s) to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of various embodiments. The embodiments discussed herein were chosen and described in order to explain the principles and the nature of various embodiments and its practical application to enable one skilled in the art to utilize the present invention(s) in various embodiments and with various modifications as are suited to the particular use contemplated. The features of the embodiments described herein may be combined in all possible combinations of methods, apparatus, modules, systems, and computer program products.
This application is a continuation of U.S. patent application Ser. No. 16/221,859, filed Dec. 17, 2018, entitled “METHOD AND APPARATUS FOR ENHANCING 3D MODEL RESOLUTION,” which is a continuation-in-part of U.S. patent application Ser. No. 15/963,937, filed Apr. 26, 2018, entitled “METHOD AND APPARATUS FOR ENHANCING 3D MODEL RESOLUTION,” which is a continuation of U.S. Pat. No. 10,002,407 issued on Jun. 19, 2018, and entitled “METHOD AND APPARATUS FOR ENHANCING 30 MODEL RESOLUTION,” which claims the benefit of U.S. Provisional Patent Application Ser. No. 62/544,608, filed on Aug. 11, 2017, and entitled “METHOD AND APPARATUS FOR ENHANCING 30 MODEL RESOLUTION,” the disclosures of which are hereby incorporated by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
5109430 | Nishihara | Apr 1992 | A |
5187754 | Currin | Feb 1993 | A |
5926581 | Pritt | Jul 1999 | A |
6272448 | Ishii | Aug 2001 | B1 |
6448544 | Stanton | Sep 2002 | B1 |
6792684 | Hyyppa | Sep 2004 | B1 |
6937775 | Gindele | Aug 2005 | B2 |
6973218 | Alderson | Dec 2005 | B2 |
6978050 | Hunter | Dec 2005 | B2 |
7043090 | Gindele | May 2006 | B2 |
7116838 | Gindele | Oct 2006 | B2 |
7187452 | Jupp | Mar 2007 | B2 |
7236649 | Fenney | Jun 2007 | B2 |
7386164 | Shragai | Jun 2008 | B2 |
7627491 | Feyen | Dec 2009 | B2 |
7639842 | Kelle | Dec 2009 | B2 |
7856312 | Coombes | Dec 2010 | B2 |
7881913 | O'Neil | Feb 2011 | B2 |
7917292 | Du | Mar 2011 | B1 |
7917346 | Sullivan | Mar 2011 | B2 |
8010294 | Dorn | Aug 2011 | B2 |
8139111 | Oldroyd | Mar 2012 | B2 |
8239179 | Duncan | Aug 2012 | B2 |
8295554 | Francini | Oct 2012 | B2 |
8369579 | Barnes | Feb 2013 | B2 |
8473264 | Frigerio | Jun 2013 | B2 |
8655595 | Green et al. | Feb 2014 | B1 |
8825456 | Vasudevan | Sep 2014 | B2 |
8855439 | Louis | Oct 2014 | B2 |
8874187 | Thomson | Oct 2014 | B2 |
9147285 | Kunath | Sep 2015 | B2 |
9299191 | Sinram | Mar 2016 | B2 |
9830690 | Chiang | Nov 2017 | B2 |
9875554 | Imber | Jan 2018 | B2 |
10002407 | Mercer | Jun 2018 | B1 |
10147057 | Zhang | Dec 2018 | B2 |
10186015 | Mercer | Jan 2019 | B1 |
10204454 | Goldman | Feb 2019 | B2 |
20030138152 | Fennye | Jul 2003 | A1 |
20040260471 | McDermott | Dec 2004 | A1 |
20080260237 | Savolainen | Oct 2008 | A1 |
20100292973 | Barnes | Nov 2010 | A1 |
20110246935 | Maeder | Oct 2011 | A1 |
20110295575 | Levine et al. | Dec 2011 | A1 |
20120330553 | Mollaei | Dec 2012 | A1 |
20130046471 | Rahmes | Feb 2013 | A1 |
20130197807 | Du | Aug 2013 | A1 |
20130342736 | Numata | Dec 2013 | A1 |
20140267250 | Tennant | Sep 2014 | A1 |
20150046139 | Iwamura | Feb 2015 | A1 |
20150154805 | Hsu | Jun 2015 | A1 |
20160148347 | Guido | May 2016 | A1 |
20170358067 | Jung | Dec 2017 | A1 |
Number | Date | Country |
---|---|---|
2008034465 | Mar 2008 | WO |
2010056643 | May 2010 | WO |
Entry |
---|
Non-Final Office Action in U.S. Appl. No. 14/825,806, dated Dec. 1, 2017, 13 pages. |
Non-Final Office Action in U.S. Appl. No. 15/963,937, dated May 31, 2018, 10 pages. |
Notice of Allowance in U.S. Appl. No. 14/825,806, dated Jul. 11, 2018, 12 pages. |
Notice of Allowance in U.S. Appl. No. 14/825,806, dated Sep. 4, 2018, 11 pages. |
Notice of Allowance in U.S. Appl. No. 15/723,154, dated Jan. 26, 2018, 9 pages. |
Notice of Allowance in U.S. Appl. No. 15/963,937, dated Sep. 19, 2018, 8 pages. |
Notice of Allowance in U.S. Appl. No. 16/221,859, dated Jan. 29, 2019, 10 pages. |
Number | Date | Country | |
---|---|---|---|
20190311461 A1 | Oct 2019 | US |
Number | Date | Country | |
---|---|---|---|
62544608 | Aug 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16221859 | Dec 2018 | US |
Child | 16398027 | US | |
Parent | 15723154 | Oct 2017 | US |
Child | 15963937 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15963937 | Apr 2018 | US |
Child | 16221859 | US |