System and method for construction estimation using aerial images

Information

  • Patent Grant
  • 11693996
  • Patent Number
    11,693,996
  • Date Filed
    Thursday, December 23, 2021
    2 years ago
  • Date Issued
    Tuesday, July 4, 2023
    11 months ago
Abstract
A system and method for construction estimation using aerial images is provided. The system receives at least one aerial image of a building. An estimation engine processes the aerial image at a plurality of angles to automatically identify a plurality (e.g., perimeter and interior) lines in the image corresponding to a plurality of features of a roof the building. The estimation engine allows users to generate two-dimensional and three-dimensional models of the roof by automatically delineating various roof features, and generates a report including information about the roof of the building.
Description
BACKGROUND OF THE INVENTION
Field of the Invention

The present invention relates generally to the field of construction estimation. More specifically, the present invention relates to a system and method for construction estimation using aerial images.


Related Art

In the construction and insurance industries, accurate and rapid estimation of construction materials and costs is important. For example, such information is often used by construction professionals to specify materials and associated costs for both newly-constructed buildings, as well as for replacing and upgrading existing structures. Further, in the insurance industry, accurate information about construction materials and costs is critical to determining the proper costs for insuring buildings/structures.


It is of particular importance in the construction and insurance industries to quickly and accurately estimate various parameters regarding roof structures, such as roof dimensions, pitches, surfaces, areas, and associated material costs. To this end, various software systems have been implemented to estimate roofing materials and roofing costs. Such systems process aerial images (e.g., using digital photographs taken from airplanes flying over buildings) and allow users to estimate construction materials and costs by delineating the bounds of a roof, and by calculating dimensions based upon the delineated bounds. However, such systems are time-consuming and difficult to use. Indeed, such systems often require a large amount of manual input by the user (e.g., by manually delineating roof boundaries using a mouse and graphical user interface) before a roof model or estimation report can be generated.


In view of existing technology in this field, what would be desirable is a system that processes aerial images of a building structure and automatically performs steps to quickly and efficiently assist a user in the roof estimation process. Specifically, what would be desirable is a system that automatically delineates roof boundaries in an image of a building or a structure with minimal user intervention, automatically creates a model of the roof structure, and automatically generates a roof estimation report which includes useful information about the roof including material costs and dimensions of the roof. Accordingly, what would be desirable, but has not yet been provided, is a system and method for construction estimation using aerial images which addresses the foregoing needs.


SUMMARY OF THE INVENTION

The present invention relates to a system and method for construction estimation using aerial images. The system includes a computer system for receiving at least one aerial image of a building, and an estimation engine for processing the aerial image to estimate one or more features relating to a roof of the building, the estimation engine histogram processing the aerial image at a plurality of angles to automatically identify a plurality of lines in the image corresponding to features of a roof the building.


A method for building estimation is also provided. The method comprises the steps of receiving at a computer system at least one aerial image of a building, histogram processing the aerial image at a plurality of angles using an estimation engine in the computer system to automatically identify a plurality of lines in the image corresponding to a plurality of feature of a roof of the building, constructing a three-dimensional wireframe model of the roof using the plurality of lines, and processing the three-dimensional model of the roof using the estimation engine to generate a report including information about the roof of the building.


A non-transitory, computer-readable medium is also provided. The computer readable medium includes computer-readable instructions stored thereon which, when executed by a computer system, cause the computer system to perform the steps comprising receiving at a computer system at least one aerial image of a building, histogram processing the aerial image at a plurality of angles using an estimation engine in the computer system to automatically identify a plurality of lines in the image corresponding to a plurality of feature of a roof of the building, constructing a three-dimensional wireframe model of the roof using the plurality of lines, and processing the three-dimensional model of the roof using the estimation engine to generate a report including information about the roof of the building.





BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing features of the invention will be apparent from the following Detailed Description of the Invention, taken in connection with the accompanying drawings, in which:



FIG. 1 is a diagram showing the system of the present invention for construction estimation using aerial images;



FIG. 2 is a flowchart showing overall processing steps carried out by the system of the present invention;



FIGS. 3A-3L are screenshots showing operation of the system of the present invention in detail;



FIGS. 4A-4B are views of a sample estimation report generated by the present invention;



FIG. 5 is a flowchart showing processing steps carried out by the present invention for processing an aerial image to automatically identify roof lines; and



FIGS. 6A-6K are views illustrating the steps of FIG. 5 for processing an aerial image to automatically identify roof lines.





DETAILED DESCRIPTION OF THE INVENTION

The present invention relates to a system and method for construction estimation using aerial images, as discussed in detail below in connection with FIGS. 1-6K.



FIG. 1 is a diagram showing the system of the present invention, indicated generally at 10. The system 10 comprises a computer system 12 (e.g., a server) having an aerial image database 14 stored therein and an estimation software engine (module) 16. The database 14 could be stored on the computer system 12, or located externally (e.g., in a separate database server in communication with the system 10). As will be discussed in greater detail below, the estimation engine 16 allows users to generate three-dimensional models of roofs and buildings/structures to estimate dimensions of such structures as well as material costs and other parameters. Importantly, the estimation engine 16 allows users to easily generate such models by automatically delineating various roof structures such as eaves, hips, ridges, valleys, etc.


The system 10 can communicate through a network 18 with one or more of a variety of image providers to obtain aerial images or photographs of a building structure 20 and can store them in the aerial image database 14 in any suitable format, such as JPEG, TIFF, GIF, etc. Network communication could be over the Internet using standard TCP/IP communications protocols (e.g., hypertext transfer protocol (HTTP), secure HTTP (HTTPS), file transfer protocol (FTP), electronic data interchange (EDI), etc.), through a private network connection (e.g., wide-area network (WAN) connection, e-mails, electronic data interchange (EDI) messages, extensible markup language (XML) messages, file transfer protocol (FTP) file transfers, etc.), or any other suitable wired or wireless electronic communications format.


Image providers that the computer system 12 could communicate with include, but are not limited to, an airplane 22 having a camera 24 capable of capturing images of the structure 20, and/or a third-party aerial image provider 26, such as Pictometry, Google, or Bing. Although images of any quality can be used, high-quality images free from obstructions (e.g., trees, shadows, snow, etc.) are preferred.


The computer system 12 could be any suitable computer server (e.g., a server with an INTEL microprocessor, multiple processors, multiple processing cores) running any suitable operating system (e.g., Windows by Microsoft, Linux, etc.). The computer system 12 includes non-volatile storage, which could include disk (e.g., hard disk), flash memory, read-only memory (ROM), erasable, programmable ROM (EPROM), electrically-erasable, programmable ROM (EEPROM), or any other type of non-volatile memory. The estimation engine 16, discussed in greater detail below, could be embodied as computer-readable instructions stored in computer-readable media (e.g., the non-volatile memory mentioned above), and programmed in any suitable programming language (e.g., C, C++, Java, etc.).


The system 10 could be web-based and could allow for remote access to the system 10 over a network 28 (e.g., Internet, WAN, LAN, etc.) by one or more devices, such as a personal computer system 30, a smart cellular telephone 32, a tablet computer 34, or other devices. It is also contemplated that at least some of the functionality of the system 10 could run locally on devices (e.g., personal computer 30, smart cellular telephone 32, tablet computer 34, etc.) programmed with software in accordance with the present invention. It is conceivable that, in such circumstances, the device could communicate with a remote aerial image database over a network 28.



FIG. 2 is a flowchart showing overall processing steps 40 carried out by the estimation engine 16 of the system 10. Reference is also made herein to FIGS. 3A-4B, which are views showing operation of the estimation engine 16. Beginning in step 42, as shown in FIG. 3A, a graphical user interface 80 is generated by the estimation engine 16 and presented to a user to allow the user to select an aerial image 82 of a building or structure, preferably from the aerial image database 14. Then, in step 44, the user can crop the image 82 by selecting a roof 86 from the image 82 using a selection tool 84 of the graphical user interface 80, e.g., by creating a box around the desired area. In step 46, as shown in FIG. 3B, the estimation engine 16 automatically rotates the roof 86 of the cropped image 82 so that the image 82 is aligned vertically and horizontally using roof lines 88, 90.


The process for aligning the image 82 is described as follows. First, the processing steps discussed in detail below in connection with FIG. 5 are carried out to automatically identify perimeter and interior roof lines in the image. Using this information, a sequence of points are identified in each direction (e.g., 360 degrees) and are defined as rays in the set {raycast}, and include top ray {top}, bottom ray {bottom}, left ray {left}, and right ray {right}. Then, for each ray in the set {raycast}, the sequence of points are divided into |raycast|−2 subsets of two sequences made up of points {{0 . . . 1}, {2 . . . |raycast|−1}}, {{0 . . . 2}, {3 . . . |raycast|−1} } to {{0−|raycast| . . . 3}, {|raycast|−2 . . . |raycast|−1}} where |raycast| is the length of the sequence and 0 . . . 1 is the indices of the first and second point and 2 . . . |raycast|−1 is the remaining points in the sequence. For each of these subsequences {{0 . . . n}, {n+1 . . . |raycast|−1}}, the two point sets are treated as plots and their linear slope is found using linear regression. The difference between slopes is found, and only the maximum difference is retained. After iterating through every subsequence in the set {raycast}, if the maximum difference is greater than a specified threshold, the sequence is split at that point and the process is recursively repeated for each of the subsequences. Once all subsequences can no longer be divided, the length and slope of each subsequence is cached for further use. Finally, slopes are identified which are similar, and their lengths are combined. The slopes are then used to identify the roof lines 88, 90, which are then used to align the image 82 vertically and horizontally.


Then in step 48, as shown in FIG. 3C, the estimation engine 16 uses image processing algorithms to automatically identify and delineate perimeter lines 92 of the roof 86 to form a closed polygon or a bounded area. In step 50, a user decides whether the estimation engine 16 has correctly identified the perimeter lines 92 or whether adjustments or alterations must be made. If alterations are required, step 52 occurs where, as shown in FIG. 3D, the user manually adjusts the perimeter lines 92 to create corrected perimeter lines 94, such as by clicking on the corners of the roof 86 that were not correctly identified automatically. As the user clicks on or moves the cursor toward a roof feature (e.g., a corner), the estimation engine 16 could automatically suggest potential changes to the outline of the perimeter, by way of dashed lines 94 as shown in FIG. 3D. The user can confirm changes by clicking on the dashed lines 94, which converts the dashed lines to solid lines.


After perimeter line alterations have been made, or if no alterations are determined to be needed in step 50, the process proceeds to step 54 where, as shown in FIG. 3E, the estimation engine 16 automatically identifies and delineates interior lines 96 (e.g., ridges and valleys) and interior intersections 98 of the roof 86. Different roof features can be identified and distinguished using lines of different colors (e.g., ridges as red lines, valleys as blue lines, etc.). The user can move the mouse cursor over certain sections of the roof, such as intersection 98, and the system evaluates that section (e.g., by evaluating the proximity and configuration of existing adjacent points and other detected and identified lines) and automatically delineates colored dashed lines 96 as suggestions for lines to be included in the model. The user can then confirm the line, change the type of line, or reject the line. Once confirmed by the user, the dashed lines 96 are converted to solid lines 100, as shown in FIG. 3F. In step 56, a determination is made as to whether all lines have been found. If a negative determination is made, step 54 is repeated. Otherwise, if all lines have been identified (e.g., including the lines 102, 104, 106 in FIG. 3F, as well as any other lines corresponding to roof structures), a two-dimensional wireframe of the roof is created based on the perimeter lines and interior lines discussed above.


The two-dimensional model 110 is shown in FIG. 3G. The model 110 can be processed, either automatically by the engine 16 or through manual user manipulation, into a three-dimensional model. Thus, in step 59, a decision is made as to whether to automatically or manually create the three-dimensional model. If automatic creation is to be performed, a three-dimensional wireframe of the roof is created in step 61, based on the perimeter lines and interior lines. In this regard, the estimation engine 16 creates closed polygon surfaces, as shown in FIG. 3G, that represent the roof faces, and then calculates and assigns an orientation (e.g., direction of slope) based on roof constraints for each face. Using a center of mass calculation, the engine 16 selects a prominent face and assigns it a default slope. The slopes of the other faces are then iteratively calculated by the engine 16 based on constraints imposed by the relationship of lines between adjacent faces and orientations of the faces. Also, the user can manually enter a slope for a specific face and the engine 16 will recalculate the slope of the adjacent faces based on the manual slope entry. A three dimensional wireframe model is then created based on the slopes, as shown in FIG. 3H. The process then proceeds to step 70 discussed below.


If manual creation of the three-dimensional model is to be performed, the process proceeds to step 60, as shown in FIG. 3I, where an oblique image 112 is selected, preferably from the aerial database 14. Then, in step 62, the wireframe 110 is superimposed over the oblique image 112 either by the user or automatically by the estimation engine 16. Then in step 64, as shown in FIGS. 3J-3K, a plurality of points on the wireframe 110 are matched with a plurality of corresponding points on the roof 86 of the oblique image 112, and the wireframe 110 is deformed as a result. The plurality of points could be on eave edges and/or at the same eave height. For example, a user could match three corners 104, 116, 118 of the wireframe 110 with three corresponding points of the roof 86 in the oblique image 112, causing the wireframe 110 to deform accordingly. Then, the user clicks and drags the interior points of the wireframe 110 to corresponding locations on the roof 86 of the oblique image 112 (e.g., the ridge lines of the roof 86), until the wireframe 110 outlines all of the perimeter lines and interior lines of the roof 86 in the oblique image 112. It is noted that this process could also be carried out automatically by the estimation engine 16. The final, deformed wireframe is shown in FIG. 3L. In step 66, the estimation engine 16 calculates roof pitches 120 based on the deformations of the superimposed wireframe 110, as shown in FIG. 3L.


In step 68, a user could optionally set eave heights to fine-tune the wireframe 110, and the estimation engine 16 would recalculate pitches based thereon. For example, the user could use an eave edit handle 122 on the midpoint of each eave line of the wireframe 110 to adjust the slope of one or more faces of the wireframe 110 by clicking and dragging. Alternatively, the user could click and drag the entire wireframe 110 until the eave line of the wireframe 110 overlays the eave line of the roof 86 of the oblique image 112.


Proceeding to step 70, as shown in FIGS. 4A-4B, the system 10 generates a report. The report could comprise a two-dimensional wireframe model 130, a three-dimensional model, and/or one or more charts 132 containing construction information (e.g., roof dimensions, pitches, surfaces, and areas, and associated material costs). Other types of information, such as waste calculations, could also be included in the report. The report could have an XML flag as a way to verify that a roof model was created by the system 10. The system 10 could also require that a model have an XML flag before including the model in the report or before automatically generating the report.



FIG. 5 is a flowchart showing more detailed processing steps 140 carried out by the estimation engine 16 for processing an aerial image to automatically identify roof lines. Reference is also made herein to FIGS. 6A-6K, which are views illustrating the steps carried out in FIG. 5. Beginning with step 142, as shown in FIG. 6A, an aerial image is selected. A portion of the image containing the roof is selected in step 144, as shown in FIG. 6B, and the image is then cropped in step 146, as shown in FIG. 6C. The image is then converted to grayscale in step 148, as shown in FIG. 6D. In step 150, as shown in FIG. 6E, a filter is applied to the grayscale image, such as a Sobel filter, although it is noted that a filter could be applied without first converting the image to grayscale. The Sobel filter calculates the gradient of image intensity at each point in the image, giving a direction and rate of change to the largest increases from light to dark. The estimation engine 16 computes the likelihood that a point in the image represents a perimeter or interior line of a roof based on the direction and magnitude of the greatest intensity increase. In step 152, as shown in FIG. 6F, a Gaussian filter or blur could be applied to the image to decrease blur and/or sharpen the image.


In step 154, as shown in FIG. 6G, the image is rotated by a first angle θ (e.g., 48 degrees), as shown in FIG. 6G, and then histogram processing is performed on the rotated image in step 156. The horizontal lines shown in FIGS. 6G-6I correspond to the direction of scan for image intensity. In step 158, a determination is made as to whether further histogram processing is required. If so, the process repeats until the image has been processed at a sufficient number of angles. It has been found that histograms taken at one degree increments (e.g., from 0 through 360 degrees) have provided excellent results, but other increments and/or number of histograms used are possible. Indeed, as shown in FIG. 6H, the image could be rotated to align vertically and then histogram processed, and/or the image could be rotated to align horizontally and then histogram processed, as shown in FIG. 6I. Of course any other angles could be used.


In step 160, as shown in FIG. 6J, grid lines could be superimposed on the image, such as at perimeter lines, interior lines, and/or points of intersection. The grid lines can be used to segment/divide the image into sub-images, each of which are processed in accordance with the present invention. The grid lines need not be shown to the user, and indeed, can be used solely for purposes of image processing.


The histograms generated at the plurality of angles indicate the probability that a given feature in the image corresponds to a line (e.g., to a perimeter line or to an interior line corresponding a roof feature). In step 162, the histogram data (e.g., all of the histograms generated in steps 156-158) is processed to identify lines (e.g., perimeter and interior lines) corresponding to roof features. For example, in step 162, cluster analysis could be performed on the histograms to identify the lines (e.g., perimeter or interior lines). Thus, by application of the foregoing image processing techniques, the estimation engine 16 can scan the entire image and detect and identify all perimeter and interior lines. Finally, in step 164, as shown in FIG. 6K, a filtered image is generated and used by the estimation engine 16 to automatically delineate perimeter lines and interior lines in the GUI 80 shown in FIGS. 3A-3F.


As noted above, the estimation engine 16 can automatically identify the type of roof feature to which a given line in the model corresponds. This is accomplished by rules-based pattern recognition performed by the estimation engine 16, which could vary according to the type of roof feature to be automatically identified. For example, a set of rules could be programmed into the estimation engine 16 such that a ridge or a valley is automatically identified for a given point on the image if specific lines are detected in proximity to a given point in the image and/or if a pre-defined angle (or, range of angles) is detected between such lines. If the estimation engine 16 is unable to identify the type of line, the line could be temporarily marked as unknown and automatically altered by the estimation engine 16 as more information becomes available. Subsequently, the user could then manually correct the lines and line types if required.


Having thus described the invention in detail, it is to be understood that the foregoing description is not intended to limit the spirit or scope thereof. What is desired to be protected is set forth in the following claims.

Claims
  • 1. A system for construction estimation, comprising: a memory for storing at least one aerial image of a building; anda processor in communication with the memory, the processor: receiving, from the memory, the at least one aerial image;processing the at least one aerial image to automatically identify a plurality of perimeter lines and interior lines in the at least one aerial image corresponding to one or more features of a roof of the building;constructing a two-dimensional wireframe model of the roof based on the identified plurality of perimeter lines and interior lines; andautomatically generating a three-dimensional wireframe model of the roof of the building based on the constructed two-dimensional wireframe model by creating at least one closed polygon representing a roof face from the two-dimensional wireframe model of the roof and assigning a three-dimensional orientation to the at least one closed polygon surface based on at least one roof constraint associated with the roof face.
  • 2. The system of claim 1, further comprising a graphical user interface for displaying the at least one aerial image of the building.
  • 3. The system of claim 1, wherein the plurality of interior lines in the at least one aerial image correspond to one or more of an eave, a hip, a ridge, a valley and an interior intersection of the roof of the building.
  • 4. The system of claim 1, wherein the processor generates a report comprising construction information associated with the generated three-dimensional wireframe model of the roof, the construction information being indicative of one or more of roof dimensions, pitches, surfaces, areas, and material costs associated with the roof.
  • 5. The system of claim 1, wherein the processor crops the at least one aerial image around the building.
  • 6. The system of claim 1, wherein the processor allows a user to adjust one or more of the automatically identified plurality of perimeter and interior lines via the graphical user interface.
  • 7. A method for construction estimation, comprising the steps of: receiving, by a processor, at least one aerial image of a building;processing the at least one aerial image to automatically identify a plurality of perimeter lines and interior lines in the at least one aerial image corresponding to one or more features of a roof of the building;constructing a two-dimensional wireframe model of the roof based on the identified plurality of perimeter lines and interior lines; andautomatically generating a three-dimensional wireframe model of the roof of the building based on the constructed two-dimensional wireframe model by creating at least one closed polygon representing a roof face from the two-dimensional wireframe model of the roof and assigning a three-dimensional orientation to the at least one closed polygon surface based on at least one roof constraint associated with the roof face.
  • 8. The method of claim 7, further comprising generating and displaying a graphical user interface for displaying the at least one aerial image of the building.
  • 9. The method of claim 7, further comprising generating a report comprising construction information associated with the generated three-dimensional wireframe model of the roof, the construction information being indicative of one or more of roof dimensions, pitches, surfaces, areas, and material costs associated with the roof.
  • 10. The method of claim 7, further comprising cropping the at least one aerial image around the building.
  • 11. The method of claim 7, further comprising allowing a user to adjust one or more of the automatically identified plurality of perimeter and interior lines via a graphical user interface.
  • 12. A non-transitory, computer-readable medium having computer-readable instructions stored thereon which, when executed by a processor, causes the processor to perform the steps comprising: receiving, by the processor, at least one aerial image of a building;processing the at least one aerial image to automatically identify a plurality of perimeter lines and interior lines in the at least one aerial image corresponding to one or more features of a roof of the building;constructing a two-dimensional wireframe model of the roof based on the identified plurality of perimeter lines and interior lines; andautomatically generating a three-dimensional wireframe model of the roof of the building based on the constructed two-dimensional wireframe model by creating at least one closed polygon representing a roof face from the two-dimensional wireframe model of the roof and assigning a three-dimensional orientation to the at least one closed polygon surface based on at least one roof constraint associated with the roof face.
  • 13. The non-transitory, computer-readable medium of claim 12, further comprising computer-readable instructions for causing the processor to perform the steps of generating and displaying a graphical user interface for displaying the at least one aerial image of the building.
  • 14. The non-transitory computer-readable medium of claim 12, further comprising computer-readable instructions for causing the processor to perform the steps of generating a report comprising construction information associated with the generated three-dimensional wireframe model of the roof, the construction information being indicative of one or more of roof dimensions, pitches, surfaces, areas, and material costs associated with the roof.
  • 15. The non-transitory computer-readable medium of claim 12, further comprising computer-readable instructions for causing the processor to perform the steps of cropping the at least one aerial image around the building.
  • 16. The non-transitory computer-readable medium of claim 12, further comprising allowing a user to adjust one or more of the automatically identified plurality of perimeter and interior lines via a graphical user interface.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of, and claims the benefit of priority to, U.S. patent application Ser. No. 16/709,112 filed on Dec. 10, 2019, which is a continuation of U.S. patent application Ser. No. 15/358,870 filed on Nov. 22, 2016, now U.S. Pat. No. 10,503,842, issued on Dec. 10, 2019, which is a continuation of U.S. patent application Ser. No. 13/397,325 filed on Feb. 15, 2012, now U.S. Pat. No. 9,501,700, issued on Nov. 22, 2016, the entire disclosures of which are expressly incorporated herein by reference.

US Referenced Citations (272)
Number Name Date Kind
3908281 Fox Sep 1975 A
4845643 Clapp Jul 1989 A
5247356 Ciampa Sep 1993 A
5259037 Plunk Nov 1993 A
5276866 Paolini Jan 1994 A
5422989 Bell et al. Jun 1995 A
5592375 Salmon et al. Jan 1997 A
5633995 McClain May 1997 A
5666441 Rao et al. Sep 1997 A
5727138 Harada Mar 1998 A
5983010 Murdock et al. Nov 1999 A
6037945 Loveland Mar 2000 A
6046745 Moriya et al. Apr 2000 A
6134338 Solberg et al. Oct 2000 A
6198431 Gibson Mar 2001 B1
6323885 Wiese Nov 2001 B1
6333749 Reinhardt et al. Dec 2001 B1
6342884 Kamen et al. Jan 2002 B1
6356280 Kamen et al. Mar 2002 B1
6385541 Blumberg et al. May 2002 B1
6396491 Watanabe et al. May 2002 B2
6434277 Kamada et al. Aug 2002 B1
6446053 Elliott Sep 2002 B1
6448964 Isaacs et al. Sep 2002 B1
6456287 Kamen et al. Sep 2002 B1
6496184 Freeman et al. Dec 2002 B1
6525728 Kamen et al. Feb 2003 B2
6556195 Totsuka et al. Apr 2003 B1
6581045 Watson Jun 2003 B1
6636803 Hartz, Jr. et al. Oct 2003 B1
6810383 Loveland Oct 2004 B1
6816819 Loveland Nov 2004 B1
6826539 Loveland Nov 2004 B2
6829584 Loveland Dec 2004 B2
6836270 Du Dec 2004 B2
6912293 Korobkin Jun 2005 B1
6980690 Taylor et al. Dec 2005 B1
6982712 Onto Jan 2006 B2
7003400 Bryant Feb 2006 B2
7006977 Attra et al. Feb 2006 B1
7098909 Hayano et al. Aug 2006 B2
7133551 Chen et al. Nov 2006 B2
7149346 Oniyama Dec 2006 B2
7164883 Rappaport et al. Jan 2007 B2
7187452 Jupp et al. Mar 2007 B2
7246044 Imamura et al. Jul 2007 B2
7305983 Meder et al. Dec 2007 B1
7324666 Zoken et al. Jan 2008 B2
7343268 Kishikawa Mar 2008 B2
7376284 Tao et al. May 2008 B2
7386164 Shragai et al. Jun 2008 B2
7421125 Rees Sep 2008 B1
7424133 Schultz et al. Sep 2008 B2
7444013 Chen Oct 2008 B2
7487114 Florance et al. Feb 2009 B2
7508977 Lyons et al. Mar 2009 B2
7509241 Guo et al. Mar 2009 B2
7515153 Jin et al. Apr 2009 B2
7519206 Mulet-Parada et al. Apr 2009 B2
7720276 Korobkin May 2010 B1
7728833 Verma et al. Jun 2010 B2
7752018 Rahmes et al. Jul 2010 B2
7787659 Schultz et al. Aug 2010 B2
7804996 Ohtomo et al. Sep 2010 B2
7869981 Pendyala et al. Jan 2011 B2
7873238 Schultz et al. Jan 2011 B2
7920963 Jouline et al. Apr 2011 B2
7961982 Sibiryakov et al. Jun 2011 B2
7991226 Schultz et al. Aug 2011 B2
7995799 Schultz et al. Aug 2011 B2
7995862 Tao et al. Aug 2011 B2
8040343 Kikuchi et al. Oct 2011 B2
8059888 Chen et al. Nov 2011 B2
8068643 Schultz et al. Nov 2011 B2
8078396 Meadow et al. Dec 2011 B2
8078436 Pershing et al. Dec 2011 B2
8081841 Schultz et al. Dec 2011 B2
8099264 Kelley et al. Jan 2012 B2
8131514 Royan et al. Mar 2012 B2
8145578 Pershing et al. Mar 2012 B2
8154633 Gloudemans et al. Apr 2012 B2
8170840 Pershing May 2012 B2
8204341 Schultz et al. Jun 2012 B2
8207964 Meadow et al. Jun 2012 B1
8209152 Pershing Jun 2012 B2
8233666 Schultz et al. Jul 2012 B2
8275194 Zebedin Sep 2012 B2
8331654 Abraham et al. Dec 2012 B2
8385672 Giuffrida et al. Feb 2013 B2
8390617 Reinhardt Mar 2013 B1
8401222 Thornberry et al. Mar 2013 B2
8452125 Schultz et al. May 2013 B2
8477190 Giuffrida et al. Jul 2013 B2
8515125 Thornberry et al. Aug 2013 B2
8515198 Giuffrida et al. Aug 2013 B2
8520079 Schultz et al. Aug 2013 B2
8531472 Freund et al. Sep 2013 B2
8542880 Thornberry et al. Sep 2013 B2
8588547 Giuffrida et al. Nov 2013 B2
8593518 Schultz et al. Nov 2013 B2
8630510 Giuffrida et al. Jan 2014 B2
8634594 Schultz et al. Jan 2014 B2
8634597 Ivanov et al. Jan 2014 B2
8643720 Schultz et al. Feb 2014 B2
8648872 Freund et al. Feb 2014 B2
8649596 Schultz et al. Feb 2014 B2
8660382 Schultz et al. Feb 2014 B2
8670961 Pershing Mar 2014 B2
8731234 Ciarcia et al. May 2014 B1
8774525 Pershing Jul 2014 B2
8818076 Shenkar et al. Aug 2014 B2
8818770 Pershing Aug 2014 B2
8823732 Adams et al. Sep 2014 B2
8825454 Pershing Sep 2014 B2
8855442 Owechko Oct 2014 B2
8897539 Stone et al. Nov 2014 B2
8938090 Thornberry et al. Jan 2015 B2
8970615 Freund et al. Mar 2015 B2
8971624 Schultz et al. Mar 2015 B2
8977520 Stephens et al. Mar 2015 B2
8995757 Ciarcia et al. Mar 2015 B1
9014415 Chen et al. Apr 2015 B2
9036861 Chen et al. May 2015 B2
9047688 Lynch Jun 2015 B2
9070018 Ciarcia et al. Jun 2015 B1
9129376 Pershing Sep 2015 B2
9135737 Pershing Sep 2015 B2
9141880 Ciarcia Sep 2015 B2
9147276 Giuffrida et al. Sep 2015 B2
9147287 Ciarcia Sep 2015 B2
9159164 Ciarcia Oct 2015 B2
9182657 Schultz et al. Nov 2015 B2
9183538 Thornberry et al. Nov 2015 B2
9460517 Fathi et al. Oct 2016 B2
9501700 Loveland et al. Nov 2016 B2
9679227 Taylor et al. Jun 2017 B2
9886774 Fathi et al. Feb 2018 B2
9904867 Fathi et al. Feb 2018 B2
9911228 Pershing Mar 2018 B2
10032310 Fathi et al. Jul 2018 B2
10140756 Van Der Zwan et al. Nov 2018 B2
10503842 Loveland et al. Dec 2019 B2
10540577 Taylor et al. Jan 2020 B2
10592765 Fathi et al. Mar 2020 B2
10896353 Taylor et al. Jan 2021 B2
11094113 Mundy et al. Aug 2021 B2
11144795 Taylor et al. Oct 2021 B2
11210433 Loveland et al. Dec 2021 B2
20010027404 Loveland Oct 2001 A1
20020061132 Furukawa May 2002 A1
20020076098 Love Jun 2002 A1
20020154174 Redlich et al. Oct 2002 A1
20020167515 Kamen et al. Nov 2002 A1
20030014224 Guo et al. Jan 2003 A1
20030023412 Rappaport et al. Jan 2003 A1
20030028393 Coulston et al. Feb 2003 A1
20030088362 Melero et al. May 2003 A1
20030115163 Moore et al. Jun 2003 A1
20030147553 Chen et al. Aug 2003 A1
20030171957 Watrous Sep 2003 A1
20030233310 Stavrovski Dec 2003 A1
20040047498 Mulet-Parada et al. Mar 2004 A1
20040105573 Neumann et al. Jun 2004 A1
20040220906 Gargi et al. Nov 2004 A1
20040263514 Jin et al. Dec 2004 A1
20040264763 Mas et al. Dec 2004 A1
20050012742 Royan Jan 2005 A1
20050102394 Loveland May 2005 A1
20050203768 Florance et al. Sep 2005 A1
20050288959 Eraker et al. Dec 2005 A1
20060056732 Holmes Mar 2006 A1
20060061566 Verma et al. Mar 2006 A1
20060136126 Coombes et al. Jun 2006 A1
20060137736 Nishitani et al. Jun 2006 A1
20060188143 Strassenburg-Kleciak Aug 2006 A1
20060200311 Arutunian et al. Sep 2006 A1
20060232605 Imamura Oct 2006 A1
20060239537 Shragai et al. Oct 2006 A1
20060262112 Shimada Nov 2006 A1
20060265287 Kubo Nov 2006 A1
20070036467 Coleman et al. Feb 2007 A1
20070067191 Loveland Mar 2007 A1
20070115284 Kim et al. May 2007 A1
20070150366 Yahiro et al. Jun 2007 A1
20070220174 Abhyanker Sep 2007 A1
20080021683 Rahmes et al. Jan 2008 A1
20080068379 Larsen et al. Mar 2008 A1
20080071604 Scanlan Mar 2008 A1
20080089610 Tao et al. Apr 2008 A1
20080103991 Moore et al. May 2008 A1
20080105045 Woro May 2008 A1
20080162380 Suga et al. Jul 2008 A1
20080204570 Schultz et al. Aug 2008 A1
20080221843 Shenkar et al. Sep 2008 A1
20080231700 Schultz et al. Sep 2008 A1
20080262789 Pershing et al. Oct 2008 A1
20080273753 Giuffrida et al. Nov 2008 A1
20080279447 Friedlander et al. Nov 2008 A1
20080298638 Miyazaki Dec 2008 A1
20080310756 Tao et al. Dec 2008 A1
20090089018 Kelley et al. Apr 2009 A1
20090110327 Chen et al. Apr 2009 A1
20090132210 Royan et al. May 2009 A1
20090132436 Pershing et al. May 2009 A1
20090141020 Freund et al. Jun 2009 A1
20090216501 Yeow et al. Aug 2009 A1
20090234692 Powell et al. Sep 2009 A1
20090271154 Coad et al. Oct 2009 A1
20090304227 Kennedy et al. Dec 2009 A1
20090310867 Matei et al. Dec 2009 A1
20100034483 Giuffrida et al. Feb 2010 A1
20100060631 Sugihara Mar 2010 A1
20100110074 Pershing May 2010 A1
20100114537 Pershing May 2010 A1
20100164953 Wouhaybi et al. Jul 2010 A1
20100179787 Pershing et al. Jul 2010 A2
20100182316 Akbari et al. Jul 2010 A1
20100201682 Quan et al. Aug 2010 A1
20100217724 Wayne et al. Aug 2010 A1
20100241406 Rahmes et al. Sep 2010 A1
20100275018 Pedersen Oct 2010 A1
20100296693 Thornberry et al. Nov 2010 A1
20100303340 Abraham et al. Dec 2010 A1
20110047048 Yahiro et al. Feb 2011 A1
20110096083 Schultz Apr 2011 A1
20110157213 Takeyama et al. Jun 2011 A1
20110164029 King et al. Jul 2011 A1
20110187713 Pershing et al. Aug 2011 A1
20110205245 Kennedy et al. Aug 2011 A1
20110222757 Yeatman, Jr. et al. Sep 2011 A1
20120007982 Giuffrida et al. Jan 2012 A1
20120026322 Malka et al. Feb 2012 A1
20120101783 Stephens et al. Apr 2012 A1
20120154446 Adams et al. Jun 2012 A1
20120170797 Pershing et al. Jul 2012 A1
20120179431 Labrie Jul 2012 A1
20120183217 Schultz et al. Jul 2012 A1
20120191424 Pershing Jul 2012 A1
20120209782 Pershing et al. Aug 2012 A1
20120223965 Pershing Sep 2012 A1
20120253725 Malka et al. Oct 2012 A1
20120253751 Malka et al. Oct 2012 A1
20120288158 Schultz et al. Nov 2012 A1
20130113831 Giuffrida et al. May 2013 A1
20130135471 Giuffrida et al. May 2013 A1
20130138401 Thornberry et al. May 2013 A1
20130170694 Thornberry et al. Jul 2013 A1
20130202157 Pershing Aug 2013 A1
20130204575 Pershing Aug 2013 A1
20130208116 Schultz et al. Aug 2013 A1
20130208996 Schultz et al. Aug 2013 A1
20130211790 Loveland et al. Aug 2013 A1
20130212536 Thornberry et al. Aug 2013 A1
20130226515 Pershing et al. Aug 2013 A1
20140064554 Coulter et al. Mar 2014 A1
20150347872 Taylor et al. Dec 2015 A1
20150370929 Pershing Dec 2015 A1
20160239976 Fathi et al. Aug 2016 A1
20160343140 Ciprari et al. Nov 2016 A1
20170154131 Loveland et al. Jun 2017 A1
20170277980 Taylor et al. Sep 2017 A1
20170353708 Petrichkovich et al. Dec 2017 A1
20180322698 Fathi et al. Nov 2018 A1
20190095694 Waizenegger et al. Mar 2019 A1
20190220711 Taylor et al. Jul 2019 A1
20190236839 Fathi et al. Aug 2019 A1
20200082168 Fathi et al. Mar 2020 A1
20200110847 Loveland et al. Apr 2020 A1
20200202183 Taylor et al. Jun 2020 A1
20210174580 Mundy et al. Jun 2021 A1
20210407188 Mundy et al. Dec 2021 A1
20220027622 Taylor et al. Jan 2022 A1
Foreign Referenced Citations (27)
Number Date Country
2008230031 Jul 2010 AU
2191954 Dec 1995 CA
2703423 Nov 2010 CA
4419359 Dec 1995 DE
19719620 Nov 1998 DE
19857667 Aug 2000 DE
1010966 Jun 2000 EP
00029806 May 2000 WO
2004044692 May 2004 WO
2005124276 Dec 2005 WO
2006040775 Apr 2006 WO
2006090132 Aug 2006 WO
2009049151 Apr 2009 WO
2009073726 Jun 2009 WO
2010017255 Feb 2010 WO
2011056402 May 2011 WO
2011094760 Aug 2011 WO
2012050648 Apr 2012 WO
2012054239 Apr 2012 WO
2012083135 Jun 2012 WO
2013116164 Aug 2013 WO
2013116165 Aug 2013 WO
2013116793 Aug 2013 WO
2013116794 Aug 2013 WO
2014149509 Sep 2014 WO
2014151122 Sep 2014 WO
2015081026 Jun 2015 WO
Non-Patent Literature Citations (280)
Entry
Office Action dated Mar. 19, 2019, issued in connection with U.S. Appl. No. 15/621,458 (10 pages).
European Search Report dated Aug. 9, 2019, issued by the European Patent Office in connection with European Patent Application No. 19171426.0 (9 pages).
Notice of Allowance dated Sep. 23, 2019, issued in connection with U.S. Appl. No. 15/621,458 (10 pages).
Hirschmuller, “Stereo Processing by Semi-Global Matching and Mutual Information,” IEEE transactions on Pattern Analysis and Machine Intelligence, vol. 30, 2, Dec. 2007 (14 pages).
Census Transform—<https://en.wikipedia.org/wiki/Census_transform>—webpage last edited on Oct. 24, 2019 (2 pages).
Dynamic Programming—<https://en.wikipedia.org/wiki/Dynamic_programming>—webpage last edited on Feb. 24, 2020 (17 pages).
Delaunay Triangulation—<https://en.wikipedia.org/wiki/Delaunay_triangulation>—webpage last edited on Feb. 13, 2020 (8 pages).
Office Action dated Feb. 27, 2020, issued by the Canadian Intellectual Property Office in connection with Canadian Patent Application No. 2,864,831 (3 pages).
Examination Report dated Mar. 17, 2020, issued by the Australian Patent Office in connection with Australian Patent Application No. 2019200259 (4 pages).
Office Action dated Sep. 15, 2020, issued in connection with U.S. Appl. No. 16/703,644 (19 pages).
Notice of Allowance dated Sep. 9, 2020, issued in connection with U.S. Appl. No. 16/365,847 (13 pages).
Office Action dated Sep. 15, 2020, issued by the Canadian Patent Office in connection with Canadian Patent Application No. 2,920,251 (4 pages).
Office Action dated Nov. 10, 2020, issued in connection with U.S. Appl. No. 16/748,174 (8 pages).
International Search Report of the International Searching Authority dated Jan. 29, 2021, issued in connection with International Application No. PCT/US2020/063004 (3 pages).
Written Opinion of the International Searching Authority dated Jan. 29, 2021, issued in connection with International Application No. PCT/US2020/063004 (7 pages).
Office Action dated Feb. 8, 2021, issued in connection with U.S. Appl. No. 16/709,112 (16 pages).
Notice of Allowance dated Apr. 6, 2021, issued in connection with U.S. Appl. No. 16/703,644 (9 pages).
Office Action dated Apr. 16, 2021, issued by the Canadian Patent Office in connection with Canadian Patent Application No. 2,864,831 (7 pages).
Office Action dated Jul. 6, 2021, issued by the Canadian Patent Office in connection with Canadian Patent Application No. 2,920,251 (3 pages).
Notice of Allowance dated Aug. 4, 2021, issued in connection with U.S. Appl. No. 16/748,174 (8 pages).
Notice of Allowance dated Aug. 19, 2021, issued in connection with U.S. Appl. No. 16/709,112 (5 pages).
Examination Report dated Apr. 4, 2022, issued by the Australian Patent Office in connection with Australian Patent Application No. 2021201579 (4 pages).
Office Action dated Apr. 22, 2022, issued in connection with U.S. Appl. No. 17/403,792 (13 pages).
Pictometry Visual Intelligence, “Pictometry—In the News, Pictometry Announces Software and Web-based Solution for Engineers, Architects, and Planners”, 2009 (3 pages).
Pictometry Visual Intelligence, “Frequently Asked Questions”, 2005 (6 pages).
Pictometry Visual Intelligence, http://web.archive.org/web/20020725232638/http://www.pictometry.com, 1995-2002 (2 pages).
Porway, et al., “A Hierarchical and Contextual Model for Aerial Image Parsing,” 2008 (53 pages).
Poullis, et al., “Photogrammetric Modeling and Image-based Rendering for Rapid Virtual Environment Creation”, 1998 (7 pages).
PrecigeoRoof, “Why precigeoRoof”, Jan. 7, 2007 (2 pages).
Zheng, et al., “A Consistent Segmentation Approach to Image-based Rendering”, Technical Report CSE-09-03-02, 2002 (8 pages).
Preciozzi, “Dense Urban Elevation Models From Stereo Images by an Affine Region Merging Approach,” Master's Thesis, Universidad de la Republica, Montevideo, Sep. 18, 2006 (93 pages).
Reddy, et al., “Frequency-Space Decomposition and Acquisition of Light Transport Under Spatially Varying Illumination”, EECV 2012 (14 pages).
RoofCAD, “Satellite Takeoff Tutorial—Pitched Roof”, 2012 (25 pages).
RoofCAD, “User Guide”, True North Estimating Systems, Ltd., 2003 (320 pages).
Rottensteiner, et al., “Automatic Generation of Building Models from Lidar Data and the Integration of Aerial Images,” ISPRS, vol. XXXIV, 2003 (7 pages).
Rupnik, et al., “Oblique Multi-Camera Systems—Orientation and Dense Matching Issues”, The International Archives of teh Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. XL-3W1, 2014 (8 pages).
San, et al., “Building Extraction from High Resolution Satellite Images Using Hough Transform,” International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, vol. XXXVIII, 2010 (6 pages).
Scholze, et al., “A Probabilistic Approach to Building Roof Reconstruction Using Semantic Labelling”, 2002 (8 pages).
Seitz, et al., “A Comparison and Evaluation of Multi-View Stereo Reconstruction Algorithms”, CVPR 2006 (8 pages).
Shan, et al., “Refractive Height Fields from Single and Multiple Images”, 2012 (8 pages).
“Sorcerer: Nobody builds roofs like this builds roofs”, retrieved from URL=http://web.archive.org/web/2006021409237/http://www.applicad.com/au/product-features . . . on Mar. 29, 2012 (2 pages).
Zheng, et al. “Parallax Photography: Creating 3D Cinematic Effects from Stills”, 2009 (8 pages).
Syed, et al., “Semi-Automatic 3D Building Model Generation From Lidar and High Resolution Imagery,” Proceedings of SSC Spatial Intelligence, Sep. 2005 (8 pages).
“Transcription of points of potential interest in the attached Youtube video titled: Pictometry Online Demo”, retrieved on Feb. 10, 2010 (1 page).
Taillandier, et al., “Automatic Building Reconstruction from Aerial Images: A Generic Bayesian Framework”, 2004 (6 pages).
Ulm, et al., “3D City Modelling with Cybercity-Modeler”, 1st EARSel workshop of the SIG Urban Remote Sensing, Mar. 2-3, 2006 (9 pages).
University of Washington, “College of Arts & Sciences Mathematics: Detailed course offerings . . . ”, retrieved from http://www.washington.edu/students/crscat/math.html on Oct. 25, 2013 (16 pages).
Verma, “3D Building Detection and Modeling from Aerial LIDAR Data,” IEEE, 2006 (8 pages).
Vosselman, “Fusion of Laser Scanning Data, Maps, and Aerial Photographs for Building Reconstruction”, 2002 (4 pages).
Vosselman, et al. “Map Based Building Reconstruction from Laser Data and Images”, 2001 (9 pages).
Vosselman, et al., “Mapping by Dragging and Fitting of Wire-Frame Models”, Photogrammetric Engineering and Remote Sensing, Jul. 1999 (8 pages).
Wang, et al., “Pictometry's Proprietary Airborne Digital Imaging System and It's Application In 3D City Modelling”, 2008 (6 pages).
Wattenberg, et al., “Area, Volume and Torque in Three Dimensions”, retrieved from http://www.math.montana.edu/frankw/ccp/multiworld/twothree/atv/learn.htm on Sep. 24, 2013 (14 pages).
Weeks, et al., “A Real Time, Multichannel System with Parallel Digital Signal Processors”, IEEE, 1990 (4 pages).
Werner, et al., “New Techniques for Automated Architectural Reconstruction from Photographs,” Department of Engineering Science, University of Oxford, 2002 (15 pages).
Wolf, Elements of Photogrammetry—Chapter 14: Aerotriangulation, 1974 (3 pages).
Wood, et al., “Surface Light Fields for 3D Photography”, SIGGRAPH 2000 (10 pages).
Zhang, et al., “Spacetime Stereo: Shape Recovery for Dynamic Scenes”, 2003 (8 pages).
Written Opinion of the International Searching Authority dated Nov. 17, 2014, issued in connection with International Application No. PCT/US14/49605 (4 pages).
Wu, et al., “Multicore Bundle Adjustment”, 2011 (8 pages).
Wu, et al., “Schematic Surface Reconstruction”, 2012 (8 pages).
www.archive.org, “Main Features: Photomodeler is Fully Loaded and Ready to Perform”, retrieved from http://www.photomodeler.com/pmpro08.html on Oct. 21, 2013 (4 pages).
Xactware Solutions, Inc., “Xactimate Sketch—Import Underlay Image,” 2008 (4 pages).
Xactware, “Roof and Property Insight”, 2015 (10 pages).
Xiao, et al., “Geo-spatial Aerial Video Processing for Scene Understanding and Object Tracking,” IEEE, 2008 (8 pages).
Ye, et al., “Automated Reconstruction of Urban House Roofs from Aerial Imagery”, IEEE 2001 (3 pages).
YouTube, “Pictometry Online Demo”, retrieved Feb. 6, 2009 (1 page).
Zhang, et al., “Rapid Shape Acquisition Using Color Structured Light and Multi-pass Dynamic Programming”, 2002 (13 pages).
Zhang, et al., “Shape and Motion Under Varying Illumination: Unifying Structure from Motion, Photometric Stereo, and Multi-view Stereo”, 2003. (8 pages).
Notice of Allowance dated Nov. 14, 2016, issued in connection with U.S. Appl. No. 14/450,861 (11 pages).
Notice of Allowance dated Jul. 25, 2016, issued in connection with U.S. Appl. No. 14/450,861 (12 pages).
Office Action dated Nov. 5, 2015, issued in connection with U.S. Appl. No. 14/450,861 (14 pages).
European Search Report dated Mar. 14, 2017, issued by the European Patent Office in connection with European Patent Application No. 14832467.6 (6 pages).
Applicad Roofing, sample report, Jul. 30, 2007 (1 page).
Applicad Roofing, sample report, Mar. 2, 2005 (28 pages).
AppliCad USA, “Linear Nesting Reports,” AppliCad Sample Reports, Nov. 25, 1999 (10 pages).
Applicad webpage 2005 snip different color lines (1 page).
Applicad, “Example Output and Brochures,” retrieved from URL=http://www.applicad.com/auiproduct-reports.html, Apr. 16, 2012 (2 pages).
AppliCad, “Product Overview—Sorcerer: Advanced Software for Roofing Modeling, Estimating, Presentation and nstallation,” Issue 5, Mar. 2, 2001 (13 pages).
AppliCad, “Roofing Software: Product Bulletin Section 1—Modeling the Roof,” Dec. 20, 2005 (3 pages).
AppliCad, “Roofing Software: Product Bulletin Section 1—Modeling the Roof,” Jan. 7, 2002 (3 pages).
AppliCad, “Roofing Software: Product Bulletin Section 2—Modifying the Model,” Dec. 20, 2005 (2 pages).
AppliCad, “RoofScape: Advanced Software for Roof Modelling and Estimating,” Learning Guide (English Units) Revision 1.1, Aug. 23, 2007 (48 pages).
AppliCad, “Tips and Tricks: Items drawn from AppliCad's Customer Service file”, Jul. 27, 2007 (11 pages).
Autodesk, “Autodesk ImageModeler-Features”, http://usa.autodesk.com/adsk/servlet/index?siteID=123112&id=115639 . . . , 2008 (1 page).
Automatic House Reconstruction, retrieved on Sep. 29, 2008, from http://www.vision.ee.ethz.ch/projects/Amobe_I/recons.html (6 pages).
Avrahami, et al., “Extraction of 3D Spatial Polygons Based on the Overlapping Criterion for Roof Extraction from Aerial Images”, International Archives of Photogrammetry, Remote Sensing & Spatial Information Sciences, Aug. 29-30, 2005 (6 pages).
Azuma, et al., “View-dependent Refinement of Multiresolution Meshes with Subdivision Connectivity”, Feb. 2003 (9 pages).
“8 Epipolar Geometry and the Fundamental Matrix”, retrieved Oct. 25, 2013 (25 pages).
Baillard, et al., “3-D Reconstruction of Urban Scenes from Aerial Stereo Imagery: A Focusing Strategy,” Computer Vision and Image Understanding, vol. 76, No. 3 pp. 244-258, Dec. 1999 (15 pages).
Baillard, et al., “Automatic Reconstruction of Piecewise Planar Models from Multiple Views”, 1999 (7 pages).
Bazaraa, et al., “Nonlinear Programming Theory and Algorithms”, Second Edition, John Wiley & Sons, Inc., New York, 1993 (646 pages).
Behley, et al., “Generation of 3D City Models using Domain-Specific Information Fusion”, Institute of Computer Science III, 2009 (10 pages).
Bernhardsen, “Geographic Information Systems, An Introduction,” 2nd Ed., 1999 (4 pages).
Bertan, et al., “Automatic 3D Roof Reconstruction Using Digital Cadastral Map, Architectural Knowledge and an Aerial Image,” 2006 (4 pages).
Bhat, et al., “A Perceptually-Motivated Optimization-Framework for Image and Video Processing”, 2008 (10 pages).
Bhat, et al., “Fourier Analysis of the 2D Screened Poisson Equation for Gradient Domain Problems”, ECCV 2008 (14 pages).
Bhat, et al., “GradientShop: A Gradient-Domain Optimization Framework for Image and Video Filtering”, 2010 (14 pages).
Bhat, et al., “Piecewise Image Registration in the Presence of Multiple Large Motions”, Jun. 2006 (7 pages).
Bhat, et al., “Using Photographs to Enhance Videos of a Static Scene”, Eurographics Symposium on Rendering, 2007 (12 pages).
Bignone, et al., “Automatic Extraction of Generic House Roofs from High Resolution Aerial Imagery”, 1996 (12 pages).
Brofferio, et al., “Interactive Detection of 3D Models of Building's Roofing for the Estimation of the Solar Energy Potential,” Sep. 4-8, 2006 (5 pages).
Burrough, et al., “Principles of Geographical Information Systems”, Spatial Information Systems and Geostatistics, 1998 (14 pages).
Capell, et al., “A Multiresolution Framework for Dynamic Deformations”, SIGGRAPH 2002 (8 pages).
Chen, et al., “Building Reconstruction from LIDAR Data and Aerial Imagery”, IEEE 2005 (4 pages).
Chen, et al., “Fusion of LIDAR Data and Optical Imagery for Building Modeling”, 2004 (6 pages).
Chen, et al., “Reconstruction of Building Models with Curvilinear Boundaries from Laser Scanner and Aerial Imagery”, 2006 (10 pages).
Chevrier, et al., “Interactive 3D Reconstruction for Urban Areas: An image based tool”, 2001 (13 pages).
Chikomo, et al., “An Integrated Approach to Level-of-Detail Building Extraction and Modelling Using Airborne LIDAR and Optical Imagery”, Sep. 19-21, 2007 (6 pages).
Chuang, et al., “A Bayesian Approach to Digital Matting”, IEEE 2001 (8 pages).
Chuang, et al., “Animating Pictures with Stochastic Motion Textures”, SIGGRAPH, 2005 (8 pages).
Chuang, et al., “Animating Pictures with Stochastic Motion Textures”, Technical Report UW-CSE-04-04-02, 2005 (7 pages).
Chuang, et al., “Environment Matting Extensions: Towards Higher Accuracy and Real-Time Capture”, SIGGRAPH 2000 (10 pages).
Chuang, et al., “Shadow Matting and Compositing”, SIGGRAPH 2003 (7 pages).
Clarke, “Getting Started with Geographic Information Systems,” Geographic Information Science, 2nd Ed., 1999 (7 pages).
Colburn, et al., “Image-Based Remodeling”, IEEE Transactions on Visualization and Computer Graphics, vol. 19, No. 1, 2012 (13 pages).
Collins, et al., “Automatic Extraction of Buildings and Terrain from Aerial Images”, Department of Computer Science, University of Massachusetts, 1995 (10 pages).
Collins, et al., “UMass Progress in 3D Building Model Acquisition”, 1996 (11 pages).
Notice of Allowance dated May 20, 2016, issued in connection with U.S. Appl. No. 13/397,325 (7 pages).
Notice of Allowance dated Sep. 20, 2016, issued in connection with U.S. Appl. No. 13/397,325 (5 pages).
Cord, et al., “Bayesian Model Identification: Application to Building Reconstruction in Aerial Imagery”, IEEE 1999 (5 pages).
Croitoru, et al., “Right-Angle Reooftop Polygon Extraction in Regularised Urban Areas: Cutting the Corners,” Technion-Israel Institute of Technology, Dec. 2004 (31 pages).
Curless, “From Range Scans to 3D Models” SIGGRAPH Computer Graphics, 1999 (8 pages).
Jaynes, et al., “Recognition and Reconstruction of Buildings from Multiple Aerial Images,” Oct. 18, 2001 (37 pages).
Johnson, et al., Surface Matching for Object Recognition in Complex 3-D Scenes, 1998 (31 pages).
Khoshelham, et al., “A Model-Based Approach to Semi-Automated Reconstruction of Buildings from Aerial Images”, The Photogrammetric Record, Dec. 2004 (18 pages).
Kolbl, et al., “Chapter 2: Scanning and State-of-the-Art Scanners”. Digital Photogrammetry: An Addendum to the Manual of Photogrammetry, 1996 (37 pages).
Kolman, “Elementary Linear Algebra: Chapter 4, Linear Transformations and Matrices”, Second Edition, Macmillan Publishing Co., 1997 (12 pages).
Korte, “The GIS Book: Understanding the Value and Implementation of Geographic Information Systems”, 4th Ed., 1997 (14 pages).
Krainin, et al., “Autonomous Generation of Complete 3D Object Models Using Next Best View Manipulation Planning”, CRA 2011 (7 pages).
Kushal, et al., “Photo Tours”, 3DimPVT, Oct. 2012 (8 pages).
Labe, et al., “Robust Techniques for Estimating Parameters of 3D Building Primitives”, International Society for Photogrammetry and Remote Sensing, vol. XXXII, Part 2, Commission II, Proceedings of the Commission II Symposium, Data Integration: Systems and Techniques, Jul. 13-17, 1998 (11 pages).
Lee, et al., “Fusion of Lidar and Imagery for Reliable Building Extraction”, Photogrammetric Engineering and Remote Sensing, Feb. 2008 (11 pages).
Levoy, “The Digital Michelangelo Project”, retrieved from http://www-graphics.stanford.edu/projects/mich/ on Oct. 25, 2013 (10 pages).
Levoy, et al., “The Digital Michelangelo Project: 3D Scanning of Large Statues”, 2000 (14 pages).
LexisNexis, “Software; New Products”, Roofing Contractor, Jan. 3, 2006 (1 page).
Li, et al., “Automated Generation of Interactive 3D Exploded View Diagrams” SIGGRAPH 2007 (7 pages).
Li, et al., “Interactive Cutaway Illustrations of Complex 3D Models”, ACM Transactions on Graphics 26(3), SIGGRAPHY, 2007 (11 pages).
Liu, et al., “Building Extraction from High Resolution Satellite Imagery Based on Multi-scale Image Segmentation and Model Matching”, IEEE 2008 (7 pages).
Lu, et al., “Automatic Building Detection Using the Dempster-Shafer Algorithm,” Photogrammetric Engineering & Remote Sensing, vol. 72, No. 4, Apr. 2006 (9 pages).
Ziegler, et al., “3D Reconstruction Using Labeled Image Regions”, 2003 (12 pages).
Lu, et al., “Stereo Image Matching Using Robust Estimation and Image Analysis Techniques for Dem Generation,” International Archives of Photogrammetry and Remote Sensing, vol. XXXIII, Part B3, Amsterdam 2000 (8 pages).
Lueders, “Infringement Allegations by Eagleview Technologies”, Feb. 10, 2009 (3 pages).
Mahajan, et al., “A Theory of Frequency Domain Invariants: Spherical Harmonic Identities for BRDF/Lighting Transfer and Image Consistency”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, No. 2, Feb. 2008 (14 pages).
Mahajan, et al., “A Theory of Spherical Harmonic Identities for BRDF/Lighting Transfer and Image Consistency”, ECCV 2006 (15 pages).
Maini, et al., “Study and Comparison of Various Image Edge Detection Techniques”, International Journal of Image Processing, vol. 3: Issue 1, 2009 (12 pages).
Mann, “Roof with a view”, Contract Journal, Nov. 23, 2005 (2 pages).
Maune, Chapter 6: DEM Extraction, Editing, Matching and Quality Control Techniques. Digital Photogrammetry: An Addendum to the Manual of Photogrammetry, 1996 (27 pages).
McGlone, “Chapters: Aerial Triangulation Adjustment and Image Registration,” Digital Photogrammetry: An Addendum to the Manual of Photogrammetry, 1996 (25 pages).
McGlone, “Sensor Modeling in Image Registration, Chapter 5:Aerial Triangulation Adjustment and Image Registration”, 1996 (9 pages).
McGlone, et al., “Projective and Object Space Geometry for Monocular Building Extraction,” School of Computer Science, Carnegie Mellon University, Jun. 20-23, 1994 (31 pages).
McKeown, Jr., et al., “Chapter 9: Feature Extraction and Object Recognition, Automatic Cartographic Feature Extraction Using Photogrammetric Principles”. Digital Photogrammetry: An Addendum to the Manual of Photogrammetry, 1996 (19 pages).
Meixner, et al., 3-Dimensional Building Details from Aerial Photography for Internet Maps, Institute for Computer Graphics and Vision, Apr. 8, 2011 (27 pages).
Mikhail, et al., “Introduction to Modern Photogrammetry”, John Wiley&Sons, Inc., New York, 2001 (487 pages).
Mikuni, “Chapter 7: Digital Orthophotos: Production, Mosaicking, and Hardcopy”. Digital Photogrammetry: An Addendum to the Manual of Photogrammetry, 1996 (11 pages).
Miller, “Pictometry in Arlington Virginia:Software gives small Arlington the big picture”, Dec. 2001 (2 pages).
Miller, et al., “Miller's Guide to Framing and Roofing”, McGraw Hill, New York, 2005 (9 pages).
Minialoff, “Introduction to Computer Aided Design”, Apr. 2000 (4 pages).
Moons, et al., “Automatic Modelling and 3D Reconstruction of Urban House Roofs from High Resolution Aerial Imagery”, 2006 (16 pages).
Mortensen, et al., “Intelligent Scissors for Image Composition”, Brigham Young University, 1995 (8 pages).
Mostafa, et al., “A Multi-Sensor System for Airborne Image Capture and Georeferencing,” Photogrammetric Engineering & Remote Sensing, vol. 66, No. 12, Dec. 2000 (7 pages).
Nizar, et al., “Reconstruction of Buildings from Airborne Laser Scanning Data”, 2006 (10 pages).
Noronha, et al., “Detection and Modeling of Buildings from Multiple Aerial Images”. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, No. 5, May 2001 (32 pages).
Ortner, et al., “Building Extraction from Digital Elevation Model”, INRIA, Jul. 2002 (51 pages).
Pedant, et al., “Scene Registration in Aerial Image Analysis”. Digital Mapping Laboratory, School of Computer Science, Carnegie Mellon University, Pittsburg PA, 1990 (13 pages).
Photogrammetric Engineering and Remote Sensing, “PE&RS, Journal of the American Society for Photogrammetry and Remote Sensing”, vol. 68, No. 9, Sep. 2002 (2 pages).
PhotoModeler Pro 4.0—The New Release, “The Premier Software for Measuring and Modeling the Real-World is even better!,” 1995-2000 (35 pages).
Photomodeler.com, “PhotoModeler Pro 5: Measuring and Modeling the Real World”, 2003 (2 pages).
“Pictometry Aerial Images and Electronic Field Study Software”, 2008 (2 pages).
Pictometry Intelligent Images, EFS Version 2.7 Release Notes, 2007 (30 pages).
Pictometry International Corp., “Electronic Field Study User Guide”. Version 2.7, Jul. 2007 (536 pages).
Pictometry Online, “Government”, Oct. 7, 2008 (3 pages).
Pictometry search results, researched on Sep. 23, 2013 (21 pages).
Applicad, “Product Bulletin—Nov. 2002: Key Features of Our Roofing Software,” Nov. 2002 (46 pages).
“3D Reconstruction”, retrieved Oct. 25, 2013 (5 pages).
Bhanu, et al. “Adaptive Integrated Image Segmentation and Object Recognition; IEEE Trans. Systems, Man, and Cybernetics—Part C” Applications and Reviews, vol. 30, No. 4, Nov. 2000; 427-441 (15 pages).
Choi, et al. “Vehicle Detection from Aerial Images Using Local Shape Information” PSIVT 2009, LNCS 5414; Springer-Verlag Berlin Heidelberg; 2009; pp. 227-236 (10 pages).
Collins, et al., “Site Model Acquisition and Extension from Aerial Images” ICCV '95 Proc. 5th Int. Conf. Computer Vision; 1995; pp. 1-6 (6 pages).
Kaufhold, et al., “Recognition and Segmentation of Scene Content Using Region-Based Classification” Proceedings of the 18th International Converence on Pattern Recognition (ICPR'06); 2006; pp. 1-6 (6 pages).
Moscatelli, et al., “Advanced Topics in Artificial Intelligence” Lecture Notes in Computer Science vol. 617, 1992, pp. 161-197 (37 pages).
Nobel, et al., “Histogram Classification Using Vector Quantization” Proc. IEEE Int. Symp. Information Theory; 1994; p. 391 (1 page).
Shafer, et al., “Recursive Region Segmentation by Analysis of Histograms” Proc. IEEE Int. Conf. Acoustics, Speech, Signal Processing, 1982, pp. 1166-1171 (6 pages).
Stilla, et al., “Reconstruction of Building Models from Maps and Laser Altimeter Data” Agouris and Stefanidis (Eds.): ISD'99, LNCS1737, pp. 34-46, 1999 (13 pages).
Ahonen, et al., “Rotation Invariant Image Description with Local Binary Pattern Histogram Fourier Features” in Text: Image Analysis, SCIA 2009 Proceedings, Lecture Notes in Computer Science 5575, 2009; pp. 61-70 (10 pages).
“Xactimate 27—Aerial Sketch,” Internet printout from http://www.xactware.com/solutions/claims-estimating/27-3/aerial-sketch, 2012 (3 pages).
International Search Report of the International Searching Authority dated Apr. 26, 2013, issued in connection with International Application No. PCT/US2013/26385 (2 pages).
Written Opinion dated Apr. 26, 2013, issued in connection with International Application No. PCT/US2013/26385 (4 pages).
Notice of Allowance dated Jan. 14, 2016, issued in connection with U.S. Appl. No. 13/397,325 (5 pages).
Office Action dated Oct. 21, 2014, issued in connection with U.S. Appl. No. 13/397,325 (36 pages).
Office Action dated Mar. 12, 2014, issued in connection with U.S. Appl. No. 13/397,325 (31 pages).
Office Action dated Jun. 16, 2015, issued in connection with U.S. Appl. No. 13/397,325 (16 pages).
Office Action dated Oct. 9, 2018, issued by the Canadian Intellectual Property Office issued in connection with Canadian Patent Application No. 2,864,831 (6 pages).
Office Action dated Mar. 4, 2019, issued in connection with U.S. Appl. No. 15/358,870 (7 pages).
Notice of Allowance dated Jul. 23, 2019, issued in connection with U.S. Appl. No. 15/358,870 (8 pages.
Gehrke, et al., “Semi-Global Matching: An Alternative to Lidar for DSM Generation?” published 2010 (6 pages).
A History of Roof Modelling Using Aerial Imagery, Sep. 1983 (4 pages).
Able Software Corp., “R2V User's Manual, Advanced Raster to Vector Conversion Software”. Publicly available Sep. 16, 2000 (167 pages).
AeroDach Web Site http://www.aerodach.de from Jun. 13, 2004 (retrieved Sep. 20, 2012) and translations to English (21 pages).
Aerodach, “Protokoll zur Dachauswertung”, Oct. 19, 2010 (12 pages).
Aerowest GmbH Logo, “Aerodach Online Roof Analysis: Standard Delivery Format and 3D Dataset”, 2002 (6 pages).
Aerowest GmbH, “AeroDach-das patentierte Dachaufmass”, retrieved from URL=http://web.archive.org/web/20060101021543/http://www.aerowest.de/aerodach.html, 2006 (2 pages).
Aerowest GmbH, “Aerowest Pricelist of Geodata”, AeroDach Online, Aeroview, Oct. 21, 2005 (2 pages).
Aerowest GmbH, “Geodata Service; AeroDach-Patented Roof Dimensions”, 2006 (2 pages).
Aerowest GmbH, “Preisliste Geodaten Aerowest”, Oct. 21, 2005 (1 page).
Aerowest GMBH, AeroDach Online Dachauswertung: Standardlieferformat und 3D-Datensatz, 2002 (6 pages).
Aerowest GmbH, AeroDach Online, Geodatenservice, 2005 (18 pages).
Aerowest Web Site http://aerowest.de/ from Feb. 6, 2006 (retrieved Sep. 20, 2012) and translated to English (61 pages).
Agarwal, et al., “Building Rome in a Day”, Communications of the ACM, vol. 54, No. 10, Oct. 2011 (8 pages).
Agarwal, et al., “Reconstructing Rome”, IEEE Computer Society, 2010 (8 pages).
Agarwala, et al., “Interactive Digital Photomontage”, SIGGRAPH 2004 (9 pages).
Agarwala, et al., “Panoramic Video Textures”, ACM SIGGRAPH 2005 (7 pages).
Ameri, et al., “Automatic 3D Building Reconstruction Using Plane-Roof Structures”, Institute for Photogrammetry, University of Stuttgart, 2000 (12 pages).
American Congress on Surveying and Mapping, “Definitions and Surveying and Associated Terms”, 1989 (2 pages).
American Society of Civil Engineering, “Glossary of the Mapping Sciences” ASCE Publications, 1994 (3 pages).
Appli-cad Australia, “Linear Nesting Reports,” AppliCad Australia, UK Sample Reports, Jul. 18, 2000 (9 pages).
Appli-cad Australia, “Roof Magician: Especially Suited to Shingle, Shake and Tile Roofing,” Sample Reports, Jun. 24, 2004 (13 pages).
Appli-cad Australia, “Roof Wizard: Advanced Software for Roof Modeling and Estimating,” Sep. 25, 2004 (10 pages).
Appli-cad Australia, “Roof Wizard: Especially Suited to Metal Roofing”, Mar. 9, 2005 (7 pages).
Appli-cad Australia, “Roof Wizard: Especially Suited to Metal Roofing,” Jul. 13, 2004 (24 pages).
Appli-cad Australia, “Roof Wizard: Especially Suited to Metal Roofing,” Sep. 14, 2006 (7 pages).
Appli-cad Australia, “Roof Wizard: Especially Suited to Metal Roofing,” Sep. 17, 2002 (12 pages).
Appli-cad Australia, “Sorcerer: Advanced Software for Roof Modeling and Estimating,” Reference Guide V. 3, Sep. 8, 1999 (142 pages).
Appli-cad Australia, “Sorcerer: The Complete Solution for Professional Roof Estimating,” Demonstration Kit, Mar. 9, 2005 (15 pages).
Curless et al., “New Methods for Surface Reconstruction from Range Images”, Dissertation, submitted to the Department of Electrical Engineering and the Committee of Graduate Studies of Stanford University, Jun. 1997 (209 pages).
Curless, et al., “A Volumetric Method for Building Complex Models from Range Images”, 1996 (10 pages).
Curless, et al., “Better Optical Triangulation through Spacetime Analysis”, 1995 (8 pages).
Curless, et al., “Computer model and 3D fax of Happy Buddha”, retrieved Oct. 25, 2013 (4 pages).
Debevec, et al., “Modeling and Rendering Architecture from Photographs: A hybrid geometry- and image-based approach,” University of California at Berkeley, 1996 (10 pages).
Delaney, “Searching for Clients from Above—More Small Businesspeople Use Aerial mapping Services to Scout Potential Customers”, The Wall Street Journal, Jul. 31, 2007 (4 pages).
Directions Magazine, “Microsoft MSN Virtual Earth: The Map is the Search Platform”, 2009 (10 pages).
Eagle View Tech v. Aerialogics LLC, “Prior Art Presentation”, Case No. 2:12-cv-00618-RAJ, Aug. 17, 2012 (61 pages).
Eagle View Technologies and Applicad Software, “AppliCad Software and EagleView Technologies Partner for Metal Roofing Contractors”, EagleView Blog, Feb. 4, 2011 (2 pages).
ECE 390 Introduction to Optimization, Spring 2004, retrieved Oct. 25, 2013 (1 page).
Elaksher, et al., “Roof Boundary Extraction Using Multiple Images”, Photogrammetric Record, Mar. 2003 (14 pages).
Elbernick, et al., “Adding the Third Dimension to a Topographic Database Using Airborne Laser Scanner Data”, 2006 (6 pages).
Falkner, et al., “Aerial Mapping: Methods and Applications—Chapter 11: Aerotriangulation” Second Edition, 2002 (23 pages).
Faugeras, “What Can Be Seen in Three Dimensions with an Uncalibrated Stereo Rig?”, 1992 (16 pages).
Faugeras, et al., “3-D Reconstruction of Urban Scenes from Sequences of Images”, Institut National De Recherche En Informatique Et En Automatique, 1995 (26 pages).
Federal Register, “Notices”, Geological Survey, vol. 64, No. 18, Jan. 28, 1999 (1 page).
Fisher, et al., “Dictionary of Computer Vision and Image Processing”, John Wiley&Sons, 2005 (337 pages).
Flamanc, et al., “3D City Models: An Operational Approach Using Aerial Images and Cadastral Maps”, Sep. 17-19, 2003 (6 pages).
Fritsch, “Introduction into Digital Aerotriangulation” Photogrammetric Week, Wichmann Verlag, Heidelberg, 1995 (7 pages).
Fritsch, et al., “Oblique Image Data Processing—Potential, Experiences and Recommendations”, Photogrammetric Week, Wichmann/VDE Verlag, Berlin and Offenbach, 2013 (16 pages).
Furukawa, et al., “Manhattan-world Stereo”, 2009 (8 pages).
Furukawa, et al., “Reconstructing Building Interiors from Images”, 2009 (8 pages).
Furukawa, et al., “Towards Internet-scale Multi-view Stereo”, 2010 (8 pages).
Georgeiv, et al., “Spatio-Angular Resolution Tradeoff in Integral Photography” Eurographics Symposium on Rendering, 2006 (10 pages).
Geospan Corporation, “Digital Geo-Referenced Oblique Aerial Imagery Solution EPP-RFP No. 8444 5/13”, 2007 (28 pages).
Getting to Know ArcView GIS: the geographic information system (GIS) for everyone, “Discover the world of desktop mapping and GIS,” 1996-1998 (4 pages).
Gleicher, et al., “Image Snapping”, Advanced Technology Group, Apple Computer, Inc., 1995 (8 pages).
Goesele, et al., “Multi-View Stereo for Community Photo Collections”, Proceedings of ICCV, 2007 (8 pages).
Goesele, et al., “Multi-View Stereo Revisited”, 2006 (8 pages).
Goldman, et al., “Interactive Video Object Annotation”, Technical Report UW-CSE-2007-04-01, 2007 (7 pages).
Gomes, et al., “A Photogrammetric Project in Brazil: the Use of the PhotoModeler Software,” 1999 (8 pages).
Gong, et al., “3D Model-Based Tree Measurement from High-Resolution Aerial Imagery”, Photogrammetric Engineering and Remote Sensing, Nov. 2002 (10 pages).
Gonzalez, et al., “Digital Image Processing”, Addison-Wesley Publishing Company, Inc., 1993 (735 pages).
Gulch, et al., “On the Performance of Semi-Automatic Building Extraction”, In the International Archives of Photogrammetry and Remote Sensing, vol. 23, 1998 (8 pages).
Gulch, et al., “On the Performance of Semi-Automatic Building Extraction,” Commission III, Working Group 4, 1998 (8 pages).
Hartley, “In Defense of the Eight-Point Algorithm”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, No. 6, Jun. 6, 1997 (14 pages).
Hartley, et al., “Invariant and Calibration-Free Methods in Scene Reconstruction and Object Recognition”, Final Technical Report, Feb. 28, 1997 (266 pages).
Hartley, et al., “Multiple View Geometry in Computer Vision”, Second Edition, Cambridge University Press, 2003 (672 pages).
Hartley, et al., “Multiple View Geometry in Computer Vision: 2.4 A Hierarchy of Transformations”, Cambridge University Press, 2003 (9 pages).
Hartley, et al., “Multiple View Geometry in computer vision: Appendix 6—Iterative Estimation Methods”, Cambridge University Press, Second Edition, 2003 (34 pages).
Henricsson, et al., “3-D Building Reconstruction with Aruba: A Qualitative and Quantitative Evaluation”, Institute of Geodesy and Photogrammetry, 2001 (12 pages).
Higgins, “A Computer Algorithm for Reconstructing a Scene from Two Projections”, Macmillan Journals Ltd article, vol. 293, Sep. 10, 1981 (3 pages).
Hill, “Pictometry: Aerial Photography on Steroids”, www.law-enforcement.com, Jul. 2002 (3 pages).
Hsieh, “Design and Evaluation of a Semi-Automated Site Modeling System”, Carnegie Mellon, Nov. 1995 (83 pages).
Hsieh, “SiteCity: A Semi-Automated Site Modelling System”, IEEE, 1996 (8 pages).
Hu, et al., “Building Modeling From LIDAR and Aerial Imagery”, 2004 (8 pages).
Hudson, “Appendix D: Merging VRML Models Extending the Use of Photomodeller”, University of Virginia, Mar. 23, 1998 (23 pages).
Zongker, et al., “Environment Matting and Compositing”, 1999 (10 pages).
International Search Report of the International Searching Authority dated Nov. 17, 2014, issued in connection with International Application No. PCT/US14/49605 (3 pages).
Jaw, et al., “Building Roof Reconstruction by Fusing Laser Range data and Aerial Images”, The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. vol. XXXVII. Part B3b. 2008 (6 pages).
Communication Pursuant to Article 94(3) EPC dated Nov. 15, 2022, issued by the European Patent Office in connection with European Patent Application No. 19171426.0 (11 pages).
Office Action dated May 9, 2022, issued by the Canadian Intellectual Property Office in connection with Canadian Patent Application No. 2,864,831 (3 pages).
Office Action dated Dec. 1, 2022, issued in connection with U.S. Appl. No. 17/403,792 (28 pages).
Office Action dated Mar. 6, 2023, issued in connection with U.S. Appl. No. 17/499,308 (11 pages).
Office Action dated Feb. 15, 2023, issued by the Canadian Intellectual Property Office in connection with Canadian Patent Application No. 2,864,831 (3 pages).
Examination Report No. 2 dated Mar. 23, 2023, issued by the Australian Patent Office in connection with Australian Patent Application No. 2021201579 (3 pages).
Office Action dated Mar. 15, 2023, issued by the Columbian Patent Office, along with an English translation, in connection with Columbian Patent Application No. NC2022/0008984 (61 pages).
Related Publications (1)
Number Date Country
20220114300 A1 Apr 2022 US
Continuations (3)
Number Date Country
Parent 16709112 Dec 2019 US
Child 17560718 US
Parent 15358870 Nov 2016 US
Child 16709112 US
Parent 13397325 Feb 2012 US
Child 15358870 US