The present invention is directed toward techniques and devices for conducting building appraisals by an inspector located remotely from the building.
Gross Living Area (GLA) is generally defined as above-grade residential living space excluding unheated areas such as porches, garages, and balconies. An accurate GLA calculation for a subject property is important when generating a residential property appraisal.
Building inspectors and appraisers typically visit a property to conduct a physical and manual inspection of the property. As part of the inspection or appraisal process, inspectors obtain GLA measurements and evaluate property conditions. Although the information gathered by building inspectors is important to real estate transactions, traditional building inspections can be time consuming and expensive.
Accordingly, an aspect of the present invention includes an apparatus for conducting a building appraisal by a remote inspector. The apparatus includes an image sensor, a communication module configured to transmit and receive wireless communication signals to and from a computer network, and a memory operatively coupled to the image sensor. A processor is operatively coupled to the image sensor, the memory, the display, and the communication module. The processor is configured to execute instructions stored in the memory to cause the processor to receive a gross floor area of the building through the communication module and store images from the image sensor of an interior room of the building. The images are processed to determine a planar surface in the images corresponding to a floor surface of the interior room. The images are additionally processed to determine a plurality of corners in the images forming vertices of a bounded floor area on the floor surface. The processor transmits an adjusted floor area of the building. The adjusted floor area includes the bounded floor area subtracted from the gross floor area.
Another example aspect of the present invention is a method of conducting a building appraisal by a remote inspector. The method includes receiving, by a processor of a computing device, a gross floor area of a building through a computer network. A capturing operation captures, with an image sensor, a plurality of images of an interior room of the building. A determining operation determines, by the processor of the computing device, a planar surface in the images corresponding to a floor surface of the interior room. An identifying operation identifies, by the processor of the computing device, a plurality of corners in the images forming vertices of a bounded floor area on the planar surface. Another determining operation determines, by the processor of the computing device, an adjusted floor area. The adjusted floor area includes the bounded floor area subtracted from the gross floor area. A transmitting operation transmits, by the processor of the computing device, the adjusted floor area.
Yet another example aspect of the present invention is a computer program product for conducting a building appraisal by a remote inspector. The computer program includes computer readable program code configured to receive a gross floor area of a building through a computer network, capture a plurality of images of an interior room of the building, determine a planar surface in the images corresponding to a floor surface of the interior room, identify a plurality of corners in the images forming vertices of a bounded floor area on the planar surface, determine an adjusted floor area including the bounded floor area subtracted from the gross floor area, and transmit the adjusted floor area.
The subject matter which is regarded as the invention is particularly pointed out and distinctly claimed in the claims at the conclusion of the specification. The foregoing and other features and advantages of the invention are apparent from the following detailed description taken in conjunction with the accompanying drawings in which:
The present invention is described with reference to embodiments of the invention. Throughout the description of the invention reference is made to
Several embodiments of the invention with reference to the appended drawings are now explained. While numerous details are set forth, it is understood that some embodiments of the invention may be practiced without these details. In other instances, well-known circuits, structures, and techniques have not been shown in detail so as not to obscure the understanding of this description.
The real estate server 110 includes a database providing information in response to building queries. The real estate server 110 may be managed by a government agency or a private entity. In one embodiment of the invention, the remote inspector obtains a gross floor area of the building 104 from the real estate server 110. The gross floor area is the entire floor area of a building. For example, the gross floor area of a building can be the building's footprint area multiplied by the number of floors in the building.
An important factor in assessing the building 120 is calculating an adjusted floor area of the building 120. The adjusted floor area is determined by subtracting certain non-qualifying areas from the gross floor area of the building. These non-qualifying areas may be rooms below grade, staircases, unheated areas such as porches and balconies, and unfinished areas such as a garage. Inspectors in different localities may have different standards for whether a building area is considered a non-qualifying area. Additionally, the type of building (i.e., residential or commercial) may change whether an area is considered non-qualifying or qualifying. In some regions, the adjusted floor area may be referred to as the gross living area (GLA) of a residential building.
The inspector 106 determines the non-qualifying areas in the building by communicating with an inspection device 112 over the computer network 108. In one embodiment, the inspection device 112 is a handheld device coupled to the computer network 108. The inspection device 112 may be carried by a device user 114 located within the building 104. The device user 114 may, for example, be the building owner or resident. It is contemplated the inspection device 112 may be a smartphone, tablet, or other handheld computer programmed to provide the functionality described herein.
As discussed in more detail below, the inspection device 112 includes an image sensor for capturing images of a real-world scene. A processor in the inspection device 112 is operatively coupled to the image sensor. The processor is configured to process the captured images to determine a planar surface in the images corresponding to a floor surface 116 in the real-world scene. The processor also uses the images to determine corners in the captured images that form vertices of a bounded floor area on the planar surface.
For example, it may be the garage 118 in the building 104 constitutes a non-qualifying area in a particular jurisdiction. The remote inspector 106 can instruct the device user 114 to enter the garage 118 and point the inspection device 112 at a corner of the garage 118. A planar surface corresponding to the floor surface of the garage 118 is determined by the processor. Using the image sensor of the inspection device 112, images of the garage 118 are captured. The remote inspector 106 can further instruct the device user 114 to move around the garage 118 such that the processor in the inspection device 112 detects the garage corners captured by the image sensor on the garage floor. Once the bounded floor area of the garage 118 is determined, the processor calculates an adjusted floor area of the building 104 by subtracting the adjusted floor area from the gross floor area.
A non-qualifying area may also be a portion of a room in a building. For example, it may be that a staircase landing 120 in the building 104 constitute a non-qualifying area in a particular jurisdiction. In such a situation, the remote inspector 106 can direct the device user 114 to aim the inspection device 112 at the staircase landing 120 such that the image sensor captures images of the staircase landing 120. A processor at the inspection device 112 again determines a bounded floor area for the staircase landing 120 by detecting staircase landing corners on the floor surface. The adjusted floor area of the building is updated by further subtracting the staircase landing's bounded floor area from the gross floor area.
Calculating the adjusted floor area may require subtracting several separate bounded floor areas from the gross floor area. Thus, it will be understood by those skilled in the art that the process of capturing images of an interior room of the building 104, identifying corners in the captured images forming vertices of a bounded floor area on a planar surface corresponding to a floor surface of an interior room, and calculating an adjusted floor area of the building 104 by subtracting the bounded floor area from the gross floor area may be iterated for each non-qualifying area in the building 104.
After the adjusted floor area of the building 104 is calculated, the adjusted floor area is transmitted to the remote inspector 106 via the computer network 108 for review and recordation. The remote inspector 106 may use the adjusted floor area to, for example, generate an appraisal report, find comparable properties, and determine the value of the building 104.
In one embodiment, the inspection device 112 can employ image stitching technology. The processor in the inspection device 112 may digitally combine the captured images to produce one or more panorama images of the real-world scene. For example, as the image sensor captures images of the interior room, the processor combines the images to form a 360-degree panoramic image of the room. The panorama image can be transmitted through the computer network 108 to the remote inspector 106.
In another embodiment, the inspection device 112 may include a Global Positioning System (GPS) receiver. The GPS receiver obtains GPS signals 124 from human-made satellites 122 orbiting the Earth. The GPS signals 124 are used to calculate the current location of the inspection device 112 on Earth by the processor. The processor also receives a historical location of the building 104 over the computer network 108. For example, the historical location of the building 104 may be stored in the real estate server 110. If the current location of the inspection device 112 does not match the historical location of the building 104, the processor transmits an alert message to the remote inspector 106. This helps ensure the building 104 inspected by the inspection device 112 is same building the remote inspector 106 intends to inspect.
The image sensor 202 is configured to capture images of a real-world scene. The image sensor 202 has a field of view and includes an array of imaging sensors, such as CMOS active-pixel image sensors or charge-coupled device sensors. The image sensor 202 may also include other known image capture technologies, such as optical lenses, a shutter, and an autofocus mechanism.
The processor 208 executes program instructions stored in memory 206. The memory 206 includes read-only memory (ROM) and random-access memory (RAM). Various technologies known to those skilled in the art may be used in combination to implement the memory 206, such as static memory, dynamic memory, and/or non-volatile memory.
Instructions stored in the memory 206 cause the processor 208 to store images from the image sensor 202 of an interior room of the building in memory 206. In one embodiment, images captured by the image sensor 202 are stored with a timestamp to ensure they are current images of the property. The instructions additionally cause the processor to process the images to determine a planar surface corresponding to a floor surface of an interior room of a building. The images are further processed to determine corners in the images forming vertices of a bounded floor area on the planar surface.
The communication module 204 is used to communicate with cellular and/or computer networks. The communication module 204 includes an antenna and a transceiver to transmit and receive information over various wireless network standards, such as GSM, CDMA, IEEE 802.11, and Bluetooth®. In one embodiment, the communication module 204 is used to conduct audio and/or video calls between the remote inspector and the device user. It is contemplated a WebRTC server can be used to provide provides real-time communication between the remote inspector and the device user.
As discussed above, the processor 208 is configured to receive a gross floor area of the building through the communication module 204. In one embodiment, the communication module 204 transmits real-time images from the image sensor 202 to a remote inspector. After the gross floor area is received and the bounded floor area is determined, the processor 208 calculates an adjusted floor area of the building. The adjusted floor area includes the bounded floor area subtracted from the gross floor area. The processor 208 then transmits an adjusted floor area to a remote inspector via the communication module 204.
The processor 208 may digitally combine images of an interior room of the building to produce a panorama image, also referred to as an equirectangular-panoramic image. The panorama image can provide an immersive experience by allowing the image viewer to virtually look around the interior room at various angles. The processor 208 may be configured to transmit the panorama image to a remote inspector via the communication module 204.
In one embodiment, the processor 208 may be configured to receive a magnification message from the communication module 204. For example, a remote inspector may transmit a magnification message to the inspection device 112. The magnification message includes a magnification factor. In response to receipt of the magnification message, the processor 208 causes the image sensor 202 to magnify the captured images. The magnification message can assist the remote inspector view and capture building details during a building examination.
The inspection device 112 may include a light source 210 to illuminate a field of view of the image sensor 202. The processor 208 may be configured to receive a light activation message from the communication module 204. For example, a remote inspector may transmit a light activation message to the inspection device 112 to illuminate images captured by the image sensor 202. In response to receipt of the magnification message, the processor 208 causes the light source 210 to activate a light in response to the light activation message. The light activation message can assist the remote inspector by illuminating scenes captured by the image sensor 202 in conditions with poor ambient lighting. Similarly, the processor 208 can deactivate the light source 210 to conserve the inspection device's battery life.
The inspection device 112 may include a GPS receiver 212. The GPS receiver 212 receives signals from GPS satellites that enable the processor 208 to compute the inspection device's current location. In one embodiment, images captured by the image sensor 202 are stored with the computed GPS location as metadata to ensure they are from the inspected property. The processor 208 compares the current location of the inspection device 112 to a historical location of the building under inspection to ensure the correct building is being inspected. The historical location of the building is received by the communication module 204. If the current location of the inspection device 112 does not match the historical location of the building, the communication module 204 transmits an alert message to the remote inspector. Other methods of obtaining the inspection device's current location are contemplated. For example, Google location services may be used to determine the inspection device's current location.
The processor 208 can receive a pointer message from a remote inspector through the communication module 204. The pointer message includes a pointer location. In response to the pointer message, the processor 208 is configured to digitally overlay a pointer image on the real-time images captured by the image sensor 202.
The pointer image guides a user of the inspection device 112 to aim the image sensor 202 at a location indicated by the pointer location. The display 214 shows the pointer image overlayed on the real-time images to guide the device user to the desired location. For example, the pointer image may initially be positioned at the edge of the display 214. As the user aims the image sensor 202 closer to the desired location, the pointer image moves closer to the center of the display 214. This indicates to the device user that the device 112 is being moved to the desired location. On the other hand, if the device user moves the device 112 away from the desired location, the pointer image is moved away from the display center in a direction opposite the device movement.
In addition to the display 214, the inspection device 112 may include a user interface 216 operatively coupled to the processor 208 for communicating with a device user. The user interface 216 can include, for example, a microphone, speakers, a touch sensor, a keypad, a fingerprint reader, and/or haptic feedback.
The inspection device 112 may also include additional sensors 218 operatively coupled to the processor 208 for conducting a building inspection. For example, the inspection device 112 may include, for example, an Inertial Measurement Unit (IMU), a proximity sensor, a LiDAR scanner, a temperature sensor, a fingerprint sensor, a moisture sensor, and/or infrared sensor.
In one embodiment, the inspection device 112 is capable of locomotion. A motor controller 220 operatively coupled to the processor 208 is used to drive motors and actuators.
For example,
As mentioned above, a remote inspector may control movement of the inspection device 112 from outside the building 104. For example, the remote inspector may drive the inspection device 112 to a non-qualifying interior room 304 of the building 104 using a live image stream from the image sensor 202 transmitted by the communication module. The processor of the inspection device 112 determines a planar surface in the images corresponding to a floor surface 306 of the interior room 304. As the inspection device 112 pans the interior room 304, the processor identifies a plurality of corners 308 in the images forming vertices of a bounded floor area on the planar surface. Furthermore, the processor determines an adjusted floor area by subtracting the bounded floor area from the building's gross floor area.
At receiving step 501, a processor of a computing device receives a historical location of a building through a computer network. The historical location of the building can be used to confirm that the computing device is at the intended building during inspection. The historical location may come from various data sources, such as official government records, commercial records, or surveying service providers. After receiving step 501 is completed, control passes to receiving step 502.
At receiving step 502, a processor of a computing device receives a gross floor area for a building through a computer network. As discussed above, the gross floor area may be computed using, for example, photogrammetry. An image of the building's roof can be analyzed to determine a square footage per floor. The number of building floors can be determined by the length of the building's shadow at a known sun angle. Alternatively, the gross floor area may be accessed through a government agency, such as a government taxing agency. In one embodiment, the gross floor area is stored at and received from a real estate server. After receiving step 502 is completed, control passes to capturing step 504.
At capturing step 504, a plurality of images of an interior room of the building is captured with an image sensor. Various image sensors know to those skilled in the art may be used to capture the images. In one embodiment, images of the interior room are captured using multiple cameras with differing optical zooms and aperture openings. During capturing step 504, control may be passed to determining step 506.
At determining step 506, the processor of the computing device determines a planar surface in the images corresponding to a floor surface of the interior room. Various techniques known in the art may be used to detect the planar surface. For example, U.S. Pat. No. 9,070,194B2, incorporated herein by reference, discloses methods for detecting one or more planar surfaces within a physical environment by obtaining a collection of three-dimensional surface points of a physical environment from one or more optical sensors. Planar surfaces are identified within the collection of three-dimensional surface points such that each planar surface is at least partially defined by a group of polygons containing at least a threshold number of polygons. After determining step 506 is completed, process flow continues to identifying step 508.
At identifying step 508, the processor of the computing device identifies a plurality of corners in the images forming vertices of a bounded floor area on the planar surface. The bounded area may be the entire interior room area, such as an entire garage room, or a portion of the interior room, such as a staircase landing area.
Various techniques known in the art for identifying corners in images may be used by the processor. U.S. Patent Application Publication No. US20200034623A1, incorporated herein by reference, discloses methods and apparatuses for corner detection. Optionally, a corner in an image is detected by applying a Harris corner detection to the image. For example, U.S. Pat. No. 9,640,142B2, incorporated herein by reference, describes a method for detecting a region of interest. A plurality of corner points of a target region is detected, and the detection may be based on the Harris corner detection algorithm. The region of interest is detected based on the plurality of corner points. Another technique that may be employed is the SUSAN (Smallest Univalue Segment Assimilating Nucleus) algorithm. U.S. Pat. No. 8,456,711B2, incorporated herein by reference, describes corner sharpening methods and apparatus based on a modification of the SUSAN algorithm for corner detection. Another approach to corner detection is the FAST (Features from Accelerated Segment Test) algorithm. For example, U.S. Pat. No. 9,977,977B2, incorporated herein by reference, discloses a FAST corner-detector computer vision feature computation block implemented in hardware. This patent further discloses that in some implementations, a vision sensor may output FAST corner locations and optionally image segments around FAST-detected corners. After identifying step 508 is completed, control passes to receiving step 510.
At receiving step 510, the processor receives an indication by a user of the computing device that the bounded floor area is a non-qualifying area and should be deducted from the gross floor area. The user input can be received, for example, as a response to a dialog box displayed on the device display or by audio confirmation through the device microphone. After receiving step 510, control continues to determining step 512.
At determining step 512, the processor determines an adjusted floor area. The adjusted floor area is computed by subtracting the bounded floor area from the gross floor area. For example, a building with a gross floor area of 1,200 ft2 and a non-qualifying garage with a floor area of 200 ft2 has an adjusted floor area of 1,000 ft2. For residential buildings, the adjusted floor area may provide a more accurate measurement of the building's livable area.
A building may include several non-qualifying areas that require to be subtracted from the gross floor area. In this case, as shown in
In one embodiment, the method includes combining step 516. At combining step 516, the images captured by the image sensor are digitally combined to produce a panorama image of the interior room. The panorama image can then be transmitted to the remote inspector for review and inclusion in an inspection report.
Various techniques known in the art can be used to stitch images captured by the inspection device. For example, images can be stitched into a mosaic using image gradient calculation. In U.S. Patent Application Publication No. US20200242786A1, incorporated herein by reference, image registration is achieved by calculating a correlation coefficient image between a reference image and a floating image and calculating a gradient image of the correlation coefficient image. Another technique that may be employed is fast image blending based on the graph cuts algorithm. U.S. Pat. No. 8,593,506B2, incorporated herein by reference, uses an iterative graph cuts approach for forming a panoramic image of a scene from a sequence of input frames captured by a camera having an optical center that translates relative to the scene. A further approach to image stitching is to implement image stitching by adopting feature-based alignment algorithm and blending algorithm to produce a panoramic image. For example, U.S. Patent Application Publication No. US20200302584A1, incorporated herein by reference, discloses sampled feature points distributed across different grid tiles within overlap regions for pairs of images to match feature points to inform the alignments of a pair with respect to each image in the pair.
In one embodiment of the invention, the processor verifies that the inspection device is located at the intended building. For example, at capturing step 504, the process may follow path “B” to computing step 518 (see
At computing step 518, the processor of the computing device computes a current location of the computing device based on GPS signals received by the GPS receiver. GPS positioning is well known to those skilled in the art and provides accurate location information of the inspection device. Other location technologies, such as cell tower triangulation, may be used by the inspection device. After computing step 518 is completed, control passes to comparing step 520.
At comparing step 520, the processor compares the computed current location of the inspection device to the historical location of the building obtained at receiving step 501. If the computed current location of the inspection device does not match the historical location of the building an alert message is displayed at displaying step 522. The alert message may be displayed at the inspection device and the remote inspector computer.
In one configuration, the remote inspector may instruct the inspection device user within the building to aim the image sensor of the inspection device at a particular location. Such functionality starts at receiving step 524, where a pointer message is received. The pointer message includes a pointer location. In response to receiving step 524, control passes to overlaying step 526.
At overlaying step 526, the processor digitally overlays a pointer image on a real-time image of the real-world scene captured by the image sensor. The pointer image guides a user of the computing device to point the image sensor at the pointer location. After overlaying step 526 is completed, the real-time image and the pointer image is displayed on a display screen of the computing device at displaying step 528.
Returning to
In one embodiment of the invention, the inspection device allows the remote inspector to magnify images captured by the image sensor. For example, at receiving step 534 (see
In response to the magnification message, the real-time image captured by the image sensor is magnified by the magnification factor at magnifying step 536. Optical magnification may be achieved by moving lens elements in the image sensor. Alternatively, the processor may digitally zoom the captured image. If the inspection device is equipped with multiple image sensors, the processor may switch to an image sensor with a higher image magnification and a narrower field of view.
After magnifying step 536 is completed, process flow may return back to determining step 506 (see
Embodiments disclosed herein allow an inspector to conduct a building inspection or appraisal without visiting the building. In particular embodiments, a gross living area (GLA) can be computed for a building without the need for an inspector to enter the building.
The inspection device can provide a comprehensive inspection platform. The platform can enable inspector driven video calling for a virtual appraisal where the inspector can operate onsite devices to capture the images, an appraisal order management system with dynamic form creation, an exterior property view using third party platform, capturing non-GLA component measurements using automatic corner/markup techniques, as well as augmented reality and artificial intelligence techniques.
Additionally, the inspection device can capture a 360-degree view of the property. Secured storage can record all captured images and the recorded video call between the remote appraiser and the inspection device user for future reference. Software can provide multilanguage support for non-English speaking users. The inspection platform can schedule inspection or appraisal appointments, provide video call links, and inspection or appraisal application links.
For example, a remote inspector and a property occupant can join a video call using a smartphone application. During the video call, the remote inspector can record property details with assistance from the property occupant without visiting the property using the smartphone application. The smartphone application enables capturing of GLA measurements and building sketches that include both interior and exterior areas of the property. The smartphone application allows capturing of images with labels and markups tagged with current geographical location and time. The smartphone application can also capture 360-degree views of the property. The remote inspector can fill inspection forms and submit appraisal orders to an appraisal management company.
In one embodiment, the system captures a whole property sketch with measurements.
Next, the system prompts the property occupant to create a GLA sketch with measurements. All non-GLA areas are then displayed, and the property occupant can drag the non-GLA areas on the property sketch one by one. Once all non-GLA areas are placed on the sketch, the system creates a GLA sketch with measurements. Next, the system subtracts non-GLA area measurements from the total property measurement to provide a total GLA area.
In some embodiments, the inspection device detects doors and windows. Detected doors and windows can be included in the building floorplan 1018 generated by the inspection device. Particularly, the inspection device uses an image sensor to capture wall images with doors and/or windows. A processor detects doors and/or windows in the captured wall images and adds door symbols 1020 and window symbols 1022 to the building floorplan 1018 accordingly.
For example, the building occupant may point the image sensor at a wall of an interior room with a door or a window. A processor in the inspection device uses a deep learning algorithm with a large database of window and door image to detect the door or window. With the help of such machine learning models, the processor can identify the captured doors and windows in the interior room. Once identified, door symbols 1020 and window symbols 1022 are added to an appropriate location in the building floorplan 1018.
In some embodiments, processing the images to determine the plurality of corners in the images includes detecting at least two wall surfaces intersecting the floor surface at floor edges, extrapolating the floor edges to an intersection point, and identifying the intersection point as a corner of the plurality of corners in the images. For example, in
The inspection device further detects a first wall edge 1106 where first planar wall surface 1102 meets the floor surface 306 and a second wall edge 1108 where second planar wall surface 1104 meets the floor surface 306. Next, the inspection device extrapolates the wall edges 1106 and 1108 to an intersection point 1110. The intersection point 1110 thus defines a corner forming a floor area vertex of the interior room 302.
In one embodiment, ARKit is used with 3D space data to extrapolate wall edges. For example, ARKit extends the first and second wall edges 1106 and 1108 in the same angle and direction, respectively, as shown in
Extrapolating the wall edges to an intersection point is particularly advantageous when an object 1112, such as a piece of furniture, obstructs a room corner. In one embodiment, the inspection device may extrapolate a wall-to-wall edge 1114 formed where the first planar wall surface 1102 meets the second planar wall surface 1104 to the floor surface 306. The extrapolated wall-to-wall edge 1114 provides a third point of reference to determine intersection point the intersection point 1110. Thus, extrapolating the wall-to-wall edge 1114 can provide additional accuracy to the corner point measurement.
The inspection device's position, orientation and distance to objects and surfaces may be calculated by means of images captured by the image sensor, motion sensors, and the ARKit. The ARKit uses a captured image to help determine the position of the image sensor in space. In one embodiment, the property occupant stands in the center of an interior room and points the image sensor of the inspection device at a wall 1102. Data from the ARKit is used to determine the wall position in the 3D space. The orientation of the device is used to get the wall's orientation.
The display of the inspection device may indicate to the building occupant when an intersection point 1110 is detected. Augmented reality may be used to digitally overlay a corner marker on the images from the image sensor in response to detection of the intersection point 1110. Augmented reality can be used to take measurements between two corners. The augmented reality position of a point in the real world can be acquired with a three-dimensional point (x, y, z). Once the positions of two points are known, the distance between these points can be determined using the Euclidean distance formula,
√{square root over ([(x1−x2)2+(y1−y2)2+(z1−z2)2])}.
After the distance between the first wall 1008 is determined, the distance is stored in device memory. The property occupant repeats the measurement process for remaining walls 1010, 1012, 1014, and 1016. Once the property occupant reaches the first corner 1004 again, the application uses the distance of all room walls to determine the room area.
It is contemplated that various methods for collecting distance information may be utilized by embodiments of the invention. By way of illustration only, a LiDAR scanner may be used. In addition, stereoscopic systems employing two image sensors, spaced slightly apart yet looking at the same scene, may be used as well. By analyzing the slight differences between the images seen by each image sensor, it is possible to determine the distance at each point in the images. In yet another embodiment, distance information may be compiled by using a single image sensor, travelling at a known velocity, to capture building details the scenery passes by. Image sensor velocity may be measured by an IMU, for example. Subsequent frames may be compared to extract the different distances between the objects and the image sensor based on the motion velocity.
Thus, if a remote inspector desires to determine a building's GLA measurement, the inspector communicates with the inspection device carried with the building occupant. The inspection device prompts the building occupant to begin taking area measurements, as described above. Exterior building measurements can be acquired using, for example, third-party aerial views and machine learning algorithms. The system can measure interior closed areas with the assistance of, for example, automatic wall and corner detection algorithms, deep learning algorithms and artificial intelligence algorithms. Measuring open areas, such as staircases, uses, for example, markup of one place to another place and calculating distance between two or more marked points using augmented reality and artificial intelligence. The interior non-GLA area and open area measurements are subtracted from the total exterior view to arrive at an accurate GLA calculation.
In one embodiment, the Cortex AI platform developed by Matterport, Inc. is used to transform interior room images into an interactive three-dimensional model. The Cortex platform may identify objects within rooms, stitch images together, and reconstruct building spaces into an immersive virtual tour. The inspection device may leverage Cortex's deep learning neural network to create three-dimensional data from various capture devices, such as LiDAR scanners, 360-degree cameras, and smartphones. The platform can provide automatic measurements for building rooms and objects in the rooms, automatic 2D-from-3D HD photo gallery creation, auto face blurring for privacy protection, custom videos, walkthroughs, auto room labeling, and object recognition.
During the inspection or appraisal process, remote inspectors can activate the inspection device's flash if they determine there is insufficient ambient light. Socket messages communicated between the remote inspector and the inspection device can be used to activate and deactivate a light at the remote device.
Inspectors can point on the display to regions of interest in the building using pointers to guide device users to the regions of interest. Socket messages communicated between the remote inspector and the inspection device can be used to display a pointer on the inspection device's display. For example, an inspector computer can send a socket message with screen coordinates for the pointer to the inspection device. After receiving the pointer socket message, the inspection device maps the screen coordinate to the device screen and displays the pointer on the device screen.
Inspectors can control image zoom on the inspection device to capture images of damage to or important details of the property. The remote inspector can also capture images of the property using the inspection device. Socket messages communicated between the remote inspector and the inspection device can be used to control the zoom factor of the remote device's image sensor.
When a remote inspector wants to capture an image from the inspection device user, the inspector clicks a button to send a socket message to the inspection device. In response to receiving the socket message, the inspection device captures an image from the image sensor. The inspection device attaches a current time to the image. The current location of the device can also be determined with the assistance of, for example, Google location services. The location data is also attached to the image. Once time and location metadata is attached, the image is sent from the inspection device to the remote inspector's computer.
The descriptions of the various embodiments of the present invention have been presented for purposes of illustration but are not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement over technologies found in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method, or computer program product. Accordingly, the present invention may be a system, a method, and/or a computer program product. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random-access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random-access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.