The present disclosure generally relates to electronic devices that use sensors to scan physical environments to generate three dimensional (3D) models such as 3D floor plans.
Existing scanning systems and techniques may be improved with respect to assessing and using the sensor data obtained during scanning processes to generate 3D representations such as 3D floor plans representing physical environments.
Various implementations disclosed herein include devices, systems, and methods that provide a 3D floor plan based on scanning a room and detecting windows, doors, and openings using 2D orthographic projection. A 3D floor plan is a 3D representation of a room or other physical environment that generally identifies or otherwise represents 3D positions of one or more walls, floors, ceilings, or other boundaries or regions of the environment. Some implementations disclosed herein generate a 3D floor plan that identifies or otherwise represents 3D positions of windows, doors, and/or openings within the 3D floor plan, e.g., on the walls, floors, ceilings, or other regions.
The 3D positions and/or other characteristics of windows, doors, and openings may be determined using 2D orthographic projection. In some implementations, a set of points, such as points of a 3D point cloud, nodes of a 3D mesh, or points of any other 3D representation, is generated to represent a room or other physical environment. The points that are close to a plane representing a wall are projected onto the plane and used to identify windows, doors, and openings on the wall. Representations of the detected windows, doors, and openings may then be positioned in a 3D floor plan based on the known position of the wall within the room, i.e., the location of the wall plane relative to the set of points is known. In some implementations, boundaries or regions corresponding to walls, floors, ceilings, etc., are detected directly from points of the set of points, windows, doors, and openings are detected indirectly using projections of the points of the set of points onto a 2D plane, and these detected aspects are combined into a single 3D floor plan. Detecting the windows, doors, and openings using a 2D projection as opposed to detecting them directly using the points of the set of points may be more accurate, more efficient, or otherwise advantageous.
In some implementations, a processor performs a method by executing instructions stored on a computer readable medium. The method identifies a set of points of a set of points (e.g., a 3D point cloud) representing a physical environment, where the set of points correspond to a wall in the physical environment. In some implementations, the set of points is identified by identifying points that are within a threshold distance of a wall plane prediction. The wall plane prediction may predict the position of a wall surface and approximate positions of wall boundaries and/or openings within the wall surface. The method projects the set of points onto a 2D plane corresponding to the wall, where each point of the set of points is projected to a location on the 2D plane. In some implementations, the set of points is projected via orthographic projection. The projected points may additionally be associated with semantics, color/RGB info, and/or normalized distance information that corresponds to distances of the points from the wall in the 3D point cloud. In some implementations, projecting the points may include producing one or more 2D data sets, for example, including a semantic map, an RGB map, and/or a point distance map (of normalized distance information). The method detects one or more windows, one or more doors, and/or one or more openings based on the set of points projected onto the 2D plane. The detecting may involve predicting parameters that parametrically define the windows, doors, and/or openings in terms of 2D location coordinates, 2D dimensions, and/or other characteristics such as open, closed, percentage open, etc. The method generates a 3D floor plan based on the detecting of a window, door, or opening. Other portions of the 3D floor plan, such as a representation of the wall (and other walls), floors, ceiling, counters, appliances, etc., may be generated based on the 3D point cloud and the windows, doors, and/or openings may be positioned based on a known spatial relationship of the wall plane relative to the set of points. For example, the 3D location of the plane corresponding to the wall within the set of points (e.g., relative to a 3D point cloud) may enable the wall and the windows, doors, and openings detected on it to be positioned relative to a 3D floor plan generated based on the set of points.
In accordance with some implementations, a device includes one or more processors, a non-transitory memory, and one or more programs; the one or more programs are stored in the non-transitory memory and configured to be executed by the one or more processors and the one or more programs include instructions for performing or causing performance of any of the methods described herein. In accordance with some implementations, a non-transitory computer readable storage medium has stored therein instructions, which, when executed by one or more processors of a device, cause the device to perform or cause performance of any of the methods described herein. In accordance with some implementations, a device includes: one or more processors, a non-transitory memory, and means for performing or causing performance of any of the methods described herein.
So that the present disclosure can be understood by those of ordinary skill in the art, a more detailed description may be had by reference to aspects of some illustrative implementations, some of which are shown in the accompanying drawings.
In accordance with common practice the various features illustrated in the drawings may not be drawn to scale. Accordingly, the dimensions of the various features may be arbitrarily expanded or reduced for clarity. In addition, some of the drawings may not depict all of the components of a given system, method or device. Finally, like reference numerals may be used to denote like features throughout the specification and figures.
Numerous details are described in order to provide a thorough understanding of the example implementations shown in the drawings. However, the drawings merely show some example aspects of the present disclosure and are therefore not to be considered limiting. Those of ordinary skill in the art will appreciate that other effective aspects and/or variants do not include all of the specific details described herein. Moreover, well-known systems, methods, components, devices and circuits have not been described in exhaustive detail so as not to obscure more pertinent aspects of the example implementations described herein.
In one example, the user 102 moves around the physical environment 100 and device 110 captures sensor data from which a 3D floor plan of the physical environment 100 is generated. The device 110 may be moved to capture sensor data from different viewpoints, e.g., at various distances, viewing angles, heights, etc. The device 110 may provide information to the user 102 that facilitates the environment scanning process. For example, the device 110 may provide a view from a camera showing the content of RGB images currently being captured, e.g., a live camera feed, during the room scanning process. As another example, the device 110 may provide a view of a live 3D point cloud or a live 3D floor plan to facilitate the scanning process or otherwise provides feedback that informs the user 102 of which portions of the physical environment 100 have already been captured in sensor data and which portions of the physical environment 100 require more sensor data in order to be represented accurately in a 3D representation and/or 3D floor plan.
Accordingly,
In alternative implementations, a 3D mesh is generated in which points of the 3D mesh have 3D coordinates such that groups of the mesh points identify surface portions, e.g., triangles, corresponding to surfaces of the physical environment 100. Such points and/or associated shapes (e.g., triangles) may be associated with color, surface normal directions, and/or semantic labels.
In the example of
In the process illustrated in
In
In the process illustrated in
Moreover, in some implementations, points on either side of the wall plane 310 are selected. For example, subset 400 of selected points includes points on one side of the wall plane 310 (i.e., the near side) such as a subset of left wall points 402, a subset of floor points 404, a subset of ceiling points 406, a subset of right wall points 408, door frame points 440, and window frame points 460. In addition, the subset 400 of selected points include points on the other side of the wall plane 310 (i.e., the far side) such as the door points 430 and a subset of the floor points 412.
In the process illustrated in
All of these points 502, 504, 506, 508, 520, 530a, 530b, 540, 550, 560 are projected onto a single 2D plane. Characteristics of the points (e.g., color, semantic labels, normal distances of the corresponding 3D points to the wall plane 310, etc.) may be retained in the points of the 2D projection 500. In the example of
In the process illustrated in
In some implementations, such a technique receives one or more 2D data structures, e.g., maps, matrices, etc., such as using a semantic map, an RGB map, and/or a point distance map and outputs 2D coordinate information or parameters defining the 2D locations of any windows or doors detected within the 2D plane represented by the 2D data structures. In the example of
In the process illustrated in
In this example, the window-door detection process determined the 2D positions of window 620 and door 610 on a 2D projection corresponding to wall plane 310. Since wall plane 310 corresponds to the wall region 790 of the 3D floor plan 700, the window 620 and 610 can be positioned within the 3D floor plan 700. In other words, the 3D positions of the door 610 and window 620 in the 3D floor plan 700 are determined based on their 2D positions within the 2D projection 500 and the position of the region 790 of the 3D floor plan 700 that corresponds to the wall plane 310 that corresponds to the 2D projection 500.
In some implementations, a similar process is repeated to detect any windows, doors, and openings in other portions of the physical environment 100 (e.g., on other walls, the ceiling, the floor, etc.) to be represented in the 3D floor plan 700.
The inclusion of windows, doors, and openings in 3D floor plans can provide various benefits. For example, the inclusion of windows may enable potential applications for lighting and energy use estimation. The inclusion of doors may be significant with respect to connecting rooms to one another, for example, in combined 3D floor plans that represent entire buildings.
Some of the techniques disclosed herein convert a 3D window/door detection process into a 2D detection problem on the surface of planes corresponding to walls or other regions of a physical environment. Orthographic detection is well suited for detecting windows, doors, and openings. Such windows, doors, and openings are often rectangular planar objects on a wall or ceiling that are well suited or orthographic detection due to their shapes and planar positioning. Moreover, a rectangular area in a 2D projection with no points is information that can be interpreted. Thus, the 2D projection may actually create or identify information that makes detection of doors, windows, and openings more accurate. Orthographic projection based on a distance threshold may also provide information that may not otherwise be represented as clearly in a 3D point-based representation such as a 3D point cloud. Orthographic projection may also reduce the impact of viewpoints and angles associated with sensor data, for example, by ensuring that rectangular objects are represented as rectangles in the 2D projections. A detection process can be configured to detect such objects without needing to account for the skewed appearance of rectangular objects that might otherwise be required. In some implementations, orthographic projection enables the fusion of 3D semantics (e.g., identifying walls, floors, ceilings, etc.), RGB data, distance data, etc. together at an input layer of the detection process. Orthographic projection may preserve structural information relevant to detecting windows, doors, and openings while reducing the complexity of the information. It may provide information from which clear patterns indicative of doors and windows can be recognized in contrast to the corresponding 3D information from which the relative scarcity of information with respect to 3D space may prevent accurate identification.
The wall projection module 840 processes these inputs 810, 820, 830 and produces a 2D projection information 850a-c, for example, by encoding semantics, RGB, and distance information onto a wall plane. The wall plane prediction may predict the position of a wall surface and approximate positions of wall boundaries and/or openings within the wall surface. The 2D projection information includes: semantic maps 850a, RGB maps 850b, and points distance maps 850c. The 2D projection information 850a-c, including semantic maps 850a, RGB maps 850b, and points distance maps 850c, are input to 2D orthographic detection module 860 that detects instances of windows, doors, and/or openings according to the techniques disclosed herein. The 2D orthographic detection module 860 produces instance bounding box information 870 output providing 2D bounding boxes around detected windows, doors, and openings. The instance bounding box information 870 is input to a projecting to the 3D wall module 880 that positions the detected 2D bounding boxes (corresponding to detected windows, doors, and openings) into a 3D coordinate system to produce final output 890 such as a floor plan.
At block 902, the method 900 identifies a set of points of a set of points (e.g., of a 3D point cloud or 3D mesh) representing a physical environment, where the set of points correspond to a wall in the physical environment. For example, this may involve identifying a subset of points of a 3D point cloud that are within a threshold distance of a wall plane as illustrated in
At block 904, the method 900 projects the set of points onto a 2D plane corresponding to the wall, where each point of the set of points is projected to a location on the 2D plane. The 2D plane corresponding to the wall may be predicted using a process that evaluates the set of points to predict the position of a wall surface and approximate positions of wall boundaries and/or openings within the wall surface. The points of the set of points may be projected onto the 2D plane by orthographic projection. For example, a subset set of points of a 3D point cloud may be orthographically projected onto a plane as illustrated in
At block 906, the method 900 detects a window, door, or opening based on the set of points projected onto the 2D plane. For example, this may involve a 2D orthographic detection process as illustrated in
At block 908, the method 900 generates a 3D floor plan based on the detecting of the window, door, or opening. Other portions of the 3D floor plan such as a representation of the wall (and other walls) may be generated based on the 3D point cloud and the window, door, or opening may be positioned based on a known spatial relationship of the wall plane used to detect the window, door, or opening and the 3D point cloud.
In some implementations, the one or more communication buses 1004 include circuitry that interconnects and controls communications between system components. In some implementations, the one or more I/O devices and sensors 1006 include at least one of an inertial measurement unit (IMU), an accelerometer, a magnetometer, a gyroscope, a thermometer, one or more physiological sensors (e.g., blood pressure monitor, heart rate monitor, blood oxygen sensor, blood glucose sensor, etc.), one or more microphones, one or more speakers, a haptics engine, one or more depth sensors (e.g., a structured light, a time-of-flight, or the like), and/or the like.
In some implementations, the one or more output device(s) 1012 include one or more displays configured to present a view of a 3D environment to the user. In some implementations, the one or more displays 1012 correspond to holographic, digital light processing (DLP), liquid-crystal display (LCD), liquid-crystal on silicon (LCoS), organic light-emitting field-effect transitory (OLET), organic light-emitting diode (OLED), surface-conduction electron-emitter display (SED), field-emission display (FED), quantum-dot light-emitting diode (QD-LED), micro-electromechanical system (MEMS), and/or the like display types. In some implementations, the one or more displays correspond to diffractive, reflective, polarized, holographic, etc. waveguide displays. In one example, the device 1000 includes a single display. In another example, the device 1000 includes a display for each eye of the user.
In some implementations, the one or more output device(s) 1012 include one or more audio producing devices. In some implementations, the one or more output device(s) 1012 include one or more speakers, surround sound speakers, speaker-arrays, or headphones that are used to produce spatialized sound, e.g., 3D audio effects. Such devices may virtually place sound sources in a 3D environment, including behind, above, or below one or more listeners. Generating spatialized sound may involve transforming sound waves (e.g., using head-related transfer function (HRTF), reverberation, or cancellation techniques) to mimic natural soundwaves (including reflections from walls and floors), which emanate from one or more points in a 3D environment. Spatialized sound may trick the listener's brain into interpreting sounds as if the sounds occurred at the point(s) in the 3D environment (e.g., from one or more particular sound sources) even though the actual sounds may be produced by speakers in other locations. The one or more output device(s) 1012 may additionally or alternatively be configured to generate haptics.
In some implementations, the one or more image sensor systems 1014 are configured to obtain image data that corresponds to at least a portion of a physical environment. For example, the one or more image sensor systems 1014 may include one or more RGB cameras (e.g., with a complimentary metal-oxide-semiconductor (CMOS) image sensor or a charge-coupled device (CCD) image sensor), monochrome cameras, IR cameras, depth cameras, event-based cameras, and/or the like. In various implementations, the one or more image sensor systems 1014 further include illumination sources that emit light, such as a flash. In various implementations, the one or more image sensor systems 1014 further include an on-camera image signal processor (ISP) configured to execute a plurality of processing operations on the image data.
The memory 1020 includes high-speed random-access memory, such as DRAM, SRAM, DDR RAM, or other random-access solid-state memory devices. In some implementations, the memory 1020 includes non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid-state storage devices. The memory 1020 optionally includes one or more storage devices remotely located from the one or more processing units 1002. The memory 1020 comprises a non-transitory computer readable storage medium.
In some implementations, the memory 1020 or the non-transitory computer readable storage medium of the memory 1020 stores an optional operating system 1030 and one or more instruction set(s) 1040. The operating system 1030 includes procedures for handling various basic system services and for performing hardware dependent tasks. In some implementations, the instruction set(s) 1040 include executable software defined by binary information stored in the form of electrical charge. In some implementations, the instruction set(s) 1040 are software that is executable by the one or more processing units 1002 to carry out one or more of the techniques described herein.
The instruction set(s) 1040 include a 3D representation instruction set 1042 configured to, upon execution, obtain sensor data, provide views/representations, select sets of sensor data, and/or generate 3D point clouds, 3D meshes, 3D floor plans, and/or other 3D representations of physical environments as described herein. The instruction set(s) 1040 further include a plane detection instruction set 1044 configured to detect planes such as walls, ceilings, floors, and the like in physical environments and/or corresponding 3D point-based representations as described herein. The instruction set(s) 1040 further include a window/door/opening detection instruction set configured to detect windows, doors, and openings in physical environments and/or corresponding 3D point-based representations. The instruction set(s) 1040 may be embodied as a single software executable or multiple software executables.
Although the instruction set(s) 1040 are shown as residing on a single device, it should be understood that in other implementations, any combination of the elements may be located in separate computing devices. Moreover, the figure is intended more as functional description of the various features which are present in a particular implementation as opposed to a structural schematic of the implementations described herein. As recognized by those of ordinary skill in the art, items shown separately could be combined and some items could be separated. The actual number of instructions sets and how features are allocated among them may vary from one implementation to another and may depend in part on the particular combination of hardware, software, and/or firmware chosen for a particular implementation.
It will be appreciated that the implementations described above are cited by way of example, and that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope includes both combinations and sub combinations of the various features described hereinabove, as well as variations and modifications thereof which would occur to persons skilled in the art upon reading the foregoing description and which are not disclosed in the prior art.
As described above, one aspect of the present technology is the gathering and use of sensor data that may include user data to improve a user's experience of an electronic device. The present disclosure contemplates that in some instances, this gathered data may include personal information data that uniquely identifies a specific person or can be used to identify interests, traits, or tendencies of a specific person. Such personal information data can include movement data, physiological data, demographic data, location-based data, telephone numbers, email addresses, home addresses, device characteristics of personal devices, or any other personal information.
The present disclosure recognizes that the use of such personal information data, in the present technology, can be used to the benefit of users. For example, the personal information data can be used to improve the content viewing experience. Accordingly, use of such personal information data may enable calculated control of the electronic device. Further, other uses for personal information data that benefit the user are also contemplated by the present disclosure.
The present disclosure further contemplates that the entities responsible for the collection, analysis, disclosure, transfer, storage, or other use of such personal information and/or physiological data will comply with well-established privacy policies and/or privacy practices. In particular, such entities should implement and consistently use privacy policies and practices that are generally recognized as meeting or exceeding industry or governmental requirements for maintaining personal information data private and secure. For example, personal information from users should be collected for legitimate and reasonable uses of the entity and not shared or sold outside of those legitimate uses. Further, such collection should occur only after receiving the informed consent of the users. Additionally, such entities would take any needed steps for safeguarding and securing access to such personal information data and ensuring that others with access to the personal information data adhere to their privacy policies and procedures. Further, such entities can subject themselves to evaluation by third parties to certify their adherence to widely accepted privacy policies and practices.
Despite the foregoing, the present disclosure also contemplates implementations in which users selectively block the use of, or access to, personal information data. That is, the present disclosure contemplates that hardware or software elements can be provided to prevent or block access to such personal information data. For example, in the case of user-tailored content delivery services, the present technology can be configured to allow users to select to “opt in” or “opt out” of participation in the collection of personal information data during registration for services. In another example, users can select not to provide personal information data for targeted content delivery services. In yet another example, users can select to not provide personal information, but permit the transfer of anonymous information for the purpose of improving the functioning of the device.
Therefore, although the present disclosure broadly covers use of personal information data to implement one or more various disclosed embodiments, the present disclosure also contemplates that the various embodiments can also be implemented without the need for accessing such personal information data. That is, the various embodiments of the present technology are not rendered inoperable due to the lack of all or a portion of such personal information data. For example, content can be selected and delivered to users by inferring preferences or settings based on non-personal information data or a bare minimum amount of personal information, such as the content being requested by the device associated with a user, other non-personal information available to the content delivery services, or publicly available information.
In some embodiments, data is stored using a public/private key system that only allows the owner of the data to decrypt the stored data. In some other implementations, the data may be stored anonymously (e.g., without identifying and/or personal information about the user, such as a legal name, username, time and location data, or the like). In this way, other users, hackers, or third parties cannot determine the identity of the user associated with the stored data. In some implementations, a user may access their stored data from a user device that is different than the one used to upload the stored data. In these instances, the user may be required to provide login credentials to access their stored data.
Numerous specific details are set forth herein to provide a thorough understanding of the claimed subject matter. However, those skilled in the art will understand that the claimed subject matter may be practiced without these specific details. In other instances, methods apparatuses, or systems that would be known by one of ordinary skill have not been described in detail so as not to obscure claimed subject matter.
Unless specifically stated otherwise, it is appreciated that throughout this specification discussions utilizing the terms such as “processing,” “computing,” “calculating,” “determining,” and “identifying” or the like refer to actions or processes of a computing device, such as one or more computers or a similar electronic computing device or devices, that manipulate or transform data represented as physical electronic or magnetic quantities within memories, registers, or other information storage devices, transmission devices, or display devices of the computing platform.
The system or systems discussed herein are not limited to any particular hardware architecture or configuration. A computing device can include any suitable arrangement of components that provides a result conditioned on one or more inputs. Suitable computing devices include multipurpose microprocessor-based computer systems accessing stored software that programs or configures the computing system from a general-purpose computing apparatus to a specialized computing apparatus implementing one or more implementations of the present subject matter. Any suitable programming, scripting, or other type of language or combinations of languages may be used to implement the teachings contained herein in software to be used in programming or configuring a computing device.
Implementations of the methods disclosed herein may be performed in the operation of such computing devices. The order of the blocks presented in the examples above can be varied for example, blocks can be re-ordered, combined, and/or broken into sub-blocks. Certain blocks or processes can be performed in parallel.
The use of “adapted to” or “configured to” herein is meant as open and inclusive language that does not foreclose devices adapted to or configured to perform additional tasks or steps. Additionally, the use of “based on” is meant to be open and inclusive, in that a process, step, calculation, or other action “based on” one or more recited conditions or values may, in practice, be based on additional conditions or value beyond those recited. Headings, lists, and numbering included herein are for ease of explanation only and are not meant to be limiting.
It will also be understood that, although the terms “first,” “second,” etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first node could be termed a second node, and, similarly, a second node could be termed a first node, which changing the meaning of the description, so long as all occurrences of the “first node” are renamed consistently and all occurrences of the “second node” are renamed consistently. The first node and the second node are both nodes, but they are not the same node.
The terminology used herein is for the purpose of describing particular implementations only and is not intended to be limiting of the claims. As used in the description of the implementations and the appended claims, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
As used herein, the term “if” may be construed to mean “when” or “upon” or “in response to determining” or “in accordance with a determination” or “in response to detecting,” that a stated condition precedent is true, depending on the context. Similarly, the phrase “if it is determined [that a stated condition precedent is true]” or “if [a stated condition precedent is true]” or “when [a stated condition precedent is true]” may be construed to mean “upon determining” or “in response to determining” or “in accordance with a determination” or “upon detecting” or “in response to detecting” that the stated condition precedent is true, depending on the context.
The foregoing description and summary of the invention are to be understood as being in every respect illustrative and exemplary, but not restrictive, and the scope of the invention disclosed herein is not to be determined only from the detailed description of illustrative implementations but according to the full breadth permitted by patent laws. It is to be understood that the implementations shown and described herein are only illustrative of the principles of the present invention and that various modification may be implemented by those skilled in the art without departing from the scope and spirit of the invention.
This Application claims the benefit of U.S. Provisional Application Ser. No. 63/247,834 filed Sep. 24, 2021 which is incorporated herein in its entirety.
Number | Date | Country | |
---|---|---|---|
63247834 | Sep 2021 | US |