Z-PLANE IDENTIFICATION AND BOX DIMENSIONING USING THREE-DIMENSIONAL TIME-OF-FLIGHT IMAGING

Information

  • Patent Application
  • 20230367017
  • Publication Number
    20230367017
  • Date Filed
    September 21, 2021
    3 years ago
  • Date Published
    November 16, 2023
    a year ago
Abstract
A sensor system that obtains and processes time-of-flight data (TOF) obtained in an arbitrary orientation is provided. A TOF sensor obtains distance data describing various surfaces. A processor identifies a horizontal Z-plane in the environment, and transforms the data to align with the Z-plane. In some embodiments, the environment includes a box, and the processor identifies a bottom and a top of the box in the transformed data. The processor can further determine dimensions of the box, e.g., the height between the top and bottom of the box, and the length and width of the box top.
Description
BACKGROUND

Man-made environments are generally endowed with a preferred direction corresponding to the local orientation of earth's gravity field. In simple terms, “up” and “down” define natural engineering directions for indoor settings (e.g., rooms) and outdoor settings (e.g., streets). Floors, walls, and ceilings are strongly constrained by the direction of local gravity. In particular, man-made environments are usually populated by horizontal Z-planes (e.g., tabletops, chair seats, floors, sidewalks).


As a person walks around a man-made environment while holding a 3D time-of-flight (TOF) imaging system in hand, the sensor's angular orientation relative to the natural “up” and “down” directions is typically unknown. Humans do not reliably align imaging systems to their environments, and having users align and re-align sensors to match their environment can be a time-consuming and frustrating process.


One potential application of TOF imaging systems is determining the dimensions of boxes. Measuring volumes of physical objects is a basic problem for various industrial and consumer markets, such as packing, shipping, and storage of objects. In typical packing and shipping contexts, humans use tape measures to measure box dimensions, which is a time-consuming process. Existing technical solutions are often fragile, expensive, and/or can only be used in certain settings. For example, some dimensioning solutions rely on fixed frames of reference, e.g., deriving the volume of a box placed on a designated surface from an image taken by a camera at a fixed position relative to the designated surface.





BRIEF DESCRIPTION OF THE DRAWINGS

To provide a more complete understanding of the present disclosure and features and advantages thereof, reference is made to the following description, taken in conjunction with the accompanying figures, wherein like reference numerals represent like parts, in which:



FIG. 1 is a block diagram of a TOF sensor system, according to some embodiments of the present disclosure.



FIG. 2 illustrates ray directions of pixels of a TOF sensor, according to some embodiments of the present disclosure.



FIG. 3 is a flow diagram showing a process for identifying a Z-plane in TOF data obtained in an arbitrary frame of reference, according to some embodiments of the present disclosure.



FIG. 4 is a flow diagram showing a process for identifying basis vectors based on TOF data, according to some embodiments of the present disclosure.



FIG. 5 is a flow diagram showing a process for identifying a Z-plane in a point cloud transformed based on the identified basis vectors, according to some embodiments of the present disclosure.



FIG. 6 is a flow diagram showing a process for determining and outputting box dimension based on TOF data, according to some embodiments of the present disclosure.



FIG. 7 is a flow diagram showing a process for identifying a box top and a box bottom, according to some embodiments of the present disclosure.



FIG. 8 is a flow diagram showing a process for calculating the length and width of the box top, according to some embodiments of the present disclosure.



FIG. 9 is an example image showing a box resting on a tabletop, according to some embodiments of the present disclosure.



FIG. 10 illustrates an example of distance data obtained by a TOF sensor, according to some embodiments of the present disclosure.



FIG. 11 illustrates an example point cloud calculated from the distance data, according to some embodiments of the present disclosure.



FIGS. 12A and 12B illustrate example angular coordinates of the surface normals of the points in the point cloud, according to some embodiments of the present disclosure.



FIG. 13 is an example histogram that bins angular coordinates of the surface normals, according to some embodiments of the present disclosure.



FIG. 14 illustrates an example of the point cloud transformed into a reference coordinate system of the identified basis vectors, according to some embodiments of the present disclosure.



FIG. 15 is an example height map obtained from the transformed point cloud, according to some embodiments of the present disclosure.



FIG. 16 is an example Z-profile of the height map with peaks indicating various horizontal surfaces, according to some embodiments of the present disclosure.



FIG. 17 illustrates four example Z-plane slices identified from the height map, according to some embodiments of the present disclosure.



FIGS. 18A-18B illustrate two sets of connected components for two different Z-plane slices, according to some embodiments of the present disclosure.



FIGS. 19A-19B illustrate two candidate box tops identified in the connected components, according to some embodiments of the present disclosure.



FIG. 20 illustrates a set of points corresponding to a connected component identified as the box top, according to some embodiments of the present disclosure.



FIG. 21 is an example profile of the box top projected along the x- and y-axes, according to some embodiments of the present disclosure.



FIG. 22 illustrates an axis-aligned box top rotated based on the profile in FIG. 21, according to some embodiments of the present disclosure.



FIGS. 23A and 23B show example box top width and length profiles, according to some embodiments of the present disclosure.



FIG. 24 illustrates identified box edges overlayed on an image obtained by the TOF sensor, according to some embodiments of the present disclosure.



FIG. 25 illustrates identified box edges and determined box dimension overlayed on an image obtained by a camera, according to some embodiments of the present disclosure.





DESCRIPTION OF EXAMPLE EMBODIMENTS OF THE DISCLOSURE

Overview


The systems, methods and devices of this disclosure each have several innovative aspects, no single one of which is solely responsible for all of the desirable attributes disclosed herein. Details of one or more implementations of the subject matter described in this specification are set forth in the description below and the accompanying drawings.


Reliable identification of Z-planes (e.g., a floor, a street, a tabletop) in an environment can be useful in many two-dimensional and three-dimensional image processing applications. In particular, it is useful to determine a time-of-flight sensor's roll and pitch angles relative to a Z-plane, as well as the sensor's height relative to Z-planes in its environment. As used herein, a Z-plane is a plane in a real-world environment that is parallel to the ground in a particular environment. Z-planes include the ground or floor, and surfaces parallel to the ground or floor. In many cases, the Z-plane is perpendicular to the direction of gravity. In some cases, e.g., on a hill or other slanted surface, Z-planes (e.g., the ground, a table resting on the ground) may be somewhat tilted with respect to the direction of gravity.


A base Z-plane is the lowest Z-plane within an image that captures an environment. For example, in an image of an environment that includes a box resting on a table placed on a floor, the box top, tabletop, and floor are all Z-planes, and the floor is the base Z-plane. If another image includes the box and the tabletop but does not include the floor, the tabletop is the base Z-plane for that image.


Methods and systems for identifying Z-planes in an environment and, in some cases, identifying a base Z-plane in an environment, are described herein. The method involves extracting parameters for the roll and pitch rotation angles relative to a Z-plane. In some embodiments, the method also extracts a parameter for the height of the sensor relative to the base Z-plane from a single input TOF depth frame. Once the two rotation parameters and translation parameters are extracted, the number of a priori unknown extrinsic camera calibration parameters is reduced from six (3 translations+3 rotation angles) to three (2 translations+1 rotation angle). Time-of-flight applications become easier and faster for processing systems to handle when the number of unknown sensor degrees of freedom is reduced in this way. In addition, aligning coordinate system axes to a Z-plane simplifies time-of-flight imagery exploitation in several applications, such as box dimensioning, object dimensioning, box packing, or obstacle detection.


Methods and systems for measuring dimensions of a box are also described herein. One method involves receiving a three-dimensional point cloud obtained from time-of-flight data and identifying a box within the point cloud. In particular, the method includes identifying a box top within the point cloud, and then identifying a surface on which the box is resting, such as a tabletop or the floor. The method then includes calculating the height of the box as the distance between the box top and the surface on which the box is resting, and identifying edges of the box top. The method then includes calculating width and length profiles for the edges, and determining a width and a length for the box based on the width and length profiles. Quantitative height, width, and length values, e.g., measured in centimeters, may be reported to a user, e.g., on a display of a TOF measurement device. In some examples, the device also generates a visualization of the identified box superimposed on an image of the box so that the user may qualitatively confirm the calculated dimensions.


Existing box dimensioning solutions are typically highly vulnerable to sunlight because sunlight creates significant noise in TOF data or image data. Prior box dimensioning systems were only suitable for indoor use or under particular lighting conditions. In some embodiments described herein, TOF measurement data is filtered to reduce the impact of visual noise, enabling the TOF sensor system to be used in a variety of ambient lighting conditions, including both indoors and outdoors. In one example, the measurement data is filtered at a first stage for identifying Z-planes in the observed environment. Because Z-planes are relatively large, an aggressive filter (e.g., a large filter window) can be used. As noted above, after the box has been identified using the Z-planes, the box edges are identified. A finer filter (e.g., a smaller filter window) may be used to filter the measurement data for finding the box edges, since more precision is needed at this stage.


One embodiment provides a method for identifying a Z-plane. The method includes receiving distance data describing distances between a sensor that captured the distance data and a plurality of surfaces in an environment of the sensor, where at least one of the surfaces is a Z-plane; generating a point cloud based on the distance data, the point cloud in a frame of reference of the sensor; identifying a basis vector representing a peak direction across the point cloud; transforming the point cloud into a frame of reference of the basis vector; and identifying a Z-plane in the transformed point cloud.


Another embodiment provides an imaging system that includes a TOF depth sensor and a processor. The TOF depth sensor obtains distance data describing distances between the TOF depth sensor and a plurality of surfaces in an environment of the TOF depth sensor. The processor receives the distance data from the TOF depth sensor; generates a point cloud based on the distance data, the point cloud in a frame of reference of the TOF depth sensor; identifies a basis vector representing a peak direction across the point cloud; transforms the point cloud into a frame of reference of the basis vector; and identifies a Z-plane in the transformed point cloud.


Yet another embodiment provides a method for determining dimensions of a physical box. The method includes receiving distance data describing distances between a sensor and a plurality of surfaces in an environment of the sensor, at least a portion of the surfaces corresponding to a box to be measured; transforming the distance data into a frame of reference of one of the surfaces in the environment of the sensor; selecting, from the plurality of surfaces in the environment of the sensor, a first surface corresponding to a top of the box and a second surface corresponding to a surface the box is resting on; calculating a height between the first surface and the second surface; and calculating a length and a width based on the selected first surface corresponding to the top of the box.


Another embodiment provides an imaging system including a TOF depth sensor and a processor. The TOF depth sensor obtains distance data describing distances between the TOF depth sensor and a plurality of surfaces in an environment of the TOF depth sensor. The processor receives the distance data from the TOF depth sensor; transforms the distance data into a frame of reference of one of the surfaces in the environment of the sensor; selects a first surface corresponding to a top of the box and a second surface corresponding to a surface the box is resting on; calculates a height between the first surface and the second surface; and calculates a length and a width based on the selected first surface corresponding to the top of the box.


As will be appreciated by one skilled in the art, aspects of the present disclosure, in particular aspects of identifying a Z-plane and determining box dimensions based on TOF imagery, described herein, may be embodied in various manners (e.g., as a method, a system, a computer program product, or a computer-readable storage medium). Accordingly, aspects of the present disclosure may take the form of a hardware embodiment, a software embodiment (including firmware, resident software, micro-code, etc.), or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Functions described in this disclosure may be implemented as an algorithm executed by one or more hardware processing units, e.g. one or more microprocessors, of one or more computers. In various embodiments, different steps and portions of the steps of each of the methods described herein may be performed by different processing units. Furthermore, aspects of the present disclosure may take the form of a computer program product embodied in one or more computer-readable medium(s), preferably non-transitory, having computer-readable program code embodied, e.g., stored, thereon. In various embodiments, such a computer program may, for example, be downloaded (updated) to the existing devices and systems (e.g. to the existing perception system devices and/or their controllers, etc.) or be stored upon manufacturing of these devices and systems.


The following detailed description presents various descriptions of specific certain embodiments. However, the innovations described herein can be embodied in a multitude of different ways, for example, as defined and covered by the claims and/or select examples. In the following description, reference is made to the drawings where like reference numerals can indicate identical or functionally similar elements. It will be understood that elements illustrated in the drawings are not necessarily drawn to scale. Moreover, it will be understood that certain embodiments can include more elements than illustrated in a drawing and/or a subset of the elements illustrated in a drawing. Further, some embodiments can incorporate any suitable combination of features from two or more drawings.


The following disclosure describes various illustrative embodiments and examples for implementing the features and functionality of the present disclosure. While particular components, arrangements, and/or features are described below in connection with various example embodiments, these are merely examples used to simplify the present disclosure and are not intended to be limiting. It will of course be appreciated that in the development of any actual embodiment, numerous implementation-specific decisions must be made to achieve the developer's specific goals, including compliance with system, business, and/or legal constraints, which may vary from one implementation to another. Moreover, it will be appreciated that, while such a development effort might be complex and time-consuming; it would nevertheless be a routine undertaking for those of ordinary skill in the art having the benefit of this disclosure.


In the Specification, reference may be made to the spatial relationships between various components and to the spatial orientation of various aspects of components as depicted in the attached drawings. However, as will be recognized by those skilled in the art after a complete reading of the present disclosure, the devices, components, members, apparatuses, etc. described herein may be positioned in any desired orientation. Thus, the use of terms such as “above”, “below”, “upper”, “lower”, “top”, “bottom”, or other similar terms to describe a spatial relationship between various components or to describe the spatial orientation of aspects of such components, should be understood to describe a relative relationship between the components or a spatial orientation of aspects of such components, respectively, as the components described herein may be oriented in any desired direction. When used to describe a range of dimensions or other characteristics (e.g., time, pressure, temperature, length, width, etc.) of an element, operations, and/or conditions, the phrase “between X and Y” represents a range that includes X and Y.


Other features and advantages of the disclosure will be apparent from the following description and the claims.


TOF System Overview



FIG. 1 is a block diagram of an example sensor system 100, according to some embodiments of the present disclosure. The sensor system 100 includes a TOF sensor 110, a processor 120, a camera 130, a display device 140, and a memory 150. In alternative configurations, different, fewer, and/or additional components may be included in the TOF sensor system from those shown in FIG. 1. Furthermore, the functionality described in conjunction with one or more of the components shown in FIG. 1 may be distributed among the components in a different manner than described. In some embodiments, some or all of the components 110-150 may be integrated into a single unit, e.g., a handheld unit having a TOF sensor 110, a processor 120 for processing the TOF data, a local memory 150, and a display device 140 for displaying an output of the processor 120 to a user. In some embodiments, some components may be located in different devices, e.g., a handheld TOF sensor 110 may transmit TOF data to an external processing system (e.g., a computer or tablet) that stores and processes the TOF data and provides one or more displays to a user. Different devices may communicate over wireless or wired connections.


The TOF sensor 110 collects distance data describing a distance between the TOF sensor 110 and various surfaces in the environment of the TOF sensor 110. The TOF sensor 110 may contain a light source, e.g., a laser, and an image sensor for capturing light reflected off the surfaces. In some embodiments, the TOF sensor 110 emits a pulse of light and capture multiple image frames at different times to determine an amount of time for the light pulse to travel to the surface and be returned to the image sensor. In other embodiments, the TOF sensor 110 detects phase shifts in the captured light, and the phase shifts indicate the distance between the TOF sensor 110 and various surfaces. In some embodiments, the TOF sensor 110 may generate and capture light at multiple different frequencies. If the TOF sensor 110 emits and captures light at multiple frequencies, this can help resolve ambiguous distances and help the TOF sensor 110 work at larger distance ranges. For example, if, for a first frequency, a first observed phase may correspond to a surface 0.5 meters, 1.5 meters, or 2.5 meters away, and, for a second frequency, a second observed phase may correspond to a surface 0.75 meters, 1.5 meters, or 2.25 meters away, by combining the two observations, the TOF sensor 110 can determine that the surface is 1.5 meters away. Using multiple frequencies may also improve robustness against noise caused by particular frequencies of ambient light, whether phase shift or pulse return time are used to measure distance. In alternate embodiments, different types of sensors may be used instead of and/or in addition to the TOF sensor 110 to obtain distance data.


The processor 120 receives distance data from the TOF sensor 110 and processes the distance data to identify various features in the environment of the TOF sensor 110, as described in detail herein, e.g., with respect to FIGS. 3-8. In some embodiments, the distance data includes the observed distances to various surfaces measured by the TOF sensor 110 using, e.g., the phase shift or pulse return time methods described above. In some embodiments, if the TOF sensor 110 measures phase shifts, the distance data received by the processor 120 from the TOF sensor 110 is the phase shift data, and the processor 120 calculates the distances to the surfaces from the phase shift data.


A camera 130 may capture image frames of the environment. The camera 130 may be a visual light camera that captures images of the environment in the visible range. In other embodiments, the camera 130 is an infrared (IR) camera captures IR intensities of the surfaces in the sensor system's environment. The field of view of the camera 130 and TOF sensor 110 are partially or fully overlapping, e.g., the field of view of the camera 130 may be slightly larger than the field of view of the TOF sensor 110. The camera 130 may pass captured images to the processor 120. In some embodiments, two processors or processing units may be included, e.g., a first processing unit for performing the Z-plane identification and box dimensioning algorithms described herein, and a second graphical processing unit that receives images from the camera 130 and generates displays based on the images and data from the first processing unit. In some embodiments, image data from the camera 130 may be used to determine a level of sunlight in the environment of the TOF sensor 110. In alternate embodiments, the sensor system 100 may include a separate light sensor for detecting sunlight or other ambient light conditions in the environment of the TOF sensor 110.


The display device 140 provides visual output for a user of the sensor system 100. For example, the display device 140 may display box dimensions and/or a box volume calculated by the processor 120 based on distance data from the TOF sensor 110. In some embodiments, the display device 140 displays an image obtained by the camera 130 and overlays visual imagery indicating one or more features identified in the field of view of the camera 130 and TOF sensor 110 based on the distance data. For example, the processor 120 may instruct the display device 140 to display an outline of a box over an image of the box obtained by the camera 130. A user can use this display to determine whether the sensor system 100 has correctly identified the box and the box's edges. The sensor system 100 may include additional or alternative input and/or output devices, e.g., buttons, a speaker, a touchscreen, etc.


The memory 150 stores data for the sensor system 100. For example, the memory 150 stores processing instructions used by the processor 120 to identify features in the environment of the TOF sensor 110, e.g., instructions to identify one or more Z-planes and/or to calculate box dimensions of an observed box. The memory 150 may temporarily store data and images obtained by the camera 130 and/or TOF sensor 110 and accessed by the processor 120. The memory 150 may further store image data accessed by the display device 140 to generate an output display.



FIG. 2 illustrates ray directions of pixels of the TOF sensor 110, according to some embodiments of the present disclosure. The distance data obtained by the TOF sensor 110 may be arranged as a set of pixels, e.g., the pixels 210a and 210b, within an image frame, e.g., the image frame 220. Each pixel 210 has an associated ray direction 215, where the ray direction 215 points outwards from the TOF sensor 110. The ray directions 215 are projected towards the image frame 220. While 25 rays and pixels are shown in FIG. 2, it should be understood that the TOF sensor 110 may have many more pixels. While the image frame 220 has a square shape in the example shown in FIG. 2, the image frame 220 may have other shapes in other embodiments. In some examples, certain pixels, e.g., pixels near the edge of the image frame 220, may not be considered valid (e.g., not sufficiently reliable), and are removed from the distance data.


For example, a first pixel 210a has a ray direction 215a that extends straight out from the TOF sensor 110; the pixel 210a is in the center of the image frame 220. A second pixel 210b at a corner of the image frame 220 is associated with a ray direction 215b that extends out from the TOF sensor 110 at, for example, a 30° angle in both an x-direction and y-direction from the center of the image frame 220, where the image frame 220 is an x-y plane in a frame of reference of the TOF sensor 110. The TOF sensor 110 returns distance data (e.g., a distance, one or more phase shifts) to a surface along each valid pixel's ray. In one example, the first pixel 210a may have a measured distance of 1 meter representing a distance to a particular point on a box, and the second pixel 210b may have a measured distance of 2 meters representing a distance to particular point on a wall behind the box.


Example Process for Identifying a Z-Plane



FIG. 3 is a flow diagram showing a process 300 for identifying a Z-plane in TOF data obtained in an arbitrary frame of reference, according to some embodiments of the present disclosure. The TOF sensor 110 captures 310 distance data of an environment, including various surfaces in the environment. It may be assumed that at least one of the surfaces is a Z-plane. The TOF sensor 110 passes the distance data to the processor 120. In some examples, the camera 130 captures an image of the environment, e.g., at the same time that the TOF sensor 110 captures the distance data, and the camera 130 passes the image to the processor 120.



FIGS. 9 and 10 show two example visual representations of the inputs from the camera 130 and the TOF sensor 110. FIG. 9 is an example image showing a box resting on a tabletop, according to some embodiments of the present disclosure. In this example, FIG. 9 shows an IR intensity image that shows a box 910 sitting on a table 920, with a chair 930 and a floor 940 to the left of the table. The IR intensity image may be used during visualization (e.g., to visualize the locations of the extracted world Z-planes, or for visualizations of other applications, such as box dimensions).



FIG. 10 illustrates an example of distance data obtained by a TOF sensor, according to some embodiments of the present disclosure. The field of view of the TOF sensor 110 shown in FIG. 10 corresponds to the field of view of the IR intensity image shown in FIG. 9. The distance data is represented by shading, where different shades represent difference distances from the TOF sensor 110 to the various surfaces in the environment of the TOF sensor 110, e.g., the box 910 is closer to the TOF sensor 110 than the floor 940. As described with respect to FIG. 2, the distance data is comprised of various pixels having the ray directions 215 shown in FIG. 2.


In some embodiments, the processor 120 filters 320 the received distance data. Ambient light in the environment of the TOF sensor 110 can create noise in the distance data. To reduce the effect of noise, a filter, e.g., an integral filter, may be applied to the distance data before further analysis is performed. Filtering the noise in this manner may be particularly useful if the TOF sensor 110 captures data in an outdoor environment, due to the noise caused by sunlight. To filter the distance data, the processor 120 may compute, for each pixel, an average pixel value based on pixel values in a region around the pixel. For example, the filtered pixel value for a given pixel may be the average value for an 11×11 or 21×21 square of pixels centered on the given pixel. In some embodiments, the processor 120 performs the filtering on phase measurement data received from TOF sensor 110, e.g., the processor 120 first filters multiple phase measurements (for different frequencies, as described above), and then processes the filtered phase data to determine a distance measurement for each pixel. Alternatively, the processor 120 may filter the distance measurements, e.g., if the pulse return method is used to obtain the distance data.


In some embodiments, the filtering step may be omitted, e.g., if, the TOF sensor 110 is intended for use in an environment with relatively low noise levels, e.g., if the TOF sensor 110 is designed for indoor use only. In some cases, the processor 120 may perform filtering in response to determining that there is a threshold level of sunlight in the environment of the TOF sensor 110. Furthermore, in some embodiments, the processor 120 may perform adaptive filtering based on the type or level of ambient light in the environment of the TOF sensor 110, e.g., using a larger filter window when brighter sunlight is detected, using a larger filter window when a greater frequency distribution is detected in the ambient light, or using a larger filter window when particular frequencies known to interact with the TOF sensor 110 are detected in the ambient light.


The processor 120 generates 330 a point cloud based on the distance data and the pixel ray directions 215. For example, for each individual pixel, the processor 120 multiplies the ray direction 215 for the pixel by the measurement distance to the surface for that pixel, e.g., the measurement distance shown in FIG. 10. The processor 120 may retrieve the ray directions 215 from the memory 150.


The point cloud is in the reference frame of the TOF sensor 110, also referred to as the ego frame. For example, if a user is holding the TOF sensor 110 at a slight angle relative to the ground in the environment, the Z-direction in the reference frame of the TOF sensor 110 does not align with the Z-planes in the environment (e.g., the Z-direction in the reference frame is angled relative to the direction of gravity, if the Z-planes are perpendicular to the direction of gravity).



FIG. 11 illustrates an example point cloud calculated from the distance data, according to some embodiments of the present disclosure. Because the TOF sensor 110 was at an arbitrary angle relative to the Z-planes in the environment when the distance data was captured, the resulting point cloud in the reference frame of the TOF sensor 110 is difficult for humans to interpret. This point cloud is also challenging for computer algorithms to work with in various applications that make use of TOF data.


The processor 120 identifies 340 basis vectors for a frame of reference of the surfaces in the environment, also referred to as a “world” frame of reference. A first basis vector corresponds to the direction perpendicular to the Z-planes in the environment observed by the TOF sensor 110. Second and third basis vectors are each orthogonal to the first basis vector. The basis vectors define a “world” coordinate system, i.e., a coordinate system in which the Z-planes are horizontal.



FIG. 4 is a flow diagram showing an example process for identifying basis vectors based on TOF distance data, according to some embodiments of the present disclosure. From the three-dimensional point cloud in the ego frame (e.g., the point cloud shown in FIG. 11), the processor 120 computes 410 surface normal vectors (also referred to as surface normals) for the points within the point cloud. To compute the surface normal for a given point, the processor 120 may fit a plane to a set of points in a region around the individual point, and the processor 120 then computes the surface normal to the fitted plane. For flat surfaces (e.g., floors, box surfaces, walls), the surface normals associated with the point cloud are fairly uniform, with some noise variation. As noted above, the filtering 320 can reduce the noise variation in the surface normals. The surface normals may be represented by polar and azimuthal angles in a polar coordinate system. In other embodiments, Cartesian coordinate systems may be used to represent the surface normal vectors.



FIGS. 12A and 12B illustrate example angular coordinates of the surface normals of the points in the point cloud shown in FIG. 11, according to some embodiments of the present disclosure. In particular, FIG. 12A shows the polar angles of each of the computed surface normals, and FIG. 12B shows the azimuthal angles of each of the computed surface normals. As these figures illustrate, the Z-planes, which correspond to the top of the box 910, the 920 table, and chair 930, have consistent surface normals across their surfaces, with some variations due to noise in the distance data. Moreover, because each of these objects are flat along a Z-plane, they each have a similar surface normal (represented by the similar shade across these Z-planes in each of the two images). The front face of the box 910, by contrast, has an orthogonal surface normal relative to the Z-planes, as represented by the darker shading in the front face of the box 910 in FIG. 12B.


Having computed the surface normals, the processor 120 extracts one or more basis vectors based on the computed surface normals. To extract the first basis vector, the processor 120 may bin 420 the coordinates of the surface normals, e.g., the processor 120 bins the polar and azimuthal angles for each of the computed surface normals. This binning results a two-dimensional distribution. The bins may be represented visually by a histogram, e.g., the two-dimensional histogram that bins angular coordinates of the surface normals shown in FIG. 13. The two-dimensional histogram shown in FIG. 13 has a strong peak 1310 (represented by the dark shading) corresponding to the Z-plane direction vector in the ego frame of the TOF sensor 110. The processor 120 identifies 430 the peak coordinates, e.g., the peak azimuthal angle and the peak polar angle in the two-dimensional distribution of the binned coordinates. The processor 120 defines 440 the first basis vector as the direction vector corresponding to the peak direction (e.g., the peak azimuthal angle and the peak polar angle) of the surface normals across the point cloud, which is the surface normal to the Z-planes in the environment of the TOF sensor 110.


Having selected the first basis vector, processor 120 selects 450 the second and third basis vectors. The second and third basis vectors are orthogonal to the first basis vector (i.e., orthogonal to the surface normal to the Z-planes). The second and third basis vectors are also orthogonal to each other. The first, second, and third basis vectors define the world frame of reference.


In some embodiments, the processor 120 calculates a projection of the TOF sensor's pointing direction (e.g., the ray direction 215b, which extends straight out from the TOF sensor 110) into a Z-plane (e.g., a plane orthogonal to the first basis vector), and the processor 120 selects this projection as the second basis vector. The processor 120 selects the vector orthogonal to the first and second basis vectors as a third basis vector; the processor 120 may compute the third basis vector as the cross product of the first basis vector and the second basis vector. In other embodiments, the second and third basis vectors may be chosen in other ways.


Returning to FIG. 3, having identified the basis vectors, the processor 120 transforms 350 the point cloud to a frame of reference of the basis vectors. For example, each point in the untransformed point cloud may be defined as a vector (e.g., the product of the ray direction and the measured distance, where the ray direction is a vector in the reference frame of the TOF sensor 110, as described above). In the transformed point cloud, each point may be defined as a linear combination of the basis vectors. In particular, the processor 120 may define each point as a sum of the basis vectors each multiplied by a scalar, where the processor 120 determines each of the scalars by computing an inner product of the point in the untransformed point cloud (in vector notation) and the basis vector. FIG. 14 illustrates an example of the point cloud transformed into a reference coordinate system of the identified basis vectors, according to some embodiments of the present disclosure. The transformed point cloud is easier for humans to interpret that the point cloud shown in FIG. 11. The box 910, table 920, chair 930, and floor 940 are visible in the transformed point cloud. Moreover, the transformed point cloud is easier for the processor 120 to work with for further computations (e.g., identifying Z-planes, identifying boxes, and determining box dimensions, as described further below) than the untransformed point cloud in the ego frame. For example, aligning the point cloud to the reference frame of the Z-planes simplifies the steps of identifying and isolating the Z-planes.


The processor 120 next identifies 360 Z-planes in the transformed point cloud. FIG. 5 is a flow diagram showing an example process for identifying a Z-plane in a point cloud transformed based on the identified basis vectors, according to some embodiments of the present disclosure. First, the processor 120 generates 510 a height map from the transformed point cloud. For example, the processor 120 distributes the points the transformed point cloud into square “chimneys,” and subsequently selects a representative height for each chimney. Each chimney may be a shape of the same size, e.g., a (0.75 cm)2 square. Other sizes or shapes may be used to construct the height map. The representative height may be, for example, the top point (maximum height) of the chimney, an average height, a median height, or another height selected or computed from the heights of the points falling within the chimney. Reducing the three-dimensional point cloud down to a two-dimensional height map simplifies data processing and increases computation speed.



FIG. 15 is an example visual representation of a height map obtained from the transformed point cloud, according to some embodiments of the present disclosure. The shading in FIG. 15 represents the height of each chimney. For example, the lighter shading of the at the top of the box 910 represents a greater height than the darker shading of the table 920.


The processor 120 then generates 520 a profile representation of the height map. For example, the processor 120 integrates the height map over the x- and y-directions to obtain a Z-profile of the height map. This profile represents the probability density of heights within the height map. Peaks within the profile correspond to various flat surfaces, i.e., surfaces that are orthogonal to the first basis vector.



FIG. 16 is an example Z-profile of the height map with peaks indicating various horizontal surfaces, according to some embodiments of the present disclosure. FIG. 16 includes four peaks 1610, 1620, 1630, and 1640. In this example, the peak 1610 at the lowest height corresponds to the floor 940. The next peak 1620 corresponds to the chair 930. The next peak 1630, which is also the tallest peak (indicating that the most points within the height map fall at this peak), corresponds to the table 920. Finally, the last peak 1640 corresponds to the top of the box 910.


The processor 120 identifies 530 the peaks in the profile representation of the height map as Z-planes. For example, for each portion of the Z-profile where the probability density falls above a given threshold (e.g., 0.01), the processor 120 identifies a Z-plane. In some embodiments, the processor 120 applies one or more other rules or heuristics to the Z-profile to identify the Z-planes, e.g., to remove spurious noise peaks while retaining genuine weak signals, such as the floor peak shown on page 9. For example, the processor 120 may identify peaks having at least a threshold number of associated points, or peaks in which the heights fall within a given range of each other.


The processor 120 may select a particular height point within a given peak (e.g., a highest point or a center point) as the Z-plane height for that peak. In some embodiments, the processor 120 selects the lowest Z-value peak as the height of the base Z-plane. The processor 120 may set the height of the base Z-plane to zero, and determine the heights of the other Z-planes based on their height relative to the base Z-plane. For example, the processor 120 sets the height of the floor peak 1610 as 0, and the processor 120 determines that the chair peak 1620 has a height of 0.569 meters, the table peak 1630 has a height of 0.815 meters, and the box top peak 1640 has a height of 0.871 meters.


Having identified the Z-planes and their associated heights, the processor 120 associates 540 various points in the point cloud with the identified Z-planes. For example, the processor 120 may associate a particular point in the point cloud with an identified Z-plane if the height of the point is within a height range of the identified Z-plane. For example, if the height of a particular point is within twice a peak's FWHM (full width at half maximum), the point is associated with the Z-plane corresponding to the peak. In other examples, other ranges around a Z-plane height are used to associate points in the point cloud to Z-planes.



FIG. 17 illustrates four example Z-slices in the point cloud identified from the height map, according to some embodiments of the present disclosure. Each set of points that the processor 120 associated with a particular Z-plane may be referred to as a Z-plane slice, or simply a Z-slice. Each of the Z-slices is represented with a different shading, where the differently shaded Z-slices correspond to Z-planes at different heights.


The transformed point cloud and identified Z-planes can be used for various further processing on the point cloud data. In some examples, the processor 120 can proceed to locate a box in the environment of the TOF sensor 110 and determine dimensions and/or the volume of the box, as described further below. In other examples, the processor 120 can perform other types of identification or analysis on other types of objects in the environment of the TOF sensor 110.


In some embodiments, the sensor system 100 displays outputs of the Z-plane identification process to a user. For example, the processor 120 may correlate the identified Z-planes to various pixels in an image obtained by the camera 130, and generate a display with visual indications of the identified Z-planes. For example, the Z-planes may be outlined or color-coded in a display output by the display device 140. The display device 140 may alternatively or additionally output the determined heights of the identified Z-planes.


Example Process for Box Dimensioning



FIG. 6 is a flow diagram showing a process 600 for determining and outputting box dimension based on TOF data, according to some embodiments of the present disclosure. A sensor system, such as the sensor system 100, receives 610 distance data describing various surfaces in the environment of the sensor system 100. For example, as described with respect to steps 310-330 in FIG. 3, the TOF sensor 110 captures distance data, the processor 120 optionally filters the distance data, and the processor 120 generates a point cloud based on the distance data.


For the box dimensioning process, the processor 120 may assume that at least a portion of the surfaces in the environment of the TOF sensor 110 correspond to a box to be measured. Several additional assumptions may be made about the box being measured. Such assumptions can improve the speed and accuracy of the box dimensioning process, particularly in applications where fast detection and measuring of the box are important, e.g., if box dimensions are calculated provided to a user in real or near-real time as the user points the TOF sensor 110 at a box. These assumptions may include that the angles between adjacent surfaces of the box are reasonably close to 90° (e.g., between 85° and 95°, or within some other range); that the box is located within a particular distance of the TOF sensor 110 (e.g., within 3 meters or within 5 meters); that each box dimension is within a particular range (e.g., at least 3 centimeters, or at least 10 centimeters; no more than 2 meters, or no more than 3 meters); that the box is closed; that the top face of the box is visible to the TOF sensor 110; and that the box rests on a flat, horizontal surface (i.e., a Z-plane) that is also visible to the TOF sensor 110.


It should be understood that, in some embodiments, one or more of these assumptions may be relaxed or removed. For example, the range between the sensor and box and the minimum and maximum box dimensions are merely exemplary, and in other embodiments, different ranges and dimensions may be used. In some embodiments, the ranges may vary on the intended uses or target users of the sensor system 100. For example, if the sensor system 100 is used to measuring boxes being loaded into a moving truck (e.g., including wardrobe boxes and boxed furniture), greater distance ranges and greater box dimensions may be used. In some embodiments, a user may be able to input a distance range and/or maximum and minimum box dimensions.


The processor 120 transforms 620 the distance data (e.g., the point cloud calculated based on distance data from the TOF sensor 110) into a frame of reference of a surface in the environment of the TOF sensor 110. For example, as described with respect to steps 340 and 350 in FIG. 3, the processor 120 identifies basis vectors for a frame of reference of the surfaces in the environment of the TOF sensor 110, and the processor 120 transforms the distance data (e.g., the point cloud) into the frame of reference of the basis vectors. A process for transforming the point cloud into the frame of reference of the basis vectors is described in greater detail with respect to FIG. 4. The processor 120 may further identify Z-planes in the transformed distance data, as described with respect to step 360 of FIG. 3 and in greater detail with respect to FIG. 5.


As noted with respect to FIGS. 5 and 17, each of the Z-planes may be represented as a Z-slice of the height map data. The processor 120 selects 630 a surface corresponding to the box top and a surface corresponding to the box bottom based on the height map data. For example, the processor 120 identifies one of the Z-slices as containing the box top, and another of the Z-slices as a surface on which the box is resting (e.g., the floor or a table), which corresponds to the box bottom.



FIG. 7 is a flow diagram showing an example process for identifying the box top and the box bottom, according to some embodiments of the present disclosure. The processor 120 generates 710 a height map based on the distance data. For example, the processor 120 may generate the height map as described with respect to step 510 of FIG. 5. The processor 120 then identifies 720 Z-slices in the height map. For example, the processor 120 generates a profile representation of the height map, identifies Z-planes as peaks in the profile representation of the height map, and associates various points in the distance data (e.g., in the point cloud) with the Z-planes, as described with respect to steps 520-540 of FIG. 5. As noted above, each set of points associated with a particular Z-plane is referred to as a Z-slice.


The processor 120 identifies 730 connected components within at least some of the Z-slices. Each of the connected components may be a candidate box top. To identify a connected component, the processor 120 finds clusters of nearby or connecting points within a Z-slice. The processor 120 may identify a connected component by finding sets of pixels in a Z-slice that may be reached by moving across the Z-slice, e.g., pixels that are within a threshold distance of each other. For example, the processor 120 may select a particular pixel in a Z-slice and recursively add neighboring pixels that are also in the Z-slice to a connected component. Each connected component has a respective height along the Z-axis in the frame of reference of the basis vectors; the height corresponds to the height of the Z-slice.



FIGS. 18A-18B illustrate two sets of connected components for two different Z-plane slices, according to some embodiments of the present disclosure. In particular, FIG. 18A illustrates the connected components within the 81.5 cm Z-slice, and FIG. 18B illustrates the connected components within the 87.1 cm Z-slice. Each respective connected component is assigned a different shading. If a box exists in the distance data obtained by the TOF sensor 110, is expected that one of the connected components corresponds to the box top.


In some embodiments, prior to identifying the connected components in the Z-slices, the processor 120 may apply one or more rules to remove one or more Z-slices from consideration as the box top. For example, the processor 120 may eliminate the lowest or base Z-slice as potentially containing the box top, since it is assumed that the box top is above the lowest surface. The processor 120 may also remove a Z-slice that does not lie sufficiently close within the height map (in the x- and y-direction) to some other, lower Z-slice (i.e., a potential surface for the box to be resting on). For example, for the Z-plane slices shown in FIG. 17, the processor 120 may remove the Z=56.9 cm slice from consideration because it is not sufficiently close to the only Z-slice that is lower than it, the Z=0 cm slice. By contrast, the 56.9 cm Z-slice is sufficiently close to the Z=81.5 cm Z-slice that the Z=81.5 cm Z-slice cannot be eliminated as potentially containing a box top, with the 56.9 cm Z-slice being the surface holding the box bottom.


Having identified the connected components representing candidate box tops, the processor 120 selects 740 one of the connected components as the box top. The processor 120 may apply various rules to the connected components to identify the box top. For example, the processor 120 may remove connected components that are very small (e.g., having a width and/or length below the threshold minimum box dimensions described above). The processor 120 may remove connected components that are highly elongated or non-compact (e.g., the connected component has a large perimeter compared to the square root of its area). The processor 120 may remove connected components for which a box bottom (i.e., the surface on which the box is resting) cannot be derived from the height map, e.g., because no other connected component or Z-slice is sufficiently close to the connected component in the height map in the x- and y-direction.


Having applied these rules to the connected components shown in FIGS. 18A and 18B, two connected components, shown in FIGS. 19A and 19B, remain as candidate box tops. FIGS. 19A and 19B illustrate two candidate box tops identified in the connected components. In particular, FIG. 19A shows a connected component 1910a in the Z=81.5 cm slice, and FIG. 19B shows a connected component 1910b in the Z=87.1 cm slice. To select one of the remaining connected components as the box top, the processor 120 applies an additional rule that considers the shape of the convex hull polygon enclosing the connected component, e.g., how closely the convex hull polygon matches an expected rectangular shape. The convex hull polygons 1920a and 1920b are drawn around each of the connected components in FIGS. 19A and 19B. The convex hull polygon 1920a in FIG. 19A strongly deviates from a rectangular shape, while the convex hull polygon 1920b in FIG. 19B is very nearly a rectangle, e.g., the convex hull polygon 1920b deviates from an expected rectangular shape by less than a threshold deviation. Thus, the processor 120 selects the rectangular connected component in the 87.1 cm Z-slice as the box top in this example.


While several example rules for identifying a box top are discussed above, the processor 120 may apply additional, fewer, or different rules to the connected components to identify the box top in different embodiments. In some embodiments, if multiple candidates pass each of the rules described above, the processor 120 may use an additional rule or rules for choosing between the possible box tops. For example, the processor 120 may select the candidate box top that is closest to the TOF sensor 110.


Having identified the box top, the processor 120 identifies 750 the surface on which the box is resting, which corresponds to the box bottom. For example, the processor 120 selects a Z-slice that has a lower height than the box top and that is closest in lateral range to the box top in the height map, e.g., the Z-slice that is closest to the identified box top in the x- and y-directions in the height map. In the example height map shown in FIG. 17, the identified box top is resting on the Z=81.5 cm slice. This also corresponds to the height of the box bottom.


Returning to FIG. 6, having identified the box top and the box bottom, the processor 120 calculates 640 the box height from the box top to the box bottom. The box height is the difference between the respective Z-slice heights of the box top and the box bottom, e.g., 87.1 cm-81.5 cm=5.6 cm.


The processor 120 further 650 calculates the length and the width of the box based on the selected box top. For example, having identified the box top, the processor 120 calculates the length and width of the box top. Because the distance data for the box top is taken at an angle and may have noise, the processor 120 may filter the box top data, rotate the box top so it is aligned with an x-axis and a y-axis, and calculate horizontal and vertical profiles of the edges to determine the length and width. For example, the trailing edge of the box (i.e., the edge of the box farthest from the TOF sensor 110) may be blurred, which can make it difficult for the processor 120 to identify the trailing edge without performing additional data processing.



FIG. 8 is a flow diagram showing an example process for calculating the length and width of the box top, according to some embodiments of the present disclosure. In some embodiments, the processor 120 filters the box top data, e.g., the distance data corresponding to at least the points in the distance data that correspond to the identified box top. In some embodiments, the processor 120 filters all of the distance data. As described with respect to FIG. 3, ambient light in the environment of the TOF sensor 110 can create noise in the distance data. To reduce the effect of noise, a filter, e.g., an integral filter, may be applied to the distance data before the box length and width are calculated. Filtering the noise in this manner may be particularly useful if the TOF sensor 110 captures data in an outdoor environment, due to the noise caused by sunlight.


To filter the distance data, the processor 120 may compute, for each pixel in the distance data, an average pixel value based on pixel values in a region around the pixel. The processor may use a different filter than the filter described with respect to step 320. In particular, the processor 120 may use a smaller filter window than the filter used to identify the Z-planes. For example, the filtered pixel value for a given pixel may be the average value for an 5×5 or 7×7 square of pixels centered on the given pixel. As described with respect to FIG. 3, the processor 120 may perform the filtering on phase measurement data received from TOF sensor 110, e.g., the processor 120 first filters multiple phase measurements (for different frequencies, as described above), and then processes the filtered phase data to determine a distance measurement for each pixel. Alternatively, the processor 120 may filter the distance measurements, e.g., if the pulse return method is used to obtain the distance data.


In some embodiments, the filtering step may be omitted, e.g., if, the TOF sensor 110 is intended for use in an environment with relatively low noise levels, e.g., if the TOF sensor 110 is designed for indoor use only. In some cases, the processor 120 may perform filtering in response to determining that there is a threshold level of sunlight in the environment of the TOF sensor 110. Furthermore, in some embodiments, the processor 120 may perform adaptive filtering based on the type or level of ambient light in the environment of the TOF sensor 110, e.g., using a larger filter window when brighter sunlight is detected, using a larger filter window when a greater frequency distribution is detected in the ambient light, or using a larger filter window when particular frequencies known to interact with the TOF sensor 110 are detected in the ambient light.


The processor 120 extracts 820 a subset of points within the transformed distance data corresponding to the box top, e.g., the connected component selected as the box top at step 740. If the filtering 810 is performed, the processor 820 may calculate a second point cloud based on the filtered data (following the process described in step 330 in FIG. 3), transform the second point cloud (following the process described with respect to steps 340-350 in FIG. 3 and with respect to FIG. 4), and extract the points corresponding to the connected component in the second point cloud. The processor 120 may use the same basis vectors selected during the Z-plane identification stage to transform the point cloud based on the newly filtered data.



FIG. 20 illustrates a set of points corresponding to a connected component identified as the box top, according to some embodiments of the present disclosure. This set of extracted points is also referred to as a box top subcloud. To simplify processing and understanding of the box top, the processor 120 may determine an angle of rotation for the extracted box top subcloud and rotate 830 the box top subcloud by the angle of rotation so that the edges of the box top are aligned with the x- and y-axes. For example, the processor 120 projects points in the box top subcloud onto the x- and y-axes as a function of subcloud rotation angle about its center. The processor 120 then calculates the sum of the x- and y-projections as a function of rotation angle to generate a profile. FIG. 21 is an example profile of the box top subcloud projected along the x- and y-axes, according to some embodiments of the present disclosure. The processor 120 identifies the azimuthal rotation angle for which the sum of the box top projections is minimized. More particularly, the selected rotation angle minimizes a sum of the projections of the edges of the box top onto a set of axes of the previously determined frame of reference, e.g., the frame of reference of the basis vectors. The processor 120 rotates the box top subcloud through the identified azimuthal angle so that the box top subcloud is axis-aligned. FIG. 22 illustrates an axis-aligned box top rotated based on the profile in FIG. 21, according to some embodiments of the present disclosure.


Having rotated the box top subcloud, the processor 120 calculates 840 a width profile and a length profile for the box top. For many TOF sensors, while the leading edges of the box closest to the sensor are sharp and easy for both humans and computers to identify, flying voxels may blur the trailing edges of the box located farther downstream from the TOF sensor 110. This causes the location of the trailing edges to be more ambiguous and difficult to identify from the distance data. The processor 120 generates boxtop width and length profiles by projecting the points of the rotated box top subcloud onto the horizontal and vertical axes. FIGS. 23A and 23B show example box top width and length profiles, respectively, according to some embodiments of the present disclosure.


The processor 120 identifies 850 the leading edges and trailing edges in the width and length profiles. For example, the processor 120 applies one or more rules to identify the edges from the profiles. The processor 120 may fit lines to each of the profiles' interiors and define a leading edge as a location where the profile equals a set percentage of the linear fit's value, e.g., 40% of the linear fit's value. The processor 120 may define the trailing edge by a location where the profile equals the same percentage or different percentage of the linear fit's value, e.g., a particular value in the range of 25% to 85%. The processor 120 may further apply one or more rules to determine the trailing edge fraction. For example, trailing edge percentage thresholds may vary with the height of the box. Alternatively, percentage thresholds can differ for the shorter and longer of the two box top edges. FIGS. 23A and 23B show examples of the trailing edges and leading edges identified based on the width and length profiles.


The processor 120 calculates 860 the width and length of the box top based on the determined leading edges and trailing edges. In particular, the width is the distance between the leading edge and trailing edge in the width projection, and the length is the distance between the leading edge and trailing edge in the length projection. FIGS. 23A and 23B indicate the width and length, respectively, between the trailing edges and leading edges.


Returning to FIG. 6, the sensor system 100 (e.g., the processor 120 and the display device 140) may display 660 the determined box dimensions to a user. For example, the processor 120 may generate a display for output on the display device 140 that includes a visual representation of the box along with the height, width, and length. For example, the display may show the identified edges and/or dimensions projected onto an image captured by the camera 130, or an image created based on the distance data from the TOF sensor 110. In one embodiment, the processor 120 generates an image of the three-dimensional box defined by the identified leading and trailing edges, the identified box top and box bottom surfaces, and/or the calculated height, width, and length. The processor 120 may also calculate the volume (length×width×height) and output the volume on the display device 140.


The processor 120 may project the image of the three-dimensional box onto a two-dimensional image plane of the camera 130 to generate an overlay image, e.g., an outline overlaying the image of the box. The calculated width, length, and height dimensions may also be reported in the graphical display, either along the edges or in a separate area. A user can view the graphical display in the display device 140 to qualitatively confirm that the sensor system 100 has correctly identified the box and correctly identified the edges and surfaces.



FIG. 24 illustrates identified box edges overlayed on an image obtained based on data from the TOF sensor, according to some embodiments of the present disclosure. The image in FIG. 24 may be an image generated by the processor 120 based on the point cloud generated from the distance data from the TOF sensor 110. FIG. 24 also includes an outline of the box edges superimposed on the point cloud image.



FIG. 25 illustrates identified box edges and determined box dimension overlayed on an image obtained by a camera, according to some embodiments of the present disclosure. In this example, the image may be an IR image obtained by an IR camera. FIG. 24 further includes an outline of the box edges superimposed on the IR image, and displays the calculated width, length, and height for the box in the upper-left of the display. FIG. 24 also displays an intersection over union (IoU) score. In some embodiments, the processor 120 calculates an IoU score that measures an overlap between the box top and a pre-defined circle 2510 appearing in the center of the field of view of the TOF sensor 110. A larger IoU score correlates with a higher-accuracy dimensioning result, and a user may adjust the view of the TOF sensor 110 to obtain a higher IoU score. In some embodiments, the sensor system 100 may set a lower IoU bound for reporting the box dimensions, e.g., the processor 120 displays the box dimensions if the IoU score is greater than 0.40 or another threshold, and displays a request to the user to move the TOF sensor 110 if the IoU score is lower than the threshold. Ensuring that a user is orienting the TOF sensor 110 relative to the box with a sufficiently high IoU can reduce errors in the box dimensions reported by the sensor system 100.


In some embodiments, the sensor system 100 may additionally or alternatively report an intensity indicator that indicates a measured intensity at a particular pixel or across a set of pixels in the distance data collected by the TOF sensor 110 and/or a measured intensity in the corresponding pixel or set of pixels collected by the camera 130. In some cases, if the measured intensity in an area of interest in the image frame 220 is too low, it may be difficult for the processor 120 to find Z-planes, determine the box top dimensions, or perform other processing of the TOF distance data. The processor 120 can analyze the intensity of at least a portion of the sensor system's field of view and report the intensity to the user. Based on the reported intensity, the user may determine whether to adjust the environment, e.g., by changing lighting conditions, by changing the angle of the TOF sensor 110 relative to the box or other area of interest, by moving the box to a different location (e.g., onto a different Z-plane, into another room), etc., in order to increase the intensity. In some embodiments, if the processor 120 determines that the intensity is too low (e.g., the intensity is below a given threshold and/or the processor 120 is having difficulty finding Z-planes or the box, e.g., none of the identified connected components satisfies the rules for identifying the box top), the processor 120 may output an instruction to the user to make a change to the environment, sensor position, or location of the box to increase the intensity.


For example, if the camera 130 is an IR camera, the processor 120 may determine an IR intensity for at least a portion of the camera's field of view, e.g., at or near the center of the image frame of the camera 130. If the camera 130 is a visible light camera, the processor 120 may determine an intensity or brightness of the visible light at or near the center of the image frame. The intensity measurement may be correlated with the reflectivity of the material(s) in a given region, e.g., a reflectivity of a box material. Since a user typically points the TOF sensor 110 at a box, Z-plane, or other area of interest, and may be encouraged to include a box top in the center of the image frame by the IoU (as described above), the center of the image frame of the camera 130 typically corresponds to the box top, other portion of a box, Z-plane, or other area of interest.


As a particular example, the camera 130 captures an image frame with an area corresponding to the image frame 220. The processor 120 may identify, in the image frame captured by the camera 130, an intensity near the center of the image frame, e.g., an intensity at a location corresponding to the pixel 215a in the center of the image frame 220 of the TOF sensor 110, or an average intensity for set of pixels including the center of the image frame. For example, the processor 120 may determine an average intensity for a set of pixels corresponding to the circle 2510 shown in FIG. 25.


Select Examples

The following paragraphs provide various examples of the embodiments disclosed herein.


Example 1 provides a method for identifying a Z-plane, the method including receiving distance data describing distances between a sensor that captured the distance data and a plurality of surfaces in an environment of the sensor, where at least one of the surfaces is a Z-plane; generating a point cloud based on the distance data, the point cloud in a frame of reference of the sensor; identifying a basis vector representing a peak direction across the point cloud; transforming the point cloud into a frame of reference of the basis vector; and identifying a Z-plane in the transformed point cloud.


Example 2 provides the method of example 1, where the sensor is a TOF sensor including a light source and an image sensor.


Example 3 provides the method of example 1, where the distance data is arranged in a plurality of pixels within an image frame of the sensor.


Example 4 provides the method of example 3, where an individual pixel has a distance to one of the plurality of surfaces in the environment of the sensor, and the individual pixel has an associated ray direction describing a direction from the sensor to the surface.


Example 5 provides the method of example 4, where generating the point cloud involves multiplying the ray direction for the individual pixel by the distance to the one of the plurality of surfaces for the individual pixel.


Example 6 provides the method of example 1, where the distance data is arranged as a plurality of pixels, the method further including filtering the distance data by computing, for an individual pixel, an average pixel value based on pixel values in a region around the individual pixel.


Example 7 provides the method of example 1, where identifying the basis vector includes computing surface normals for points in the point cloud; and extracting the basis vector based on the computed surface normals, the basis vector representing the peak direction of the surface normals across the point cloud.


Example 8 provides the method of example 7, where computing the surface normal for points in the point cloud includes computing angular coordinates of the surface normals of the points in the point cloud.


Example 9 provides the method of example 8, where extracting the basis vector includes binning the angular coordinates of the surface normals; identifying a peak angle of each of the angular coordinates; and identifying the basis vector based on the identified peak angles.


Example 10 provides the method of example 7, where computing a surface normal for an individual point in the point cloud includes fitting a plane to a set of points in a region around the individual point.


Example 11 provides the method of example 1, where the basis vector is a first basis vector, the method further including selecting a second basis vector orthogonal to the first basis vector and a third basis vector orthogonal to the first basis vector and the second basis vector, where the frame of reference of the basis vector is a frame of reference of the first basis vector, the second basis vector, and the third basis vectors.


Example 12 provides the method of example 11, where the second basis vector is selected as a projection of a pointing direction of the sensor into a Z-plane, and the third basis vector is set equal to a cross product of the first basis vector and the second basis vector.


Example 13 provides the method of example 1, where identifying the Z-plane in the transformed point cloud includes generating a height map of the transformed point cloud; generating a profile representation of the height map, the profile representation having a peak corresponding to each of a plurality of Z-planes; and identifying the Z-plane in the profile representation.


Example 14 provides the method of example 13, where the identified Z-plane is a base Z-plane, the method further including setting a height of the base Z-plane to zero.


Example 15 provides the method of example 13, further including associating a point in the transformed point cloud with the identified Z-plane based on determining that a height of the point is within a height range associated with the identified Z-plane.


Example 16 provides an imaging system including a TOF depth sensor to obtain distance data describing distances between the TOF depth sensor and a plurality of surfaces in an environment of the TOF depth sensor; and a processor to receive the distance data from the TOF depth sensor; generate a point cloud based on the distance data, the point cloud in a frame of reference of the TOF depth sensor; identify a basis vector representing a peak direction across the point cloud; transform the point cloud into a frame of reference of the basis vector; and identify a Z-plane in the transformed point cloud.


Example 17 provides the system of example 16, where the TOF depth sensor includes a light source to illuminate the environment of the TOF depth sensor and an image sensor to sense reflected light.


Example 18 provides the system of example 16, where the TOF depth sensor has an image frame, and the distance data is arranged in a plurality of pixels within the image frame.


Example 19 provides the system of example 18, where an individual pixel has a distance to one of the plurality of surfaces in the environment of the TOF depth sensor, and the individual pixel has an associated ray direction describing a direction from the TOF depth sensor to the surface.


Example 20 provides the system of example 19, where, to generate the point cloud, the processor multiplies the ray direction for the individual pixel by the distance to the one of the plurality of surfaces for the individual pixel.


Example 21 provides the system of example 16, further including a camera to capture an image of the environment of the TOF depth sensor.


Example 22 provides the system of example 21, further including a display screen, the processor to display, on the display screen, the image captured by the camera and a visual indication of the identified Z-plane.


Example 23 provides the system of example 16, further including a light sensor for detecting sunlight in the environment of the TOF depth sensor, where the processor applies a filter to the distance data in response to detecting at least a threshold level of sunlight.


Example 24 provides a method for determining dimensions of a physical box, the method including receiving distance data describing distances between a sensor and a plurality of surfaces in an environment of the sensor, at least a portion of the surfaces corresponding to a box to be measured; transforming the distance data into a frame of reference of one of the surfaces in the environment of the sensor; selecting, from the plurality of surfaces in the environment of the sensor, a first surface corresponding to a top of the box and a second surface corresponding to a surface the box is resting on; calculating a height between the first surface and the second surface; and calculating a length and a width based on the selected first surface corresponding to the top of the box.


Example 25 provides the method of example 24, where the distance data is a point cloud in a frame of reference of the sensor.


Example 26 provides the method of examples 25, where transforming the distance data into the frame of reference of one of the surfaces in the environment of the sensor includes identifying a basis vector representing a peak direction across the point cloud; and transforming the point cloud into a frame of reference of the basis vector.


Example 27 provides the method of example 26, where identifying the basis vector includes computing angular coordinates of surface normals for points in the point cloud; and extracting the basis vector based on the computed angular coordinates of the surface normals, the basis vector representing the peak direction of the surface normals across the point cloud.


Example 28 provides the method of example 24, where the sensor is a TOF sensor including a light source and an image sensor.


Example 29 provides the method of example 24, where the one of the surfaces used as the frame of reference for transforming the distance data is a Z-plane.


Example 30 provides the method of example 24, where selecting the first surface includes identifying a plurality of connected components within the transformed distance data, each connected component having a respective height along a Z-axis in the frame of reference of the one of the surfaces; and selecting, as the first surface, one of the plurality of connected components by applying a set of rules to the plurality of connected components.


Example 31 provides the method of example 30, where identifying the plurality of connected components includes identifying a plurality of Z-slices of the transformed distance data, each of the plurality of Z-slices having a respective height along the Z-axis; and identifying, within each of the plurality of Z-slices, at least one connected component of height map pixels.


Example 32 provides the method of example 31, where identifying the plurality of Z-slices includes generating a height map of the distance data; generating a profile representation of the height map, the profile representation having a peak corresponding to each Z-slice; and identifying the plurality of Z-slices from the profile representation.


Example 33 provides the method of example 31, where selecting the second surface corresponding to the surface the box is resting on includes selecting a Z-slice of the plurality of Z-slices within a lateral range of the selected first surface.


Example 34 provides the method of example 30, where the set of rules applied to the plurality of connected components includes removing a connected component having a width or length less than a threshold minimum width or length; removing a connected component at least a threshold distance from another connected component; and removing a connected component having an enclosing convex hull polygon that deviates from an expected rectangular shape by at least a threshold deviation.


Example 35 provides the method of example 24, where calculating the length and the width based on the selected first surface involves extracting a subset of the transformed distance data corresponding to the selected first surface; calculating a length profile and a width profile of the subset; identifying, within the width profile, a first leading edge and a first trailing edge of the box; identifying, within the length profile, a second leading edge and a second trailing edge of the box; and calculating the width of the box between the first leading edge and the second leading edge and calculating the length of the box between the second leading edge and the second trailing edge.


Example 36 provides the method of example 24, further including determining an angle of rotation for the extracted subset corresponding to the first selected surface, the determined angle selected to minimize a sum of projections of edges of the first selected surface onto a set of axes of the frame of reference of one of the surfaces in the environment of the sensor; and rotating the extracted subset corresponding to the first selected surface by the determined angle.


Example 37 provides the method of example 24, where the transformed distance data includes a plurality of pixels, and calculating the length and the width based on the selected first surface corresponding to the top of the box includes, for at least pixels in the selected first surface, filtering the pixels by computing, for an individual pixel, an average pixel value based on pixel values in a region around the individual pixel; and calculating the length and width based on the filtered pixels in the selected first surface.


Example 38 provides the method of example 24, further including generating a visual representation of the box, the visual representation indicating the height, width, and length of the box.


Example 39 provides the method of example 24, further including calculating an IoU score based on an overlap between the first surface corresponding to the top of the box and a circle in a field of view of the sensor; and generating a display including the calculated IoU score.


Example 40 provides the method of example 24, further including receiving camera data from a camera, the camera having a camera field of view that at least partially overlaps with a field of view of the sensor; determining, based on the camera data, an intensity of at least portion of the camera field of view; and generating a display including the determined intensity.


Example 41 provides an imaging system including a TOF depth sensor to obtain distance data describing distances between the TOF depth sensor and a plurality of surfaces in an environment of the TOF depth sensor; and a processor to receive the distance data from the TOF depth sensor; transform the distance data into a frame of reference of one of the surfaces in the environment of the sensor; select a first surface corresponding to a top of the box and a second surface corresponding to a surface the box is resting on; calculate a height between the first surface and the second surface; and calculate a length and a width based on the selected first surface corresponding to the top of the box.


Example 42 provides the system of example 41, where the TOF depth sensor includes a light source to illuminate the environment of the depth sensor and an image sensor to sense reflected light.


Example 43 provides the system of example 41, where the TOF sensor has an image frame, and the distance data is arranged in a plurality of pixels within the image frame.


Example 44 provides the system of example 43, where an individual pixel has a distance to one of the plurality of surfaces in the environment of the TOF depth sensor, and the individual pixel has an associated ray direction describing a direction from the sensor to the TOF depth surface.


Example 45 provides the system of example 41, further including a camera to capture an image of the environment of the TOF depth sensor.


Example 46 provides the system of example 45, further including a display screen, the processor to display, on the display screen, the image captured by the camera and the calculated width, length, and height.


Example 47 provides the system of example 45, further including a display screen, the processor to display, on the display screen, the image captured by the camera and an overlaid depiction of the selected first surface.


Example 48 provides the system of example 47, the processor further to display, on the display screen, a plurality of box edges below the selected first surface.


Other Implementation Notes, Variations, and Applications


It is to be understood that not necessarily all objects or advantages may be achieved in accordance with any particular embodiment described herein. Thus, for example, those skilled in the art will recognize that certain embodiments may be configured to operate in a manner that achieves or optimizes one advantage or group of advantages as taught herein without necessarily achieving other objects or advantages as may be taught or suggested herein.


In one example embodiment, any number of electrical circuits of the figures may be implemented on a board of an associated electronic device. The board can be a general circuit board that can hold various components of the internal electronic system of the electronic device and, further, provide connectors for other peripherals. More specifically, the board can provide the electrical connections by which the other components of the system can communicate electrically. Any suitable processors (inclusive of digital signal processors, microprocessors, supporting chipsets, etc.), computer-readable non-transitory memory elements, etc. can be suitably coupled to the board based on particular configuration needs, processing demands, computer designs, etc. Other components such as external storage, additional sensors, controllers for audio/video display, and peripheral devices may be attached to the board as plug-in cards, via cables, or integrated into the board itself. In various embodiments, the functionalities described herein may be implemented in emulation form as software or firmware running within one or more configurable (e.g., programmable) elements arranged in a structure that supports these functions. The software or firmware providing the emulation may be provided on non-transitory computer-readable storage medium comprising instructions to allow a processor to carry out those functionalities.


It is also imperative to note that all of the specifications, dimensions, and relationships outlined herein (e.g., the number of processors, logic operations, etc.) have only been offered for purposes of example and teaching only. Such information may be varied considerably without departing from the spirit of the present disclosure, or the scope of the appended claims. The specifications apply only to one non-limiting example and, accordingly, they should be construed as such. In the foregoing description, example embodiments have been described with reference to particular arrangements of components. Various modifications and changes may be made to such embodiments without departing from the scope of the appended claims. The description and drawings are, accordingly, to be regarded in an illustrative rather than in a restrictive sense.


Note that with the numerous examples provided herein, interaction may be described in terms of two, three, four, or more components. However, this has been done for purposes of clarity and example only. It should be appreciated that the system can be consolidated in any suitable manner. Along similar design alternatives, any of the illustrated components, modules, and elements of the FIGS. may be combined in various possible configurations, all of which are clearly within the broad scope of this Specification.


Note that in this Specification, references to various features (e.g., elements, structures, modules, components, steps, operations, characteristics, etc.) included in “one embodiment”, “example embodiment”, “an embodiment”, “another embodiment”, “some embodiments”, “various embodiments”, “other embodiments”, “alternative embodiment”, and the like are intended to mean that any such features are included in one or more embodiments of the present disclosure, but may or may not necessarily be combined in the same embodiments.


Numerous other changes, substitutions, variations, alterations, and modifications may be ascertained to one skilled in the art and it is intended that the present disclosure encompass all such changes, substitutions, variations, alterations, and modifications as falling within the scope of the appended claims. Note that all optional features of the systems and methods described above may also be implemented with respect to the methods or systems described herein and specifics in the examples may be used anywhere in one or more embodiments.


In order to assist the United States Patent and Trademark Office (USPTO) and, additionally, any readers of any patent issued on this application in interpreting the claims appended hereto, Applicant wishes to note that the Applicant: (a) does not intend any of the appended claims to invoke paragraph (f) of 35 U.S.C. Section 112 as it exists on the date of the filing hereof unless the words “means for” or “step for” are specifically used in the particular claims; and (b) does not intend, by any statement in the Specification, to limit this disclosure in any way that is not otherwise reflected in the appended claims.

Claims
  • 1. A method for identifying a Z-plane, the method comprising: receiving distance data describing distances between a sensor that captured the distance data and a plurality of surfaces in an environment of the sensor, wherein at least one of the surfaces is a Z-plane;generating a point cloud based on the distance data, the point cloud in a frame of reference of the sensor;identifying a basis vector representing a peak direction across the point cloud;transforming the point cloud into a frame of reference of the basis vector; andidentifying a Z-plane in the transformed point cloud.
  • 2. The method of claim 1, wherein the sensor is a time-of-flight (TOF) sensor comprising a light source and an image sensor.
  • 3. The method of claim 1, wherein the distance data is arranged in a plurality of pixels within an image frame of the sensor.
  • 4. The method of claim 3, wherein an individual pixel comprises a distance to one of the plurality of surfaces in the environment of the sensor, and the individual pixel has an associated ray direction describing a direction from the sensor to the surface.
  • 5. The method of claim 4, wherein generating the point cloud comprises multiplying the ray direction for the individual pixel by the distance to the one of the plurality of surfaces for the individual pixel.
  • 6. The method of claim 1, wherein the distance data is arranged as a plurality of pixels, the method further comprising: filtering the distance data by computing, for an individual pixel, an average pixel value based on pixel values in a region around the individual pixel.
  • 7. The method of claim 1, wherein identifying the basis vector comprises: computing surface normals for points in the point cloud; andextracting the basis vector based on the computed surface normals, the basis vector representing the peak direction of the surface normals across the point cloud.
  • 8. The method of claim 7, wherein computing the surface normal for points in the point cloud comprises computing angular coordinates of the surface normals of the points in the point cloud.
  • 9. The method of claim 8, wherein extracting the basis vector comprises: binning the angular coordinates of the surface normals;identifying a peak angle of each of the angular coordinates; andidentifying the basis vector based on the identified peak angles.
  • 10. The method of claim 7, wherein computing a surface normal for an individual point in the point cloud comprises fitting a plane to a set of points in a region around the individual point.
  • 11. The method of claim 1, wherein the basis vector is a first basis vector, the method further comprising: selecting a second basis vector orthogonal to the first basis vector and a third basis vector orthogonal to the first basis vector and the second basis vector, wherein the frame of reference of the basis vector is a frame of reference of the first basis vector, the second basis vector, and the third basis vectors.
  • 12. The method of claim 11, wherein the second basis vector is selected as a projection of a pointing direction of the sensor into a Z-plane, and the third basis vector is set equal to a cross product of the first basis vector and the second basis vector.
  • 13. The method of claim 1, wherein identifying the Z-plane in the transformed point cloud comprises: generating a height map of the transformed point cloud;generating a profile representation of the height map, the profile representation having a peak corresponding to each of a plurality of Z-planes; andidentifying the Z-plane in the profile representation.
  • 14. The method of claim 13, wherein the identified Z-plane is a base Z-plane, the method further comprising setting a height of the base Z-plane to zero.
  • 15. The method of claim 13, further comprising associating a point in the transformed point cloud with the identified Z-plane based on determining that a height of the point is within a height range associated with the identified Z-plane.
  • 16. An imaging system comprising: a time-of-flight (TOF) depth sensor to obtain distance data describing distances between the TOF depth sensor and a plurality of surfaces in an environment of the TOF depth sensor; anda processor to: receive the distance data from the TOF depth sensor;generate a point cloud based on the distance data, the point cloud in a frame of reference of the TOF depth sensor;identify a basis vector representing a peak direction across the point cloud;transform the point cloud into a frame of reference of the basis vector; andidentify a Z-plane in the transformed point cloud.
  • 17. The system of claim 16, wherein the TOF depth sensor comprises a light source to illuminate the environment of the TOF depth sensor and an image sensor to sense reflected light.
  • 18. The system of claim 16, wherein the TOF depth sensor has an image frame, and the distance data is arranged in a plurality of pixels within the image frame.
  • 19. The system of claim 18, wherein an individual pixel comprises a distance to one of the plurality of surfaces in the environment of the TOF depth sensor, and the individual pixel has an associated ray direction describing a direction from the TOF depth sensor to the surface.
  • 20. The system of claim 19, wherein, to generate the point cloud, the processor multiplies the ray direction for the individual pixel by the distance to the one of the plurality of surfaces for the individual pixel.
  • 21. The system of claim 16, further comprising a camera to capture an image of the environment of the TOF depth sensor.
  • 22. The system of claim 21, further comprising a display screen, the processor to display, on the display screen, the image captured by the camera and a visual indication of the identified Z-plane.
  • 23. The system of claim 16, further comprising a light sensor for detecting sunlight in the environment of the TOF depth sensor, wherein the processor applies a filter to the distance data in response to detecting at least a threshold level of sunlight.
  • 24. A method for determining dimensions of a physical box, the method comprising: receiving distance data describing distances between a sensor and a plurality of surfaces in an environment of the sensor, at least a portion of the surfaces corresponding to a box to be measured;transforming the distance data into a frame of reference of one of the surfaces in the environment of the sensor;selecting, from the plurality of surfaces in the environment of the sensor, a first surface corresponding to a top of the box and a second surface corresponding to a surface the box is resting on;calculating a height between the first surface and the second surface; andcalculating a length and a width based on the selected first surface corresponding to the top of the box.
  • 25. The method of claim 24, wherein the distance data comprises a point cloud in a frame of reference of the sensor.
  • 26. The method of claim 25, wherein transforming the distance data into the frame of reference of one of the surfaces in the environment of the sensor comprises: identifying a basis vector representing a peak direction across the point cloud; andtransforming the point cloud into a frame of reference of the basis vector.
  • 27. The method of claim 26, wherein identifying the basis vector comprises: computing angular coordinates of surface normals for points in the point cloud; andextracting the basis vector based on the computed angular coordinates of the surface normals, the basis vector representing the peak direction of the surface normals across the point cloud.
  • 28. The method of claim 24, wherein the sensor is a time-of-flight (TOF) sensor comprising a light source and an image sensor.
  • 29. The method of claim 24, wherein the one of the surfaces used as the frame of reference for transforming the distance data is a Z-plane.
  • 30. The method of claim 24, wherein selecting the first surface comprises: identifying a plurality of connected components within the transformed distance data, each connected component having a respective height along a Z-axis in the frame of reference of the one of the surfaces; andselecting, as the first surface, one of the plurality of connected components by applying a set of rules to the plurality of connected components.
  • 31. The method of claim 30, wherein identifying the plurality of connected components comprises: identifying a plurality of Z-slices of the transformed distance data, each of the plurality of Z-slices having a respective height along the Z-axis; andidentifying, within each of the plurality of Z-slices, at least one connected component of height map pixels.
  • 32. The method of claim 31, wherein identifying the plurality of Z-slices comprises: generating a height map of the distance data;generating a profile representation of the height map, the profile representation having a peak corresponding to each Z-slice; andidentifying the plurality of Z-slices from the profile representation.
  • 33. The method of claim 31, wherein selecting the second surface corresponding to the surface the box is resting on comprises selecting a Z-slice of the plurality of Z-slices within a lateral range of the selected first surface.
  • 34. The method of claim 30, wherein the set of rules applied to the plurality of connected components comprises: removing a connected component having a width or length less than a threshold minimum width or length;removing a connected component at least a threshold distance from another connected component; andremoving a connected component having an enclosing convex hull polygon that deviates from an expected rectangular shape by at least a threshold deviation.
  • 35. The method of claim 24, wherein calculating the length and the width based on the selected first surface comprises: extracting a subset of the transformed distance data corresponding to the selected first surface;calculating a length profile and a width profile of the subset;identifying, within the width profile, a first leading edge and a first trailing edge of the box;identifying, within the length profile, a second leading edge and a second trailing edge of the box; andcalculating the width of the box between the first leading edge and the second leading edge and calculating the length of the box between the second leading edge and the second trailing edge.
  • 36. The method of claim 24, further comprising: determining an angle of rotation for the extracted subset corresponding to the first selected surface, the determined angle selected to minimize a sum of projections of edges of the first selected surface onto a set of axes of the frame of reference of one of the surfaces in the environment of the sensor; androtating the extracted subset corresponding to the first selected surface by the determined angle.
  • 37. The method of claim 24, wherein the transformed distance data comprises a plurality of pixels, and calculating the length and the width based on the selected first surface corresponding to the top of the box comprises: for at least pixels in the selected first surface, filtering the pixels by computing, for an individual pixel, an average pixel value based on pixel values in a region around the individual pixel; andcalculating the length and width based on the filtered pixels in the selected first surface.
  • 38. The method of claim 24, further comprising generating a visual representation of the box, the visual representation indicating the height, width, and length of the box.
  • 39. The method of claim 24, further comprising: calculating an intersection over union (IoU) score based on an overlap between the first surface corresponding to the top of the box and a circle in a field of view of the sensor; andgenerating a display including the calculated IoU score.
  • 40. The method of claim 24, further comprising: receiving camera data from a camera, the camera having a camera field of view that at least partially overlaps with a field of view of the sensor;determining, based on the camera data, an intensity of at least portion of the camera field of view; andgenerating a display including the determined intensity.
  • 41. An imaging system comprising: a time-of-flight (TOF) depth sensor to obtain distance data describing distances between the TOF depth sensor and a plurality of surfaces in an environment of the TOF depth sensor; anda processor to: receive the distance data from the TOF depth sensor;transform the distance data into a frame of reference of one of the surfaces in the environment of the sensor;select a first surface corresponding to a top of the box and a second surface corresponding to a surface the box is resting on;calculate a height between the first surface and the second surface; andcalculate a length and a width based on the selected first surface corresponding to the top of the box.
  • 42. The system of claim 41, wherein the TOF depth sensor comprises a light source to illuminate the environment of the depth sensor and an image sensor to sense reflected light.
  • 43. The system of claim 41, wherein the TOF sensor has an image frame, and the distance data is arranged in a plurality of pixels within the image frame.
  • 44. The system of claim 43, wherein an individual pixel comprises a distance to one of the plurality of surfaces in the environment of the TOF depth sensor, and the individual pixel has an associated ray direction describing a direction from the sensor to the TOF depth surface.
  • 45. The system of claim 41, further comprising a camera to capture an image of the environment of the TOF depth sensor.
  • 46. The system of claim 45, further comprising a display screen, the processor to display, on the display screen, the image captured by the camera and the calculated width, length, and height.
  • 47. The system of claim 45, further comprising a display screen, the processor to display, on the display screen, the image captured by the camera and an overlaid depiction of the selected first surface.
  • 48. The system of claim 47, the processor further to display, on the display screen, a plurality of box edges below the selected first surface.
CROSS-REFERENCE TO RELATED APPLICATION(S)

This application claims priority to U.S. provisional patent application Nos. 63/081,742, filed Sep. 22, 2020 and entitled “BOX DIMENSIONING USING THREE-DIMENSIONAL TIME-OF-FLIGHT IMAGING,” and 63/081,775, filed Sep. 22, 2020 and entitled “WORLD Z-PLANE IDENTIFICATION IN TIME-OF-FLIGHT IMAGERY,” which are hereby incorporated by reference in their entireties.

PCT Information
Filing Document Filing Date Country Kind
PCT/US2021/051238 9/21/2021 WO
Provisional Applications (2)
Number Date Country
63081742 Sep 2020 US
63081775 Sep 2020 US