Mobile automation apparatuses are increasingly being used in retail environments to perform inventory tracking. Such mobile automation apparatuses may be equipped with various sensors to autonomously navigate paths within the retail environment and perform such tasks. During such autonomous navigation, the mobile automation apparatuses may perform ancillary tasks that include data capture, avoiding obstacles, stopping, docking, and the like. Hence, a key issue during such autonomous navigation in a retail environment is to be able to detect static obstacles and/or moving obstacles. Static obstacles include boxes, pillars, reflective stanchions, baskets, and other items found in retail environments. Moving obstacles include shopping carts, retail associates, cleaning equipment, and shoppers. Especially with obstacles such as humans and/or obstacles associated with humans, such as shopping carts, the mobile automation apparatuses must be able to detect these obstacles from a safe distance in order to stop or avoid the obstacles. These challenges are compounded in retail environments where lighting varies in different areas. Furthermore, highly reflective and/or shiny obstacle objects found in the retail environment, such as mirrors, stanchions, see-through glass doors, aisle siding, and other reflective materials, may also cause variable lighting that compounds the challenges associated with obstacle avoidance. In particular, the reflective metal lattice on shopping carts, for example, can be difficult to detect. Indeed, a single sensor by itself does not exist, at low cost, that can detect all of the complex obstacles listed above.
The accompanying figures, where like reference numerals refer to identical or functionally similar elements throughout the separate views, together with the detailed description below, are incorporated in and form part of the specification, and serve to further illustrate embodiments of concepts that include the claimed invention, and explain various principles and advantages of those embodiments.
Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help improve understanding of embodiments of the present invention.
The apparatus and method components have been represented where appropriate by conventional symbols in the drawings, showing only those specific details that are pertinent to understanding the embodiments of the present invention so as not to obscure the disclosure with details that will be readily apparent to those of ordinary skill in the art having the benefit of the description herein.
An aspect of the specification provides a mobile automation apparatus comprising: a support structure extending in an upward direction; a navigation module configured to move the mobile automation apparatus in an environment; a structured light camera positioned on the mast, the structured light camera configured to detect at least a top of a shopping cart in a region adjacent to the mobile automation apparatus; and, a controller configured to: when the structured light camera detects the top of the shopping cart, control the navigation module to implement an obstacle avoidance protocol.
In some embodiments, the structured light camera is located at a height on the mast where the structured light camera acquires images of the top of a shopping cart in the region.
In some embodiments, the structured light camera is at an angle of 45°±5° to the mast, and at a height of between 1.3 meters to 1.8 meters up the mast.
In some embodiments, the mobile automation apparatus further comprises a base, the support structure extending in the upward direction from the base; and one or more of a depth sensing device and a LIDAR device located at the base, and configured to sense depth at an ankle height.
In some embodiments, the mobile automation apparatus further comprises one or more second structured light cameras between a base of the mast and the structured light camera, the one or more second structured light cameras configured to detect objects over respective angles on either side of the region.
In some embodiments, the mobile automation apparatus further comprises one or more ultrasound arrays on the mast, each of the one or more ultrasound arrays configured to detect objects at least in the region. In some embodiments, the controller is further configured to: when the one or more ultrasound arrays detect an object, and the structured light camera simultaneously does not detect the object, control the navigation module to implement the obstacle avoidance protocol. In some embodiments, the controller is further configured to: when the one or more ultrasound arrays detect the object, and the structured light camera simultaneously does not detect the object, control the navigation module to implement the obstacle avoidance protocol only when the object is within a threshold distance.
In some embodiments, the navigation module configured to move the mobile automation apparatus in the environment in a forward direction, and the structured light camera is further configured to detect at least the top of the shopping cart in the region in the forward direction.
In some embodiments, the obstacle avoidance protocol comprises one or more of: stopping the navigation module, slowing the navigation module and controlling the navigation module to avoid the shopping cart.
A further aspect of the specification provides a method comprising: at a mobile automation apparatus comprising: a support structure extending in an upward direction; a navigation module configured to move the mobile automation apparatus in an environment; a structured light camera positioned on the mast, the structured light camera configured to detect at least a top of a shopping cart in a region adjacent to the mobile automation apparatus; and, a controller, when the structured light camera detects the top of the shopping cart, controlling, using the controller, the navigation module to implement an obstacle avoidance protocol.
A further aspect of the specification provides a mobile automation apparatus comprising: a support structure extending in an upward direction; a navigation module configured to move the mobile automation apparatus in an environment; at least one structured light camera located on the mast, the at least one structured light camera configured to detect objects over a region; one or more ultrasound arrays on the mast, each of the one or more ultrasound arrays configured to detect the objects over the region; and, a controller configured to: when the one or more ultrasound arrays detect an object, and the at least one structured light camera simultaneously does not detect the object, control the navigation module to implement an obstacle avoidance protocol.
In some embodiments, the controller is further configured to: when the one or more ultrasound arrays detect the object, and at least one structured light camera simultaneously does not detect the object, control the navigation module to implement the obstacle avoidance protocol only when the object is within a threshold distance.
In some embodiments, the navigation module is configured to move the mobile automation apparatus in the environment in a forward direction, and at least one structured light camera and the one or more ultrasound arrays are further configured to detect the objects in the forward direction.
In some embodiments, the obstacle avoidance protocol comprises one or more of: stopping the navigation module, slowing the navigation module and controlling the navigation module to avoid the object.
In some embodiments, the structured light camera is mounted at an angle of 45°±5° to the mast, and at a height of between 1.3 meters to 1.8 meters up the mast.
In some embodiments, the one or more ultrasound arrays are further configured to detect the objects over a 360° field of view.
A further aspect of the specification provides a method comprising: at a mobile automation apparatus comprising: a support structure extending in an upward direction; a navigation module configured to move the mobile automation apparatus in an environment; at least one structured light camera located on the mast, the at least one structured light camera configured to detect objects over a region; one or more ultrasound arrays on the mast, each of the one or more ultrasound arrays configured to detect the objects over the region; and, a controller, when the one or more ultrasound arrays detect an object, and the at least one structured light camera simultaneously does not detect the object, controlling, using the controller, the navigation module to implement an obstacle avoidance protocol.
A further aspect of the specification provides a mobile automation apparatus comprising: a support structure extending in an upward direction; a navigation module configured to move the mobile automation apparatus in an environment; at least one structured light camera located on the mast, the at least one structured light camera configured to detect objects over a region; one or more ultrasound arrays on the mast, each of the one or more ultrasound arrays configured to detect the objects over the region; and, a controller, wherein when the one or more ultrasound arrays detect an object and the at least one structured light camera simultaneously does not detect the object, the controller controls the navigation module to implement an obstacle avoidance protocol.
A further aspect of the specification provides a mobile automation apparatus comprising an ultrasound sensor; a structured light camera positioned higher than the ultrasound sensor; a controller configured to compare ultrasound sensor data to structured light data; and a navigation module configured to alter the navigation of the mobile automation apparatus based on the comparison. In some embodiments, the mobile automation apparatus further comprises a communication interface configured to transmit a notification of the alteration to a mobile device. In some embodiments, the notification includes a location of the mobile automation apparatus.
The server 101 includes a controller 120, interconnected with a non-transitory computer readable storage medium, such as a memory 122. The memory 122 includes any suitable combination of volatile (e.g. Random Access Memory or RAM) and non-volatile memory (e.g. read only memory or ROM, Electrically Erasable Programmable Read Only Memory or EEPROM, flash memory, hard drive). In general, the controller 120 and the memory 122 each comprise one or more integrated circuits. The controller 120, for example, includes any suitable combination of central processing units (CPUs) and graphics processing units (GPUs) and/or any suitable combination of field-programmable gate arrays (FPGAs) and/or application-specific integrated circuits (ASICs) configured to implement the specific functionality of the server 101. In an embodiment, the controller 120, further includes one or more central processing units (CPUs) and/or graphics processing units (GPUs). In an embodiment, a specially designed integrated circuit, such as a Field Programmable Gate Array (FPGA), is designed to perform specific functionality of the server 101, either alternatively or in addition to the controller 120 and the memory 122. As those of skill in the art will realize, the mobile automation apparatus 103 also includes one or more controllers or processors and/or FPGAs, in communication with the controller 120, specifically configured to control navigational and/or data capture aspects of the mobile automation apparatus 103.
The server 101 also includes a communications interface 124 interconnected with the controller 120. The communications interface 124 includes any suitable hardware (e.g. transmitters, receivers, network interface controllers and the like) allowing the server 101 to communicate with other computing devices—particularly the apparatus 103 and the mobile device 105—via the links 107. In some embodiments, a link 107 may be a direct wired link, such as a cable or data bus. It some embodiments a link 107 may be a direct wireless link, such as a radio signal, infrared communication, or sound communication. In some embodiments, a link 107 may be a networked communication that traverses one or more networks, such as a local area network or wide-area network. The specific components of the communications interface 124 are selected based on the type of communication links (e.g. networks) enabled by the server 101. In the present example, a wireless local-area network is implemented within the retail environment via the deployment of one or more wireless access points. The links 107 therefore include both wireless links between the apparatus 103 and the mobile device 105 and the above-mentioned access points, and a wired link (e.g. an Ethernet-based link) between the server 101 and the access point.
The memory 122 stores a plurality of applications, each including a plurality of computer readable instructions executable by the controller 120. The execution of the above-mentioned instructions by the controller 120 configures the server 101 to perform various actions discussed herein. The applications stored in the memory 122 include a control application 128, which may also be implemented as a suite of logically distinct applications. In general, via execution of the control application 128 or subcomponents thereof, the controller 120 is configured to implement various functionality. The controller 120, as configured via the execution of the control application 128, is also referred to herein as the controller 120. As will now be apparent, some or all of the functionality implemented by the controller 120 described below may also be performed by preconfigured hardware elements (e.g. one or more ASICs) rather than by execution of the control application 128 by the controller 120.
In general, the controller 120 is configured to control the apparatus 103 to capture data, and to obtain the captured data via the communications interface 124 and store the captured data in a repository 132 in the memory 122. The server 101 is further configured to perform various post-processing operations on the captured data, and to detect the status of the products 112 on the modules 110. When certain status indicators are detected, the server 101 is also configured to transmit status notifications to the mobile device 105.
For example, in some embodiments, the server 101 is configured via the execution of the control application 128 by the controller 120, to process image and depth data captured by the apparatus 103 to identify portions of the captured data depicting a back of a module 110, and to detect gaps between the products 112 based on those identified portions. Furthermore, the server 101 is configured to provide a map and/or paths and/or path segments and/or navigation data and/or navigation instructions to the apparatus 103 to enable the apparatus 103 to navigate the modules 110.
Attention is next directed to
With reference to
The apparatus 103 is further provisioned with various sensors for obstacle avoidance which include, but are not limited to: a structured light camera 218 located on the mast 214, the structured light camera 218 configured to detect objects over a region, for example in a field of view of the structured light camera 218; one or more ultrasound arrays 220-1, 220-2 on the mast 214, each of the one or more ultrasound arrays 220-1, 220-2 configured to detect objects over the region in which the structured light camera 218 is configured to detect objects; one or more of a depth sensing device and a LIDAR device 221 located at the base 210, and configured to sense depth at an ankle height (e.g. over the region in which the structured light camera 218 is configured to detect objects); and one or more second structured light cameras 228 between a bottom end of the mast 214 and the structured light camera 218, the one or more second structured light cameras 218 configured to detect objects over respective angles on either side of the region (e.g. the region in which the structured light camera 218 is configured to detect objects).
It is further understood that detection of an object as described herein includes, but is not limited to, detection of a portion of an object and further understood that different sensors may detect different portions of the same object.
While not depicted, the base 210 and the mast 214 may be provisioned with other various navigation sensors for navigating in the environment in which the modules 110 are located and various further data acquisition sensors for acquiring electronic data and/or digital data associated with the modules 110, products 112, or shelf merchandising materials 114. Such data may be used for one or more of stock detection, low stock detection, price label verification, plug detection, planogram compliance and the like.
As depicted, the location and geometry of the structured light camera 218 on the mast 214 is selected such that the structured light camera 218 is configured to detect at least a top of a shopping cart in a region adjacent to the mobile automation apparatus 103. For example, the structured light camera 218 is located at a height on the mast 214 where the structured light camera 218 acquires images of a top of a shopping cart in a region adjacent the mobile automation apparatus 103. Indeed, the structured light camera 218 is generally configured to detect at least a top of a shopping cart in a region adjacent to the mobile automation apparatus 103. For example, shopping carts are generally about 0.8 m high. Hence, the structured light camera 218 may be mounted as depicted at an angle of 45°±5° to the mast 214, and at a height of between 1.3 meters to 1.8 meters up the mast 214, for example, as depicted, at about 1.5 m from a bottom of the wheels 212 or above the floor 213.
Other geometries of the structured light camera 218 are within the scope of present embodiments. However, as the structured light camera 218 is generally pointing towards the floor 213, the structured light camera 218 is generally configured to detect objects in the field of view of structured light camera 218 including objects on the floor 213 that may not be detectable by the other sensors.
The structured light camera 218 is generally configured to detect objects by projecting structured light (including, but not limited to, infrared light) onto a region, including, but not limited to one or more of vertical stripes, horizontal stripes, dots, and the like, and acquiring images of the structure light projected onto the region (including, but not limited to, infrared images). Discontinuities in the structured light detected in the acquired images indicate objects in the field of view of the structured light camera 218. In other words, the region in which objects are detected is the field of view of the structured light camera 218 that is illuminated by the projected images. Hence, the structured light camera 218 is configured to detect at least a top of a shopping cart in a field of view of the structured light camera 218, which depends on the angle and height of the structured light camera 218.
The two or more structured light cameras 228 operate in a similar fashion, but are aimed about perpendicularly away from the mast 214, however at different angles. In a particular embodiment, the two or more structured light cameras 218 are aimed without an overlap in their fields of view, with a gap therebetween being filled in by the field of view of the structured light camera 218. An example geometry of the two or more structured light cameras 228 is described below with respect to
Furthermore, while the one or more ultrasound arrays 220-1, 220-2 are at least configured to detect objects over the region in which objects are detected by the structured light camera 218, in some embodiments, the one or more ultrasound arrays 220-1, 220-2 are further configured to detect the objects circumferentially around the mast 214 and/or the apparatus 103. An ultrasound array 220 may be configured to detect objects in any direction from the mast (360° detection) or only in certain angular regions around the mast 214 and/or the apparatus 103. The one or more ultrasound arrays 220-1, 220-2 are interchangeably referred to hereafter, collectively, as ultrasound arrays 220 and, generically, as an ultrasound array 220. Furthermore, in some embodiments, the ultrasound arrays 220 comprise phased ultrasound arrays.
The ultrasound arrays 220 generally operate by transmitting an ultrasound pulse and receiving reflections of the ultrasound pulse from objects in a range of the transmitted ultrasound pulse, for example in a phased manner. Indeed, each of the ultrasound arrays 220 generally comprises a plurality of ultrasound transducers arranged circumferentially around the mast 214 which transmit respective ultrasound pulses in a phased manner. Overlaps in the fields of view of adjacent ultrasound transducers can assist with pinpointing a general location of an object reflecting a respective ultrasound pulse. For example, if an object is detected by a first ultrasound transducer, but not detected by a second adjacent ultrasound transducer, the object is located in an angular region associated with only the first ultrasound transducer; however, an object detected by both the first and the second ultrasound transducers is located in an angular region within the overlapping fields of view of the two adjacent ultrasound transducers.
The LIDAR device 221 generally operates by transmitting laser pulses (e.g. an infrared laser pulse) at a plurality of known angles, and a sensor (e.g. a light sensor and/or an infrared sensor) at the LIDAR device 221 measures a distance to a point of interaction with a surface using a time-of-flight measurement. The LIDAR device 221 is generally at an ankle height, for example between about 5 cm to about 10 cm from the floor 213 and is generally configured to scan over a plane horizontal to the floor 213. Hence, the LIDAR device 221 is configured to detect wheels of shopping carts, legs of modules 110, and the like. However, other depth detecting devices may be used in place of the LIDAR device 221 in other embodiments.
Furthermore, one or more combinations of sensor data from the sensors (e.g. the structured light cameras 218, 228, the one or more ultrasound arrays 220, and the LIDAR device 221) are useable by the apparatus 103 to determine a distance of an object from each respective sensor. For example, structured light images can indicate distance of objects, as can data from LIDAR devices and ultrasound arrays.
In addition, in some embodiments, the apparatus 103 is generally configured to move in a forward direction, for example, in a direction being imaged by the LIDAR device 221. In some of these embodiments, the structured light camera 218 is further configured to detect at least a top of a shopping cart in a region in the forward direction and/or a movement direction.
Data from the various obstacle sensors at the apparatus 103 may be processed at the apparatus 103 and used to avoid obstacles including, but not limited to shopping carts.
For example, attention is next directed to
The apparatus 103 also includes a communications interface 324 interconnected with the controller 320. The communications interface 324 includes any suitable hardware (e.g. transmitters, receivers, network interface controllers and the like) allowing the apparatus 103 to communicate with other devices—particularly the server 101 and, optionally, the mobile device 105—for example via the links 107, and the like. The specific components of the communications interface 324 are selected based on the type of network or other links over which the apparatus 103 is required to communicate, including, but not limited to, the wireless local-area network of the retail environment of the system 100.
The memory 322 stores a plurality of applications, each including a plurality of computer readable instructions executable by the controller 320. The execution of the above-mentioned instructions by the controller 320 configures the apparatus 103 to perform various actions discussed herein. The applications stored in the memory 322 include a control application 328, which may also be implemented as a suite of logically distinct applications. In general, via execution of the control application 328 or subcomponents thereof, the controller 320 is configured to implement various functionality. The controller 320, as configured via the execution of the control application 328, is also referred to herein as the controller 320. As will now be apparent, some or all of the functionality implemented by the controller 320 described below may also be performed by preconfigured hardware elements (e.g. one or more ASICs) rather than by execution of the control application 328 by the controller 320.
The apparatus 103 further includes sensors 330, including, but not limited to, the structured light camera 218, the one or more ultrasound arrays 220, the LIDAR device 221 (and/or any other depth sensing device at the same location), the one or more second structured light cameras 228, and other navigation and/or data acquisition sensors.
The apparatus 103 further includes a navigation module 340 configured to move the apparatus 103 in an environment, for example the environment of the modules 110. The navigation module 340 comprises any suitable combination of motors, electric motors, stepper motors, and the like configured to drive and/or steer the wheels 212, and the like, of the apparatus 103. The apparatus 103 generally includes sensors (e.g. a wheel odometer, an IMU (inertial measurement unit), and the like for measuring distances travelled by the apparatus 103.
Hence, in general, the controller 320 is configured to control the apparatus 103 to: navigate the environment of the module 110 using data from the navigation sensors (which can include, but is not limited to, the structured light cameras 218, 228, the LIDAR device 221 and the one or more ultrasound arrays 220) to control the navigation module 340; and control the data acquisition sensors to acquire sensor data, such as electronic data and/or digital data used for one or more of stock detection, low stock detection, price label verification, plug detection, planogram compliance and the like. For example, while not depicted, the memory 322 further stores a map, and the like, of the environment, which may be used for navigation in conjunction with the navigation sensors. The memory may also store sensor data from the data acquisition sensors that may be transmitted to the server 101 for processing.
In the present example, in particular, the apparatus 103 is configured via the execution of the control application 328 by the controller 320, to avoid obstacles using at least camera data from the structured light camera 218.
Turning now to
As depicted the application 328 is depicted as receiving camera data 428 (e.g. images) from the structured light camera 218, as well as ultrasound data 420 from the one or more ultrasound arrays 220, LIDAR data 421 from the LIDAR device 221 and further camera data 428 (e.g. images) from the one or more second structured light cameras 228.
The control application 328 includes an obstacle avoider 401. In brief, the obstacle avoider 401 is configured to process one or more of the camera data 418, the ultrasound data 420, the LIDAR data 421 and the further camera data 428 to avoid obstacles.
The obstacle avoider 401 includes: an obstacle detector 410 (also referred to herein simply as a detector 410), a distance determiner 411, and an obstacle avoidance protocol 412. In brief, the detector 410 is configured to process camera data 418 captured by the structured light camera 218 to determine whether an object has been detected or not and, depending on a distance to the object, as determined by the distance determiner 411, the obstacle avoidance protocol 412 is used to control the navigation module 340 to avoid the object. As depicted, the obstacle detector 410 includes a shopping cart detector 410a configured to process the camera data 418 captured by the structured light camera 218 to determine whether the object includes a shopping cart and, in particular, a top of a shopping cart.
In some embodiments, each of the detector 410 and shopping cart detector 410a comprises signature data (e.g., sensor output signatures indicative of detecting a shopping cart object) and/or threshold data and/or algorithms respectively associated with detecting objects, including shopping carts, using, for example, each of the camera data 418, the ultrasound data 420, the LIDAR data 421 and the further camera data 428.
One or more of the camera data 418, the ultrasound data 420, the LIDAR data 421 and the further camera data 428 is used by the distance determiner 411 to determine a distance of an object to the apparatus 103 (and/or a respective sensor), which implements any number of distance determining techniques.
Attention is now directed to
Furthermore, the example method 500 of
At block 501, the controller 320 determines whether the structured light camera 218 has detected a top of a shopping cart. For example, the block 501 occurs using the shopping cart detector 410a in conjunction with camera data 418 from the structured light camera 218, such as by comparing the camera data 418 with sensor output signature that is indicative of detecting a top of a shopping cart object.
At the block 503, when the structured light camera 218 detects the top of the shopping cart, the controller 320 controls the navigation module 340 to implement the obstacle avoidance protocol 412. For example, the obstacle avoidance protocol 412 comprises one or more of: stopping the navigation module 340 (e.g. stopping the apparatus 103), slowing the apparatus 103, and controlling the navigation module 340 to avoid the shopping cart (e.g. steer the apparatus 103 around the shopping cart 603).
The method 500 repeats each time the controller 320 determines that the structured light camera 218 has detected a top of a shopping cart at block 501. Hence, the method 500 generally runs in parallel with other methods at the apparatus 103, including, for example other data detection methods (e.g. for detecting prices, barcodes, and the like) and/or navigation methods.
The method 500 is now described with reference to
Furthermore, as depicted, the shopping cart 603 is located in a region being imaged by the structured light cameras 218, 228, and the LIDAR device 221. The behavior of the ultrasound arrays 220 is described below.
As depicted, the LIDAR device 221 is attempting to depth scan the wheels 612 (as indicated by the dotted line extending from the LIDAR device 221), and the two or more further structured light cameras 228 are attempting to image the metal lattice 605. However, as the wheels 612 are generally inset from an outside perimeter of the shopping cart 603 (e.g. due to the angle of the lattice 605, which widens from bottom to top) and/or the wheels 612 occupy only a portion of the space under the shopping cart 603, a depth detection of the wheels 612 by the LIDAR device 221 is generally a poor indication of the position of the shopping cart 603; in other words, a front edge of the shopping cart could collide with the apparatus 103 prior to the wheels 612 being detected.
Similarly, the two or more structured light cameras 228 are imaging (as indicated by the dotted lines extending from the two or more structured light cameras 228) the metal lattice 605, which generally is both specularly reflective of light (e.g. the structured light projected by the two are more structured light cameras 228) and includes various air gaps between the metal portions of the lattice which provide no reflections. Hence, the imaging of the metal lattice by the two or more structured light cameras 228 is generally prone to error and/or can be unreliable.
However, due to the location and orientation of the structured light camera 218, the structured light camera 218 is configured to detect at least a top of the shopping cart 603, and in particular the non-reflective bumper 621, in a region adjacent to the mobile automation apparatus 103. As the bumper 621 is not specularly reflective, and is also a solid piece (e.g. lacking air gaps), the camera data 418 from the structured light camera 218 provides a more reliable indication of whether there is a shopping cart adjacent the apparatus 103.
In any event, once the shopping cart 603 is detected by the structured light camera 218, as determined using, for example, the shopping cart detector 410a, the controller 320 controls the navigation module 340 to stop and/or slow down and/or navigate around the shopping cart 603. For example, the navigation module 340 moves the apparatus 103 to a new path that goes around the shopping cart 603 (e.g. as depicted in
The region where the top of the shopping cart 603 is detectable by the structured light camera 218 is generally controlled by adjusting the height and orientation of the structured light camera 218. For example, by raising the structured light camera 218 on the mast 214, and/or increasing the angle of the structured light camera 218 outwards from the mast 214, the structured light camera 218 detects the shopping cart 603 when it is further from the apparatus 103; similarly, by lowering the structured light camera 218 on the mast 214, and/or decreasing the angle of the structured light camera 218 inwards, the structured light camera 218 detects the shopping cart 603 when it is closer to the apparatus 103. The structured light camera 218 is, however, at a height and orientation where the bumper 621 of the shopping cart 603 is in a field of view of the structured light camera 218. Hence, in general, the height and orientation of the structured light camera 218 is based on a height of a shopping cart and a distance at which the shopping cart is to be detected. In the depicted embodiments, the shopping cart 603 is about 0.8 m high, and the structured light camera 218 is at a height of about 1.5 m, and an angle of about 45°±5°; hence the shopping cart 603 is detectable about 1 m away from the apparatus 103, which is a sufficient distance for the apparatus 103 to navigate around the shopping cart 603 when detected. However other heights and orientations are within the scope of present embodiments.
In some embodiments, the one or more ultrasound arrays 220 are used to determine a false negative at the structured light camera 218 and/or the second structured light cameras 228. For example, in some instances, including when a reflective metal lattice 605 is within their fields of view, the structured light cameras 218, 228 may not detect a shopping cart and/or other object and/or obstacle. As such, the one or more ultrasound arrays 220 are used at least as a secondary level of obstacle detection.
Hence, attention is now directed to
Furthermore, the example method 700 of
At block 701, the controller 320 determines when the one or more ultrasound arrays 220 detect a reflective object and/or an object having discontinuous planar surfaces (e.g., a reflective metal lattice 605 of the cart 603), and the structured light cameras 218, 228 simultaneously do not detect the object. For example, the block 701 occurs using the obstacle detector 410 in conjunction with one or more sets of camera data 418, 428 from at least one of the structured light cameras 218, 228.
At the block 702, the controller 320 determines whether the reflective object is within a threshold distance using, for example, the distance determiner 411 in conjunction with one or more sets of the ultrasound data 420 and/or the LIDAR data 421 (e.g. presuming that the one or more sets of camera data 418, 428 indicate that the object is not detected). In some embodiments, the threshold distance is set to a minimum distance at which the apparatus 103 is navigable around the object.
Furthermore, the block 702 is, in some embodiments, optional, and whether or not the block 702 is implemented is configurable at the apparatus 103.
When the object is not within the threshold distance (e.g. a “No” decision at the block 702), the block 701 repeats whenever an object is detected by the one or more ultrasound arrays 220. In particular, at the block 701, a determination of whether an object and the like is detected occurs periodically (for example, about every 100 ms) and, in some of these embodiments, a local “map” around the apparatus 103 is determined based on data from the structured light cameras 218, 228, the lidar device 221, the ultrasound arrays 220 and/or any other sensors used for detecting obstacles. The local “map” generally represents an obstacle landscape around the apparatus 103. In some of these embodiments, each cell of an occupancy grid (e.g. with cells of about 5 cm×5 cm) of the local map is reviewed by the controller 320 to determine whether or not an obstacle is present in each cell, according to the data received from the from the structured light cameras 218, 228, the lidar device 221, the ultrasound arrays 220. In these embodiments, the data from the ultrasound arrays 220 indicate an obstacle in a given cell, while the data from one or more of the structured light cameras 218, 228 indicate that there is no obstacle in the same given cell. However, other techniques for determining whether obstacles are detected by one or more of the ultrasound arrays 220, but not by at least one of the structured light cameras 218, 228 are within the scope of present embodiments.
However, when the object is within the threshold distance (e.g. a “Yes” decision at the block 702), at the block 703, the controller 320 controls the navigation module 340 to implement the obstacle avoidance protocol 412 as described above with respect to the block 503 of the method 500. Indeed, put another way, the ultrasound arrays 220 are used to detect a false negative of the structured light cameras 218, 228 including, but not limited to, a false negative when detecting a shopping cart, or any other object having reflective and/or discontinuous surfaces (e.g., mesh surfaces) within a field of view of the structured light cameras 218, 228.
In some embodiments, when the structured light cameras 218, 228 provide an indication that an object is detected, corresponding data from the one or more ultrasound arrays 220 is ignored and/or not used. Indeed, as the ultrasounds arrays 220 generally provide a lower resolution of detection than the structured light cameras 218, 228 (e.g. ultrasound images are generally lower resolution than structured light camera images), in some embodiments, the controller 320 ignores ultrasound data when the structured light cameras 218, 228 detect an obstacle (e.g., to reduce use of processing resources and/or reduce processing overhead in obstacle detection).
In some embodiments, the control of the navigation module 340 further depends on the distance to the object. For example, when the object is within a first threshold distance, where the apparatus 103 is not navigable around the object and/or where a collision with the object may occur, the navigation module 340 is controlled to stop and/or slow down the apparatus 103. Similarly, when the object is within a second threshold distance, greater than the first threshold distance, where the apparatus 103 is navigable around the object, the navigation module 340 is controlled to navigate the apparatus 103 around the object. When the embodiment of block 703 causes the apparatus 103 to stop and/or slow down, once the object is at least at the second threshold distance (presuming the object is moving), the navigation module 340 is controlled to navigate the apparatus 103 around the object.
However, if the distance to the object does not change, and the apparatus 103 is stopped, in some embodiments the controller 320 controls the interface 324 to transmit a notification to the server 101 that the apparatus 103 has stopped. In an embodiment, the notification includes a location of the apparatus 103. Upon receipt of the notification, the server 101 transmits a notice to the mobile device 105 that the apparatus 103 is stopped and may also notify the mobile device 105 of the location of the apparatus 103; this can cause a user of the mobile device 105 to investigate why the apparatus 103 has stopped.
Alternatively, if the distance to the object does not change, and the apparatus 103 is stopped, after a given time period, the navigation module 340 is controlled to navigate the apparatus 103 to a position, relative to the object, where the object is at least at the second threshold distance (for example by moving the apparatus 103 in an opposite direction to a movement direction where the object was detected), and the apparatus 103 navigates around the object.
The method 700 repeats each time the controller 320 determines that the one or more ultrasound arrays 220 detects an object, and the structured light cameras 218, 228 simultaneously do not detect the object. Hence, the method 700 generally runs in parallel with other methods at the apparatus 103, including, for example other data detection methods (e.g. for detecting prices, barcodes, and the like) and/or navigation methods.
Indeed, the method 700 is generally executed in parallel with the method 500 and, in some embodiments, the method 500 is modified to include aspects of the method 700. For example, in some embodiments, the method 500 is modified such that when the one or more ultrasound arrays 220 detects an object, for example the shopping cart 603, and the structured light camera 218 simultaneously does not detect the object, for example the shopping cart 603, the navigation module 340 is controlled to implement the obstacle avoidance protocol 412 and/the navigation module 340 is controlled to implement the obstacle avoidance protocol only when the object is within a threshold distance as described above.
The method 700 is now described with reference to
As the structured light cameras 218, 228 do not detect the object 803, but one or more of the ultrasound arrays 220 do detect the object 803, the obstacle avoidance protocol 412 is implemented, as described above. For example, attention is directed to
In yet further embodiments, the controller 320 can be configured to recognize obstacle types; for example, when the object 803 comprises a shopping cart, a person, a display, and/or any object having a particular physical configuration that can be determined from data from the one or more of the ultrasound arrays 220 and/or the lidar device 221, the obstacle avoidance protocol can be object specific. For example, when the object 803 includes a person, the apparatus 103 can stop; however, when the object 803 includes a shopping cart (e.g. without a person), display, a pallet, and the like, the apparatus 803 can go around the object 803.
In the foregoing specification, specific embodiments have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the invention.
For example, the method 700 is modified, in some embodiments, to control the navigation module 340 to implement the obstacle avoidance protocol 412 when one or more of the structured light cameras 218, 228 and/or the LIDAR device 221 detect an object regardless of whether the one or more ultrasound arrays 220 detect the object. In other words, in some embodiments, the navigation module 340 to implement the obstacle avoidance protocol 412 whenever one or more of the structured light cameras 218, 228, the ultrasound arrays and the LIDAR device 221 detects an object, regardless of whether the other sensors detect an object.
For example, while each of the structured light cameras 218, 228 are described as facing in a forward direction (e.g. in a direction of movement of the apparatus 103), in another example, one or more of the structured light cameras 218, 228 are pointed in other directions, and/or the apparatus 103 further comprises other structured light cameras facing in other directions.
Furthermore, the fields of view of the structured light cameras 218, 228 are arranged, in some embodiments, such that a field of view of the structured light camera 218 is in a dead zone of the fields of view of the structured light cameras 228. For example, attention is next directed to
Indeed, such an arrangement can ensure that objects in a region adjacent to the apparatus 103 are detectable by the structured light camera 218, and that objects on either side of this region are detectable by the structured light cameras 228, Indeed, the relative orientations of the structured light cameras 218, 228, as well as the orientations of one or more LIDAR devices and/or depth sensing devices, and/or ultrasound arrays, are generally selected to increase a range over which objects around the apparatus 103 are detectable.
Furthermore, as the structured light camera 218 is pointed towards the floor 213, the structured light camera 218 further detects objects that are above or below the fields of view of each of the LIDAR device 221, the ultrasound arrays 220, and the structured light camera 228.
Indeed, as the apparatus 103 is, in some embodiments, configured to navigate in many directions, the sensors at the apparatus 103 are generally arranged to sense objects over the directions in which the apparatus 103 is configured to navigate.
For example, in some embodiments, the apparatus 103 is configured to move only longitudinally (e.g. forward and backward, as well as steering and/or turning in each of these directions) but not sideways (e.g. non-holonomic). In these embodiments, the sensors at the apparatus 103 are generally arranged to cover a range around the apparatus 103 over which the apparatus 103 is configured to move.
For example, the sensors described herein do not necessarily detect an entirety of an object and/or an obstacle, but rather the sensors are arranged to detect a portion and/or a region of the obstacle in order to infer structure sufficient to implement the obstacle avoidance protocol. Indeed, the sensors described herein are generally low cost readily available sensors, which are arranged to work together so that forward facing regions (e.g. in a movement direction) are generally covered by the sensors. For example, if LIDAR does not detect any part of an obstacle, then a structured light camera may detect part of the obstacle; however, if the structured light camera does not detect part of the obstacle, then the ultrasound arrays should detect at least a part of the obstacle. In other words, as the apparatus 103 is provisioned with three types of sensors for obstacle detection, the obstacle avoider 401 includes primary, secondary, and tertiary levels of obstacle detection.
The LIDAR devices described herein may have difficulties detecting light absorbing (black/dark) materials or highly reflective materials. Furthermore, the LIDAR devices described herein may be 2D (two-dimensional) planar sensors, which hence detect obstacles over a slice of the environment in which the apparatus 103 is navigating, and at ankle level.
Structured light cameras described herein may have difficulties detecting light absorbing (black/dark) materials or highly reflective materials. However, structured light cameras can otherwise detect a large percentage of the obstacles found in the retail environment. Ultrasound arrays and/or sensors may have issues detecting obstacles with high resolution since the sound waves are directed over a wide angle. However, ultrasound arrays and/or sensors very accurately detect highly reflective materials or very small objects.
Hence, in the apparatus described herein, a LIDAR device mounted at a front of the apparatus at ankle height is used to detect a large majority of the obstacles found at ankle height. Any obstacle found above or below ankle height is to be detected using a different sensor modality. Multiple structured light cameras (e.g. structured light cameras 228) are used to detect shoppers, aisle shelves, doors, baskets, shopping carts, items protruding from shelf, and any obstacle that is more than 1 m away from the apparatus 103 (and/or outside of a field of view of the structured light camera 218) and is in the range of the structured light cameras 228. However, in order to detect a shopping cart, the structured light camera 218 is placed at a height of about 1.5 m angled downwards at about 45 degrees, which allows for detecting a plastic protector ring on the top of a shopping cart. From the camera data captured by the structured light camera 228, at least the plastic protector ring of the shopping cart is detectable which allows the shopping cart to be avoided by the apparatus 103, as per the method 500.
However, as a failsafe, as per the method 700, the one or more ultrasound arrays 220 are used to detect obstacles within close proximity to the apparatus 103, for example, which, when detected, can cause the apparatus to stop and/or slow down (and/or prepare to stop and/or slow down), for example, when the obstacle is about 1 m from the apparatus 103. Indeed, the ultrasound sensors of the one or more ultrasound arrays 220 are, in some embodiments, positioned to form a phased ultrasound array which can detect shiny reflective obstacles such as reflective shopping cart meshes and/or lattices, within 1 m from the apparatus 103. Indeed, phased ultrasound array can further increase the obstacle detection resolution.
In the foregoing specification, specific embodiments have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of present teachings.
The benefits, advantages, solutions to problems, and any element(s) that may cause any benefit, advantage, or solution to occur or become more pronounced are not to be construed as a critical, required, or essential features or elements of any or all the claims. The invention is defined solely by the appended claims including any amendments made during the pendency of this application and all equivalents of those claims as issued.
Moreover, in this document, language of “at least one of X, Y, and Z” and “one or more of X, Y and Z” can be construed as X only, Y only, Z only, or any combination of two or more items X, Y, and Z (e.g., XYZ, XY, YZ, XZ, and the like). Similar logic can be applied for two or more items in any occurrence of “at least one . . . ” and “one or more . . . ” language.
Moreover in this document, relational terms such as first and second, top and bottom, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. The terms “comprises,” “comprising,” “has”, “having,” “includes”, “including,” “contains”, “containing” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises, has, includes, contains a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. An element proceeded by “comprises . . . a”, “has . . . a”, “includes . . . a”, “contains . . . a” does not, without more constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that comprises, has, includes, contains the element. The terms “a” and “an” are defined as one or more unless explicitly stated otherwise herein. The terms “substantially”, “essentially”, “approximately”, “about” or any other version thereof, are defined as being close to as understood by one of ordinary skill in the art, and in one non-limiting embodiment the term is defined to be within 10%, in another embodiment within 5%, in another embodiment within 1% and in another embodiment within 0.5%. The term “coupled” as used herein is defined as connected, although not necessarily directly and not necessarily mechanically. A device or structure that is “configured” in a certain way is configured in at least that way, but may also be configured in ways that are not listed.
It will be appreciated that some embodiments may be comprised of one or more generic or specialized processors (or “processing devices”) such as microprocessors, digital signal processors, customized processors and field programmable gate arrays (FPGAs) and unique stored program instructions (including both software and firmware) that control the one or more processors to implement, in conjunction with certain non-processor circuits, some, most, or all of the functions of the method and/or apparatus described herein. Alternatively, some or all functions could be implemented by a state machine that has no stored program instructions, or in one or more application specific integrated circuits (ASICs), in which each function or some combinations of certain of the functions are implemented as custom logic. Of course, a combination of the two approaches could be used.
Moreover, an embodiment can be implemented as a computer-readable storage medium having computer readable code stored thereon for programming a computer (e.g., comprising a processor) to perform a method as described and claimed herein. Examples of such computer-readable storage mediums include, but are not limited to, a hard disk, a CD-ROM, an optical storage device, a magnetic storage device, a ROM (Read Only Memory), a PROM (Programmable Read Only Memory), an EPROM (Erasable Programmable Read Only Memory), an EEPROM (Electrically Erasable Programmable Read Only Memory) and a Flash memory. Further, it is expected that one of ordinary skill, notwithstanding possibly significant effort and many design choices motivated by, for example, available time, current technology, and economic considerations, when guided by the concepts and principles disclosed herein will be readily capable of generating such software instructions and programs and ICs with minimal experimentation.
The Abstract of the Disclosure is provided to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in various embodiments for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.
This application claims the benefit of priority to U.S. Provisional Application No. 62/492,670 entitled “Product Status Detection System,” filed on May 1, 2017, by Perrella et al., and to U.S. Provisional Application No. 62/492,712 entitled “Obstacle Detection For A Mobile Automation Apparatus,” filed on May 1, 2017, by Arandorenko et al., all of which are incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
5209712 | Ferri | May 1993 | A |
5214615 | Bauer | May 1993 | A |
5408322 | Hsu et al. | Apr 1995 | A |
5414268 | McGee | May 1995 | A |
5423617 | Marsh et al. | Jun 1995 | A |
5534762 | Kim | Jul 1996 | A |
5566280 | Fukui et al. | Oct 1996 | A |
5704049 | Briechle | Dec 1997 | A |
5953055 | Huang et al. | Sep 1999 | A |
5988862 | Kacyra et al. | Nov 1999 | A |
6026376 | Kenney | Feb 2000 | A |
6034379 | Bunte et al. | Mar 2000 | A |
6075905 | Herman et al. | Jun 2000 | A |
6115114 | Berg et al. | Sep 2000 | A |
6141293 | Amorai-Moriya et al. | Oct 2000 | A |
6304855 | Burke | Oct 2001 | B1 |
6442507 | Skidmore et al. | Aug 2002 | B1 |
6549825 | Kurata | Apr 2003 | B2 |
6580441 | Schileru-Key | Jun 2003 | B2 |
6711293 | Lowe | Mar 2004 | B1 |
6721723 | Gibson et al. | Apr 2004 | B1 |
6721769 | Rappaport et al. | Apr 2004 | B1 |
6836567 | Silver et al. | Dec 2004 | B1 |
6995762 | Pavlidis et al. | Feb 2006 | B1 |
7090135 | Patel | Aug 2006 | B2 |
7137207 | Armstrong et al. | Nov 2006 | B2 |
7245558 | Willins et al. | Jul 2007 | B2 |
7248754 | Cato | Jul 2007 | B2 |
7277187 | Smith et al. | Oct 2007 | B2 |
7373722 | Cooper et al. | May 2008 | B2 |
7474389 | Greenberg et al. | Jan 2009 | B2 |
7487595 | Armstrong et al. | Feb 2009 | B2 |
7493336 | Noonan | Feb 2009 | B2 |
7508794 | Feather et al. | Mar 2009 | B2 |
7527205 | Zhu et al. | May 2009 | B2 |
7605817 | Zhang et al. | Oct 2009 | B2 |
7647752 | Magnell | Jan 2010 | B2 |
7693757 | Zimmerman | Apr 2010 | B2 |
7726575 | Wang et al. | Jun 2010 | B2 |
7751928 | Antony et al. | Jul 2010 | B1 |
7783383 | Eliuk et al. | Aug 2010 | B2 |
7839531 | Sugiyama | Nov 2010 | B2 |
7845560 | Emanuel et al. | Dec 2010 | B2 |
7885865 | Benson et al. | Feb 2011 | B2 |
7925114 | Mai et al. | Apr 2011 | B2 |
7957998 | Riley et al. | Jun 2011 | B2 |
7996179 | Lee et al. | Aug 2011 | B2 |
8009864 | Linaker et al. | Aug 2011 | B2 |
8049621 | Egan | Nov 2011 | B1 |
8091782 | Cato et al. | Jan 2012 | B2 |
8094902 | Crandall et al. | Jan 2012 | B2 |
8094937 | Teoh et al. | Jan 2012 | B2 |
8132728 | Dwinell et al. | Mar 2012 | B2 |
8134717 | Pangrazio et al. | Mar 2012 | B2 |
8189855 | Opalach et al. | May 2012 | B2 |
8199977 | Krishnaswamy et al. | Jun 2012 | B2 |
8207964 | Meadow et al. | Jun 2012 | B1 |
8233055 | Matsunaga et al. | Jul 2012 | B2 |
8260742 | Cognigni et al. | Sep 2012 | B2 |
8265895 | Willins et al. | Sep 2012 | B2 |
8277396 | Scott et al. | Oct 2012 | B2 |
8284988 | Sones et al. | Oct 2012 | B2 |
8423431 | Rouaix et al. | Apr 2013 | B1 |
8429004 | Hamilton et al. | Apr 2013 | B2 |
8463079 | Ackley et al. | Jun 2013 | B2 |
8479996 | Barkan et al. | Jul 2013 | B2 |
8520067 | Ersue | Aug 2013 | B2 |
8542252 | Perez et al. | Sep 2013 | B2 |
8571314 | Tao et al. | Oct 2013 | B2 |
8599303 | Stettner | Dec 2013 | B2 |
8630924 | Groenevelt et al. | Jan 2014 | B2 |
8660338 | Ma et al. | Feb 2014 | B2 |
8743176 | Stettner et al. | Jun 2014 | B2 |
8757479 | Clark et al. | Jun 2014 | B2 |
8812226 | Zeng | Aug 2014 | B2 |
8923893 | Austin et al. | Dec 2014 | B2 |
8939369 | Olmstead et al. | Jan 2015 | B2 |
8954188 | Sullivan et al. | Feb 2015 | B2 |
8958911 | Wong et al. | Feb 2015 | B2 |
8971637 | Rivard | Mar 2015 | B1 |
8989342 | Liesenfelt et al. | Mar 2015 | B2 |
9007601 | Steffey et al. | Apr 2015 | B2 |
9037287 | Grauberger et al. | May 2015 | B1 |
9064394 | Trundle | Jun 2015 | B1 |
9070285 | Ramu et al. | Jun 2015 | B1 |
9072929 | Rush et al. | Jul 2015 | B1 |
9120622 | Elazary et al. | Sep 2015 | B1 |
9129277 | Macintosh | Sep 2015 | B2 |
9135491 | Morandi et al. | Sep 2015 | B2 |
9159047 | Winkel | Oct 2015 | B2 |
9171442 | Clements | Oct 2015 | B2 |
9247211 | Zhang et al. | Jan 2016 | B2 |
9329269 | Zeng | May 2016 | B2 |
9349076 | Liu et al. | May 2016 | B1 |
9367831 | Besehanic | Jun 2016 | B1 |
9380222 | Clayton et al. | Jun 2016 | B2 |
9396554 | Williams et al. | Jul 2016 | B2 |
9400170 | Steffey | Jul 2016 | B2 |
9424482 | Patel et al. | Aug 2016 | B2 |
9517767 | Kentley et al. | Dec 2016 | B1 |
9542746 | Wu et al. | Jan 2017 | B2 |
9549125 | Goyal et al. | Jan 2017 | B1 |
9562971 | Shenkar et al. | Feb 2017 | B2 |
9565400 | Curlander et al. | Feb 2017 | B1 |
9589353 | Mueller-Fischer et al. | Mar 2017 | B2 |
9600731 | Yasunaga et al. | Mar 2017 | B2 |
9600892 | Patel et al. | Mar 2017 | B2 |
9612123 | Levinson | Apr 2017 | B1 |
9639935 | Douady-Pleven et al. | May 2017 | B1 |
9697429 | Patel et al. | Jul 2017 | B2 |
9766074 | Roumeliotis et al. | Sep 2017 | B2 |
9778388 | Connor | Oct 2017 | B1 |
9779205 | Namir | Oct 2017 | B2 |
9791862 | Connor | Oct 2017 | B1 |
9805240 | Zheng et al. | Oct 2017 | B1 |
9811754 | Schwartz | Nov 2017 | B2 |
9827683 | Hance et al. | Nov 2017 | B1 |
9880009 | Bell | Jan 2018 | B2 |
9928708 | Lin et al. | Mar 2018 | B2 |
9953420 | Wolski et al. | Apr 2018 | B2 |
9980009 | Jiang et al. | May 2018 | B2 |
9994339 | Colson et al. | Jun 2018 | B2 |
9996818 | Ren et al. | Jun 2018 | B1 |
10019803 | Venable et al. | Jul 2018 | B2 |
10111646 | Nycz et al. | Oct 2018 | B2 |
10121072 | Kekatpure | Nov 2018 | B1 |
10127438 | Fisher et al. | Nov 2018 | B1 |
10133951 | Mendonca et al. | Nov 2018 | B1 |
10197400 | Jesudason et al. | Feb 2019 | B2 |
10210603 | Venable et al. | Feb 2019 | B2 |
10229386 | Thomas | Mar 2019 | B2 |
10248653 | Blassin et al. | Apr 2019 | B2 |
10262294 | Hahn et al. | Apr 2019 | B1 |
10265871 | Hance et al. | Apr 2019 | B2 |
10289990 | Rizzolo et al. | May 2019 | B2 |
10336543 | Sills et al. | Jul 2019 | B1 |
10349031 | DeLuca | Jul 2019 | B2 |
10352689 | Brown et al. | Jul 2019 | B2 |
10373116 | Medina et al. | Aug 2019 | B2 |
10394244 | Song et al. | Aug 2019 | B2 |
20010031069 | Kondo et al. | Oct 2001 | A1 |
20010041948 | Ross et al. | Nov 2001 | A1 |
20020006231 | Jayant et al. | Jan 2002 | A1 |
20020059202 | Hadzikadic et al. | May 2002 | A1 |
20020097439 | Braica | Jul 2002 | A1 |
20020146170 | Rom | Oct 2002 | A1 |
20020158453 | Levine | Oct 2002 | A1 |
20020164236 | Fukuhara et al. | Nov 2002 | A1 |
20030003925 | Suzuki | Jan 2003 | A1 |
20030094494 | Blanford et al. | May 2003 | A1 |
20030174891 | Wenzel et al. | Sep 2003 | A1 |
20040021313 | Gardner et al. | Feb 2004 | A1 |
20040084527 | Bong et al. | May 2004 | A1 |
20040131278 | Imagawa et al. | Jul 2004 | A1 |
20040240754 | Smith et al. | Dec 2004 | A1 |
20050016004 | Armstrong et al. | Jan 2005 | A1 |
20050114059 | Chang et al. | May 2005 | A1 |
20050174351 | Chang | Aug 2005 | A1 |
20050213082 | DiBernardo et al. | Sep 2005 | A1 |
20050213109 | Schell et al. | Sep 2005 | A1 |
20060032915 | Schwartz | Feb 2006 | A1 |
20060045325 | Zavadsky et al. | Mar 2006 | A1 |
20060064286 | Fink et al. | Mar 2006 | A1 |
20060106742 | Bochicchio et al. | May 2006 | A1 |
20060279527 | Zehner et al. | Dec 2006 | A1 |
20060285486 | Roberts et al. | Dec 2006 | A1 |
20070036398 | Chen | Feb 2007 | A1 |
20070074410 | Armstrong et al. | Apr 2007 | A1 |
20070272732 | Hindmon | Nov 2007 | A1 |
20080002866 | Fujiwara | Jan 2008 | A1 |
20080025565 | Zhang et al. | Jan 2008 | A1 |
20080027591 | Lenser | Jan 2008 | A1 |
20080077511 | Zimmerman | Mar 2008 | A1 |
20080159634 | Sharma et al. | Jul 2008 | A1 |
20080164310 | Dupuy et al. | Jul 2008 | A1 |
20080175513 | Lai et al. | Jul 2008 | A1 |
20080181529 | Michel et al. | Jul 2008 | A1 |
20080183730 | Enga | Jul 2008 | A1 |
20080238919 | Pack | Oct 2008 | A1 |
20080294487 | Nasser | Nov 2008 | A1 |
20090009123 | Skaff | Jan 2009 | A1 |
20090024353 | Lee et al. | Jan 2009 | A1 |
20090057411 | Madej et al. | Mar 2009 | A1 |
20090059270 | Opalach et al. | Mar 2009 | A1 |
20090060349 | Linaker et al. | Mar 2009 | A1 |
20090063306 | Fano et al. | Mar 2009 | A1 |
20090063307 | Groenovelt et al. | Mar 2009 | A1 |
20090074303 | Filimonova et al. | Mar 2009 | A1 |
20090088975 | Sato et al. | Apr 2009 | A1 |
20090103773 | Wheeler et al. | Apr 2009 | A1 |
20090125350 | Lessing et al. | May 2009 | A1 |
20090125535 | Basso et al. | May 2009 | A1 |
20090152391 | McWhirk | Jun 2009 | A1 |
20090160975 | Kwan | Jun 2009 | A1 |
20090192921 | Hicks | Jul 2009 | A1 |
20090206161 | Olmstead | Aug 2009 | A1 |
20090236155 | Skaff | Sep 2009 | A1 |
20090252437 | Li et al. | Oct 2009 | A1 |
20090287587 | Bloebaum et al. | Nov 2009 | A1 |
20090323121 | Valkenburg et al. | Dec 2009 | A1 |
20100017407 | Beniyama et al. | Jan 2010 | A1 |
20100026804 | Tanizaki et al. | Feb 2010 | A1 |
20100066587 | Yamauchi | Mar 2010 | A1 |
20100070365 | Siotia et al. | Mar 2010 | A1 |
20100082194 | Yabushita et al. | Apr 2010 | A1 |
20100091094 | Sekowski | Apr 2010 | A1 |
20100118116 | Tomasz et al. | May 2010 | A1 |
20100131234 | Stewart et al. | May 2010 | A1 |
20100141806 | Uemura et al. | Jun 2010 | A1 |
20100161569 | Schreter | Jun 2010 | A1 |
20100171826 | Hamilton et al. | Jul 2010 | A1 |
20100208039 | Setettner | Aug 2010 | A1 |
20100214873 | Somasundaram et al. | Aug 2010 | A1 |
20100235033 | Yamamoto et al. | Sep 2010 | A1 |
20100241289 | Sandberg | Sep 2010 | A1 |
20100257149 | Cognigni et al. | Oct 2010 | A1 |
20100295850 | Katz et al. | Nov 2010 | A1 |
20100315412 | Sinha et al. | Dec 2010 | A1 |
20100326939 | Clark et al. | Dec 2010 | A1 |
20110047636 | Stachon et al. | Feb 2011 | A1 |
20110052043 | Hyung et al. | Mar 2011 | A1 |
20110093306 | Nielsen et al. | Apr 2011 | A1 |
20110137527 | Simon et al. | Jun 2011 | A1 |
20110168774 | Magal | Jul 2011 | A1 |
20110172875 | Gibbs | Jul 2011 | A1 |
20110188759 | Filimonova et al. | Aug 2011 | A1 |
20110216063 | Hayes | Sep 2011 | A1 |
20110242286 | Pace et al. | Oct 2011 | A1 |
20110246503 | Bender et al. | Oct 2011 | A1 |
20110254840 | Halstead | Oct 2011 | A1 |
20110286007 | Pangrazio et al. | Nov 2011 | A1 |
20110288816 | Thierman | Nov 2011 | A1 |
20110310088 | Adabala et al. | Dec 2011 | A1 |
20120017028 | Tsirkin | Jan 2012 | A1 |
20120019393 | Wolinsky et al. | Jan 2012 | A1 |
20120022913 | VolKmann et al. | Jan 2012 | A1 |
20120051730 | Cote et al. | Mar 2012 | A1 |
20120069051 | Hagbi et al. | Mar 2012 | A1 |
20120075342 | Choubassi et al. | Mar 2012 | A1 |
20120133639 | Kopf et al. | May 2012 | A1 |
20120307108 | Forutanpour | Jun 2012 | A1 |
20120169530 | Padmanabhan et al. | Jul 2012 | A1 |
20120179621 | Moir et al. | Jul 2012 | A1 |
20120185112 | Sung et al. | Jul 2012 | A1 |
20120194644 | Newcombe et al. | Aug 2012 | A1 |
20120197464 | Wang et al. | Aug 2012 | A1 |
20120201466 | Funayama et al. | Aug 2012 | A1 |
20120209553 | Doytchinov et al. | Aug 2012 | A1 |
20120236119 | Rhee et al. | Sep 2012 | A1 |
20120249802 | Taylor | Oct 2012 | A1 |
20120250978 | Taylor | Oct 2012 | A1 |
20120269383 | Bobbitt et al. | Oct 2012 | A1 |
20120287249 | Choo et al. | Nov 2012 | A1 |
20120323620 | Hofman et al. | Dec 2012 | A1 |
20130030700 | Miller et al. | Jan 2013 | A1 |
20130076586 | Karhuketo et al. | Mar 2013 | A1 |
20130090881 | Janardhanan et al. | Apr 2013 | A1 |
20130119138 | Winkel | May 2013 | A1 |
20130132913 | Fu et al. | May 2013 | A1 |
20130134178 | Lu | May 2013 | A1 |
20130138246 | Gutmann et al. | May 2013 | A1 |
20130138534 | Herwig | May 2013 | A1 |
20130142421 | Silver et al. | Jun 2013 | A1 |
20130144565 | Miller | Jun 2013 | A1 |
20130154802 | O'Haire et al. | Jun 2013 | A1 |
20130156292 | Chang et al. | Jun 2013 | A1 |
20130162806 | Ding et al. | Jun 2013 | A1 |
20130169681 | Rasane et al. | Jul 2013 | A1 |
20130176398 | Bonner et al. | Jul 2013 | A1 |
20130178227 | Vartanian et al. | Jul 2013 | A1 |
20130182114 | Zhang et al. | Jul 2013 | A1 |
20130226344 | Wong et al. | Aug 2013 | A1 |
20130228620 | Ahem et al. | Sep 2013 | A1 |
20130232039 | Jackson et al. | Sep 2013 | A1 |
20130235165 | Gharib et al. | Sep 2013 | A1 |
20130235206 | Smith et al. | Sep 2013 | A1 |
20130236089 | Litvak et al. | Sep 2013 | A1 |
20130278631 | Border et al. | Oct 2013 | A1 |
20130299306 | Jiang et al. | Nov 2013 | A1 |
20130299313 | Baek, IV et al. | Nov 2013 | A1 |
20130300729 | Grimaud | Nov 2013 | A1 |
20130303193 | Dharwada et al. | Nov 2013 | A1 |
20130321418 | Kirk | Dec 2013 | A1 |
20130329013 | Metois et al. | Dec 2013 | A1 |
20130341400 | Lancaster-Larocque | Dec 2013 | A1 |
20130342363 | Paek et al. | Dec 2013 | A1 |
20140002597 | Taguchi et al. | Jan 2014 | A1 |
20140003655 | Gopalkrishnan et al. | Jan 2014 | A1 |
20140003727 | Lortz et al. | Jan 2014 | A1 |
20140006229 | Birch et al. | Jan 2014 | A1 |
20140016832 | Kong et al. | Jan 2014 | A1 |
20140019311 | Tanaka | Jan 2014 | A1 |
20140025201 | Ryu et al. | Jan 2014 | A1 |
20140028837 | Gao et al. | Jan 2014 | A1 |
20140047342 | Breternitz et al. | Feb 2014 | A1 |
20140049616 | Stettner | Feb 2014 | A1 |
20140052555 | MacIntosh | Feb 2014 | A1 |
20140086483 | Zhang et al. | Mar 2014 | A1 |
20140098094 | Neumann et al. | Apr 2014 | A1 |
20140100813 | Shaowering | Apr 2014 | A1 |
20140104413 | McCloskey et al. | Apr 2014 | A1 |
20140112537 | Frank et al. | Apr 2014 | A1 |
20140129027 | Schnittman | May 2014 | A1 |
20140156133 | Cullinane et al. | Jun 2014 | A1 |
20140161359 | Magri et al. | Jun 2014 | A1 |
20140192050 | Qiu et al. | Jul 2014 | A1 |
20140195095 | Flohr et al. | Jul 2014 | A1 |
20140195374 | Bassemir et al. | Jul 2014 | A1 |
20140214547 | Signorelli et al. | Jul 2014 | A1 |
20140214600 | Argue et al. | Jul 2014 | A1 |
20140267614 | Ding et al. | Sep 2014 | A1 |
20140267688 | Aich et al. | Sep 2014 | A1 |
20140277691 | Jacobus et al. | Sep 2014 | A1 |
20140277692 | Buzan et al. | Sep 2014 | A1 |
20140279294 | Field-Darragh et al. | Sep 2014 | A1 |
20140300637 | Fan et al. | Oct 2014 | A1 |
20140316875 | Tkachenko et al. | Oct 2014 | A1 |
20140330835 | Boyer | Nov 2014 | A1 |
20140344401 | Varney et al. | Nov 2014 | A1 |
20140351073 | Murphy et al. | Nov 2014 | A1 |
20140369607 | Patel et al. | Dec 2014 | A1 |
20150015602 | Beaudoin | Jan 2015 | A1 |
20150019391 | Kumar et al. | Jan 2015 | A1 |
20150029339 | Kobres et al. | Jan 2015 | A1 |
20150032304 | Nakamura et al. | Jan 2015 | A1 |
20150039458 | Reid | Feb 2015 | A1 |
20150052029 | Wu et al. | Feb 2015 | A1 |
20150088618 | Basir et al. | Mar 2015 | A1 |
20150088701 | Desmarais et al. | Mar 2015 | A1 |
20150088703 | Yan | Mar 2015 | A1 |
20150092066 | Geiss et al. | Apr 2015 | A1 |
20150106403 | Haverinen et al. | Apr 2015 | A1 |
20150117788 | Patel et al. | Apr 2015 | A1 |
20150139010 | Jeong et al. | May 2015 | A1 |
20150154467 | Feng et al. | Jun 2015 | A1 |
20150161793 | Takahashi | Jun 2015 | A1 |
20150170256 | Pettyjohn et al. | Jun 2015 | A1 |
20150181198 | Baele et al. | Jun 2015 | A1 |
20150212521 | Pack | Jul 2015 | A1 |
20150235157 | Avegliano et al. | Aug 2015 | A1 |
20150245358 | Schmidt | Aug 2015 | A1 |
20150262116 | Katircioglu et al. | Sep 2015 | A1 |
20150279035 | Wolski et al. | Oct 2015 | A1 |
20150298317 | Wang et al. | Oct 2015 | A1 |
20150310601 | Rodriguez et al. | Oct 2015 | A1 |
20150332368 | Vartiainen et al. | Nov 2015 | A1 |
20150352721 | Wicks et al. | Dec 2015 | A1 |
20150353280 | Brazeau et al. | Dec 2015 | A1 |
20150355639 | Versteeg et al. | Dec 2015 | A1 |
20150363625 | Wu et al. | Dec 2015 | A1 |
20150363758 | Wu | Dec 2015 | A1 |
20150365660 | Wu et al. | Dec 2015 | A1 |
20150379704 | Chandrasekar et al. | Dec 2015 | A1 |
20160026253 | Bradski et al. | Jan 2016 | A1 |
20160042223 | Suh et al. | Feb 2016 | A1 |
20160044862 | Kocer | Feb 2016 | A1 |
20160061591 | Pangrazio et al. | Mar 2016 | A1 |
20160070981 | Sasaki et al. | Mar 2016 | A1 |
20160092943 | Vigier et al. | Mar 2016 | A1 |
20160012588 | Taguchi et al. | Apr 2016 | A1 |
20160104041 | Bowers et al. | Apr 2016 | A1 |
20160107690 | Oyama et al. | Apr 2016 | A1 |
20160112628 | Super et al. | Apr 2016 | A1 |
20160114488 | Mascorro Medina | Apr 2016 | A1 |
20160129592 | Saboo | May 2016 | A1 |
20160132815 | Itoko et al. | May 2016 | A1 |
20160150217 | Popov | May 2016 | A1 |
20160156898 | Ren et al. | Jun 2016 | A1 |
20160163067 | Williams et al. | Jun 2016 | A1 |
20160171336 | Schwartz | Jun 2016 | A1 |
20160171429 | Schwartz | Jun 2016 | A1 |
20160171707 | Schwartz | Jun 2016 | A1 |
20160185347 | Lefevre et al. | Jun 2016 | A1 |
20160191759 | Somanath et al. | Jun 2016 | A1 |
20160221549 | Tanase | Aug 2016 | A1 |
20160224927 | Pettersson | Aug 2016 | A1 |
20160253735 | Scudillo et al. | Sep 2016 | A1 |
20160253844 | Petrovskaya et al. | Sep 2016 | A1 |
20160259329 | High et al. | Sep 2016 | A1 |
20160260051 | Wu et al. | Sep 2016 | A1 |
20160260054 | High et al. | Sep 2016 | A1 |
20160271795 | Vicenti | Sep 2016 | A1 |
20160313133 | Zeng et al. | Oct 2016 | A1 |
20160328618 | Patel et al. | Nov 2016 | A1 |
20160328767 | Bonner et al. | Nov 2016 | A1 |
20160353099 | Thomson et al. | Dec 2016 | A1 |
20160364634 | Davis et al. | Dec 2016 | A1 |
20170004649 | Collet Romea et al. | Jan 2017 | A1 |
20170011281 | Dijkman et al. | Jan 2017 | A1 |
20170011308 | Sun et al. | Jan 2017 | A1 |
20170032311 | Rizzolo et al. | Feb 2017 | A1 |
20170041553 | Cao et al. | Feb 2017 | A1 |
20170054965 | Raab et al. | Feb 2017 | A1 |
20170066459 | Singh | Mar 2017 | A1 |
20170074659 | Giurgiu et al. | Mar 2017 | A1 |
20170083774 | Solar et al. | Mar 2017 | A1 |
20170109940 | Guo et al. | Apr 2017 | A1 |
20170147966 | Aversa et al. | May 2017 | A1 |
20170150129 | Pangrazio | May 2017 | A1 |
20170178060 | Schwartz | Jun 2017 | A1 |
20170178227 | Gornish | Jun 2017 | A1 |
20170178310 | Gornish | Jun 2017 | A1 |
20170193434 | Shah et al. | Jul 2017 | A1 |
20170219338 | Brown et al. | Aug 2017 | A1 |
20170219353 | Alesiani | Aug 2017 | A1 |
20170227645 | Swope et al. | Aug 2017 | A1 |
20170227647 | Baik | Aug 2017 | A1 |
20170228885 | Baumgartner | Aug 2017 | A1 |
20170261993 | Venable et al. | Sep 2017 | A1 |
20170262724 | Wu et al. | Sep 2017 | A1 |
20170280125 | Brown et al. | Sep 2017 | A1 |
20170286773 | Skaff et al. | Oct 2017 | A1 |
20170286901 | Skaff et al. | Oct 2017 | A1 |
20170297478 | Sherman et al. | Oct 2017 | A1 |
20170323253 | Enssle et al. | Nov 2017 | A1 |
20170323376 | Glaser et al. | Nov 2017 | A1 |
20170337508 | Bogolea et al. | Nov 2017 | A1 |
20180001481 | Shah et al. | Jan 2018 | A1 |
20180005035 | Bogolea et al. | Jan 2018 | A1 |
20180005176 | Williams et al. | Jan 2018 | A1 |
20180020145 | Kotfis et al. | Jan 2018 | A1 |
20180051991 | Hong | Feb 2018 | A1 |
20180053091 | Savvides et al. | Feb 2018 | A1 |
20180053305 | Gu et al. | Feb 2018 | A1 |
20180075403 | Mascorro Medina et al. | Mar 2018 | A1 |
20180089613 | Chen et al. | Mar 2018 | A1 |
20180101813 | Paat et al. | Apr 2018 | A1 |
20180108120 | Venable et al. | Apr 2018 | A1 |
20180108134 | Venable et al. | Apr 2018 | A1 |
20180114183 | Howell | Apr 2018 | A1 |
20180129201 | Douglas et al. | May 2018 | A1 |
20180130011 | Jacobsson | May 2018 | A1 |
20180143003 | Clayton et al. | May 2018 | A1 |
20180174325 | Fu et al. | Jun 2018 | A1 |
20180190160 | Bryan et al. | Jul 2018 | A1 |
20180197139 | Hill | Jul 2018 | A1 |
20180201423 | Drzewiecki et al. | Jul 2018 | A1 |
20180204111 | Zadeh et al. | Jul 2018 | A1 |
20180218218 | Madan et al. | Aug 2018 | A1 |
20180251253 | Taira et al. | Sep 2018 | A1 |
20180276596 | Murthy et al. | Sep 2018 | A1 |
20180281191 | Sinyayskiy et al. | Oct 2018 | A1 |
20180293442 | Fridental et al. | Oct 2018 | A1 |
20180293543 | Tiwari | Oct 2018 | A1 |
20180306958 | Goss et al. | Oct 2018 | A1 |
20180313956 | Rzeszutek et al. | Nov 2018 | A1 |
20180314260 | Jen et al. | Nov 2018 | A1 |
20180314908 | Lam | Nov 2018 | A1 |
20180315007 | Kingsford et al. | Nov 2018 | A1 |
20180315065 | Zhang et al. | Nov 2018 | A1 |
20180315173 | Phan et al. | Nov 2018 | A1 |
20180315865 | Haist et al. | Nov 2018 | A1 |
20180321692 | Castillo-Effen et al. | Nov 2018 | A1 |
20180370727 | Hance et al. | Dec 2018 | A1 |
20190049962 | Ouellette et al. | Feb 2019 | A1 |
20190057588 | Savvides et al. | Feb 2019 | A1 |
20190065861 | Savvides et al. | Feb 2019 | A1 |
20190073554 | Rzeszutek | Mar 2019 | A1 |
20190073559 | Rzeszutek et al. | Mar 2019 | A1 |
20190073627 | Nakdimon et al. | Mar 2019 | A1 |
20190077015 | Shibasaki et al. | Mar 2019 | A1 |
20190087663 | Yamazaki et al. | Mar 2019 | A1 |
20190094876 | Moore et al. | Mar 2019 | A1 |
20190108606 | Komiyama | Apr 2019 | A1 |
20190160675 | Paschall, II et al. | May 2019 | A1 |
20190178436 | Mao et al. | Jun 2019 | A1 |
20190180150 | Taylor et al. | Jun 2019 | A1 |
20190197439 | Wang | Jun 2019 | A1 |
20190197728 | Yamao | Jun 2019 | A1 |
20190236530 | Cantrell et al. | Aug 2019 | A1 |
20190271984 | Kingsford | Sep 2019 | A1 |
20190304132 | Yoda et al. | Oct 2019 | A1 |
20190392212 | Sawhney et al. | Dec 2019 | A1 |
20200053325 | Deyle et al. | Feb 2020 | A1 |
20200314333 | Liang et al. | Oct 2020 | A1 |
Number | Date | Country |
---|---|---|
2835830 | Nov 2012 | CA |
3028156 | Jan 2018 | CA |
102214343 | Oct 2011 | CN |
104200086 | Dec 2014 | CN |
105989512 | Oct 2016 | CN |
107067382 | Aug 2017 | CN |
206952978 | Feb 2018 | CN |
766098 | Apr 1997 | EP |
1311993 | May 2007 | EP |
2309378 | Apr 2011 | EP |
2439487 | Apr 2012 | EP |
2472475 | Jul 2012 | EP |
2562688 | Feb 2013 | EP |
2662831 | Nov 2013 | EP |
2693362 | Feb 2014 | EP |
3400113 | Nov 2018 | EP |
3001567 | Aug 2014 | FR |
2323238 | Sep 1998 | GB |
2330265 | Apr 1999 | GB |
2014170431 | Sep 2014 | JP |
2016194834 | Nov 2016 | JP |
2017016539 | Jan 2017 | JP |
101234798 | Jan 2009 | KR |
1020190031431 | Mar 2019 | KR |
WO 9923600 | May 1999 | WO |
WO 2003002935 | Jan 2003 | WO |
WO 2003025805 | Mar 2003 | WO |
WO 2006136958 | Dec 2006 | WO |
WO 2007042251 | Apr 2007 | WO |
WO 2008057504 | May 2008 | WO |
WO 2008154611 | Dec 2008 | WO |
WO 2012103199 | Aug 2012 | WO |
WO 2012103202 | Aug 2012 | WO |
WO 2012154801 | Nov 2012 | WO |
WO 2013165674 | Nov 2013 | WO |
WO 2014066422 | May 2014 | WO |
WO 2014092552 | Jun 2014 | WO |
WO 2014181323 | Nov 2014 | WO |
WO 2015127503 | Sep 2015 | WO |
WO 2016020038 | Feb 2016 | WO |
WO 2017175312 | Oct 2017 | WO |
WO 2017187106 | Nov 2017 | WO |
WO 2018018007 | Jan 2018 | WO |
WO 2018204308 | Nov 2018 | WO |
WO 2018204342 | Nov 2018 | WO |
WO 2019023249 | Jan 2019 | WO |
Entry |
---|
“Fair Billing with Automatic Dimensioning” pp. 1-4, undated, Copyright Mettler-Toledo International Inc. |
“Plane Detection in Point Cloud Data” dated Jan. 25, 2010 by Michael Ying Yang and Wolfgang Forstner, Technical Report 1, 2010, University of Bonn. |
“Swift Dimension” Trademark Omniplanar, Copyright 2014. |
Ajmal S. Mian et al., “Three-Dimensional Model Based Object Recognition and Segmentation in Cluttered Scenes”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, No. 10, Oct. 2006. |
Batalin et al., “Mobile robot navigation using a sensor network,” IEEE, International Conference on robotics and automation, Apr. 26, May 1, 2004, pp. 636-641. |
Bazazian et al., “Fast and Robust Edge Extraction in Unorganized Point clouds,” IEEE, 2015 International Conference on Digital Image Computing: Techniques and Applicatoins (DICTA), Nov. 23-25, 2015, pp. 1-8. |
Biswas et al. “Depth Camera Based Indoor Mobile Robot Localization and Navigation” Robotics and Automation (ICRA), 2012 IEEE International Conference on IEEE, 2012. |
Bohm, Multi-Image Fusion for Occlusion-Free Façade Texturing, International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, pp. 867-872 (Jan. 2004). |
Bristow et al., “A Survey of Iterative Learning Control”, IEEE Control Systems, Jun. 2006, pp. 96-114. |
Buenaposada et al. “Realtime tracking and estimation of plane pose” Proceedings of the ICPR (Aug. 2002) vol. II, IEEE pp. 697-700. |
Carreira et al., “Enhanced PCA-based localization using depth maps with missing data,” IEEE, pp. 1-8, Apr. 24, 2013. |
Notice of allowance for U.S. Appl. No. 14/068,495 dated Apr. 25, 2016. |
Notice of allowance for U.S. Appl. No. 14/518,091 dated Apr. 12, 2017. |
Notice of allowance for U.S. Appl. No. 15/211,103 dated Apr. 5, 2017. |
Olson, Clark F., etal. “Wide-Baseline Stereo Vision for terrain Mapping” in Machine Vision and Applications, Aug. 2010. |
Oriolo et al., “An iterative learning controller for nonholonomic mobile Robots”, the international Journal of Robotics Research, Aug. 1997, pp. 954-970. |
Ostafew et al., “Visual Teach and Repeat, Repeat, Repeat: Iterative learning control to improve mobile robot path tracking in challenging outdoor environment”, IEEE/RSJ International Conference on Intelligent robots and Systems, Nov. 2013, p. 176-181. |
Park et al., “Autonomous mobile robot navigation using passive rfid in indoor environment,” IEEE, Transactions on industrial electronics, vol. 56, issue 7, pp. 2366-2373 (Jul. 2009). |
Perveen et al. (An overview of template matching methodologies and its application, International Journal of Research in Computer and Communication Technology, v2nl0, Oct. 2013) (Year: 2013). |
Pivtoraiko et al., “Differentially constrained mobile robot motion planning in state lattices”, journal of field robotics, vol. 26, No. 3, 2009, pp. 308-333. |
Pratt W K Ed: “Digital Image processing, 10-image enhancement, 17-image segmentation”, Jan. 1, 2001, Digital Image Processing: PIKS Inside, New York: John Wily & Sons, US, pp. 243-258, 551. |
Puwein, J., et al.“Robust Multi-view camera calibration for wide-baseline camera networks” in IEEE Workshop on Applications of computer vision (WACV), Jan. 2011. |
Rusu, et al. “How to incrementally register pairs of clouds,” PCL Library, retrieved from internet on Aug. 22, 2016 [http://pointclouds.org/documentation/tutorials/pairwise_incremental_registration.php. |
Rusu, et al. “Spatial Change detection on unorganized point cloud data,” PCL Library, retrieved from internet on Aug. 19, 2016 [http://pointclouds.org/documentation/tutorials/octree_change.php]. |
Schnabel et al. “Efficient RANSAC for Point-Cloud Shape Detection”, vol. 0, No. 0, pp. 1-12 (1981). |
Senthilkumaran, et al., “Edge Detection Techniques for Image Segmentation—A Survey of Soft Computing Approaches”, International Journal of Recent Trends in Engineering, vol. 1, No. 2 (May 2009). |
Szeliski, “Modified Hough Transform”, Computer Vision. Copyright 2011, pp. 251-254. Retrieved on Aug. 17, 2017 [http://szeliski.org/book/drafts/SzeliskiBook_20100903_draft.pdf]. |
Tahir, Rabbani, et al., “Segmentation of point clouds using smoothness constraint,”International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences 36.5 (Sep. 2006): 248-253. |
Trevor et al., “Tables, Counters, and Shelves: Semantic Mapping of Surfaces in 3D,” Retrieved from Internet Jul. 3, 2018 @ http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.703.5365&rep=rep1&type=p. |
Tseng, et al., “A Cloud Removal Approach for Aerial Image Visualization”, International Journal of Innovative Computing, Information & Control, vol. 9, No. 6, pp. 2421-2440 (Jun. 2013). |
Uchiyama, et al., “Removal of Moving Objects from a Street-View Image by Fusing Multiple Image Sequences”, Pattern Recognition, 2010, 20th International Conference On, IEEE, Piscataway, NJ pp. 3456-3459 (Aug. 23, 2010). |
United Kingdom Intellectual Property Office, “Combined Search and Examination Report” for GB Patent Application No. 1813580.6 dated Feb. 21, 2019. |
United Kingdom Intellectual Property Office, Combined Search and Examination Report dated Jan. 22, 2016 for GB Patent Application No. 1417218.3. |
United Kingdom Intellectual Property Office, Combined Search and Examination Report dated Jan. 22, 2016 for GB Patent Application No. 1521272.3. |
United Kingdom Intellectual Property Office, Combined Search and Examination Report dated Mar. 11, 2015 for GB Patent Application No. 1417218.3. |
United Kingdom Intellectual Property Office, Combined Search and Examination Report dated May 13, 2020 for GB Patent Application No. 1917864.9. |
Varol Gul et al.: “Product placement detection based on image processing”, 2014 22nd Signal Processing and Communication Applications Conference (Siu), IEEE, Apr. 23, 2014. |
Varol Gul et al.: “Toward Retail product recognition on Grocery shelves”, Visual Communications and image processing; Jan. 20, 2004; San Jose, (Mar. 4, 2015). |
Weber et al., “Methods for Feature Detection in Point clouds,” visualization of large and unstructured data sets—IRTG Workshop, pp. 90-99 (2010). |
Zhao Zhou et al.: “An Image contrast Enhancement Algorithm Using PLIP-based histogram Modification”, 2017 3rd IEEE International Conference on Cybernetics (Cybcon), IEEE, (Jan. 21, 2017). |
Ziang Xie et al., “Multimodal Blending for High-Accuracy Instance Recognition”, 2013 IEEE RSJ International Conference on Intelligent Robots and Systems, p. 2214-2221. |
Fan Zhang et al., “Parallax-tolerant Image Stitching”, 2014 Computer Vision Foundation, pp. 4321-4328. |
Kaimo Lin et al., “SEAGULL: Seam-guided Local Alignment for Parallax-tolerant Image Stitching”, Retrieved on Nov. 16, 2020 [http://publish.illinois.edu/visual-modeling-and-analytics/files/2016/08/Seagull.pdf]. |
Julio Zaragoza et al., “As-Projective-As-Possible Image Stitching with Moving DLT”, 2013 Computer Vision Foundation, pp. 2339-2346. |
Zeng et al., Multi-view Self Supervised Deep Learning for 6D Pose Estimation in the Amazon Picking Challenge, May 7, 2017. Retrieved on Nov. 16, 2019 [https://arxiv.org/pdf/1609.09475.pdf]. |
Zhang et al., “Mobile computing and industrial augmented reality for real-time data access”, Emerging Technologies and Factory Automation, 2001, 8th IEEE International Conference on Oct. 15-18, 2001, pp. 583-588, vol. 2. |
Castorena et al., “Autocalibration of lidar and optical cameras via edge alignment”, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (Mar. 20, 2016). |
Chen et al. “Improving Octree-Based Occupancy Maps Using Environment Sparsity with Application to Aerial Robot Navigation” Robotics and Automation (ICRA), 2017 IEEE International Conference on IEEE, pp. 3656-3663, 2017. |
Cleveland Jonas et al.: “Automated System for Semantic Object Labeling with Soft-Object Recognition and Dynamic Programming Segmentation”, IEEE Transactions on Automation Science and Engineering, IEEE Service Center, New York, NY (Apr. 1, 2017). |
Cook et al., “Distributed Ray Tracing” ACM SIGGRAPH Computer Graphics, vol. 18, No. 3, ACM pp. 137-145, 1984. |
Datta, A., et al. “Accurate camera calibration using iterative refinement of control points,” in Computer Vision Workshops (ICCV Workshops), 2009. |
Deschaud, et al., “A Fast and Accurate Place Detection algoritm for large noisy point clouds using filtered normals and voxel growing,” 3DPVT, May 2010, Paris, France. |
Douillard, Bertrand, et al. “On the segmentation of 3D LIDAR point clouds.” Robotics and Automation (ICRA), 2011 IEEE International Conference on IEEE, 2011. |
Dubois, M., et al., “A comparison of geometric and energy-based point cloud semantic segmentation methods,” European Conference on Mobile Robots (ECMR), p. 88-93, Sep. 25-27, 2013. |
Duda, et al., “Use of the Hough Transformation to Detect Lines and Curves in Pictures”, Stanford Research Institute, Menlo Park, California, Graphics and Image Processing, Communications of the ACM, vol. 15, No. 1 (Jan. 1972). |
F.C.A. Groen et al., “The smallest box around a package,” Pattern Recognition, vol. 14, No. 1-6, Jan. 1, 1981, pp. 173-176, XP055237156, GB, ISSN: 0031-3203, DOI: 10.1016/0031-3203(81(90059-5 p. 176- p. 178. |
Federico Tombari et al. “Multimodal cue integration through Hypotheses Verification for RGB-D object recognition and 6DOF pose estimation”, IEEE International Conference on Robotics and Automation, Jan. 2013. |
Flores, et al., “Removing Pedestrians from Google Street View Images”, Computer Vision and Pattern Recognition Workshops, 2010 IEEE Computer Society Conference On, IEE, Piscataway, NJ, pp. 53-58 (Jun. 13, 2010). |
Glassner, “Space Subdivision for Fast Ray Tracing.” IEEE Computer Graphics and Applications, 4.10, pp. 15-24, 1984. |
Golovinskiy, Aleksey, et al. “Min-Cut based segmentation of point clouds.” Computer Vision Workshops (ICCV Workshops), 2009 IEEE 12th International Conference on. IEEE, 2009. |
Hackel et al., “Contour Detection in unstructured 3D point clouds,”IEEE, 2016 Conference on Computer vision and Pattern recognition (CVPR), Jun. 27-30, 2016, pp. 1-9. |
Hao et al., “Structure-based object detection from scene point clouds,” Science Direct, V191, pp. 148-160 (2016). |
Hu et al., “An improved method of discrete point cloud filtering based on complex environment,” International Journal of Applied Mathematics and Statistics, v48, i18 (2013). |
International Search Report and Written Opinion for International Patent Application No. PCT/US2013/070996 dated Apr. 2, 2014. |
International Search Report and Written Opinion for International Patent Application No. PCT/US2013/053212 dated Dec. 1, 2014. |
International Search Report and Written Opinion for corresponding International Patent Application No. PCT/US2016/064110 dated Mar. 20, 2017. |
International Search Report and Written Opinion for corresponding International Patent Application No. PCT/US2017/024847 dated Jul. 7, 2017. |
International Search Report and Written Opinion for International Application No. PCT/CN2017/083143 dated Feb. 11, 2018. |
International Search Report and Written Opinion for International Application No. PCT/US2018/030419 dated Aug. 31, 2018. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2018/030345 dated Sep. 17, 2018. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2018/030360 dated Jul. 9, 2018. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2018/030363 dated Jul. 9, 2018. |
International Search Report and Written Opinion for International Application No. PCT/US2019/025859 dated Jul. 3, 2019. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2019/025849 dated Jul. 9, 2019. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2019/049761 dated Nov. 15, 2019. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2019/051312 dated Nov. 15, 2019. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2019/054103 dated Jan. 6, 2020. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2019/064020 dated Feb. 19, 2020. |
International Search Report and Written Opinion for International Patent Application No. PCT/US2020/027948 dated Jul. 16, 2020. |
International Search Report and Written Opinion for International Patent Application No. PCT/US2020/028133 dated Jul. 24, 2020. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2020/029134 dated Jul. 27, 2020. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2020/028183 dated Jul. 24, 2020. |
International Search Report and Written Opinion from International Patent Application No. PCT/US2020/035285 dated Aug. 27, 2020. |
International Search Report and Written Opinion for International Patent Application No. PCT/US2021/024805 dated Aug. 2, 2021. |
International Search Report and Written Opinion for International Patent Application No. PCT/US2021/057388 dated Feb. 2, 2022. |
Jadhav et al. “Survey on Spatial Domain dynamic template matching technique for scanning linear barcode,” International Journal of science and research v 5 n 3, Mar. 2016)(Year: 2016). |
Jian Fan et al: “Shelf detection via vanishing point and radial projection”, 2014 IEEE International Conference on image processing (ICIP), IEEE, (Oct. 27, 2014), pp. 1575-1578. |
Kaikai Liu et al., “Enabling Context-Aware Indoor Augmented Reality via Smartphone Sensing and Vision Tracking”, ACM Transactions on Multimedia Computing Communications and Applications, Association for Computer Machinery, US, vol. 12, No. |
Kang et al., “Kinematic Path-Tracking of Mobile Robot Using Iterative learning Control”, Journal of Robotic Systems, 2005, pp. 111-121. |
Kay et al. “Ray Tracing Complex Scenes.” ACM SIGGRAPH Computer Graphics, vol. 20, No. 4, ACM, pp. 269-278, 1986. |
Kelly et al., “Reactive Nonholonomic Trajectory Generation via Parametric Optimal Control”, International Journal of Robotics Research, vol. 22, No. 7-8, pp. 583-601 (Jul. 30, 2013). |
Lari, Z., et al., “An adaptive approach for segmentation of 3D laser point cloud.” International Archives of the Photogrammertry, Remote sensing and spatial information Sciences, vol. XXXVIII-5/W12, 2011, ISPRS Calgary 2011 Workshop, Aug. 29-31, 2011, Calgary, Canada. |
Lecking et al.: “Localization in a wide range of industrial environments using relative 3D ceiling features”, IEEE, pp. 333-337 (Sep. 15, 2008). |
Lee et al. “Statistically Optimized Sampling for Distributed Ray Tracing.” ACM SIGGRAPH Computer Graphics, vol. 19, No. 3, ACM, pp. 61-67, 1985. |
Li et al., “An improved RANSAC for 3D Point cloud plane segmentation based on normal distribution transformation cells,” Remote sensing, V9: 433, pp. 1-16 (2017). |
Likhachev, Maxim, and Dave Ferguson. “Planning Long dynamically feasible maneuvers for autonomous vehicles.” The international journal of Robotics Reasearch 28.8 (2009): 933-945. (Year:2009). |
Marder-Eppstein et al., “The Office Marathon: robust navigation in an indoor office environment,” IEEE, 2010 International conference on robotics and automation, May 3-7, 2010, pp. 300-307. |
McNaughton, Matthew, et al. “Motion planning for autonomous driving with a conformal spatiotemporal lattice.” Robotics and Automation (ICRA), 2011 IEEE International Conference on. IEEE, 2011. (Year: 2011). |
Meyersohn, “Walmart turns to robots and apps in stores”, https://www.cnn.com/2018/12/07/business/walmart-robot-janitors-dotcom-store/index.html, Oct. 29, 2019. |
Mitra et al., “Estimating surface normals in noisy point cloud data,” International Journal of Computational geometry & applications, Jun. 8-10, 2003, pp. 322-328. |
N.D.F. Campbell et al. “Automatic 3D Object Segmentation in Multiple Views using Volumetric Graph-Cuts”, Journal of Image and Vision Computing, vol. 28, Issue 1, Jan. 2010, pp. 14-25. |
Ni et al., “Edge Detection and Feature Line Tracing in 3D-Point Clouds by Analyzing Geometric Properties of Neighborhoods,” Remote Sensing, V8 I9, pp. 1-20 (2016). |
Norriof et al., “Experimental comparison of some classical iterative learning control algorithms”, IEEE Transactions on Robotics and Automation, Jun. 2002, pp. 636-641. |
Notice of allowance for U.S. Appl. No. 13/568,175 dated Sep. 23, 2014. |
Notice of allowance for U.S. Appl. No. 13/693,503 dated Mar. 11, 2016. |
Number | Date | Country | |
---|---|---|---|
20180314261 A1 | Nov 2018 | US |
Number | Date | Country | |
---|---|---|---|
62492670 | May 2017 | US | |
62492712 | May 2017 | US |