Robotic systems and methods for identifying and processing a variety of objects

Information

  • Patent Grant
  • 12159192
  • Patent Number
    12,159,192
  • Date Filed
    Wednesday, September 21, 2022
    2 years ago
  • Date Issued
    Tuesday, December 3, 2024
    18 days ago
Abstract
A robotic system is disclosed that include an articulated arm and a first perception system for inspecting an object, as well as a plurality of additional perception systems, each of which is arranged to be directed toward a common area in which an object may be positioned by the robotic arm such that a plurality of views within the common area may be obtained by the plurality of additional perception systems.
Description
BACKGROUND

The invention generally relates to robotic and other sortation systems, and relates in particular to robotic and other sortation systems that are intended to be used in dynamic environments requiring the system to accommodate processing a variety of objects.


Many order fulfillment operations achieve high efficiency by employing a process called wave picking. In wave picking, orders are picked from warehouse shelves and placed into bins containing multiple orders that are sorted downstream. At the sorting stage individual articles are identified, and multi-article orders are consolidated into a single bin or shelf location so that they may be packed and then shipped to customers. The process of sorting these articles has been done by hand. A human sorter picks an article from an incoming bin, finds the barcode on the object, scans the barcode with a handheld barcode scanner, determines from the scanned barcode the appropriate bin or shelf location for the article, and then places the article in the so-determined bin or shelf location where all articles for that order are placed.


Manually operated barcode scanners are generally either fixed or handheld systems. With fixed systems, such as those used at point-of-sale systems, the operator holds the article and places it in front of the scanner so that the barcode faces the scanning device's sensors, and the scanner, which scans continuously, decodes any barcodes that it can detect. If the article is not immediately detected, the person holding the article typically needs to vary the position or rotation of the article in front of the fixed scanner, so as to make the barcode more visible to the scanner. For handheld systems, the person operating the scanner looks for the barcode on the article, and then holds the scanner so that the article's barcode is visible to the scanner, and then presses a button on the handheld scanner to initiate a scan of the barcode.


Other ways of identifying items by barcode scanning require that the barcode location be controlled or constrained so that a fixed or robot-held barcode scanner can reliably see the barcode. Automatic barcode scanners also involve either fixed or hand-held systems, and the same principles apply. In the case of barcode scanners typically used in industrial applications, the possible positions of barcodes must be tightly controlled so that the barcodes are visible to the one or more scanners. For example, one or more barcode scanners may be placed in fixed locations relative to a conveyor or series of moving trays so that the scanners may scan objects, typically boxes, as they pass by the scanners. In these installations the range of placement of the barcodes is comparatively limited as they must be on labels affixed to one of four sides or top of a box, which also needs to be presented at orientations suitable for scanning. The detected barcode is then associated with the immediate section of the conveyor or is associated with the particular moving tray in which the object had been placed prior to scanning.


In all of these cases, the systems employ sensors, cameras or laser reflectivity sensors, as well as software to detect barcodes and decode them. These methods have inherent limitations that include the range of distances of orientations relative to the detection system, over which they are able to reliably scan barcodes. Firstly, the barcode must be facing the scanner; secondly the range to the barcode must be such that individual elements can be reliably distinguished; and, thirdly, the tilt and skew of the barcode must be such that individual elements can be reliably distinguished. The types of sensors employed, and the robustness of the software detection and decoding schemes determine these performance parameters. There remains a need, therefore, for an object identification system for a robotic system that is able to accommodate the automated identification and processing of a variety of objects in a variety of orientations.


SUMMARY

In accordance with an embodiment, the invention provides a robotic system that include an articulated arm and a first perception system for inspecting an object, as well as a plurality of additional perception systems, each of which is arranged to be directed toward a common area in which an object may be positioned by the robotic arm such that a plurality of views within the common area may be obtained by the plurality of additional perception systems.


In accordance with another embodiment, the invention provides a method of identifying an object in a robotic system including an articulated arm. The method includes the steps of inspecting the object using a first perception system and providing an inspection response signal representative of whether the item has been identified, moving the object to a plurality of additional perception systems responsive to the inspection response signal, and inspecting the object using the plurality of additional perception systems, each of which is arranged to be directed toward the object in a common area from a different view of a plurality of views of the object within the common area that may be obtained by the plurality of additional perception systems.


In accordance with a further embodiment, the invention provides a perception system for assisting in identifying an object. The perception system includes a plurality of perception units that are each positioned to be directed toward a plurality of locations along an object path that an object may take as the object travels through the perception system.





BRIEF DESCRIPTION OF THE DRAWINGS

The following description may be further understood with reference to the accompanying drawings in which:



FIGS. 1A and 1B show an illustrative diagrammatic view of a system in accordance with an embodiment of the present invention;



FIG. 2 shows an illustrative diagrammatic view of a portion of the system of FIG. 1A that includes a plurality of perception units;



FIG. 3 shows an illustrative diagrammatic view of perception image data of a set of objects to be sorted as presented at an infeed station in a system in accordance with an embodiment of the present invention;



FIG. 4 shows an illustrative diagrammatic view of a single simulated hold of an object to be identified in accordance with an embodiment of the present invention;



FIG. 5 shows an illustrative diagrammatic view of multiple overlapping simulated holds of the object of FIG. 4;



FIG. 6 shows an illustrative diagrammatic view of simulated overlapping positions of many barcodes;



FIG. 7 shows an illustrative diagrammatic view of a configuration of perception units based on the modelling of FIGS. 4-6;



FIG. 8 shows an illustrative diagrammatic view of a system in accordance with a further of the present invention;



FIG. 9 shows an illustrative diagrammatic isometric view of a drop perception unit for use in accordance with the system of FIG. 8; and



FIG. 10 shows an illustrative diagrammatic top isometric view of the drop perception unit of FIG. 9.





The drawings are shown for illustrative purposes only.


DETAILED DESCRIPTION

In accordance with an embodiment, the invention provides an object perception system for the purposes of automatically sorting individual objects in a set. In applications such as order fulfillment, objects are collected into heterogeneous sets and need to be sorted. Individual objects need to be identified and then routed to object-specific locations. The described system reliably automates the identification of such objects by employing both automated barcode scanners and a robotic arm. In accordance with certain embodiments, for example, the system uses the robotic arm to hold the item in front of one or more barcode scanners so that the object may be scanned. In accordance with various embodiments, the perception units (e.g., cameras or scanners) may look for a variety of codes such as indicia (e.g., barcodes, radio frequency tags, Stock Keeping Unit (SKU), Universal Product Code (UPC), low wavelength IR (LWIR), as well as invisible barcodes and digital watermarks such as Digimarc DWCode, etc.).


Whereas fixed industrial scanners require that the object's barcode be situated so that its barcode is visible to a scanner, the robotic arm of the present invention may pick an object out of a heterogeneous collection of objects where the barcode is not visible and scan the item. The result is an automated barcode scanning system for arbitrary items in a heterogeneous stream of objects that may be used to accurately and reliably identify items.


Sorting for order fulfillment is one application for automatically identifying objects from a heterogeneous object stream. Barcode scanners have a wide variety of uses including identifying the stock keeping unit of an article, or tracking parcels. The described system may have many uses in the automatic identification and sortation of objects.


Operating in conjunction with a robotic pick and place system, such a system of an embodiment of the present invention automates part of the sorting process, in particular the step of identifying picked objects. Instead of a person picking the object from a bin, a robotic arm may pick an object from a bin, placed the object in front of a barcode scanner, and then, having obtained identification codes for the object, placed the object in the appropriate bin or shelf location. Since barcode scanners employ cameras or lasers to scan 1D or 2D symbologies printed on labels affixed to articles, the barcodes must be visible to the scanner's sensors for successful scanning in order to automatically identify items in a heterogeneous stream of arbitrary articles, as in a jumbled set of articles found in a bin.


In accordance with various embodiments, therefore, the invention provides a method for determining the identity of an object from a collection of objects, as well as a method for scanning the barcode of an object employing one or more barcode scanners and a robotic arm with end-effector holding the object. The invention further provides a method for determining the placement of fixed barcode scanners so as to maximize the probability of successfully scanning an object held by a robot end-effector in accordance with certain embodiments, as well as a method for determining a sequence of placements of a robot end-effector so as to minimize the time it takes a configuration of one or more barcode scanners to successfully scan an object, and a method for scanning the barcode of an object by employing a barcode scanner as an end-effector on a robotic arm.


An important aspect is the ability to identify objects via barcode or other visual markings of objects by employing a robot arm to pick up individual objects and place or drop them in front of one or more scanners. Automated scanning systems would be unable to see barcodes on objects that are presented in a way that their barcodes are not exposed or visible. Since the system uses a robot to hold the object, it can either maneuver the object so as to make the barcode visible, or employ multiple scanners that view the article from multiple viewpoints to acquire the barcode, irrespective of how the object is held.


An automated article identification system in accordance with an embodiment of the present invention is shown in FIG. 1A. FIG. 1A shows a robotic system 10 that includes an articulated arm 12 that includes an end effector 14 and articulated sections 16, 18 and 20. The articulated arm 12 selects items from a conveyor 22, that are either in a bin on the conveyor 22 or are on the conveyor itself. A stand 24 includes an attached primary perception unit 26 that is directed toward the conveyor from above the conveyor. The primary perception unit 26 may be, for example, a camera, or a scanner such as a laser reflectivity scanner or other type of bar-code reader, or a radio frequency ID scanner. A plurality of additional perception units are provided on a secondary perception unit 28 (as will be discussed in more detail below with reference to FIG. 2).


The robotic system 10 may further include the robotic environment, a target station 30 that includes a number of bins 32 into which objects may be placed after identification. FIGS. 1A and 1B further show a central computing and control system 34 that may communicate with the primary perception unit 26 and the secondary perception unit 28 as well as the articulated arm 12 via wireless communication, or, in certain embodiments, the central computing and control system 34 may be provided within the base section 20 of the articulated arm.



FIG. 2 shows the secondary perception unit 28 that includes a plurality of perception units 40, 42, 44, 46 and 48, as well as a plurality of illumination sources 50, 52, 54 and 56 for use in certain embodiments of the present invention. Each of the perception units 40, 42, 44, 46 and 48 may be, for example, a camera (e.g., 2D or 3D), or a scanner such as a laser reflectivity scanner or other type of barcode reader (e.g., 1D or 2D barcode scanners), or a radio frequency ID scanner together with the associated software to process the perceived data.


Generally, the system provides in a specific embodiment, an automated article identification system that includes a robotic pick and place system that is able to pick articles up, move them in space, and place them. The system also includes the set of objects themselves to be identified; the manner in which inbound objects are organized, commonly in a heterogeneous pile in a bin or in a line on a conveyor; the manner in which outbound objects are organized, commonly in an array of outbound bins, or shelf cubbies; the manner in which objects are labeled with barcodes or radio-frequency identification tags; a fixed primary scanner operating above the incoming stream of objects; a barcode scanning station where one or more barcode scanners and illuminators are activated when the object is held at the station; and a central computing and control system determines the appropriate location for placing the object, which is dependent on the object's decoded barcode.


As noted, the robotic pick and place system is typically a robotic arm equipped with sensors and computing, that when combined is assumed herein to exhibit the following capabilities: (a) it is able to pick objects up from a specified class of objects, and separate them from a stream of heterogeneous objects, whether they are jumbled in a bin, or are singulated on a motorized or gravity conveyor system, (b) it is able to move the object to arbitrary places within its workspace, (c) it is able to place objects in an outgoing bin or shelf location in its workspace, and (d) it is able to generate a map of objects that it is able to pick, represented as a candidate set of grasp points in the workcell, and as a list of polytopes enclosing the object in space.


The allowable objects are determined by the capabilities of the robotic pick and place system. Their size, weight and geometry are assumed to be such that the robotic pick and place system is able to pick, move and place them. These may be any kind of ordered goods, packages, parcels, or other articles that benefit from automated sorting. In certain embodiments, each object is associated with a stock keeping unit (SKU), which identifies the item.


The manner in which inbound objects arrive may, for example, be in one of two configurations: (a) inbound objects arrive piled in bins of heterogeneous objects as shown in FIG. 3; or (b) inbound articles arrive by a moving conveyor. As shown in FIG. 3, the collection of objects includes some that have exposed bar codes as shown at 60, 62, 64, 66, 68, 70 and 72, and other objects that do not have exposed bar codes. The robotic pick and place system is assumed to be able to pick items from the bin or conveyor. The stream of inbound objects is the sequence of objects as they are unloaded either from the bin or the conveyor.


The manner in which outbound objects are organized is such that articles are placed in a bin, shelf location or cubby or other destination location at which all objects corresponding to a given order are consolidated. These outbound destinations may be arranged in vertical arrays, horizontal arrays, grids, or some other regular or irregular manner, but which arrangement is known to the system. The robotic pick and place system is assumed to be able to place objects into all of the outbound destinations, and the correct outbound destination is determined from the SKU of the object.


It is assumed that the objects are marked in one or more places on their exterior with a visually distinctive mark such as a barcode or radio-frequency identification (RFID) tag or other unique identifier so that they may be identified by a scanner. The type of marking depends on the type of scanning system used, but may include 1D or 2D barcode symbologies. Multiple symbologies or labeling approaches may be employed. The types of scanners employed are assumed to be compatible with the marking approach. The marking, either by barcode, RFID tag, or other means, encodes a symbol string, which is typically a string of letters and numbers. The symbol string uniquely associates the object with a SKU.


The primary perception unit 26 is a device mounted above the area where the inbound objects arrive, which scans all inbound objects for identifying features. When inbound objects arrive in bins, the primary perception unit is mounted above the bin. The primary perception unit consists of one or more devices that are able to recognize visually distinctive marks, which may include barcodes or other identifying features, or RFID tags on the objects. Unit components may include cameras, RFID scanners, illuminators, and decoding microprocessors. The primary perception unit makes the first pass at recognizing items in the bin. The primary perception unit localizes any codes that it has found, and the robotic pick and place system is assumed to be able to use that location to pick the item with the given code.


The secondary perception unit 28 (e.g., the secondary scanners) is an assembly that scans any objects that the primary perception unit is unable to scan among the inbound objects, or that scans the objects to confirm the results of the primary perception unit. In further embodiments, the secondary scanners may be used to detect any additional indicia, which may either confirm the identification or may detect that more than one object has been grasped, in which case, both are returned to the input inbound area. Inbound objects in a bin or on a conveyor may have labels or tags partially or completely occluded by other objects, or the labels or tags may not be facing the scanner. The secondary perception unit is mounted in the workspace of the robotic pick and place system so that unidentified articles may be presented to the secondary scanner. Like the primary perception unit, the secondary perception unit consists of one or more perception devices, which may include cameras, RFID scanners, illuminators, and decoding microprocessors.


The operations of the systems described above are coordinated by the central control system 34. This system determines from perception data (e.g., symbol strings) the SKU associated with an object, as well as the outbound destination for the object. The central control system is comprised of one or more workstations or central processing units (CPUs). The correspondence between SKUs and outbound destinations is maintained by the central control system in a database called a manifest. The central control system maintains the manifest by communicating with a warehouse management system (WMS).


During operation, the broad flow of work may be generally as follows. First, the system is equipped with a manifest that provides the outbound destination for each inbound object. Next, the system waits for inbound objects to arrive either in a bin or on a conveyor. When the robotic pick and place system recognizes that one or more inbound objects are present, the central control system instructs the primary perception unit to scan the inbound objects. The primary perception unit creates a list of detected markings, which includes their symbol strings, as well as their positions in the workspace. The primary perception unit transmits this list to the central control system.


The central control system receives from the primary perception unit the perceived information including the detected markings, and from the robotic pick and place system it receives a list of unidentified but pick-able objects. The position coordinates in both lists are registered to one another by employing calibration information about the primary scanner and the robotic pick and place system. The central control system employs the map generated by the robotic pick and place system to determine by simple geometric means the objects that enclose each detected marking. Thus, the central control system associates for each marking the object to which it corresponds. This step is a marking-based data association. The central control system ranks each of the candidate articles to pick based on a heuristic, such as choosing the top-most object in a pile, and so generates a candidate pick list.


Again, FIG. 3 shows an example of a marking-based data association between objects identified by the robotic pick and place system and markings detected by a perception unit. In this instance barcode symbols (bounding quadrilaterals) are associated with the unidentified objects (shaded segments) in which the barcode symbols lie.


If there is at least one candidate pick in the candidate pick list that is associated with a marking, the system picks the highest-ranking pick. The expectation is that the picked object will correspond to the marking previously associated by the central control system, and detected by the primary perception unit. Given that this association may potentially be erroneous, the central control system runs a check. After the object has been removed and separated from the set of inbound objects, the system instructs the primary perception unit to scan the inbound articles again. If the correct object was removed, then the marking associated with it should no longer be in the list of markings detected by the primary perception unit. If the marking associated with the picked item is still there however, then it must be that it picked the wrong item. If it picked the wrong item, then it puts the item back and repeats the process of generating pick candidates from the robotic pick and place system and primary scanner.


If there are no candidate picks associated with a marking, then it picks the object associated with the highest-ranking pick. Since there is no marking associated with the object, it is an unidentified object. After the robotic pick and place system picks the item out of the set of inbound objects, the central control system instructs the robotic pick and place system to move the object to the secondary perception unit for scanning. The central control system instructs the secondary perception unit to scan the unidentified object.


If the secondary perception unit successfully recognizes a marking on the object, then the object is then identified and the central control system commands the robotic pick and place system to transfer the item to the outbound destination determined from the SKU, itself determined from the detected marking.


If the secondary perception unit is unable to recognize a marking on the object, then depending on the configuration of the scanner, the central control system may command the robotic pick and place system to rotate the object to try to make markings visible to the scanners, and to scan the object again. This may occur a number of times to guarantee that if the marking were on the object, then it would be detected. The sequence of locations and orientations of the objects are chosen so as to minimize the average or maximum amount of time that secondary scanning takes. If the object cannot be identified or if the secondary perception unit detects non-matching product identifiers (possibly indicating a double pick), the object may be transferred to a special outbound destination for unidentified objects, or it may be returned to the inbound stream.


This entire procedure operates in a loop until all of the objects in the inbound set are depleted. The objects in the inbound stream are automatically identified, sorted, and routed to outbound destinations.


In accordance with an embodiment therefore, the invention provides a system for sorting objects that arrive by inbound bins and that need to be placed into a shelf of outbound bins, where sorting is to be based on a barcode symbol. In this embodiment, the primary and secondary perception units are able to detect and decode barcode symbologies.


Key specializations in this embodiment are the specific design of the primary and secondary perception units so as to maximize the probability of a successful scan, while simultaneously minimizing the average scan time. The probability of a successful scan and the average scan time make up key performance characteristics. These key performance characteristics are determined by the configuration and properties of the primary and secondary perception units, as well as the object set and how they are marked.


The two key performance characteristics may be optimized for a given item set and method of barcode labeling. Parameters of the optimization for a barcode system include how many barcode scanners to include where and in what orientation to place them, and what sensor resolutions and fields of view for the scanners to use. Optimization can, in certain embodiments, be done by simulation with models of the object.


Optimization through simulation employs a barcode scanner performance model. A barcode scanner performance model is the range of positions, orientations and barcode element size that a barcode symbol can be detected and decoded by the barcode scanner, where the barcode element size is the size of the smallest feature on the barcode. These are typically rated at a minimum and maximum range, a maximum skew angle, a maximum pitch angle, and a minimum and maximum tilt angle.


If a barcode scanner and symbol are held upright, and the barcode symbol is facing the scanner such that the symbol is parallel to the sensor-side of the scanner, then the barcode symbol is in what is called the fronto-parallel plane. The angle between the fronto-parallel plane and a plane that rotates about the vertical axis is the skew angle. The angle between the fronto-parallel plane and a plane that rotates about the horizontal axis is the pitch axis. The angle a feature on the fronto-parallel makes as it rotates about an axis perpendicular to the fronto-parallel plane is the tilt axis.


Typical performances for camera-based barcode scanners are that they are able to detect barcode symbols within some range of distances as long as both pitch and skew of the plane of the symbol are within the range of plus or minus 45 degrees, while the tilt of the symbol can be arbitrary (between 0 and 360 degrees). The barcode scanner performance model predicts whether a given barcode symbol in a given position and orientation will be detected.


The barcode scanner performance model is coupled with a model of where barcodes would expect to be positioned and oriented. A barcode symbol pose model is the range of all positions and orientations, in other words poses, in which a barcode symbol will expect to be found. For the secondary scanner, the barcode symbol pose model is itself a combination of an article gripping model, which predicts how objects will be held by the robotic pick and place system, as well as a barcode-item appearance model, which describes the possible placements of the barcode symbol on the object. For the primary scanner, the barcode symbol pose model is itself a combination of the barcode-item appearance model, as well as an inbound-object pose model, which models the distribution of poses over which inbound articles are presented to the primary scanner. These models may be constructed empirically, modeled using an analytical model, or approximate models can be employed using sphere models for objects and a uniform distribution over the sphere as a barcode-item appearance model.


In an embodiment for example, two objects, a bottle and a toothpaste container, represent the object set, and the barcode symbols are placed in fixed and known locations on all instances of these objects. With a 3D model of these two objects, the known capabilities of the robotic pick and place system are used to generate a random sample of poses of the objects. These poses are with respect to the end-effector of the robotic pick and place system, typically a gripper.



FIG. 4 shows an example of a single simulated hold of a bottle 80, as held by an end effector 82 of a vacuum-gripper-based robotic pick and place system. FIG. 5 shows at 84 overlapping samples of the bottle 80 shown in FIG. 4. With these samples, and because the barcode symbols are at fixed positions on the two articles, a set of poses of barcode symbols are generated. FIG. 6 shows at 86 an example where the quadrilaterals represent 100 sampled barcode symbols. This sample represents the barcode symbol pose model for a secondary scanner. It is an approximation of a probability distribution over where barcodes would expect to be found when the article is held by the robotic pick and place system.


With these models, the barcode symbol pose model and the barcode scanner performance model, optimization over all of the parameters of the system is possible. FIG. 7 shows the resulting configuration, determined by simulation of the above-described models, of a secondary scanner system determined for the instance where the articles are a bottle and toothpaste container. As shown in FIG. 7, the system includes scanners 90, 92, 94, 96 and 98, each of which is generally directed toward an object area 88, while the position of each of the scanners 90-98 is selected to provide optimum planes, angles, tilts and views for the objects in question. FIG. 7 shows a mechanical model of the actualized secondary scanner. In this instance the optimization criteria was the probability of scan success. In instances where only one scanner can be employed, the optimization criteria may be average scan time, in which case the optimization is over the sequence of poses in which to present articles to a secondary scanner so that the average scan time as a whole is decreased.


In accordance with a further embodiment therefore, the invention may be used in connection with an object sortation system that yields a large (and very flexible) number of total collection bins, very low divert costs per bin, throughput as high as that of a manual system, and a far smaller need for manual labor to operate.



FIG. 8, for example, shows a system 200 that includes an articulated arm 202 with an end effector 204, an input area 206 in which objects are presented for sortation, a primary perception unit 214 such as a camera for identifying objects to be sorted, and a receiving conveyor 208 for receiving objects to be sorted from any of a human worker, another conveyor, or an input pan. The system 200 also includes a non-sortable output chute 210 that leads to a non-sortable output bin 212 for providing objects that the primary perception unit 214 either could not identify or could not sort for any other reason (e.g., could not grasp or pick up).


In addition to the primary perception unit 214, the system also includes a drop perception unit 216 for use as a secondary perception unit, which includes an open top and an open bottom, and a plurality of perception units (e.g., cameras or sensors as discussed above with reference to the previous embodiments) positioned within the drop perception unit 216 that are aimed at the top, mid and lower central regions of the interior of the drop perception unit 216. The plurality of perception units, e.g., cameras, record perception data, e.g., images, of an object when it is dropped by the end effector through the drop perception unit 216. The drop perception unit 216 may also include one or more sensors (e.g., laser sensors) at the top of the drop perception unit 216 that detect when an object is dropped into the drop perception unit 216. The plurality of perception units are designed to collect a plurality of images of each object from multiple views to aid in identifying or confirming the identity of the dropped object.


The dropped object then falls into a first carriage 218 which is provided on a track 220 on which the conveyor 218 may be moved automatically between a first sortation stage 222 and a second sortation stage 224 on either side of the area in which the object was dropped.


The first sortation stage 222 includes a second carriage 226 that may receive objects from the first carriage 218, and which travels along a track between two rows of collection bins 228 into which objects may be dumped along guide walls 230. The second sortation stage 224 includes a third carriage 232 that may receive objects from the first carriage 218, and which travels along a track between two rows of collection bins 234 into which objects may be dumped along guide walls 236.


The system of FIG. 8 shows a system with two shuttle sort wings. When an object is picked from the infeed conveyor, it is dropped onto the first shuttle sorter 218. That shuttle sorter carries the object to one of two wings, drops the object in the carrier for that wing, and then moves back to home. Due to the limited travel, this back and forth operation may be performed in the time it takes the articulated arm to pick another object (assuming the articulated arm is picking objects at approximately a human rate of throughput).


The drop perception unit 216 includes a plurality of detection units (e.g., cameras or scanners as discussed above) that are directed toward a central path such that as an object falls through the drop perception unit 216, multiple views of the object will be captured by the multiple detection units. The drop perception unit may also include lights as discussed above with reference to the system of FIG. 2.



FIGS. 9 and 10 show the detection units 300 directed toward the central path of the drop perception unit 216 at varying angles. As also shown, a laser source bar 302 may direct laser illumination toward a sensor bar 304 such that the drop perception unit 216 may detect exactly when an object enters the drop perception unit 216. FIG. 10 shows at 306 diagrammatic illustrations of the multiple detector angles and fields of view for each of the multiple detectors.


The plurality of additional detection units 300 may be positioned such that every surface of the object may be perceived regardless of the orientation of the object. In certain embodiments, the primary perception unit 214 may provide perception data regarding a unique object identifier, and the plurality of additional detection units 300 may provide additional perception data regarding any of confirming the object identifier or any additional object identifiers. In further embodiments, the primary perception unit 214 may provide perception data regarding an object shape, and the plurality of additional detection units 300 may provide additional perception data regarding a unique object identifier that is confirmed as matching the object shape.


Those skilled in the art will appreciate that numerous modifications and variations may be made to the above disclosed embodiments without departing from the spirit and scope of the present invention.

Claims
  • 1. A method of processing objects using an end-effector of a programmable motion device, said method comprising: generating perception data regarding a plurality of objects at an input area by a primary perception unit, the primary perception unit including at least one camera or at least one scanner;generating a list of detected markings from the perception data by the primary perception unit, each marking being associated with marking position coordinates;generating a list of unidentified objects by a robotic pick and place system that includes the programmable motion device, each unidentified object being associated with object position coordinates;registering the marking position coordinates with the object position coordinates to associate with each marking with an object of the plurality of objects by a central control system, the central control system including at least one workstation or at least one central processing unit;generating a candidate pick list that ranks objects for picking by the central control system; andusing the end-effector of the programmable motion device to grasp a selected object of the plurality of objects based on the candidate pick list.
  • 2. The method of claim 1, wherein the plurality of objects are provided at the input area in a bin.
  • 3. The method of claim 1, wherein the selected object is grasped by the end-effector on the marking associated with the selected object.
  • 4. The method of claim 1, wherein the method further includes generating a map of the plurality of objects by the central control system.
  • 5. The method of claim 1, wherein the method further includes rotating the selected object while grasping the selected object by the end-effector proximate a perception unit.
  • 6. The method of claim 1, wherein the method further includes dropping the selected object into a secondary perception unit, the secondary perception unit being a drop perception unit with an open top and an open bottom using the end-effector of the programmable motion device, the drop perception unit including a plurality of cameras or sensors that generate perception data for identifying the selected object falling through the drop perception unit.
  • 7. The method of claim 1, wherein the method further includes generating further perception data regarding the plurality of objects at the input area following removal of the selected object by the primary perception unit to confirm whether the marking associated with the selected object is still present at the input area.
  • 8. The method of claim 7, wherein the method further includes returning the selected object to the input area using the end-effector of the programmable motion device if the marking associated with the selected object is present in the further perception data.
  • 9. The method of claim 1, wherein the method further includes receiving the selected object from the end-effector of the programmable motion device in a carriage that is adapted for movement toward a plurality of destination locations.
  • 10. The method of claim 9, wherein the method further includes dropping the selected object from the carriage into a selected destination location.
  • 11. A method of processing objects using an end-effector of a programmable motion device, said method comprising: generating perception data regarding a plurality of objects at an input area by a primary perception unit, the primary perception unit including at least one camera or at least one scanner;generating a list of unidentified objects by a robotic pick and place system that includes the programmable motion device, each unidentified object being associated with object position coordinates;generating a map of the plurality of objects at the input area by a central control system, said map including shaded segments associated with each of the plurality of objects, the central control system including at least one workstation or at least one central processing unit;selecting a grasping location of an unidentified object of the plurality of objects based on the map by the central control system for grasping the unidentified object; andusing the end-effector of the programmable motion device to grasp the unidentified object at the selected grasping location.
  • 12. The method of claim 11, wherein the plurality of objects are provided at the input area in a bin.
  • 13. The method of claim 11, wherein the method further includes rotating the selected unidentified object while grasping the selected object by the end-effector proximate a perception unit.
  • 14. The method of claim 11, wherein the method further includes dropping the unidentified selected object into a secondary perception unit, the secondary perception unit being a drop perception unit with an open top and an open bottom using the end-effector of the programmable motion device, the drop perception unit including a plurality of cameras or sensors that generate perception data for identifying the selected object falling through the drop perception unit.
  • 15. The method of claim 11, wherein the method further incudes moving the unidentified selected object toward a secondary perception unit comprising a plurality of cameras or scanners arranged in a bowl-shaped arrangement.
  • 16. The method of claim 11, wherein the method further includes receiving the selected object from the end-effector of the programmable motion device in a carriage that is adapted for movement toward a plurality of destination locations.
  • 17. The method of claim 16, wherein the method further includes dropping the selected object from the carriage into a selected destination location.
  • 18. An object processing system for processing objects using an end-effector of a programmable motion device, said object processing system comprising: an input area for receiving a plurality of objects;a primary perception unit including at least one camera or scanner for generating perception data regarding the plurality of objects at the input area;a central control system including at least one workstation or at least one central processing unit for generating a list of detected markings, each marking being associated with marking position coordinates, for generating a list of unidentified objects, each unidentified object being associated with object position coordinates, and for registering the marking position coordinates with the object position coordinates to associate each marking with an object of the plurality of objects; andwherein the central control system further generates a candidate pick list that ranks objects for picking and for using the end-effector of the programmable motion device to grasp a selected object of the plurality of objects based on the candidate pick list.
  • 19. The object processing system as claimed in claim 18, wherein the plurality of objects are provided at the input area in a bin.
  • 20. The object processing system as claimed in claim 18, wherein the central control system further generates a map of the plurality of objects.
  • 21. The object processing system as claimed in claim 18, wherein the object processing system further includes a secondary perception unit, the secondary perception unit being a drop perception unit with an open top and an open bottom into which the selected object is dropped, the drop perception unit including a plurality of cameras or sensors that generate perception data for identifying the selected object falling through the drop perception unit.
  • 22. The object processing system as claimed in claim 18, wherein the object processing system further includes a carriage that is adapted for movement toward a plurality of destination locations.
  • 23. The object processing system as claimed in claim 22, wherein the plurality of destination locations are provided adjacent a track along which the carriage is adapted to travel.
PRIORITY

The present application is a continuation of U.S. patent application Ser. No. 16/800,587, filed Feb. 25, 2020, now U.S. Pat. No. 11,494,575, issued Nov. 8, 2022, which is a continuation of U.S. patent application Ser. No. 15/982,238, filed May 17, 2018, now U.S. Pat. No. 10,621,402, issued Apr. 14, 2020, which is a continuation of U.S. patent application Ser. No. 15/260,837, filed Sep. 9, 2016, now U.S. Pat. No. 10,007,827, issued Jun. 26, 2018, which claims priority to U.S. Provisional Patent Application Ser. No. 62/217,200, filed Sep. 11, 2015, as well as U.S. Provisional Patent Application Ser. No. 62/269,640, filed Dec. 18, 2015, the disclosures of which are hereby incorporated by reference in their entireties.

US Referenced Citations (190)
Number Name Date Kind
3734286 Simjian May 1973 A
3864566 Simpson et al. Feb 1975 A
4186836 Wassmer et al. Feb 1980 A
4704694 Czerniejewski Nov 1987 A
4722653 Williams et al. Feb 1988 A
4759439 Hartlepp Jul 1988 A
4819784 Sticht Apr 1989 A
4846335 Hartlepp Jul 1989 A
4895242 Michel Jan 1990 A
5190162 Hartlepp Mar 1993 A
5495097 Katz et al. Feb 1996 A
5648709 Maeda Jul 1997 A
5713473 Satake et al. Feb 1998 A
5742420 Peng Apr 1998 A
5794788 Massen Aug 1998 A
5794789 Payson et al. Aug 1998 A
5839566 Bonnet Nov 1998 A
5875434 Motsuoka et al. Feb 1999 A
5996316 Kirschner Dec 1999 A
6059092 Jerue et al. May 2000 A
6060677 Ulrichsen et al. May 2000 A
6076023 Sato Jun 2000 A
6079570 Oppliger et al. Jun 2000 A
6087608 Schlichter et al. Jul 2000 A
6124560 Roos et al. Sep 2000 A
6208908 Boyd et al. Mar 2001 B1
6246023 Kugle Jun 2001 B1
6323452 Bonnet Nov 2001 B1
6390756 Isaacs et al. May 2002 B1
6401936 Isaacs et al. Jun 2002 B1
6505093 Grams et al. Jan 2003 B1
6579053 Grams et al. Jun 2003 B1
6685031 Takizawa Feb 2004 B2
6688459 Bonham et al. Feb 2004 B1
6705528 Good et al. Mar 2004 B2
6721444 Gu et al. Apr 2004 B1
6762382 Danelski Jul 2004 B1
6946612 Morikawa Sep 2005 B2
7313464 Perreault et al. Dec 2007 B1
7347376 Biss et al. Mar 2008 B1
7474939 Oda et al. Jan 2009 B2
7516848 Shakes et al. Apr 2009 B1
8662314 Jones et al. Mar 2014 B2
8718814 Clark et al. May 2014 B1
8776694 Rosenwinkel et al. Jul 2014 B2
8811722 Perez Cortes Aug 2014 B2
8823801 Jacobson Sep 2014 B2
8874270 Ando Oct 2014 B2
8972049 Tidhar et al. Mar 2015 B2
9102053 Suzuki Aug 2015 B2
9102055 Konolige et al. Aug 2015 B1
9102336 Rosenwinkel Aug 2015 B2
9120622 Elazary et al. Sep 2015 B1
9227323 Konolige Jan 2016 B1
9256775 Yasunaga Feb 2016 B1
9259844 Xu et al. Feb 2016 B2
9266237 Nomura Feb 2016 B2
9283680 Yasuda et al. Mar 2016 B2
9364865 Kim Jun 2016 B2
9381645 Yarlagadda Jul 2016 B1
9481518 Neiser Nov 2016 B2
9486926 Kawano Nov 2016 B2
9492923 Wellman et al. Nov 2016 B2
9604363 Ban Mar 2017 B2
9751693 Battles et al. Sep 2017 B1
9821464 Stiernagle et al. Nov 2017 B2
9878349 Crest et al. Jan 2018 B2
9926138 Brazeau et al. Mar 2018 B1
9937532 Wagner et al. Apr 2018 B2
9962743 Bombaugh et al. May 2018 B2
9975148 Zhu et al. May 2018 B2
10007827 Wagner Jun 2018 B2
10029865 McCalib, Jr. et al. Jul 2018 B1
10048697 Theobald Aug 2018 B1
10058896 Hicham et al. Aug 2018 B2
10127514 Napoli Nov 2018 B2
10737299 Wagner Aug 2020 B2
11046530 Koga Jun 2021 B2
11986859 Wagner et al. May 2024 B2
20010056313 Osborne, Jr. Dec 2001 A1
20020092801 Dominguez Jul 2002 A1
20020147568 Wenzel et al. Oct 2002 A1
20020169698 Chien Nov 2002 A1
20020179502 Cerutti et al. Dec 2002 A1
20030029946 Lieber et al. Feb 2003 A1
20030034281 Kumar Feb 2003 A1
20030038065 Pippin et al. Feb 2003 A1
20040261366 Gillet et al. Dec 2004 A1
20050002772 Stone Jan 2005 A1
20050268579 Natterer Dec 2005 A1
20060022824 Olsen, III et al. Feb 2006 A1
20060045672 Maynard et al. Mar 2006 A1
20060070929 Fry et al. Apr 2006 A1
20060182543 Schaefer Aug 2006 A1
20060190356 Nemet Aug 2006 A1
20070043468 Schaefer Feb 2007 A1
20070209976 Worth et al. Sep 2007 A1
20080046116 Khan et al. Feb 2008 A1
20080181485 Beis et al. Jul 2008 A1
20080181753 Bastian et al. Jul 2008 A1
20100125361 Mougin et al. May 2010 A1
20100260380 Kaeser et al. Oct 2010 A1
20100292841 Wickham Nov 2010 A1
20100318216 Faivre et al. Dec 2010 A1
20110144798 Freudelsperger Jun 2011 A1
20110184555 Kosuge et al. Jul 2011 A1
20110238207 Bastian, II et al. Sep 2011 A1
20110243707 Dumas et al. Oct 2011 A1
20110320036 Freudelsperger Dec 2011 A1
20120118699 Buchmann et al. May 2012 A1
20120165986 Fuhlbrigge et al. Jun 2012 A1
20120177465 Koholka Jul 2012 A1
20120219397 Baker Aug 2012 A1
20130006423 Ito et al. Jan 2013 A1
20130051696 Garrett Feb 2013 A1
20130051969 Takeuchi et al. Feb 2013 A1
20130202195 Perez Cortes et al. Aug 2013 A1
20130232039 Jackson et al. Sep 2013 A1
20130235372 Voss Sep 2013 A1
20130245824 Barajas et al. Sep 2013 A1
20140067127 Gotou Mar 2014 A1
20140105719 Mueller et al. Apr 2014 A1
20140166549 Ito et al. Jun 2014 A1
20140305847 Kudrus Oct 2014 A1
20140360924 Smith et al. Dec 2014 A1
20150057793 Kawano Feb 2015 A1
20150073589 Khodl et al. Mar 2015 A1
20150081090 Dong Mar 2015 A1
20150081091 Blomquist et al. Mar 2015 A1
20150217937 Marquez Aug 2015 A1
20150224650 Xu et al. Aug 2015 A1
20150244650 Xu et al. Aug 2015 A1
20150283586 Dante et al. Oct 2015 A1
20150306634 Maeda Oct 2015 A1
20150352721 Wicks et al. Dec 2015 A1
20150375398 Penn Dec 2015 A1
20150378345 Winkler Dec 2015 A1
20160136816 Pistorino May 2016 A1
20160221762 Schroader Aug 2016 A1
20160228921 Doublet et al. Aug 2016 A1
20160243704 Vakanski et al. Aug 2016 A1
20160244262 O'Brien et al. Aug 2016 A1
20160379076 Nobuoka et al. Dec 2016 A1
20170024896 Houghton Jan 2017 A1
20170043953 Battles et al. Feb 2017 A1
20170050315 Henry et al. Feb 2017 A1
20170066597 Hiroi Mar 2017 A1
20170080566 Stubbs et al. Mar 2017 A1
20170087731 Wagner et al. Mar 2017 A1
20170121113 Wagner et al. May 2017 A1
20170128986 Sterkel May 2017 A1
20170136632 Wagner et al. May 2017 A1
20170137232 Messner May 2017 A1
20170157648 Wagner et al. Jun 2017 A1
20170157649 Wagner et al. Jun 2017 A1
20170173638 Wagner et al. Jun 2017 A1
20170225330 Wagner et al. Aug 2017 A1
20180075406 Kingston et al. Mar 2018 A1
20180085788 Engel et al. Mar 2018 A1
20180127219 Wagner et al. May 2018 A1
20180148272 Wagner et al. May 2018 A1
20180178255 Wagner et al. Jun 2018 A1
20180186572 Issing Jul 2018 A1
20180265291 Wagner et al. Sep 2018 A1
20180265298 Wagner et al. Sep 2018 A1
20180265311 Wagner et al. Sep 2018 A1
20180268184 Wagner et al. Sep 2018 A1
20180273295 Wagner et al. Sep 2018 A1
20180273296 Wagner et al. Sep 2018 A1
20180273297 Wagner et al. Sep 2018 A1
20180273298 Wagner et al. Sep 2018 A1
20180282065 Wagner et al. Oct 2018 A1
20180282066 Wagner et al. Oct 2018 A1
20180312336 Wagner et al. Nov 2018 A1
20180327198 Wagner et al. Nov 2018 A1
20180330134 Wagner et al. Nov 2018 A1
20180333749 Wagner et al. Nov 2018 A1
20190022702 Vegh et al. Jan 2019 A1
20190047786 Suzuki Feb 2019 A1
20190102712 Duca Apr 2019 A1
20190262868 Wagner et al. Aug 2019 A1
20190329979 Wicks et al. Oct 2019 A1
20200124484 Fuller et al. Apr 2020 A1
20200218865 Wagner et al. Jul 2020 A1
20200324973 Edwards et al. Oct 2020 A1
20200368785 Wagner et al. Nov 2020 A1
20210061563 Ueda Mar 2021 A1
20220250121 Wagner et al. Aug 2022 A1
20230390806 Lakawicz et al. Dec 2023 A1
20240207898 Wagner et al. Jun 2024 A1
Foreign Referenced Citations (66)
Number Date Country
2006204622 Mar 2007 AU
2998544 May 2023 CA
3009102 Jun 2023 CA
1033604 Jul 1989 CN
1671489 Sep 2005 CN
101482879 Jul 2009 CN
101971221 Feb 2011 CN
102363354 Feb 2012 CN
102430530 May 2012 CN
102621155 Aug 2012 CN
202539084 Nov 2012 CN
103129783 Jun 2013 CN
103942518 Jul 2014 CN
203830302 Sep 2014 CN
104093650 Oct 2014 CN
104137051 Nov 2014 CN
104858150 Aug 2015 CN
108351637 Jul 2018 CN
108700869 Oct 2018 CN
113635319 Nov 2021 CN
114970574 Aug 2022 CN
3919865 Dec 1990 DE
19510392 Sep 1996 DE
102004001181 Aug 2005 DE
102004013353 Oct 2005 DE
102005061309 Jul 2007 DE
102007023909 Nov 2008 DE
102007038834 Feb 2009 DE
102010002317 Aug 2011 DE
102012102333 Sep 2013 DE
102014111396 Feb 2016 DE
0235488 Sep 1987 EP
837415 Apr 1998 EP
1995192 Nov 2008 EP
2053350 Apr 2009 EP
2511054 Oct 2012 EP
2511653 Oct 2012 EP
2823899 Jan 2015 EP
3640582 Apr 2020 EP
3391159 May 2023 EP
3347780 Jun 2023 EP
2084531 Apr 1982 GB
2356383 May 2001 GB
2507707 May 2014 GB
S54131278 Oct 1979 JP
S63310406 Dec 1988 JP
H0395001 Apr 1991 JP
H05324662 Dec 1993 JP
H08157016 Jun 1996 JP
2002175543 Jun 2002 JP
2003150230 May 2003 JP
2007182286 Jul 2007 JP
2008037567 Feb 2008 JP
101413393 Jun 2014 KR
201004781 Feb 2010 TW
2005022076 Mar 2005 WO
2007009136 Jan 2007 WO
2010034044 Apr 2010 WO
2010099873 Sep 2010 WO
2011038442 Apr 2011 WO
2014166650 Oct 2014 WO
2015118171 Aug 2015 WO
2015162390 Oct 2015 WO
2017036780 Mar 2017 WO
2017044747 Mar 2017 WO
2017106423 Jun 2017 WO
Non-Patent Literature Citations (47)
Entry
Extended European Search Report issued by the European Patent Office in related European Patent Application No. 23177582.6 filed Aug. 30, 2023, 12 pages.
Extended European Search Report issued by the European Patent Office in related European Patent Application No. 23177584.2 filed Aug. 30, 2023, 10 pages.
Notice on Second Office Action issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 202110913420.9 on Sep. 29, 2023, 18 pages.
Notice on First Office Action issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 202110913420.9 on Mar. 30, 2023, 19 pages.
Bohg, Jeannette, et al., “Data-Driven Grasp Synthesis—A Survey,” Transactions on Robotics, pp. 289-309, Apr. 14, 2016.
Cipolla, Roberto et al., “Visuallly Guided Grasping in Unstructured Environments,” Journal of Robotics and Autonomous Systems, pp. 337-346, Mar. 3, 2001.
Communication pursuant to Article 94(3) EPC issued by the European Patent Office in related European Patent Application No. 16778496.6 on Feb. 24, 2021, 6 pages.
Communication pursuant to Article 94(3) EPC issued by the European Patent Office in related European Patent Application No. 16826518.9 on Mar. 4, 2021, 10 pages.
Communication pursuant to Article 94(3) EPC issued by the European Patent Office in related European Patent Application No. 16826518.9 on Mar. 22, 2022, 8 pages.
Communication pursuant to Article 94(3) EPC issued by the European Patent Office in related European Patent Application No. 16778496.6 on Mar. 24, 2022, 6 pages.
Communication pursuant to Rules 161(1) and 162 EPC issued by the European Patent Office on Apr. 18, 2018, in related European Patent Application No. 16778496.6, 3 pages.
Communication pursuant to Rules 161(1) and 162 EPC issued by the European Patent Office on Aug. 9, 2018, in related European Patent Application No. 16826518.9, 3 pages.
Examiner's Report issued by the Canadian Intellectual Property Office on Jan. 6, 2020 in related Canadian Patent Application No. 2,998,544, 4 pages.
Examiner's Report issued by the Canadian Intellectual Property Office on Mar. 6, 2020 in related Canadian Patent Application No. 3,009,102, 3 pages.
Examiner's Report issued by the Canadian Intellectual Property Office on Dec. 7, 2020 in related Canadian Patent Application No. 2,998,544, 4 pages.
Examiner's Report issued by the Innovation, Science and Economic Development Canada in related Canadian Patent Application No. 3,009,102 on Jan. 12, 2021, 3 pages.
Examiner's Report issued by the Innovation, Science and Economic Development Canada (Canadian Intellectual Property Office) on Nov. 5, 2021 in related Canadian Patent Application No. 2,998,544, 4 pages.
Examiner's Report issued by the Innovation, Science and Economic Development Canada (Canadian Intellectual Property Office) in related Canadian Patent Application No. 3,009,102 on Dec. 31, 2021, 4 pages.
Final Office Action issued by the U.S. Patent and Trademark Office on Nov. 7, 2019 in related U.S. Appl. No. 15/982,238, 15 pages.
Final Office Action issued by the U.S. Patent and Trademark Office on Oct. 11, 2017 in related U.S. Appl. No. 15/260,837, 32 pages.
Final Office Action issued by the U.S. Patent and Trademark Office on Oct. 22, 2019 in related U.S. Appl. No. 15/901,656, 13 pages.
First Examiner's Report issued by the Canadian Intellectual Property Office on Jan. 21, 2019 in related Canadian Patent Application No. 2,998,544, 4 pages.
First Examiner's Report issued by the Canadian Intellectual Property Office on Apr. 15, 2019 in related Canadian Patent Application No. 3,009,102, 3 pages.
First Office Action, and its English translation, issued by the China National Intellectual Property Administration on May 6, 2020 in related Chinese Patent Application No. 201680065881.2, 27 pages.
First Office Action, and its English translation, issued by the China National Intellectual Property Administration on Jun. 24, 2020 in related Chinese Patent Application No. 201680081764.5, 18 pages.
International Preliminary Report on Patentability issued by the International Bureau of WIPO in related International Application No. PCT/US2016/050949 on Mar. 13, 2018, 10 pages.
International Preliminary Report on Patentability issued by the International Bureau of WIPO in related International Application No. PCT/US2016/066786 on Jun. 19, 2018, 11 pages.
International Search Report and Written Opinion issued by the International Searching Authority in related International Application No. PCT/US2016/050949 on Dec. 8, 2016, 13 pages.
International Search Report and Written Opinion issued by the International Searching Authority in related International Application No. PCT/US2016/066786 on Mar. 20, 2017, 14 pages.
Klingbeil, Ellen et al., “Grasping with Application to an Autonomous Checkout Robot,” Proceedings—IEEE Int'l Conf. on Robotics and Automation, pp. 2837-2844, May 9, 2011.
Non-Final Office Action issued by the U.S. Patent and Trademark Office on May 22, 2017 in related U.S. Appl. No. 15/228,692, 13 pages.
Non-Final Office Action issued by the U.S. Patent and Trademark Office on Jun. 18, 2019 in related U.S. Appl. No. 15/982,238, 27 pages.
Non-Final Office Action issued by the U.S. Patent and Trademark Office on Apr. 25, 2017 in related U.S. Appl. No. 15/260,837, 28 pages.
Non-Final Office Action issued by the U.S. Patent and Trademark Office on Mar. 21, 2019 in related U.S. Appl. No. 15/901,656, 10 pages.
Non-Final Office Action issued by the United States Patent and Trademark Office in related U.S. Appl. No. 16/800,587 on Feb. 11, 2022, 27 pages.
Notice on Grant of Patent Right for Invention and Search Report, along with its English translation, issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 201680081764.5 on Mar. 3, 2022, 8 pages.
Notice on the Second Office Action, and the Second Office Action, issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 201680065881.2 on Jan. 29, 2021, 7 pages.
Notice on the Second Office and the Second Office Action, along with its English translation, issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 201680081764.5 on Apr. 13, 2021, 14 pages.
Notice on the Third Office and the Third Office Action, along with its English translation, issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 201680081764.5 on Oct. 20, 2021, 12 pages.
Rembold, Derk et al., “Object Turning for Barcode Search,” Proceedings of the 2000 IEEE/RSK—Int'l Conf. on Intelligent Robots and Systems, pp. 1267, Oct. 31, 2000.
Wikipedia, “Automatic Identification and Data Capture,” Wikipedia.org, Mar. 10, 2017 (https://en.wikipedia.org/w/index.php?title=Automatic-idenification_and_data_capture&oldid=0769563714).
Wikipedia, “Machine Vision,” Wikipedia.org, Mar. 1, 2017 (https://en.wikipedia.org/w/index.php?title=Machine_vision&oldid=768036938).
Non-Final Office Action issued by the United States Patent and Trademark Office in related U.S. Appl. No. 18/384,875 on May 30, 2024, 7 pages.
Decision on Rejection, along with its English translation, issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 202110913420.9 on Feb. 8, 2023, 9 pages.
Examiner's Report issued by the Innovation, Science and Economic Development Canada (Canadian Intellectual Property Office) in related Canadian Patent Application No. 3,192,499 on Aug. 1, 2024, 3 pages.
Notice on the Third Office Action, along with its English translation, issued by the China National Intellectual Property Administration in related Chinese Patent Application No. 202110913240.9 on Jul. 2, 2024, 8 pages.
Examiner's Report issued by the Innovation, Science and Economic Development Canada (Canadian Intellectual Property Office) issued in related Canadian Patent Application No. 3,192,629 on Aug. 1, 2024, 5 pages.
Related Publications (1)
Number Date Country
20230019431 A1 Jan 2023 US
Provisional Applications (2)
Number Date Country
62269640 Dec 2015 US
62217200 Sep 2015 US
Continuations (3)
Number Date Country
Parent 16800587 Feb 2020 US
Child 17949936 US
Parent 15982238 May 2018 US
Child 16800587 US
Parent 15260837 Sep 2016 US
Child 15982238 US