The present invention relates to a novel means of teaching a robot program. More particularly, this invention discloses a means to create a wide variety of robot programs through intuitive human input combined with a 3D scan of representative work pieces.
The programming of industrial robotics is a widely known challenge. This is because robot programming requires a rare combination of skills: knowledge of the industrial process (e.g. welding, material removal, painting), knowledge of programming (e.g. variables, control flow), and knowledge of robotics (e.g. speeds, joint limits, singularities). What is required is an intuitive method that provides ease of use, highly accurate robot motion commands, and a consistent user experience.
The present invention comprises a robot programming means that uses 3D scanning and a process agnostic pointing device that is used in conjunction with user input to create a robot program. Other embodiments include the devices, methods, hardware and software components, and combinations thereof to create and execute a robot program. Further embodiments, aspects, and details are provided in the figures and detailed description of the invention.
And
The descriptions and figures used herein are for the purposes of disclosing particular embodiments of the invention only and is not intended to be limiting of the invention. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
Referring now to
The scanning device 11 can perform 3D scans by using various techniques, such as binocular vision, photogrammetry, structured light or other 3D sensing technologies known to those who are skilled in the art. It can also be capable of collecting 2D RGB information and/or fusing 2D and 3D data into RGB-D representations, i.e. RGB plus depth representations as known to those skilled in the art.
Referring now to
Referring now to
During the teaching operation, the user can indicate that the pointing device's 12 current pose should be recorded and used to generate the robot program. The generated robot program is comprised of robot motion commands and/or other commands, such as setting IOs, or other typical industrial robot instructions known to those skilled in the art. The robot program could be either a complete stand-alone program or a series of commands that will be incorporated in a larger program at a later time. As one embodiment, the user can press one of the buttons 120 on the pointing device. Additional buttons, dials, scrollable wheels, or other input means can be used to input robot program parameters and/or robot program states, such as increasing the speed, turning a weld gun on or off, when to close a gripper, if the motion is a process motion or a non-process motion (e.g. a weld motion or simply moving in the air without welding), and so forth. These robot program values, such as robot parameters and/or robot program states, will be shown to the user on a graphical user interface. Other robot actions or robot parameters can be associated with these types of user inputs in keeping with the spirit of this invention. The buttons 120 in
Referring now to
Referring now to
At step 602, the workpiece 15 is placed in view of the scanning device 11. This may or may not include putting the workpiece 15 in a fixture, depending on the intended robotic process to be performed. It should be noted that the scanned workpiece 15 can be an object that either is the actual workpiece 15 on which the robot will perform the operation, or accurately represents the dimensions and shape of the actual workpiece.
At step 604, a 3D scan will be performed on the workpiece 15. The 3D scan will create a 3D representation of the object in the computation device 12. The 3D scan can be performed in a number of ways, as known to those who are skilled in the art. This includes using a 3D vision system which combines multiple views of the workpiece and thereby eliminating gaps caused by occlusions from one or more viewpoints. The multiple views can be achieved through several means, such as having more than one stationary vision systems, by moving a single vision system to multiple viewpoints, or by a combination of stationary and moving vision systems. The vision systems could use binocular vision, photogrammetry, structured light or other 3D sensing technologies known to those who are skilled in the art. The end result of the scan is that a 3D representation of the workpiece 15 is saved in the computation device 12. Optional post processing can be performed on the 3D representation, such as smoothing, plane fitting, hole filling and other steps that are known to those who are skilled in the art of 3D scanning.
As another alternative embodiment, a human user holds the scanning device 11 and scans the workpiece 15 by moving the scanning device to one or more positions from which the scanning device can capture 3D information through means known to those who are skilled in the art. In one such embodiment, the scanning device 11 is integrated with a tablet computer or smart phone. In another such embodiment, the scanning device 11 is removably attached to a tablet computer or smart phone. In another such embodiment, the scanning device 11 is a stand-alone device.
After scanning the workpiece the human user can mount the scanning device 11 to a fixed or movable device. In one such embodiment, the fixed device is a tripod. In another such embodiment, the movable device is a robot.
At step 606, after the 3D scan of the workpiece takes place, a process agnostic pointing device 16 will be held and positioned by the user 17 at one or more positions to indicate where the robot 14 is intended to perform its operation(s). The pose of the pointing device will be observed by a scanning device 11, and the computation device 12 will calculate the relationships between the pointing device pose, the 3D representation of the workpiece 15, and the robot coordinate system using calibration techniques known to those skilled in the art.
In one embodiment, the calibration means will be that fiducials will be placed on the workpiece 15 and/or its surrounding area. The robot 14 will move to a known offset from a known feature on the fiducials, such as but not limited to, touching a known point on the fiducials with a tool held by the robot 14. The computation device 12 will record the robot 14 poses while touching three or more points on three or more fiducials. These same points will be found in scanning device's 11 images. The calibration transformation between the robot 14 coordinate system and the scanning device coordinate system can be calculated from the aforementioned recorded robot 14 poses and the aforementioned points in the scanning device's 11 images using transformation calculation means known to those skilled in the art.
As an alternate embodiment, the scanning device 11 for scanning the workpiece 15 may be different from the scanning device 11 used to record the pose of the pointing device 16. For example, without limitation, a 3D point cloud generating device can be used as the scanning device 11 to collect the 3D information of the workpiece 15 while a high resolution 2D camera can be used to track one or more fiducials 101, 102 and/or 103 on pointing device 16 in 3D space using means well known to those skilled in the art. Other combinations or embodiments are possible while keeping with the spirit of the present invention.
Due to possible occlusions between a single scanning device 11 and the pointing device 16, multiple views of the pointing device 16 may be needed. Similar to what was described above for providing multiple views of the workpiece 15, multiple views of the pointing device can be achieved by using a 3D vision system which combines multiple views of the workpiece and thereby eliminating gaps caused by occlusions from one or more viewpoints. The multiple views can be achieved through several means, such as having more than one stationary scanning devices 11, by moving a single scanning device 11 to multiple viewpoints, or by a combination of stationary and moving scanning devices 11. In the case of a moving scanning device 11, the scanning device 11 can be mounted on a robot and the robot can then be used to move the scanning device 11 by following the pointing device 16, by the user jogging a robot, by the user selecting one of several predefined viewing locations, or any other robot moving means known to those skilled in the art. The robot motion used to move the robot mounted scanning device 11 can be achieved while the operator is collocated with the robot in the robot cell by using a safety monitored stop, speed and separation monitoring, power and force limits, or other safety means known to those skilled in the art. Alternatively, a moving scanning device 11 can be mounted on a track, pan-and-tilt system or similar device.
At step 608, after moving the pointing device 16 to the appropriate pose, the user will indicate that the pose of the pointing device 16 should be recorded by the computation device 12. This can be done via buttons 120 on the pointing device 16 or other standard input means. Later in the process, the computation device 12 will then use the recorded pose(s) of the user guided pointing device 16 to generate robot commands that perform operation(s) on the workpiece 15.
At step 610, the pose selected by the user in step 608 will be stored in the computation device 12.
At step 612, a decision will be made whether or not more poses will be recorded. In one embodiment, the default decision will be to continue collecting more poses until the user indicates that no more poses are needed by pressing one of the buttons 120 on the process agnostic pointing device 16 or an alternative input means known to those skilled in the art. To collect more poses, the process will return to step 606. If the user indicates no more poses are needed, then the process will continue to step 614.
At step 614, the computation device 12 will process both the 3D scan information collected in step 604 and all the recorded poses from the combination of steps 606 through 612.
In one embodiment, the computation device 12 will analyze the shapes created by scanning the workpiece 15 during step 604. For example, without limitation, the computation device would perform a best fit for predefined shapes, including, without limitation, planes, partial cylinders or partial spheres in the scanned representation of the workpiece 15 by using geometry fitting algorithms known by those skilled in the art. The computation device 12 will then create robot motion instructions based on locating the closest point to the pose taught from the pointing device 16 and the intersections of two or more of the best fit geometries. This approach will create high quality points on the workpiece 15 even if the user did not point accurately to the workpiece 15. In an alternative embodiment, the computation device 12 will create robot motion instructions based on the closest point to the pose taught from the pointing device 16 and the center of the closest geometric shape in the 3D scan information, such as, without limitation, a rectangle, circle, or triangle. In an alternative embodiment, the computation device 12 will create robot motion instructions based on locating the closest point to the pose taught from the pointing device 16 and the intersection of an edge of one of the best fit geometries.
The orientations contained in the robot motion instructions can also be calculated by the computation device 12 by, without limitation, using the best fit geometries. For example, the angle of intersection between two best fit geometries can be halved to determine one or more orientation angles for the robot motion instruction. In an alternate embodiment, the normal of a best fit geometry could be used.
In an alternate embodiment, the process rules and/or heuristics known by those skilled in the art for welding, painting, material removal, and so on can be combined with the best fit geometries and the recorded process agnostic pointing device 16 poses and sequences of poses. For example, without limitation, the computation device can use a previously specified robot tool angle for the welding process, combined with halving the intersection angle of two planes, to calculate a very accurate tool orientation used in the robot motion instructions. This, combined with calculating the closest projection of the recorded pointing device's 16 location on the intersection line of two best fit planes, would provide a highly accurate position and orientation for the robot motion instructions. Other rules and/or heuristics for other applications are possible in keeping with the spirit of the present invention.
In an alternate embodiment, the computation device 12 will create additional robot instructions based on knowledge of the process. For example, for a welding path, the computation device 12 will add additional points at the start and end of the welding path robot motion instructions with a predefined distance offset to enable the robot to safely approach and depart from the workpiece 15 before and after the welding motion instructions, respectively. In another example, for a machining path, the computation device 12 will add points at the start of the milling path robot motion instructions with a predefined distance offset to enable the robot to start the milling tool and then safely approach the workpiece to perform the milling work. Likewise, the computing device 12 will add points at the end of the milling path robot instructions with a predefined offset to enable the robot to depart from the workpiece 15 safely and stop the milling tool. As can be readily appreciated, many more process-based rules can be used to generate additional robot motion and non-motion instructions automatically in keeping with the spirit of this invention.
In an alternate embodiment, the computation device 12 will set choose robot instructions and parameters of robot instructions based on knowledge of the process. For example, for a welding path, the computation device 12 will choose a weld command and a weld speed based on the type of weld, type of metal and thickness of material. In another example, for a machining path, the computation device 12 will choose a tool speed based on what tool is being used and the type of material being removed.
In an additional embodiment, traceability data will be stored in a database before, during or after the user records desired poses. Examples of traceability data include, without limitation, user name, date, part type, part ID, recorded poses, and/or associated scan information. Traceability data will be retrieved for quality checks as needed.
Referring now to
Referring now to
Referring now to
Referring now to
It should be appreciated that some workpieces 15 will have multiple options for intersecting planes and other shapes. In an alternative embodiment, a tablet computer, laptop, augmented reality, or similar device will be used to allow the user to select which planes and/or shapes to use as a basis for projecting poses 22 into the proper relationship with the scanned information of workpiece 15.
It should be appreciated that the same algorithms described above to calculate the closest projected poses 23 of each of the recorded poses 22 to an intersection line 35 can be calculated in real time prior to the recording, i.e. immediately, as the user moves pointing device 16. These projections can be shown to the user on various displays, including, but not limited to, a tablet computer, a laptop, computer screen, or augmented reality. The user can adjust the pose of the pointing device 16 based on the observation of the projected pose in order to get the desired projected pose. In cases where there are multiple possible intersection lines, such as but not limited to three planes being close together, multiple potential projections can be shown to the user in real time on a graphical user interface. In such cases, the user would use the buttons 120 on the pointing device to select which of the possible projections they want to record. In an alternative embodiment, the user would use the buttons 120 on the pointing device to select to not use a projection for the pose they are recording, and just use the actual pointing device 16 pose as it is. In an alternative embodiment, the orientation of the pointing device 16 would be used to select which of the possible projections, such as but not limited to, selecting the projected point that is closest to the majority of the body of the pointing device 16. In an alternative embodiment, the projected point would be selected using rules, such as but not limited to, selecting the projection that is colinear to previously selected projected points for the most recent series of poses being taught. For example, without limitation, the previous two points would be used as a reference to determine what is colinear.
It should be appreciated that in addition to showing each of the projected poses 23 in real time, that the robot path motions and robot poses and configurations for the projected poses 23 can also be shown in real time. For example, the robot path motions can be shown in real time as lines connecting the various projected poses 23 that show the intended path of the robot tool. Additionally, the robot pose and configuration at one of the projected poses 23 can be shown in real time by showing a representation of the robot or part of the robot at that said one of the projected poses 23.
Referring now to
At step 902, the equivalent step is performed as described above for step 602.
At step 904, the equivalent step is performed as described above for step 604.
At step 906, the equivalent step is performed as described above for step 606.
At step 908, the equivalent step is performed as described above for step 608.
At step 910, the equivalent step is performed as described above for step 610.
At step 912, a decision will be made whether or not more poses will be recorded. In one embodiment, the default decision will be to continue collecting more poses until the user indicates that no more poses are needed by pressing one of the buttons 120 on the process agnostic pointing device 16. To collect more poses, the process will return to step 906. If the user indicates no more poses are needed, then the process will continue to step 913.
At step 913, an additional 3D scan will be performed on the workpiece 15. The 3D scan will create a more accurate 3D representation of all of or portions of the workpiece 15 in the computation device 12 by performing a more accurate 3D scan at locations near the poses recorded by the user using the process agnostic pointing device 16 in steps 906, 908, 910 and 912. The more accurate 3D scan can be performed in a number of ways, as known to those who are skilled in the art, as described above for step 604. In one embodiment, the more accurate scan will be performed by using a 2D laser line scanner to scan the regions within a predefined range around the poses recorded in steps 906, 908, 910 and 912. In this embodiment, the scan would result in a one or more point clouds that can be registered with the initial point cloud generated in step 904 using registration means known to those skilled in the art. The points generated by the second, more accurate scan, can then be used to replace the less accurate points generated earlier in step 904. Thus, the highly accurate points generated in the scan in step 913 would be saved for the regions of interest of the program as defined by the poses recorded in steps 906, 908, 910 and 912. As can be appreciated by those skilled in the art, alternative scanning and replacing means are also possible in keeping with the spirit of this invention. Alternate embodiments of 3D scanning would include, without limitation, collecting one or more pieces of 3D information about the workpiece 15 by using a single point laser distance sensor, or using a contact sensing means, such as by using electric or force based measurement, when a portion of a robot tool comes in contact with the workpiece 15.
At step 914, the equivalent step is performed as described above for step 614 except that the 3D information about the workpiece 15 will be based on the output of step 913, which is a merge of data from steps 904 and the scan performed in step 913 as described above.
Referring now to
At step 1002, the equivalent step is performed as described above for step 602.
At step 1006, the equivalent step is performed as described above for step 606 with the exception that there is no preceding scan of the workpiece 15 and the relationship of the workpiece 15 and the other coordinate systems is not defined at this point.
At step 1008, the equivalent step is performed as described above for step 608.
At step 1010, the equivalent step is performed as described above for step 610.
At step 1012, a decision will be made whether or not more poses will be recorded. In one embodiment, the default decision will be to continue collecting more poses until the user indicates that no more poses are needed by pressing one of the buttons 120 on the process agnostic pointing device 16. To collect more poses, the process will return to step 1006. If the user indicates no more poses are needed, then the process will continue to step 1013.
At step 1013, a 3D scan will be performed on the workpiece 15. The 3D scan will create a partial 3D representation of the object in the computation device 12 by performing a 3D scan at locations near the poses recorded by the user using the process agnostic pointing device 16 during steps 1006, 1008, 1010, and 1012. The 3D scan can be performed in a number of ways, as known to those who are skilled in the art, as described above for step 604. The computation device 12 will calculate the relationships between the recorded pointing device poses, the 3D representation of the scanned workpiece 15, and the robot coordinate system using calibration techniques known to those skilled in the art.
At step 1014, the equivalent step is performed as described above for step 614 except that the 3D information about the workpiece will be based on scan performed during step 1013.
Referring now to
In one embodiment, the computation device 12 will analyze the 3D scan from the scanning device 11. The computation device would perform a best fit for predefined shapes, including, without limitation, planes, partial cylinders or partial spheres in the scanned representation of the workpiece 15 by using geometry fitting algorithms known by those skilled in the art. The results of the geometry fitting algorithms would be used as the 3D representation 1102 of the workpiece 15 on a tablet computer 1101.
In one embodiment, the tablet computer 1101 includes a touch screen and the user will input locations 1111 on the screen by touching locations with their finger 1112. As can be appreciated by those skilled in the art, alternative input means such as, without limitation, using a stylus on a touch screen or a mouse on a laptop.
In one embodiment, the input locations 1111 are moved by the computational device 12 to touch the closest point on an intersecting line 1110 that is the intersection of two closest planes of the 3D representations 1102 of workpiece 15. The calculation of intersecting lines and the closest point on the intersecting lines can be performed by algorithms known by those skilled in the art. After the user selects all the points required for the operation on the workpiece, the computational device will use the locations of the closest points and process rules and/or heuristics known by those skilled in the art for welding, painting, material removal, and so on to generate robot instructions to perform work on the workpiece 15. For example, without limitation, the locations of the closest points will be combined with halving the intersection angle of two planes to generate the position and orientation of robot motion commands.
As can be appreciated by those skilled in the art, the computational device 12 can be a single physical device or can be a distributed system. A single physical device could be, without limitation, a tablet computer, PC, edge computer or server on the cloud. A distributed system could perform some calculations across multiple physical devices such as, without limitation, a tablet computer, PC(s), edge computer(s) or server(s) on the cloud.
In an alternative embodiment, after poses are taught using the process agnostic teaching device 16 as described elsewhere in this document, a 3D scan will be performed just before work is performed on the workpiece 15. The 3D scan will include part or all of the workpiece 15. The 3D scan can be performed by the same scanning device 11 that was used during the method to create robot instructions to perform work on workpiece 15, or a difference scanning device. The 3D information collected during this pre-work 3D scan will be registered to the 3D point information collected in the method to create robot instructions to perform work on workpiece 15. The registration method would be based on iterative closest point algorithms, or other methods known by those skilled in the art. The registration transformation generated by the registration method would be used to adjust the poses in the robot instructions in order to be accurately calibrated before the robot performs work on workpiece 15.
In an additional embodiment, a 3D scan of the workpiece 15 will be performed after the work is performed on the workpiece, producing a post work 3D scan. The said post work 3D scan would be compared with the 3D scan of an ideal workpiece that had work performed on it earlier. The two 3D scans will be registered using 3D registration means known to those skilled in the state of the art. After registering the two 3D scans, the 3D scans will be compared by calculating all the distances of points in the 3D scan from the workpiece 15 that just had work performed on it and the 3D scan of the ideal workpiece. The sum of the distances will be calculated and compared to a threshold value to determine if the differences are sufficiently small to ensure a good quality of the work performed on the workpiece 15 that just had work performed on it. In an alternative embodiment, the comparison is limited to portions of the workpiece 15, for example, regions within a predefined distance from locations where work is performed on workpiece 15. In a further embodiment, the distances can be shown to the user on a graphical user interface. This can be done by various means such as, without limitation, color coding the point cloud based on the distance values. One color coding scheme would be making the points with the largest distances red and gradually transitioning the color all the way to green as the points have the smallest distances. As can be easily appreciated, other color coding schemes could be used, as well as other graphical means, to show the distance values, such as histograms, bar graphs, or other means.
In another alternative embodiment, the threshold for determining if the sum of the distances between 3D scans is a good or bad quality is learned using machine learning techniques known by those skilled in the art. For example, without limitation, 3D scans are made after every workpiece 15 is produced. The sum of the distances between each 3D scan and the scan of an ideal workpiece are recorded and a quality indicator is also recorded with each sum of the distances. The sum of the distances can be stored as a single value, or as values based on regions of the 3D scans. Support vector machines, or other machine learning techniques, can be trained using this data to predict the quality of future workpieces based on the differences between their 3D scans and the 3D scan of the ideal workpiece.
In an additional embodiment, a visual scan (i.e. color scan) and 3D scan of the workpiece 15 will be performed after the work is performed on the workpiece, producing a post work RGB-D scan, i.e. color and depth scan. The said post work RGB-D scan would be compared with an RGB-D scan of an ideal workpiece that had work performed on it earlier. The two RGB-D scans will be registered using 3D registration means known to those skilled in the state of the art. After registering the two RGB-D scans, the 3D portion of the scans will be compared by calculating all the distances of points in the 3D scan from the workpiece 15 that just had work performed on it and the 3D scan of the ideal workpiece. The RGB (i.e. color) values for the closest points between the two scans will be compared to a threshold value to determine if the differences in color and/or intensity are sufficiently small to ensure a good quality of the work performed on the workpiece 15 that just had work performed on it. In an alternative embodiment, the RGB comparison is limited to portions of the workpiece 15, for example, regions within a predefined distance from locations where work is performed on workpiece 15. In a further embodiment, the regions with significant color differences as determined by a predefined or user defined threshold can be shown to the user on a graphical user interface by means of using a special color, such as red or white, or other visually recognizable attribute. In an alternative embodiment, the user interface can highlight areas that have both a significant RGB difference as well as 3D distance, as defined by respective predefined or user defined thresholds. The highlighting can be done by various means such as, without limitation, color coding the point cloud based on the distance values and adding a predefined color, such as red or white, for regions where the color of the two RGB-D scans is significantly different as defined by the color difference threshold. As can be appreciated, there are several additional variations of comparing and displaying RGB-D differences in keeping with the spirit of this invention.
Referring now to
Referring now to
In an alternative embodiment, additional 3D information is not collected by the scanning device 11B after the external axis has moved the workpiece 15 into the workspace of the robot 14. In this alternative embodiment, the relationship between the workpiece 15 and the fiducials in the workspace is known before and after the workpiece is moved into the workspace of the robot 14. In this embodiment, the scanning device 11B detects the fiducials and, since the relationship between the fiducials and the workpiece 15 are known, the location of the workpiece 15 can be calculated using transformation and calibration means known to those skilled in the state of the art.
Referring now to
As can be appreciated by those skilled in the art of robot programming, a benefit of the present invention is the generation of highly accurate robot motions because they are generated by the computation device 12 directly from the actual workpiece 15 dimensions. Additionally, robot motion commands are derived from poses collected from a highly accurate pointing device 16 that are further refined by projecting those poses on the shapes of the real workpiece 15.
As can be further appreciated by those skilled in the art of robot programming, one benefit of using the highly accurate, process agnostic pointing device 16 is that the same device can be used for programming a wide variety of processes. For example, even though a robotic cell may have dozens of tools for deburring, milling, cutting, and drilling, a user can teach all the deburring, milling, cutting, and drilling operations with the same process agnostic pointing device 16. They can also teach painting, dispensing, pick and place, and other processes with the same process agnostic pointing device 16. Other applications and options are possible in keeping with the spirit of the present invention.
One benefit of this approach is that it reduces the training burden of the end user who may need to teach more than one process type and may need to work with more than one robot tool when they create more than one robot program, such as, without limitation, milling tools, drills, and saws. Another benefit of this approach is that there is only one pointing device 16. Cost and time are thereby kept low in creating and calibrating the pointing device. Another benefit is that the pointing device is very consistent and accurate, no matter which user teaches the program. The pose of the pointing device 16 can be tracked consistently, reducing risks of introducing variations and inaccuracies in the teaching process. The reduction of these enables the high accuracy to be maintained, as described above.
A further benefit of this approach is that the accuracy and teaching flexibility is maintained in a robot programming system that does not require detailed understanding of robot programming. The robot teaching system 10, provides a way for the user to specify robot motion poses by using a simple pointing device 16 and without the user knowing how to write robot motion commands themselves.
It is important to note that the robot 14 described in this invention could be, but not limited to, any mechanical device with more than one degree of freedom whose motions can be programmed via software. This includes, but is not limited to, 6 degree of freedom serial industrial robot arms, CNC machines, and 4 degree of freedom parallel structure robot. The operations performed by the robot could be, but not limited to welding, deburring, milling, painting, inspecting, and pick and place.
It is also important to note that there are a wide variety of possible communication means between the pointing device 16 to the computation device 12, such as IR, Bluetooth, or other means known by those skilled in the state of the art.
Number | Name | Date | Kind |
---|---|---|---|
2019024 | Emerson | Oct 1935 | A |
5662566 | Marxrieser | Sep 1997 | A |
6421048 | Shih | Jul 2002 | B1 |
6542925 | Brown | Apr 2003 | B2 |
6724364 | Tani | Apr 2004 | B2 |
7069108 | Saarela | Jun 2006 | B2 |
7236854 | Pretlove | Jun 2007 | B2 |
7353082 | Pretlove | Apr 2008 | B2 |
7856285 | Carbonera | Dec 2010 | B2 |
8032605 | Brown | Oct 2011 | B2 |
8050782 | Faellman | Nov 2011 | B2 |
8660300 | Svajda | Feb 2014 | B2 |
8751049 | Linder | Jun 2014 | B2 |
8918208 | Hickman | Dec 2014 | B1 |
9102055 | Konolige | Aug 2015 | B1 |
9128530 | Yin | Sep 2015 | B2 |
9188437 | Kurahashi | Nov 2015 | B2 |
9333649 | Bradski | May 2016 | B1 |
9423879 | Chen | Aug 2016 | B2 |
9669544 | Buehler | Jun 2017 | B2 |
9671777 | Aichele | Jun 2017 | B1 |
9737371 | Romo | Aug 2017 | B2 |
9772394 | Nagalla | Sep 2017 | B2 |
9811074 | Aichele | Nov 2017 | B1 |
9818231 | Coffey | Nov 2017 | B2 |
9895841 | Page | Feb 2018 | B2 |
9919421 | Rossano | Mar 2018 | B2 |
10078325 | Gunnarsson | Sep 2018 | B2 |
10162329 | Ndip-Agbor | Dec 2018 | B2 |
10228428 | Gustafsson | Mar 2019 | B2 |
10383654 | Yilmaz | Aug 2019 | B2 |
10406686 | Boca | Sep 2019 | B2 |
10427300 | Boca | Oct 2019 | B2 |
10448692 | Hsu | Oct 2019 | B2 |
10737396 | Rossano | Aug 2020 | B2 |
10864633 | Chipalkatty | Dec 2020 | B2 |
10956739 | Thomasson | Mar 2021 | B2 |
11059076 | Bauer | Jul 2021 | B2 |
20010028339 | Tani | Oct 2001 | A1 |
20040111178 | Saarela | Jun 2004 | A1 |
20050149231 | Pretlove | Jul 2005 | A1 |
20050256611 | Pretlove | Nov 2005 | A1 |
20080065243 | Fallman | Mar 2008 | A1 |
20080177410 | Carbonera | Jul 2008 | A1 |
20090132088 | Taitler | May 2009 | A1 |
20100150399 | Svajda | Jun 2010 | A1 |
20100185328 | Kim | Jul 2010 | A1 |
20110288964 | Linder | Nov 2011 | A1 |
20130222580 | Kurahashi | Aug 2013 | A1 |
20130249784 | Gustafson | Sep 2013 | A1 |
20150002391 | Chen | Jan 2015 | A1 |
20150049081 | Coffey | Feb 2015 | A1 |
20150177846 | Yin | Jun 2015 | A1 |
20150290802 | Buehler | Oct 2015 | A1 |
20150309316 | Osterhout | Oct 2015 | A1 |
20150314442 | Boca | Nov 2015 | A1 |
20150321427 | Gunnarsson | Nov 2015 | A1 |
20150324490 | Page | Nov 2015 | A1 |
20160016363 | Smith | Jan 2016 | A1 |
20160143693 | Yilmaz | May 2016 | A1 |
20160167232 | Takeshita | Jun 2016 | A1 |
20160184032 | Romo | Jun 2016 | A1 |
20160257000 | Guerin | Sep 2016 | A1 |
20160260261 | Hsu | Sep 2016 | A1 |
20160303737 | Rossano | Oct 2016 | A1 |
20160375524 | Hsu | Dec 2016 | A1 |
20170028557 | Battisti | Feb 2017 | A1 |
20170130648 | Jochman | May 2017 | A1 |
20170143442 | Tesar | May 2017 | A1 |
20180092698 | Chopra | Apr 2018 | A1 |
20180154518 | Rossano | Jun 2018 | A1 |
20180173200 | Atherton | Jun 2018 | A1 |
20180311818 | Chipalkatty | Nov 2018 | A1 |
20180345495 | Aldridge | Dec 2018 | A1 |
20180350056 | Cardenas Bernal | Dec 2018 | A1 |
20190184582 | Namiki | Jun 2019 | A1 |
20190275744 | Tsoutsos | Sep 2019 | A1 |
20190291277 | Oleynik | Sep 2019 | A1 |
20200167886 | Cho | May 2020 | A1 |
20200250490 | Ozawa | Aug 2020 | A1 |
20200368904 | Aldridge | Nov 2020 | A1 |
20210129339 | Pipe-Mazo | May 2021 | A1 |
20210138646 | Matsushima | May 2021 | A1 |
20210220991 | Rajkumar | Jul 2021 | A1 |
20210316449 | Wang | Oct 2021 | A1 |
20220152833 | Lee | May 2022 | A1 |
20230043994 | Takahashi | Feb 2023 | A1 |
Number | Date | Country |
---|---|---|
102009323 | Jan 2013 | CN |
102016120132 | Apr 2018 | DE |
2018005053 | Jan 2018 | WO |
Entry |
---|
Office Action (Non-Final Rejection) dated Mar. 13, 2023 for U.S. Appl. No. 17/301,670 (pp. 1-6). |
International Search Report and Written Opinion issued in App. No. PCT/US22/42715, dated Feb. 27, 2023, 33 pages. |
Hsien-Chung Lin, “Embedding intelligence into Robotic Systems-Programming, Learning, and Planning”, University of California, Berkeley, Graduate Division, Engineering—Mechanical Engineering, Doctoral Thesis, (online), Nov. 3, 2018 [retrieved on Dec. 16, 2022] retrieved from the Internet <https://escholarship.org/content/qt5z62k45g/qt5z62k45g.pdf> 91 pages. |
Office Action dated Mar. 30, 2023 for U.S. Appl. No. 17/301,716 (pp. 1-16). |
Number | Date | Country | |
---|---|---|---|
20230150125 A1 | May 2023 | US |
Number | Date | Country | |
---|---|---|---|
62704194 | Apr 2020 | US |