The disclosure relates to techniques for iterative and rule-based position control of an end effector of a robotic system, and more particularly, to address complex situations where the end effector encounters virtual constraints.
In recent years, robotic systems have become an integral part of many orthopaedic surgeries. Higher accuracy, reproducibility and the ability to execute complex milling procedures are just some of the advantages of robotic systems. On the one hand, they perform milling procedures automatically and increase the final accuracy. On the other hand, they can produce increasing costs and consume additional training time since the surgical workflow has to be adapted to the robot. Therefore, such systems are suitable for interventions that are performed frequently and that need minimal interaction by the surgeon.
Certain robotic devices have to consider the human dynamics (human-in-the-loop control). The frequency of the on/off switching of the tool power may present problems. Moreover, prior techniques are not successful in controlling the end effector with respect to arbitrarily shaped constraints, such as a surface and not a closed volume. For example, in prior techniques, when the end effector is constrained by a surface, the end effector may stick to the virtual constraint for an extended period of time until the robotic device is moved a sufficient distance away from the constraint. This problem limits range of tool movement relative to constraints, and hence, decreases smoothness of tool control and user experience. Accordingly, there is a need to address at least the aforementioned issues.
This Summary introduces a selection of concepts in a simplified form that are further described below in the Detailed Description below. This Summary is not intended to limit the scope of the claimed subject matter nor identify key features or essential features of the claimed subject matter.
One example of a system is disclosed. The system includes a robotic device comprising an end effector and a plurality of actuators to position the end effector in a plurality of degrees of freedom. A navigation system is configured to track an actual position of the end effector. A controller is configured to determine a home position of the end effector. The one or more controllers determine that an anticipated movement of the end effector from an actual position to the home position would cause a collision between the end effector and a virtual constraint. The controller configured to compute a target position of the end effector that avoids the collision. The target position is spaced apart from the actual position and computed with respect to the virtual constraint. The plurality of actuators of the robotic device move the end effector from the actual position to target position along the virtual constraint.
One example of a method of operating a system for preparing a bone for a surgical implant is disclosed. The system includes a robotic device comprising an end effector and a plurality of actuators to position the end effector in a plurality of degrees of freedom, a navigation system is configured to track an actual position of the end effector, and a controller. The method comprising: determining a home position of the end effector, determining that an anticipated movement of the end effector from an actual position to the home position would cause a collision between the end effector and a virtual constraint, computing a target position of the end effector that avoids the collision, where the target position is spaced apart from the actual position and computed with respect to the virtual constraint, and moving the end effector from the actual position to the target position with one or more of the plurality of actuators.
Technical solutions provided by the disclosure include addressing complex situations where the end effector encounters virtual constraints. In order to prevent the end effector from sticking on the virtual constraint after contact, the techniques define the target position to enable the end effector to slide along the constraint thereby evading any further collisions and allowing continued tool movement. This solution expands range of tool movement relative to constraints, and hence, increases smoothness of tool control and user experience. Also, using a position-controlled robotic device over a speed-controlled milling device results in smoother surfaces and a faster execution. The technical solutions can be realized for hand-held robotic devices as well as larger robotic systems.
The systems, methods, and techniques may exhibit advantages and provide technical solutions other than those described herein.
Other aspects, features, and advantages of the present invention will become more fully apparent from the detailed description, the appended claims, and the accompanying drawings wherein like reference numerals identify similar or identical elements.
Described herein are iterative and rule-based position techniques for controlling a robotic device (RD) to address complex situations where an end effector 20 of the robotic device (RD) encounters virtual constraints (VC).
In one example, as shown in
When the robotic device (RD) is handheld, it can be similar to that shown in U.S. Patent Application Publication No. 2013/0060278, entitled, “Surgical Instrument Including Housing, a Cutting Accessory that Extends from the Housing and Actuators that Establish the Position of the Cutting Accessory Relative to the Housing”, and similar to that shown in U.S. Patent Application Publication No. 2014/0276949, filed on Mar. 15, 2014, entitled, “End Effector of a Surgical Robotic Manipulator,” the disclosures of which are hereby incorporated by reference.
The robotic device (RD) comprises an end effector 20. In many examples descried herein, the end effector 20 is a substantially spherically shaped burr. However, the end effector 20 can have other configurations, such as a drill, drill guide, saw blade, saw blade guide, screw driver, or the like. The end effector 20 can be dynamically relocated using one or several joints (J) integrated with the robotic device (20).
In
In other examples, the robotic device (RD) can be a larger robotic manipulator that has a base and plurality of links. A manipulator cart supports the manipulator such that the manipulator is fixed to the manipulator cart. The links collectively form one or more arms of the manipulator. The end effector 20 is attached to a distal end of the arm. The manipulator may have a serial arm configuration, a parallel arm configuration, or any other suitable manipulator configuration. In other examples, more than one manipulator may be utilized in a multiple arm configuration. One example of using this version of the robotic device (RD) is described in U.S. application Ser. No. 16/000,498, entitled “Robotic Surgical System And Method For Producing Reactive Forces To Implement Virtual Boundaries”, the entire disclosure of which is hereby incorporated by reference in its entirety.
The robotic device (RD) is movable relative to the patient (P) to interact with the anatomy in certain modes. Examples of these modes are described in U.S. Pat. No. 9,119,655, entitled, “Surgical Manipulator Capable of Controlling a Surgical Instrument in Multiple Modes,” the disclosure of which is hereby incorporated by reference.
In
The end effector 20 may comprise a TCP, which in one example, is a predetermined reference point. The TCP has known position in its own coordinate system. In one example, the TCP is assumed to be located at the center of a spherical feature of the end effector 20 such that only one point is tracked. The TCP may relate to a burr having a specified diameter, in one example. The TCP may be defined according to various manners other than those described herein.
As shown in
The navigation system 32 includes a cart assembly 34 that houses a navigation computer 36, and/or other types of control units. A navigation interface is in operative communication with the navigation computer 36. The navigation interface includes one or more displays 38. The navigation system 32 is capable of displaying a graphical representation of the relative states of the tracked objects to the operator using the one or more displays 38. Input devices 40, 42 may be used to input information into the navigation computer 36 or otherwise to select/control certain aspects of the navigation computer 36. As shown in
The navigation system 32 also includes a navigation localizer 44 (hereinafter “localizer”) coupled to the navigation computer 36. In one example, the localizer 44 is an optical localizer and includes a camera unit 46. The camera unit 46 has an outer casing 48 that houses one or more optical sensors 50.
In one example, the navigation system 32 includes one or more trackers. The trackers include one or more tool trackers 52, a first patient tracker 54, and a second patient tracker 56, or the like. In the example of
Any one or more of the trackers may include active markers 58. The active markers may include light emitting diodes (LEDs). Alternatively, the trackers 52, 54, 56 may have passive markers, such as reflectors, which reflect light emitted from the camera unit 46. Other suitable markers not specifically described herein may be utilized.
The localizer 44 tracks the trackers 52, 54, 56 to determine a state of each of the trackers 52, 54, 56, which correspond respectively to the state of the object respectively attached thereto. The localizer 44 provides the state of the trackers 52, 54, 56 to the navigation computer 36. In one example, the navigation computer 36 determines and communicates the state of one or more of the trackers 52, 54, 56 to a computing device. As used herein, the state of an object includes, but is not limited to, data that defines the position and/or orientation of the tracked object or equivalents/derivatives of the position and/or orientation. For example, the state may be a pose of the object, and may include linear data, and/or angular velocity data, and the like.
The navigation computer 36 may include a navigation controller 62 for controlling the navigation system 32. The navigation controller 62 may also communicate states of the patient (P) and end effector 20 to the operator by displaying an image of the femur F and/or tibia T, and/or the end effector 20, on the one or more displays 38.
Although one example of the navigation system 32 is shown in the Figures, the navigation system 32 may have any other suitable configuration for tracking the end effector 20 and the patient (P). In one example, the navigation system 32 and/or localizer 44 are ultrasound-based. For example, the navigation system 32 may comprise an ultrasound imaging device coupled to the navigation computer 36. The ultrasound imaging device images any of the aforementioned objects, e.g., the end effector 20, and the patient (P), and generates state signals to the navigation computer 36 based on the ultrasound images. The ultrasound images may be 2-D, 3-D, or a combination of both. The navigation computer 36 may process the images in near real-time to determine states of the objects. The ultrasound imaging device may have any suitable configuration and may be different than the camera unit 46 as shown in
In another example, the navigation system 32 and/or localizer 44 are radio frequency (RF)-based. For example, the navigation system 32 may comprise an RF transceiver coupled to the navigation computer 36. The end effector 20 and the patient (P) may comprise RF emitters or transponders attached thereto. The RF emitters or transponders may be passive or actively energized. The RF transceiver transmits an RF tracking signal and generates state signals to the navigation computer 36 based on RF signals received from the RF emitters. The navigation computer 36 may analyze the received RF signals to associate relative states thereto. The RF signals may be of any suitable frequency. The RF transceiver may be positioned at any suitable location to track the objects using RF signals effectively. Furthermore, the RF emitters or transponders may have any suitable structural configuration that may be much different than the trackers 52, 54, 56 as shown in
In yet another example, the navigation system 32 and/or localizer 44 are electromagnetically (EM) based. For example, the navigation system 32 may comprise an EM transceiver coupled to the navigation computer 36. The end effector 20 and the patient (P) may comprise EM components attached thereto, such as any suitable magnetic tracker, electromagnetic tracker, inductive tracker, or the like. The trackers may be passive or actively energized. The EM transceiver generates an EM field and generates state signals to the navigation computer 36 based upon EM signals received from the trackers. The navigation computer 36 may analyze the received EM signals to associate relative states thereto. Again, such navigation system 32 examples may have structural configurations that are different than the navigation system 32 configuration as shown throughout the Figures.
Based on the tracker data, the navigation system 32 tracks states of the end effector 20 relative to the target site (e.g., bone) of the patient, both of which may move during the surgical procedure. A process is performed wherein the target site (e.g., bone surface) is registered relative to a coordinate system of its respective tracker such that the navigation system 32 knows the state of the target site. Planning data (e.g., preoperative or intraoperative images of the patient) relating to the target site is utilized in conjunction with registration such that the robotic device (RD) can be controlled relative to the target site. Anatomical registration can occur according to various techniques. In one technique, a pointer is tracked by the navigation system 32 and the pointer is utilized to touch points on the surface of the bone. In another example, anatomical registration occurs by imaging the target site using, e.g., a fluoroscopic imaging apparatus. Registration may also occur according to ultrasound or electromagnetic techniques.
The navigation system 32 and/or localizer 44 may have any other suitable components or structure not specifically recited herein. Furthermore, any of the techniques, methods, and/or components described above with respect to the camera-based navigation system 32 shown throughout the Figures may be implemented or provided for any of the other examples of the navigation system 32 described herein. For example, the navigation system 32 may utilize solely inertial tracking or any combination of tracking techniques.
Described herein are computer-implemented algorithms for advanced control of the robotic device (RD) to move the end effector 20 relative to virtual constraints (VC). This is realized by controlling the joints of the robotic device (RD). The speed of the robotic device (RD) can be controlled as well. By using DoF of the robotic device (RD), the frequency of the on/off switching of the tool power is eliminated.
Virtual constraints (VC) may be generic, patient specific, arbitrarily shaped (convex, concave), open volumes or closed volumes. The virtual constraints (VC) may be set pre- or intraoperatively using an interactive planning software. The virtual constraints (VC) can be cutting boundaries that delineate portions of an anatomy that should be removed (e.g., to receive an implant) from portions of the anatomy that should be protected. By tracking the anatomy of the patient (P), and by virtue of anatomical registration, the system 24 can associate the virtual constraint (VC) with the target site such that as the target site moves, so does the virtual constraint (VC). In turn, the end effector 20 can engage the target site within the boundary of the virtual constraint (VC). Alternatively, the virtual constraints (VC) can be defined relative to areas that do not involve the patient (P), such as areas in the workspace of the robotic device (RD) that the robotic device (RD) should avoid to prevent collisions, etc. In other examples, the virtual constraint (VC) can be utilized to guide the end effector 20 to a location, such as a start location for milling or sawing. Any number of virtual constraints (VC) can be utilized with the techniques described herein.
Referring now to
The control system 28 may be executed on one or more computing devices comprising one or more controllers configured to provide output to the robotic device (RD) in a wired or wireless manner. The computing device includes at least one processor and at least one memory device. The processor may include a microprocessor, a microcontroller, or the like. The memory device is a non-transitory computer-readable storage medium that stores computer-readable and executable instructions embodied in one or more programs or modules. The memory device may include, for example, non-volatile memory such as a hard disk or flash memory, and may also include random access memory (RAM), which can include non-volatile RAM (NVRAM), magnetic RAM (MRAM), ferroelectric RAM (FeRAM), or any other suitable memory. The control system 28 may comprise one or more controllers.
The control system 28 software can comprise a Generic Intraoperative Planning and Execution of Orthopaedic Surgery (GIPEOS) for planning and visualization and the Active Tool Controller (ATC) to control the robotic device (RD).
GIPEOS may be used to generate the virtual constraints (VC) needed, to calibrate the robotic device (RD), and to show the visualizations related to the same. The GIPEOS software may enable the surgeon to plan orthopaedic surgeries intra-operatively. Subsequently, the plan may be executed. If necessary, the surgeon may switch between planning and execution as often as necessary. The planning is generic which means that the software offers different elementary objects such as points, lines and planes. Subsequently, these objects may be assembled to more complex structures. Surfaces may be recorded and modified. Distance and angle measurements may be offered. Objects may be digitized relative to other each other (snapping functionality, e.g. place perpendicular or parallel to another object). Alternatively, virtual constraints (VC) can be defined preoperatively and generically based on a selected implant model, or the like. One example of such software that can be utilized to create virtual constraints (VC) with the techniques described herein is described in U.S. application Ser. No. 15/952,810, filed Apr. 13, 2018, entitled “Surgical Systems and Methods For Facilitating Ad-Hoc Intraoperative Planning of Surgical Procedures”, the entire disclosure of which is hereby incorporated by reference in its entirety.
The ATC ensures that the end effector 20 does not operate outside the virtual constraint (VC). By defining objects (planes, wedges, triangle meshes in general) the region may be constrained. In one example, the ATC may comprise a Behavior Controller (BC), a Kinematic Controller (KC) and a Motor Controller (MC), which are responsible for controlling the robotic device (RD). Additional to the controllers there may be two components for reading input data and providing data to components outside the ATC. The behavior controller (BC) is independent from the hardware and can be applied to tools with any degrees of freedom. The behavior controller (BC) may receive planning data from GIPEOS and pass the computed speed and target position to the kinematic controller (KC). The behavior controller (BC) receives the planning data and the current transformations and computes the target position as well as the end effector 20 speed. These parameters are passed to the kinematic controller (KC) that calculates the joint parameters according to a kinematic model (inverse kinematics). The kinematic model of the kinematic controller (KC) may be adapted to the hardware. These joint parameters are sent to the motor controller (MC) that communicates with the motors. Besides this forwarding pipeline there may also be one for feedback. The motor controller (MC) reads the current joint parameters and passes them to the kinematic controller (KC). The kinematic controller (KC) computes the current position and passes it to the behavior controller (BC) (forward kinematics). The motor controller (MC) updates the robotic device (RD) position. A calibration provider may import the calibrations and update the controllers. A parameter provider may collect calculated parameters and supply them to the visualization on request.
The end effector 20 can interact or collide with the virtual constraints (VC) during operation. There are different ways that the control system 28 reacts in case that the end effector 20 collides with one of these virtual constraints (VC). In one example, as long as possible, the end effector 20 will retreat in order not to intersect the virtual constraints (VC). If this is not possible anymore, the end effector's 20 power is turned off thus bringing the end effector 20 into a safe state.
Referring now to
The home position can be defined according to various configurations and considering various factors. In one example, the home position is set based on any of the following: the mechanics and kinematics of the robotic device (RD), the position of the tracker on the robotic device (RD), the type of joints (J), the ergonomics of the robotic device (RD), the type or nature of surgical procedure or surgical site, or the like. For instance, the home position can be set as the position of the end effector 20 when all joints (J) are in their halfway position. In
Referring now to
The position control algorithms perform a collision detection to find a valid position in which the end effector 20 does not intersect any virtual constraint (VC). Additional to the forward and the inverse kinematics it may also perform plausibility checks for the reachability of a given position. The position control algorithms ensure that the computed parameters are within a range that the hardware can reach. In case that these parameters are outside the reachable range, the motor controller (MC) turns off the end effector 20.
The following sections refer to these coordinate systems: (1) the coordinate system A of the tracker attached to the robotic device (RD); (2) the coordinate system R of the reference tracker attached to the patient's bone; and (3) the coordinate system W of the navigation system 32. The units of all coordinate systems are given in millimeters.
Collision Detection Algorithm:
The collision detection component finds any collision between the end effector 20 movement and the defined virtual constraints (VC). In one example, the virtual constraint (VC) may be a surface consisting of one or more planar triangles. Alternatively, or additionally, the virtual constraint (VC) can be modeled according to geometry of any surface or shape. The virtual constraint (VC) can be modeled using a freeform surface, splines, non-uniform rational basis spline (NURBS), point clouds, or the like.
In one example, the end effector 20 is modeled as a spherical burr and the ATC expects spherical burrs. However, it will be readily understood that differently shaped burs may also be used with the collision detection algorithm. For instance, rotary tools can be modeled to have a surface of revolution about an axis, saws can be modeled as a volume that envelopes the range of oscillating motion of the saw, etc.
In one example, the control system 28 employs a discrete approach (a posteriori) whereby a collision between the end effector 20 and the virtual constraint (VC) is detected after it occurred. On the other hand, the control system 28 can employ a continuous approach (a priori) whereby the control system 28 predicts the movement of the end effector 28 and finds a collision before it occurs. In one instance, a discrete approach is used because there are two movements: the movement of the human hand and the movement of the end effector 20 (caused by the joints of the robotic device). Whereas the movement of the end effector 20 is controlled, the human movement is difficult to predict when there is only the navigation system 32 without additional sensors (tremor prediction normally uses high-speed 6-DoF gyroscopes). These sensors can be utilized as well. Therefore, a discrete approach may be used to calculate the collision for a given movement of the end effector 20. When the end effector 20 is modeled as a sphere, the convex hull of a linear sphere movement trajectory is a capsule.
Referring now to
For simplicity and optimization, this collision detection may be performed in the virtual constraint (VC) base coordinate system T. In this coordinate system the first triangle corner coincides with the origin, the second one lies on the positive x-axis and the third one lies in the xy-plane with a positive y-coordinate. The z-coordinate is always 0, {right arrow over (w)}0 and {right arrow over (w)}1 lie on the x-axis and {right arrow over (w)}0 is located in the origin of the coordinate system. The plane in which a triangle lies is split into seven regions by nine line segments. The lines are defined by their normals {circumflex over (n)}0 to {circumflex over (n)}8. The region in which a projected point {right arrow over (u)} lies is indicated by the sign of the distances to the lines. If a point lies inside the triangle the distance to line 0, 1 and 2 is positive. If a point lies under line 0 the distance to line 0 and 4 is negative and the one to line 3 is positive.
Referring now to
Further parameters are:
{right arrow over (m)}: The move vector from {right arrow over (p)} to {right arrow over (q)}, {right arrow over (m)}={right arrow over (q)}−{right arrow over (p)}
{right arrow over (n)}: The normal vector of the virtual constraint (VC) triangle
{right arrow over (i)}: The intersection point of the sphere with the virtual constraint (VC) triangle
{right arrow over (b)}: The “best guess” position is the position of the sphere center where it touches the virtual constraint (VC) triangle
{right arrow over (c)}: The nearest point to {right arrow over (p)} on the virtual constraint (VC) triangle
{right arrow over (a)}: The “away vector” from {right arrow over (c)} to {right arrow over (p)}. This vector is independent from the movement and only depends on the initial position {right arrow over (p)}, {right arrow over (a)}={right arrow over (p)}−{right arrow over (c)}
{right arrow over (η)}: The vector from {right arrow over (i)} to {right arrow over (b)} that is called “move plane normal”.
The move plane normal is called so since it defines the plane in which the sphere can move without intersecting the virtual constraint (VC). In certain cases, {right arrow over (η)} only equals the virtual constraint (VC) normal {right arrow over (n)}. The move plane normal depends on the movement of the sphere.
Referring now to
In a first example, as shown in a 3D view of
The collision detection algorithm may comprise of two parts. In one example, the nearest point on the triangle {right arrow over (c)} is calculated in order to know if the movement {right arrow over (m)} points towards the virtual constraint (VC). This is given if {right arrow over (a)}·{right arrow over (m)}>0 where {right arrow over (a)}={right arrow over (p)}−{right arrow over (c)}. In case the movement points towards the virtual constraint (VC), the second part of the algorithm can be executed to find smin and {right arrow over (η)}.
Referring now to
Nearest Point on Triangle Algorithm:
In order to calculate the nearest point to the sphere center on the virtual constraint (VC), vector {right arrow over (c)}, the sphere center can be transformed into base coordinates (ux,uy,uz)T and dropped into the xy-plane (ux,uy)T. Then, the control system 28 can decide in which of seven possible regions the sphere center lies, i.e., inside the triangle, perpendicular above resp. below one of the triangle sides or such that one of the triangle corners is nearest. These seven regions are defined by nine straight lines (see
di={circumflex over (n)}iT·{right arrow over (x)}−ci
where {circumflex over (n)}i is the unit normal vector of the line, ci the corresponding Hesse constant and {right arrow over (x)} an arbitrary point in the xy-plane, can be used to determine which region is the right one. In T coordinates, this results in:
d0=y
d1=n1x·(x−x1)+n1y·y
d2=n2x·x+n2y·y
d3=x
d4=x−x1
d5=n1y·(x−x1)−n1x·y
d6=n1y·(x−x2)−n1x·(y−y2)
d7=n2y·(x−x2)−n2x·(y−y2)
d8=n2y·x−n2x·y
Depending on the determined region, the nearest point on the triangle {right arrow over (c)} is calculated:
(d0≥0,d1≥0,d2≥0)→{right arrow over (c)}=(ux,uy,0)T
(d0<0,d3≥0,d4<0)→{right arrow over (c)}=(ux,0,0)T
(d4≥0,d5<0)→{right arrow over (c)}=(u1,0,0)T
(d1<0,d5≥0,d6<0)→{right arrow over (c)}=(ux−d1n1x,uy−d1n1y,0)T
(d6≥0,d7<0)→{right arrow over (c)}=(x2,y2,0)T
(d2<0,d7≥0,d8<0)→{right arrow over (c)}=(ux−d2n2x,uy−d2n2y,0)T
(d3<0,d8≥0)→{right arrow over (c)}=(0,0,0)T
Triangle Intersection Algorithm:
The goal of this algorithm is to calculate how far the sphere can move from {right arrow over (p)} to {right arrow over (q)} without intersecting the virtual constraint (VC). The best guess position in which the sphere touches the virtual constraint (VC) is defined as {right arrow over (b)}={right arrow over (p)}+{right arrow over (s)}·{right arrow over (m)} with {right arrow over (m)}={right arrow over (q)}−{right arrow over (p)}. In addition to these parameters, the move plane normal {right arrow over (η)} is determined.
a) Surface Intersection:
Intersecting the line from {right arrow over (p)} to {right arrow over (q)} with the plane z=r′ (with r′=r for mz<0 and r′=−r otherwise) results in
b) Edge Intersection:
Let {right arrow over (w)}i and {right arrow over (w)}j be the start and end position of the currently considered triangle edge. Then the edge itself is described by {right arrow over (w)}i+t·{right arrow over (d)} with {right arrow over (d)}={right arrow over (w)}i−{right arrow over (w)}j and 0≤t≤1. In case the sphere touches the edge the distance of a point {right arrow over (p)}+s·{right arrow over (m)} to the edge equals r:
With {right arrow over (h)}1=({right arrow over (p)}−{right arrow over (w)}))×{right arrow over (d)} and {right arrow over (h)}2={right arrow over (m)}×{right arrow over (d)} this results in
The best guess position {right arrow over (b)}E1,2={right arrow over (p)}+sE1,2·{right arrow over (m)} projected onto the edge {right arrow over (w)}i+t·{right arrow over (d)} results in the intersection point {right arrow over (i)}={right arrow over (w)}i+tL·{right arrow over (d)} with
c) Corner Intersection:
The distance of a corner IV to a point on the line {right arrow over (p)}+s·{right arrow over (m)} is =|{right arrow over (p)}+s·{right arrow over (m)}−{right arrow over (w)}|. In case the sphere intersects with the corner this distance equals r:
|{right arrow over (p)}+sC·{right arrow over (m)}−{right arrow over (w)}|=r
With {right arrow over (h)}={right arrow over (p)}−{right arrow over (w)} this results in
The best guess is {right arrow over (b)}C1,2={right arrow over (p)}+sC1,2·{right arrow over (m)} and the intersection point coincides with the corner point {right arrow over (i)}C={right arrow over (w)}.
s=min{{sF|0≤sF≤1}∪{sE|0≤sE≤1∧0≤tL≤1}∪{sC≤0<sC≤1}}
The corresponding best guess and intersection point are chosen and the move plane normal is {right arrow over (η)}={right arrow over (b)}−{right arrow over (i)}.
Position Control Algorithm:
Finding and correcting the end effector 20 position to avoid collisions with virtual constraints (VC) is an elementary task of this component. As long as there is no virtual constraint (VC) close to the end effector 20 the target position {right arrow over (t)} and the home position coincide. If the user holds the robotic device (RD) in such way that the home position collides with the virtual constraint (VC) the target position must be adapted to correct the condition. The distance between target and home position is minimized without intersecting any virtual constraint (VC). The calculated target position is saved and is used as the last valid position {right arrow over (l)} in the next iteration. As utilized herein, this last valid position is referred to as the “actual position” because it is the current position of the end effector 20 at the time the system is determining where the end effector 20 should go next.
The techniques described in this section address situations wherein the control system 28 detects that the home position of the end effector 20 breaches the virtual constraint (VC) and thereafter (in a first iteration) the actual position of the end effector 20 is adjusted to a target position to avoid collision with the virtual constraint (VC). Subsequently, the home position might still be behind the virtual constraint (VC) but differs from the previous one (due to user movement). Because the home position remains behind the virtual constraint (VC), the (updated) actual position of the end effector 20 contacts the virtual constraint (VC). Since the home position remains behind the constraint (VC) in the first iteration, the end effector 20 will have a tendency to stick to the virtual constraint (VC) at the same position.
Accordingly, to mitigate this problem, the control system 28 evaluates an anticipated movement of the end effector 20 from the actual position to the home position (behind the constraint). The control system 28 detects that the anticipated movement would cause an anticipated collision between the end effector 20 and the virtual constraint 20. A subsequent target position of the end effector 20 is computed that evades the anticipated collision. The target position is spaced apart from the actual position and also contacts the virtual constraint. These subsequent iterations are executed in order to prevent the end effector 20 from sticking on the virtual constraint (VC). The one or more joints (J) of the robotic device (RD) are controlled to slide the end effector 20 from the actual position to target position along the virtual constraint (VC). In one example, the control occurs until the distance between the home position {right arrow over (r)} and the actual position {right arrow over (l)} is minimized.
When the end effector 20 is modeled as a sphere, the actual position, home position and target positions can also be based on the sphere model. For example, these positions can be based on the outer surface of the sphere, a center of the sphere, or the like.
Furthermore, after the first iteration, wherein the actual position of the end effector 20 contacts the virtual constraint (VC), the actual position may, in practice, be an intersection between the sphere and the virtual constraint (VC) due to the fast movement of the robotic device (RD) by the operator. The intersection may be very small (˜10s to 100s of microns). In other examples, the actual position of the end effector 20 is positioned such that it tangentially contacts the virtual constraint (VC).
Referring now to
In some examples, a first virtual constraint (VC1) and a second virtual constraint (VC) are provided. The first and second virtual constraints (VC1, VC2) are connected and are each planar. In
Referring now to
Referring now to
Referring now to
Referring now to
In one example, movements between the actual and target position can be based on smallest allowable movement, described below. In another example, the target position is defined based on a preset distance from the actual position (e.g., 1-5 mm). In another example, the end effector 20 comprises an axis (A) (
Since the projected movement could intersect other virtual constraints (or the same constraint in other points), the collision detection has to be performed once in each phase. Performing a collision detection up to three times in one iteration guarantees that the robotic device (RD) does not execute a movement that is not previously checked against all virtual constraints (VC). This multi-phased architecture is attributed to the special nature of a robotic device (RD). In each iteration the home position and the actual position may be considered since they are moving relative to each other.
As described above, the target position is spaced apart from the actual position when there is a collision with the virtual constraint (VC). This is usually the case when the end effector 20 approaches the virtual constraint (VC) at an angle. In rare instances, continued movement {right arrow over (l)}→{right arrow over (r)} from the actual position to the home position is exactly normal to the surface of the virtual constraint (VC). In these instances, the target position may remain in the same position on the virtual constraint (VC) causing the end effector 20 to stick to the virtual constraint (VC) until the home positon is angled off normal. The control system 28 may employ supplemental or alternative techniques to address this rare situation.
Referring now to
In this example, inputs to the algorithm include the home position {right arrow over (r)} and the actual position {right arrow over (l)}. The output is the target position {right arrow over (t)}. Every collision of the current movement from {right arrow over (l)} to the projected target position {right arrow over (τ)}k produces one pair of {right arrow over (η)}k,i with corresponding sk,i. After the smallest allowed movement is found only the move plane normals with a s of smin,k are stored in Δk. This iterative position control algorithm can be summarized in the following steps:
(1) Calculate the projected target position {right arrow over (τ)}k according to the intersected triangles from the previous phases and the movement from {right arrow over (l)} to {right arrow over (r)}. In the first phase {right arrow over (τ)} equals {right arrow over (r)}.
(2) In case the projected target position {right arrow over (τ)}k equals the actual position {right arrow over (l)} the algorithm is terminated.
(3) Otherwise, a collision detection for the sphere movement from the actual position to the projected target position {right arrow over (l)}→{right arrow over (τ)}k is performed.
(4) The smallest allowed movements with the same factor smin, neglecting the triangles from the previous phases, is determined.
(5) If there is a possible movement (smin>0), the target position is set and the algorithm is terminated.
(6) Otherwise, start the next phase.
Move Plane Normal versus Triangle Normal:
In one example, every intersected triangle of the virtual constraint (VC) defines a move plane normal {right arrow over (η)} and a factor s which indicates how far the end effector 20 can move from {right arrow over (l)} to {right arrow over (τ)}k. The move plane normal defines the plane in which the end effector 20 is allowed to move relative to the triangle without intersecting it. Every move plane normal restricts the space in which the end effector 20 can move. If there are two different move plane normals with s=0 the space is restricted to the intersection line of the two planes defined by the move plane normals and the actual position {right arrow over (l)}. In this case, the end effector 20 could subsequently slide along this line. This case occurs in
When there are three different move plane normals (all with s=0) they intersect each other in one point. This means that the end effector 20 is completely restricted. In this case the projected target {right arrow over (τ)}k equals the actual position {right arrow over (l)}. One example would be a movement through the tip of a tetrahedron while touching the three faces of the tip.
The position control algorithm may use the move plane normal to compute in which direction the end effector 20 is allowed to move. The move plane normal {right arrow over (η)} is the vector from the contact point {right arrow over (i)} to the sphere center {right arrow over (b)}. Therefore, it is parallel the triangle's normal when the intersection point lies inside the triangle and differs when the intersection point lies on one of the triangle's edges or corners. Movement of the end effector 20 is limited to the move plane, when utilized.
Referring now to
Finding the Smallest Allowed Movement:
The control system 28 can identify a direction within the plane that results in the least amount of movement for the end effector 20 relative to the actual position and compute the target position based on the identified direction and the least amount of movement.
In one example, the collision detection algorithm returns a move plane normal {right arrow over (η)}k,i with corresponding factor sk,i for every collision (with 0≤i≤Ck). Different triangles can be intersected at different distances. Considering only the closest collisions, the control system 28 searches for move plane normals corresponding to the smallest allowed movement defined by smin,k. It is possible that several triangles can be intersected at the same distance. Hence, the result of this search is a matrix:
Δk=({right arrow over (η)}k,0 . . . {right arrow over (η)}k,M
containing the move plane normals corresponding to smin,k. Moreover, the hit triangles from the previous phase, represented by Δk−1, may be ignored since they are implicitly included in the projected movement (the projection was calculated according to the previous triangles). In the first phase there are no previous triangles.
Defining the Movement: Eventually the case occurs that {right arrow over (l)}→{right arrow over (τ)}k collides with a virtual constraint (VC). If smin>0, the algorithm calculates the target position {right arrow over (t)}={right arrow over (l)}+smin·({right arrow over (τ)}k−{right arrow over (l)}) and finishes the iteration (see
Projection Algorithm:
As mentioned above the current move vector {right arrow over (l)} to {right arrow over (r)} can be projected onto planes, lines, or points depending on the found virtual constraints (VC). Referring now to
Before calculating the projection, the control system 28 determines how restricted the space already is. This determination depends on the current phase.
In the first phase, k=0, the space is not yet restricted, therefore there is no projection necessary: {right arrow over (τ)}k={right arrow over (r)}
In the second phase, k=1, the space is restricted by the move plane normals found in the first phase Δ0=({right arrow over (η)}0,0 . . . {right arrow over (η)}0,M
Matrix V0 is a 3×3 matrix whose column vectors are the right-singular vectors of Δ0. Each right-singular vector {right arrow over (υ)}0,i is assigned to one singular value σ0,i, i=0, 1, 2. The vectors that belong to non-zero singular values define how the movement will be restricted. They are stored in matrix N0=({right arrow over (υ)}0,0 . . . {right arrow over (υ)}0,W
The vectors of N0 can form three different restrictions with three different projections ({right arrow over (m)}={right arrow over (r)}−{right arrow over (l)} and k=0):
(1) There is one right-singular vector {right arrow over (υ)}k,0 with σk,0>0 that defines a plane onto which the move vector is projected (see
(2) There are two right-singular vectors {right arrow over (υ)}k,0 and {right arrow over (υ)}k,1 with σk,i>0 that define a line onto which the move vector is projected (see
(3) There are three right-singular vectors with σk,i>0 that define a point, the move vector is {right arrow over (0)}
{right arrow over (τ)}k+1={right arrow over (l)}
The third phase, k=2, is similar to the second one. Additional to the move plane normals from the second phase Δ1=({right arrow over (η)}1,0 . . . {right arrow over (η)}1,M
SVD(({right arrow over (η)}1,0 . . . {right arrow over (η)}1,M
Since the vectors belonging to non-zero singular values N1 are the result of the singular value decomposition of matrix (N0Δ1), they implicitly contain the vectors of matrix Δ1. Therefore, the vectors of N1 can form just two different restrictions with two different projections ({right arrow over (m)}={right arrow over (r)}−{right arrow over (l)} and k=1):
(1) There are two right-singular vectors {right arrow over (υ)}k,0 and {right arrow over (υ)}k,i with σk,i>0 that define a line onto which the move vector is projected (see
(2) There are three right-singular vectors with σk,i>0 that define a point, the move vector is {right arrow over (0)}. The move plane normals from the previous phase Δ0 are added to the current ones Δ1.
Special Projection Cases:
Referring now to
Additionally, valid projections exist that are calculated by using the move plane normal, which may be preferred in some situations. In the example of
As explained above the number of right-singular vectors in Nk−1 indicates how to project. Additional to the standard cases from the previous section, different movement projections are calculated and added to a list of possible projections:
(1) If there are two right-singular vectors, the movement can be projected onto the virtual constraints (VC) themselves (the move plane normals {right arrow over (η)} of the virtual constraints (VC) are retrieved from Δk−1).
2) If there are three right-singular vectors, the movement {right arrow over (l)}→{right arrow over (r)} is projected onto all possible lines calculated by combining the move plane normals {right arrow over (η)} in Δk−1. As in the previous case, the movement is also projected onto all virtual constraints (VC). Now there is a list of possible projections that may be tested against all move plane normals in Δk−1. Invalid ones may be removed from the list. A projection may be invalid if the angle between the move plane normal and the projected movement is >90° where 90° is a parallel movement to the plane defined by the move plane normal. An angle <90° means that the movement points away from the plane defined by the move plane normal. It may be preferable to check several possibilities in the collision detection and choose the possibility with the longest allowed movement.
Burr Speed Control Algorithm:
The speed controller determines the speed of the end effector 20 depending on the feedback of the kinematic controller (KC). As long as the robotic device (RD) can deflect/evade the virtual constraint (VC), the speed may be set. As soon as one of the joints (J) reached its maximum, the speed may be set to 0. The burr speed may depend on the target position defined according to the techniques herein and can be varied depending on factors, such as the distance between the actual and target positions, etc.
To test the algorithms presented herein, a xyz-linear stage system was used in place of an robotic device (RD). The system comprised four linear stages driven by three motors and an orthopaedic milling device mounted on the end effector 20. Four translation stages driven by three motors were utilized. The communication between the motors and the PC was realized via CAN. A milling device was be used that allows the surgeon to easily exchange the end effector 20.
The calibration of the translational system comprised of three steps:
(1) Homing the translation stages
(2) Calibration of the transformation FCA from tracker coordinates A to linear stages coordinates C
(3) Pivoting of the end effector 20 to find the home position {right arrow over (r)}A
The homing of the device finds its maximum deflection and its center position (home position). Since the tracker coordinates A do not coincide with the axes orientation of the system, the transformation FCA between coordinate system A to C is found. In this example, the reference tracker may be rigidly attached to the tool shaft, whereas the tool tracker stays untouched. The tool shaft is moved to the maximum deflection in all three directions while recording the poses of the trackers (FWR, FWA). This results in 3×N transformations where N is the number of recorded transformation per axes. The transformation from the reference tracker to the end effector 20 tracker is calculated by FAR=FWA−1·FWR. The origin is extracted from these transformations. For all three records the extracted origins are used to calculate a regression line. The three calculated directions are part of the rotation matrix from A and C. The translation is zero. Finally, the end effector 20 is pivoted to find the home position {right arrow over (r)}1. During the pivoting calibration the end effector 20 is placed in a tracked mold. The user rotates the tool while keeping the end effector 20 in the mold. After the calibration, the end effector 20 position can be calculated from the recorded transformations. As a last step, its radius may be specified manually. In order to calibrate the radius (or the shape) of the end effector 20, an optical calibration may be applied.
Three different geometries were milled to illustrate the accuracy of the system, using a rigid mode versus an evasive mode. The evasive mode employs the position control techniques described herein whereas the rigid mode does not. For these tests, a burr with a radius of 2.5 mm was used to mill a cuboid (30 mm×30 mm×15 mm) and a half-sphere (radius=15 mm).
Referring now to
In another example, in
Whereas the rigid mode produced ragged/frayed edges and surfaces, the evasive mode produced clearer edges and smoother surfaces. The rigid mode causes cratered surfaces. The results show that the evasive mode is more accurate than the rigid mode. The results using the evasive mode could be even further improved with another mode that allows the end effector 20 to oscillate around its home position. This mode still evades from the virtual constraints (VC) but additionally moves around the position defined by the user. The milling speed of the device has a strong influence on the accuracy of the rigid mode. The higher the speed, the more accurate the results. Additionally, a semi-autonomous mode could further shorten the milling time. The user/surgeon moves the end effector 20 close to the milling volume and the tool starts to remove the bone material in its working volume.
As shown by the results, the presented control algorithms produce similar results as stationary robots and improved results over conventional manual interventions. The control algorithms work robustly and accurately on a stationary milling device.
Although specific features of various embodiments of the disclosure may be shown in some drawings and not in others, this is for convenience only. In accordance with the principles of the disclosure, any feature of a drawing or other embodiment may be referenced and/or claimed in combination with any feature of any other drawing or embodiment.
This written description uses examples to describe embodiments of the disclosure and also to enable any person skilled in the art to practice the embodiments, including making and using any devices or systems and performing any incorporated methods. The patentable scope of the disclosure is defined by the claims, and may include other examples that occur to those skilled in the art. Such other examples are intended to be within the scope of the claims if they have structural elements that do not differ from the literal language of the claims, or if they include equivalent structural elements with insubstantial differences from the literal language of the claims.
The subject application is a continuation application of U.S. Nonprovisional patent application Ser. No. 16/122,201, filed on Sep. 5, 2018, which claims priority to and all the benefits of U.S. provisional patent application Ser. No. 62/554,755, filed Sep. 6, 2017, the contents of which are hereby incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5408409 | Glassman et al. | Apr 1995 | A |
6602042 | Roy | Aug 2003 | B2 |
7035716 | Harris et al. | Apr 2006 | B2 |
7626569 | Lanier | Dec 2009 | B2 |
8961536 | Nikou et al. | Feb 2015 | B2 |
9008757 | Wu | Apr 2015 | B2 |
9060794 | Kang et al. | Jun 2015 | B2 |
9119638 | Schwarz et al. | Sep 2015 | B2 |
9119655 | Bowling et al. | Sep 2015 | B2 |
9561082 | Yen et al. | Feb 2017 | B2 |
9753542 | Chizeck | Sep 2017 | B2 |
10357315 | Otto | Jul 2019 | B2 |
10368878 | Lavallee et al. | Aug 2019 | B2 |
10441366 | Tabandeh | Oct 2019 | B2 |
10492870 | Shalayev et al. | Dec 2019 | B2 |
10828786 | Shoham | Nov 2020 | B2 |
20040024311 | Quaid | Feb 2004 | A1 |
20040128026 | Harris et al. | Jul 2004 | A1 |
20060142657 | Quaid et al. | Jun 2006 | A1 |
20110130761 | Plaskos et al. | Jun 2011 | A1 |
20120330429 | Axelson, Jr. et al. | Dec 2012 | A1 |
20130060278 | Bozung et al. | Mar 2013 | A1 |
20140058406 | Tsekos | Feb 2014 | A1 |
20140276949 | Staunton et al. | Sep 2014 | A1 |
20140320392 | Chizeck et al. | Oct 2014 | A1 |
20150320514 | Ahn et al. | Nov 2015 | A1 |
20160278754 | Todorov et al. | Sep 2016 | A1 |
20170024014 | Chizeck | Jan 2017 | A1 |
20170157361 | Barrish et al. | Jun 2017 | A1 |
20170245946 | Tabandeh | Aug 2017 | A1 |
20180023946 | Elliot | Jan 2018 | A1 |
20180333207 | Moctezuma De La Barrera | Nov 2018 | A1 |
20180353253 | Bowling | Dec 2018 | A1 |
20190070731 | Bowling et al. | Mar 2019 | A1 |
20190388099 | Zuhars et al. | Dec 2019 | A1 |
20220047329 | Lavallee | Feb 2022 | A1 |
Entry |
---|
Becker, B.C. et al., “Active Guidance of a Handheld Micromanipulator Using Visual Servoing”, IEEE International Conference on Robotics and Automation, vol. 2009, 2009, pp. 229-344, http://www.pubmedcentral.nih.gov/articlerender.fcgi? . . . . |
Brisson, G. et al., “Precision Freehand Sculpting of Bone”, Medical Image Computing and Computer-Assisted Intervention, 2004, pp. 105-112, http://www.springerlink.com/index/qfw0gd6Omyc9geq5.pdf . . . , 1 page. |
Gilbertson, M.W. et al., “Abstract of Force and Position Control System for Freehand Ultrasound”, IEEE Transactions on Robotics, vol. 31, No. 4, 2015, http://ieeexplore.ieee.org/Ipdocs/epic03/wrapper.htm?arnumber=7123185, 1 page. |
Klemm, M. et al., “Abstract of Flexible Three-Dimensional Camera-Based Reconstruction and Calibration of Tracked Instruments”, 19th International Conference on Information Fusion (FUSION), 2016, 1 page. |
Kneissler, M. et al., “Abstract of Concept and Clinical Evaluation of Navigated Control in Spine Surgery”, IEEE/ASME International Conference on Advanced Intelligent Mechatronis, AIM, vol. 2, 2003, 1 page. |
Lang, J.E. et al., “Robotic Systems in Orthopaedic Surgery”, Bone & Joint Journal, vol. 93-B, No. 10, 2011, pp. 1296-1922, http;//bjj.boneandjoint.org.uk/content/93-B/10/1296. |
Lonner, J.H. et al., “Abstract of Robotically Assisted Unicompartimental Knee Arthroplasty with a Handheld Image-Free Sculpting Tool”, 2016, 1 page. |
Pereira, R. et al., “Abstract of Hand-Held Robotic Device for Laparoscopic Surgery and Training”, IEEE, 3rd International Conference on Serious Games and Applications for Health, Books of Proceedings, 2014, 1 page. |
Shi, J. et al., “Abstract of Accuracy Study of a New Assistance System Under the Application of Navigated Control for Manual Milling on a Head Phantom”, Conference Proceedings, Annual International Conference of the IEEE Engineering in Medicine and Biology Society, vol. 2010, 2010, pp. 2296-2299; http://www.ncbi.nlm.nih.gov/pubmed/21097019, 1 page. |
U.S. Appl. No. 15/952,810, filed Apr. 13, 2018. |
Brisson, G. et al., “Precision Freehand Sculpting of Bone”, Medical Image Computing and Computer-Assisted Intervention, 2004, pp. 105-112, http://www.springerlink.com/index/qfw0gd60myc9geq5.pdf . . . , 1 page. |
Number | Date | Country | |
---|---|---|---|
20210260769 A1 | Aug 2021 | US |
Number | Date | Country | |
---|---|---|---|
62554755 | Sep 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16122201 | Sep 2018 | US |
Child | 17317140 | US |