Repositioning method of input device for robotic surgical system

Information

  • Patent Grant
  • 11925429
  • Patent Number
    11,925,429
  • Date Filed
    Tuesday, May 17, 2022
    a year ago
  • Date Issued
    Tuesday, March 12, 2024
    a month ago
Abstract
A robotic surgical system includes a linkage, an input handle, and a processing unit. The linkage moveably supports a surgical tool relative to a base. The input handle is moveable in a plurality of directions. The processing unit is in communication with the input handle and is operatively associated with the linkage to move the surgical tool based on a scaled movement of the input handle. The scaling varies depending on whether the input handle is moved towards a center of a workspace or away from the center of the workspace. The workspace represents a movement range of the input handle.
Description
BACKGROUND

Robotic surgical systems have been used in minimally invasive medical procedures. During a medical procedure, the robotic surgical system is controlled by a surgeon interfacing with a user interface. The user interface allows the surgeon to manipulate an end effector that acts on a patient. The user interface includes an input controller or handle that is moveable by the surgeon to control the robotic surgical system.


Robotic surgical systems typically used a scaling factor to scale down the motions of the surgeon's hands to determine the desired position of the end effector within the patient so that the surgeon could more precisely move the end effector inside the patient. However, the larger the scaling factor, the farther the surgeon had to move the input device handle to move the end effector the same distance. Since the input device handle has a fixed range of motion, this meant that for larger scaling factors the surgeon may have reached an end of the range of motion of an input handle more often. The surgeon then had to “clutch” the handle to decouple the motion of the input handles from the end effector so that the surgeon could move the handles to a new position within the workspace of the user interface away from the end of the range motion while the instruments remain stationary. Once the input handle was moved sufficiently away from the end of the range of motion, the surgeon “reclutched” the input handle with the end effector to recouple the motion of the input handle to motion of the end effector to complete the desired movement of the end effector. This clutching process is time consuming and distracting to surgeons.


There is a need for robotic surgical system that is able to scale down input handle movements of the surgeon while also reducing or eliminating the need for a surgeon to clutch or move the input handles away from an end of their range of motion during robotic surgical procedures.


SUMMARY

In an aspect of the present disclosure, a robotic surgical system includes a linkage, an input handle, and a processing unit. The linkage moveably supports a surgical tool relative to a base. The input handle is moveable in a plurality of directions. The processing unit is in communication with the input handle and is operatively associated with the linkage to move the surgical tool based on a scaled movement of the input handle. The scaling varies depending on whether the input handle is moved towards a center of a workspace or away from the center of the workspace. The workspace represents a movement range of the input handle.


In aspects, the processing unit is configured to scale a first movement of the input handle towards the center of the workspace by a first scaling factor and to scale a second movement of the input handle away from the center of the workspace by a second scaling factor different from the first scaling factor.


In some aspects, the processing unit is configured to scale the input distance by dividing a distance of the first movement by the first scaling factor and a distance of the second movement by the second scaling factor. The first and second scaling factors may be in a range of about 1.0 to about 10.0 and the second scaling factor may be larger than the first scaling factor. The first scaling factor may be between 0.70 and 1.40 times as large as the second scaling factor.


In certain aspects, the processing unit is configured to vary the scaling based on a distance of the input handle from the center of the workspace. The processing unit may be configured to linearly or exponentially scale the input distance based on a distance of the input handle from at least one of the center of the workspace and a limit of movement of the input handle. The processing unit may be configured to increase a movement of the surgical tool as the location of the input handle is further from the center of the workspace. The processing unit may be configured to decrease a movement of the surgical tool as the location of the input handle is further from the center of the workspace.


In particular aspects, the workspace includes a first section that is located a predetermined distance from the center of the workspace. The first and second scaling factors may be constant when the input handle is in the first section and at least one of the first or second scaling factors varies when the input handle is outside of the first section.


In another aspect of the present disclosure, a robotic surgical system includes a linkage, an input handle, and a processing unit. The linkage moveably supports a surgical tool relative to a base. The input handle is moveable a first input distance in a first input direction and a second input distance in a second input direction that is opposite the first input direction. The second input distance is different from the first input distance. The processing unit is in communication with the input handle and is operatively associated with the linkage to move the surgical tool. The processing unit is configured to move the surgical tool an output distance in a first output direction in response to the first input distance and to move the surgical tool the same output distance in a second out direction opposite the first input direction in response to the second input distance.


In another aspect of the present disclosure, a method of operating a surgical robot includes detecting a plurality of movements of an input handle that is moveable in a plurality of directions, scaling a detected movement of the input handle towards a center of a workspace by a first scaling factor, scaling a detected movement of the input handle away from the center of the workspace by a second scaling factor different from the first scaling factor, and actuating a linkage based on the scaled detected movements to move a surgical tool that is moveably supported by the linkage.


In aspects, the method includes dividing a distance of the detected movement of the input handle toward the center of the workspace by the first scaling factor and dividing a distance of the detected movement of the input handle away from the center of the workspace by the second scaling factor.


In some aspects, the method includes varying the scaling of the detected movements based on a distance of the input handle from the center of the workspace. The method may include linearly or exponentially varying the scaling of the detected movements based on a distance of the input handle from at least one of the center of the workspace and a limit of movement of the input handle.


In certain aspects, the method may include adjusting the actuation of the linkage to increase or decrease a movement of the surgical tool as the location of the input handle is further from the center of the workspace. The method may include setting the first and second scaling factors as constant when the input handle is in a section of the workspace located within a predetermined distance of the center of the workspace and varying at least one of the first and second scaling factors when the input handle is outside the section.


In another aspect of the present disclosure, a robotic surgical system includes an arm having an end, a tool supported on the end of the arm, an input handle, and a processing unit. The input handle is moveable an input distance and includes a repositioning control that has activated and deactivated states. The processing unit is in communication with the input handle and is operatively associated with the arm to move the tool. The processing unit is configured to scale the input distance of the input handle by a first scaling factor when the repositioning control is in the activated state and to sale the input distance by a second scaling factor when the repositioning control is in the deactivated state. The first scaling factor is larger than the second scaling factor.


In aspects, the first scaling factor may be in a range of about 100 to about 1000 and the second scaling factor may be in a range of about 1 to about 10. The processing unit may be configured to scale the input distance by dividing the input distance by the first scaling factor when the repositioning control is in the activated state and by dividing the input distance by the second scaling factor when the reposition control is in the deactivated state.


In another aspect of the present disclosure, a method of repositioning an input handle of a robotic surgical system includes detecting movement of an input handle of the robotic surgical system an input distance with a repositioning control of the input handle in a deactivated state, detecting whether a repositioning control of the input handle is in an activated or deactivated state, scaling the input distance to a first output distance or a different second output distance depending on the state of the repositioning control, and actuating movement of a tool supported on an end of a moveable arm of a robotic surgical system the respective scaled first or second output distance.


In aspects, the method includes dividing the input distance by a first scaling factor in a range of 1 to 10 as part of the scaling of the input distance to the first output distance and dividing the input distance by a second scaling factor in a range of 100 to 1000 as part of scaling the input distance to the second output distance. The method may include changing an orientation of the tool as part of the actuating movement when the detected movement of the input handle includes a detected orientation change of the input handle.


In another aspect of the present disclosure, a robotic surgical system including an arm having an end, a tool supported on the end of the arm, an input handle moveable an input distance, an imaging device, and a processing unit. The imaging device is configured to capture images of the tool within a surgical site of a patient. The processing unit is in communication with the input handle. The processing unit is also operatively associated with the arm to move the tool and is in communication with the imaging arm. The processing unit is configured to scale the input distance of the input handle to an output distance based on the position of the imaging device relative to the surgical site of the patient.


In aspects, the system includes an imaging arm with the imaging device positioned at the end of the imaging arm. The input handle may be moveable within a workspace having a center and a limit of movement. The input distance may be defined by the distance between the center and the limit of movement. The imaging device may be configured to zoom out the captured images as the input handle approaches the limit of movement. The processing unit may be operatively associated with the imaging arm and may be configured to move the imaging device away from the surgical site to zoom out the captured images.


In some aspects, the system includes a display that is configured to display the captured images from the imaging device. The processing unit may be configured to scale the input distance to the output distance such that the movement of the tool on the display is substantially equal to the input distance.


In certain aspects, the system includes a switch that is configured to selectively switch the association of the input handle between the tool and the imaging arm. The switch may be a foot switch or disposed on the input handle.


In another aspect of the present disclosure, a method of scaling movement of a tool supported on an end of an arm based on the position of an imaging device relative to a surgical site includes determining the position of the imaging device relative to the surgical site with a processing unit and moving an input handle of a robotic surgical system an input distance to move the tool on output distance. The tool is operatively associated to the input handle by the processing unit. The processing unit scales the input distance to the output distance based on the position of the imaging device relative to the surgical site.


In aspects, moving the input handle includes moving the input handle towards a limit of movement of the input handle. The method may include the processing unit moving the imaging device relative to the surgical site as the input handle approaches the limit of movement. The method may include the processing unit varying a scaling factor as the imaging device is moved relative to the surgical site. The processing unit may scale the input distance to the output distance by the scaling factor.


In another aspect of the present disclosure, a robotic surgical system includes an arm having an end, a tool supported on the end of the arm, an input handle, and a processing unit. The input handle is moveable an input distance in a first input direction and a second input direction opposite the first input direction. The processing unit is in communication with the input handle and is operatively associated with the arm to move the tool. The processing unit is configured to scale the input distance of the input handle to a first output distance of the tool in response to movement of the input handle in the first input direction and to a second output distance of the tool in response to movement of the input handle in the second input direction, the first output distance being different than the second output distance.


In aspects, the processing unit is configured to scale the input distance of the input handle by a first scaling factor in response to movement of the input handle in the first direction and to scale the input distance of the input handle by a second scaling factor in response to movement of the input handle in the second direction. The second scaling factor may be different than the first scaling factor. The processing unit may be configured to scale the input distance by dividing the input distance by one of the first or second scaling factors. The first scaling factor may be in a range of about 1 to about 10 and the second scaling factor may be in a range of about 1 to about 10.


In some aspects, the input handle is moveable within a workspace having a center and a limit of movement. The movement of the input handle in the first direction may be towards the limit of movement and the movement of the input handle in the second direction may be towards the center. The processing unit may be configured to further scale the input distance in response to a location of the input handle between the center and the limit of movement. The processing unit may be configured to linearly or exponentially scale the input distance in response to the location of the input handle between the center and the limit of movement. The processing unit may be configured to increase the first output distance as the location of the input handle is further from the center. The processing unit may be configured to decrease the second output distance as the location of the input handle is further from the center.


In certain aspects, the input handle is configured to scale the input distance of the input handle by a first scaling factor in response to movement of the input handle in the first direction and to scale the input distance of the input handle by a second scaling factor in response to movement of the input handle in the second direction. The second scaling factor may be different than the first scaling factor. The work space may include a first point between the center and the limit of movement with the first and second scaling factor being constant when the input handle is between the center and the first point. One of the first or second scaling factors may vary when the input handle is between the first point and the limit of movement based on the location of the handle.


In another aspect of the present disclosure, a robotic surgical system includes an arm having an end, a tool supported on the end of the arm, and input handle, and a processing unit. The input handle being moveable a first input distance in a first input direction and a second input distance in a second input direction opposite the first input direction, the second input distance being different than the first input distance. The processing unit is in communication with the input handle and operatively associated with the arm to move the tool. The processing unit is configured to move the tool an output distance in a first output direction in response to the first input distance and to move the tool the output distance in a second output direction opposite the first input direction in response to the second input distance.


In yet another aspect of the present disclosure, a method of operation a surgical robot includes moving an input handle of a robotic surgical system an input distance in a first input direction to move a tool supported on an end of an arm of the robotic surgical system a first output distance in response to the movement of the input handle in the first input direction and moving the input handle of the robotic surgical system, the input distance, in a second input direction opposite the first input direction to move the tool a second out distance in response to the movement of the input handle in the second input direction, the second output distance being different than the first output distance.


In aspects, moving the input handle of the robotic surgical system in the first input direction includes moving the input handle towards a limit of movement of a workspace. Moving the input handle of the robotic surgical system in the second input direction may include moving the input handle towards a center of a workspace.


In some aspects, moving the input handle of the robotic surgical system in the first direction includes transmitting an input signal to a processing unit of the robotic surgical system including a location of the input handle relative to a center of a workspace. The method may include the processing unit controlling the first output distance based on the input distance, the first input direction, and the location of the input handle relative to the center of the workspace.


In certain aspects, moving the input handle of the robotic surgical system in the second input direction includes transmitting an input signal to a processing unit of the robotic surgical system including a location of the input handle relative to a center of a workspace. The method may include the processing unit controlling the second output distance based on the input distance, the second input direction, and the location of the input handle relative to the center of the workspace.


In still another aspect of the present disclosure, a method of operating a surgical robot includes moving an input handle of a robotic surgical system a first input distance away from a center of a workspace to move a tool supported on an end of an arm of a surgical robot an output distance in a first output direction and moving the input handle of the robotic surgical system a second input distance towards the center of the workspace to move the tool, the output distance, in a second output direction opposite the first output direction, the second input distance being different than the first input distance.


In other aspects of the present disclosure, the robotic surgical system includes an arm having an end, a tool supported on the end of the arm, an input handle, and a processing unit. The input handle movable an input distance and including a repositioning control having an activated state and a deactivated state. The processing unit may be in communication with the input handle and operatively associated with the arm to move the tool. The processing unit is configured to scale the input distance of the input handle by a first scaling factor when the repositioning control is in the activated state and to scale the input distance by a second scaling factor when the repositioning control is in the deactivated state. The first scaling factor is larger than the second scaling factor.


In aspects, the first scaling factor is in a range of about 100 to about 1000 and the second scaling factor is in a range of about 1 to about 10. The processing unit may be configured to scale the input distance by dividing the input distance by the first scaling factor when the repositioning control in in the activated state and by dividing the input distance by the second scaling factor when the reposition control is in the deactivated state.


In another aspect of the present disclosure, a method of repositioning an input handle of a robotic surgical system includes moving an input handle of a robotic surgical system an input distance with a repositioning control of the input handle in a deactivated state to move a tool supported on an end of an arm of the robotic surgical system a first output distance, and moving the input handle of the robotic surgical system, the input distance, with the repositioning control of the input handle in an activated state to move the tool a second output distance. The input handle being in communication with a processing unit to operatively associate the tool of the robotic surgical system with the input handle. The processing unit scaling the input distance to a first output distance when the input handle is in the deactivated state and scaling the input distance to a second output distance when the input handle is in the activated state.


In aspects, scaling the input distance to the first output distance includes dividing the input distance by a first scaling factor in a range of about 1 to about 10 and scaling the input distance to the second output distance includes dividing the input distance by a second scaling factor in a range of about 100 to about 1000. Moving the input handle the input distance with the repositioning control in the activated state may include an orientation of the input handle operatively associated with an orientation of the tool.


Further details and aspects of exemplary embodiments of the present disclosure are described in more detail below with reference to the appended figures.





BRIEF DESCRIPTION OF THE DRAWINGS

Various aspects of the present disclosure are described hereinbelow with reference to the drawings, which are incorporated in and constitute a part of this specification, wherein:



FIG. 1 is a schematic illustration of a user interface and a robotic system in accordance with the present disclosure;



FIG. 2 is a plan view of an arm of the user interface of FIG. 1 within a two-dimensional workspace;



FIG. 3 is a plan view of a workspace of the user interface of FIG. 1; and



FIG. 4 is a schematic of a method for scaling movement of a user interface in accordance with the present disclosure.



FIG. 5 is a flow chart of a method for scaling movement of a user interface in accordance with the present disclosure.





DETAILED DESCRIPTION

Embodiments of the present disclosure are now described in detail with reference to the drawings in which like reference numerals designate identical or corresponding elements in each of the several views. As used herein, the term “clinician” refers to a doctor, a nurse, or any other care provider and may include support personnel. Throughout this description, the term “proximal” refers to the portion of the device or component thereof that is closest to the clinician and the term “distal” refers to the portion of the device or component thereof that is farthest from the clinician.


Referring to FIG. 1, a robotic surgical system 1 in accordance with the present disclosure is shown generally as a robotic system 10, a processing unit 30, and a user interface 40. The robotic system 10 generally includes linkages 12 and a robot base 18. The linkages 12 moveably support an end effector or tool 20 which is configured to act on tissue. The linkages 12 may be in the form of arms each having an end 14 that supports an end effector or tool 20 which is configured to act on tissue. In addition, the ends 14 of the arms 12 may include an imaging device 16 for imaging a surgical site “S”. The user interface 40 is in communication with robot base 18 through the processing unit 30.


The user interface 40 includes a display device 44 which is configured to display three-dimensional images. The display device 44 displays three-dimensional images of the surgical site “S” which may include data captured by imaging devices 16 positioned on the ends 14 of the arms 12 and/or include data captured by imaging devices that are positioned about the surgical theater (e.g., an imaging device positioned within the surgical site “S”, an imaging device positioned adjacent the patient “P”, imaging device 56 positioned at a distal end of an imaging arm 52). The imaging devices (e.g., imaging devices 16, 56) may capture visual images, infra-red images, ultrasound images, X-ray images, thermal images, and/or any other known real-time images of the surgical site “S”. The imaging devices transmit captured imaging data to the processing unit 30 which creates three-dimensional images of the surgical site “S” in real-time from the imaging data and transmits the three-dimensional images to the display device 44 for display.


The user interface 40 also includes input handles 42 which allow a clinician to manipulate the robotic system 10 (e.g., move the arms 12, the ends 14 of the arms 12, and/or the tools 20). Each of the input handles 42 is in communication with the processing unit 30 to transmit control signals thereto and to receive feedback signals therefrom. Additionally or alternatively, each of the input handles 42 may include control interfaces (not shown) which allow the surgeon to manipulate (e.g., clamp, grasp, fire, open, close, rotate, thrust, slice, etc.) the tools 20 supported at the ends 14 of the arms 12.


With additional reference to FIG. 2, each of the input handles 42 is moveable through a predefined workspace “W” to move the ends 14 of the arms 12 within a surgical site “S”. It will be appreciated that while the workspace “W” is shown in two-dimensions in FIG. 2 that the workspace “W” is a three-dimensional workspace. The three-dimensional images on the display device 44 are orientated such that the movement of the input handle 42 moves the ends 14 of the arms 12 as viewed on the display device 44. It will be appreciated that the orientation of the three-dimensional images on the display device may be mirrored or rotated relative to view from above the patient “P”. In addition, it will be appreciated that the size of the three-dimensional images on the display device 44 may be scaled to be larger or smaller than the actual structures of the surgical site permitting the surgeon to have a better view of structures within the surgical site “S”. As the input handles 42 are moved, the tools 20 are moved within the surgical site “S” as detailed below. As detailed herein, movement of the tools 20 may also include movement of the ends 14 of the arms 12 which support the tools 20.


For a detailed discussion of the construction and operation of a robotic surgical system 1, reference may be made to U.S. Patent Publication No. 2012/0116416, entitled “Medical Workstation”, the entire contents of which are incorporated herein by reference.


The movement of the tools 20 is scaled relative to the movement of the input handles 42. When the input handles 42 are moved within a predefined workspace “W”, the input handles 42 send control signals to the processing unit 30. The processing unit 30 analyzes the control signals to move the tools 20 in response to the control signals. The processing unit 30 transmits scaled control signals to the robot base 18 to move the tools 20 in response the movement of the input handles 42. The processing unit 30 scales the control signals by dividing an Inputdistance (e.g., the distance moved by one of the input handles 42) by a scaling factor SF to arrive at a scaled Outputdistance (e.g., the distance that one of the ends 14 is moved). The scaling factor SF is in a range between about 1 and about 10 (e.g., 3). This scaling is represented by the following equation:

Outputdistance=Inputdistance/SF.

It will be appreciated that the larger the scaling factor Sf. the smaller the movement of the tools 20 relative to the movement of the input handles 42.


During a surgical procedure, if the clinician reaches the edge or limit of the predefined range of motion of an input handle 42, the clinician must clutch the input handle 42 (i.e., decouple the motion of the input handle 42 from the motion of the tool 20 of the respective arm 12) to reposition the input handle 42 back within the predefined workspace “W” before continuing to move the input handle 42 in the same direction. As the scaling factor SF is increased, the clinician may be required to clutch the input handle 42 more frequently, which increases the number of steps and thus, the time and/or costs of the surgical procedure.


In addition, when the input handle 42 is clutched from the tool 20, the orientation (e.g., roll, pitch, and yaw) of the tool 20 is also decoupled from the orientation of the input handle 42. When the input handle 42 is declutched or decoupled, the processing unit 30 may be programmed to align the orientation of the tool 20 with the orientation of the input handle 42, which may result in unintended movement of the tool 20 when the input handle 42 is recoupled. Alternatively, when the input handle 42 is reclutched or recoupled, the processing unit 30 may recalibrate the orientation of the input handle 42 when it is recoupled to the current orientation of the processing unit 30, which may result in the orientation of the input handle 42 being misrepresented by the tool 20.


To reduce or eliminate the need for a clinician to clutch the input handles 42 during a surgical procedure, each of the input handles 42 may include a repositioning control 43 that sends a signal to the processing unit 30 to switch the scaling factor SF between a procedural scaling factor PSF and a repositioning scaling factor RSF. The procedural scaling factor PSF is in a range of about 1.0 to about 10.0 (e.g., 3.0) and the repositioning scaling factor RSF is significantly larger in a range of about 100.0 to about 1000.0 (e.g., 500.0). The two scaling factors allow a clinician to perform a surgical procedure using the procedural scaling factor PSF and when one of the input handles 42 approaches an edge or a limit of movement of the predefined workspace “W”, the clinician activates the repositioning control 43 to change to the repositioning scaling factor RSF to move the input handle 42 to a desired position within the predefined workspace “W” without clutching the input handle 42. Once the input handle 42 is at the desired position within the predefined workspace “W”, the clinician deactivates the repositioning control 43 to switch back to the procedural scaling factor PSF to continue the surgical procedure. It will be appreciated that by activating and deactivating the repositioning control 43, to reposition the input handle 42 within the predefined workspace “W”, the orientational relationship between the input handle 42 and the end 14 of the arm 12 is maintained. It is contemplated a repositioning control 43 on each input handle 42 is activatable independent of a repositioning control 43 on another input handle 42. While the repositioning control 43 is represented as a button, it is contemplated that the repositioning control 43 may be operated by, but not limited to, a switch, a lever, a trigger, an optical sensor, or a voice command.


Additionally or alternatively, the processing unit 30 may vary the scaling factor SF based on the direction of movement of the input handle 42 within the predefined workspace “W” to keep the input handle 42 substantially centered within the predefined workspace “W”. As detailed below with reference to FIGS. 2-4, a method 300 of varying the scaling factor SF based on the direction of movement of the input handle 42 is detailed with respect to the “X” axis; however, it will be appreciated that this method 300 may be applied to each of the “X”, “Y”, and “Z” axes of the predefined workspace “W”. When the input handle 42 is moved away from a center “C” (Step 330), of the predefined workspace “W”, towards a limit of movement of the predefined workspace “W”, the processing unit 30 assigns a first scaling factor SF1 to the movement of the input handle 42, records the direction of movement represented by arrow “A”, and identifies the limit of movement represented by the border of the workspace “W”. When the input handle 42 is moved towards the center “C” of the predefined workspace “W” (Step 335), the processing unit 30 assigns a second scaling factor SF2 to the movement of the input handle 42 that is larger than the first scaling factor SF1, and records the direction of movement represented by arrow “B”. For example, the first scaling factor SF1 may be about 3.0 and the second scaling factor SF2 may be about 4.5. It is contemplated that first scaling factor SF1 may be about 0.70 to about 1.4 times the size of the second scaling factor SF2. Varying the scaling factor SF in this manner keeps the input handle 42 substantially centered by requiring the clinician to move the input handle 42 a greater distance when moving the input handle 42 towards the center “C” of the predefined workspace “W” as compared to the distance that the clinician moved the input handle 42 away from the center “C” to move the tool 20 the same distance in each direction.


Accordingly, the center “C” of the predefined workspace “W” continually shifts relative to the surgical site “S” (FIG. 1) during the surgical procedure. It is contemplated that the shifting of the center “C” of the predefined workspace “W” relative to the surgical site “S” may be imperceptible to the clinician. With particular reference to FIG. 3, it will be appreciated that as detailed above, the direction of arrow “A” is representative of the direction of movement of the input handle 42 away from the center “C” towards a point “F”; however, when the input handle 42 is moved from the center “C” towards a point “H”, the arrow “B” is representative of movement away from the center “C”. In embodiments, the first and second scaling factors SF1, SF2 are the same in each axis (e.g., the “X”, “Y”, and “Z” axes). In some embodiments, the first and second scaling factor SF1, SF2 in one axis (e.g., the “X” axis) is different from the first and second scaling factors SF1, SF2 in the other axes (e.g., the “Y” and “Z” axes).


With continued reference to FIGS. 2-4, the method may include the processing unit 30 varying the first and second scaling factors SF1, SF2 based on the location of the input handle 42 within the predefined workspace “W” (Steps 350, 355). As detailed below, the method of varying a scaling factor SF based on the location of the input handle 42 is detailed with respect to the “X” axis; however, it will be appreciated that this method may be applied to each of the “X”, “Y”, and “Z” axes of the predefined workspace “W”. Specifically, the scaling factors SF1, SF2 are implemented by differentiating the movement of the input handle 42 to calculate a handle velocity. This handle velocity is multiplied by the scaling factor (e.g., scaling factors SF1, SF2). This scaled velocity is then integrated with the current position of the tool 20 to determine a new position of the tool 20 such that movement of the tool 20 is smooth. The scaled velocity is continuously integrated such that the movement of the tool 20 is smooth even if the scaling factor is varied (e.g., when the scaling factor is smoothly or discreetly varied as detailed below).


In accordance with this method and represented as Steps 345 and 355, the second scaling factor SF2 may increase as the location of the input handle 42 moves away from the center “C” of the predefined workspace “W” (e.g., the second scaling factor SF2 may be larger when the input handle 42 is at point “D” than when the input handle 42 is at center “C”, the second scaling factor SF2 may be larger when the input handle 42 is at point “E” than when the input handle 42 is at point “D”, and the second scaling factor SF2 may be larger when the input handle 42 is at point “F” than when the input handle 42 is at point “E”). For example, the second scaling factor SF2 may vary in a linear manner based on the location of the input handle 42 from the center “C” such that at point “F” the second scaling factor SF2 is about 4.5, at point “E” the second scaling factor SF2 is about 4.0, at point “D” the second scaling factor SF2 is about 3.5, and at center “C” the second scaling factor SF2 is about 3.0.


Alternatively, the second scaling factor SF2 may vary as a function of the location of the input handle 42 from the center “C”. For example, the second scaling factor SF2 may vary in an exponential manner based on the location of the input handle 42 from the center “C” such that at point “F” the second scaling factor SF2 is about 6.5, at point “E” the second scaling factor SF2 is about 4.5, at point “D” the second scaling factor SF2 is about 3.5, and at the center “C” the second scaling factor SF2 is about 3.0.


In addition, it is contemplated that the second scaling factor SF2 may be constant as the input handle 42 is within a first section Si close to the center “C” and linearly or exponentially increase as the location of the input handle 42 is moved away from the center “C” beyond the first section Si.


It is within the scope of this disclosure that the second scaling factor SF2 may increase, in a relatively smooth manner, as the location of the input handle 42 away from the center “C” based on a linear or exponential formula or that the second scaling factor SF2 may change discretely at each of a plurality of points (e.g., points “D”, “E”, and “F”) creating discontinuities in the second scaling factor SF2.


As detailed above, increasing the second scaling factor SF2 requires the clinician to move the input handle 42 a greater distance towards the center “C” when compared to movement of the input handle 42 away from the center “C” to move the tool 20 an equal distance in each direction as the location of the input handle 42 away from the center “C” increases, which shifts the center “C” relative to the surgical site “S” to reduce or eliminate the need to clutch the input handle 42.


In addition as represented in steps 340 and 350, the first scaling factor SF1 may decrease as the location of the input handle 42 moves away from the center “C” of the predefined workspace “W” (e.g., the first scaling factor SF1 may be smaller when the input handle 42 is at point “D” than when the input handle 42 is at the center “C”, the first scaling factor SF1 may be smaller when the input handle 42 is at point “E” than when the input handle 42 is at point “D”, and the first scaling factor SF1 may be smaller when the input handle 42 is at point “F” than when the input handle 42 is at point “E”). For example, the first scaling factor SF1 may vary in a linear manner based on the location of the input handle 42 from the center “C” such that at the center “C” the first scaling factor SF1 is about 3.0, at point “D” the first scaling factor SF1 is about 2.75, at point “E” the first scaling factor SF1 is about 2.5, and at point “F” the first scaling factor SF1 is about 2.25.


Alternatively, the first scaling factor SF1 may vary as a function of the location of the input handle 42 from the center “C”. For example, the first scaling factor SF1 may vary in an exponential manner based on the location of the input handle 42 from the center “C” such that at the center “C” the first scaling factor SF1 is about 3.0, at point “D” the first scaling factor SF1 is about 2.75, at point “E” the first scaling factor SF1 is about 2.25, and at point “F” the first scaling factor SF1 is about 1.25.


In addition, it is contemplated that the first scaling factor SF1 may be constant as the input handle 42 is at a location near the center “C” (e.g., when the input handle is between point “D” and a point “G” the first scaling factor SF1 is constant) and linearly or exponentially decreased as the location of the input handle 42 is moved beyond point “D” or point “G”. It is within the scope of this disclosure that the first scaling factor SF1 may increase, in a relatively smooth manner, as the location of the input handle 42 moves away from the center “C” based on a linear or exponential formula or that the first scaling factor SF1 may change discretely at each of a plurality of points (e.g., points “D”, “E”, and “F”) creating discontinuities in the first scaling factor SF1.


As detailed above, increasing the first scaling factor SF1 allows the clinician to move the input handle 42 a lesser distance away from the center “C” as the location of the input handle 42 away from the center “C” is increased to result in the same movement of the tool 20, which in turn shifts the center “C” relative to the surgical site “S” to reduce or eliminate the need to clutch the input handle 42.


It is contemplated that each input handle 42 may vary the respective scaling factors SF1, SF2 in a similar manner or may vary the respective scaling factors SF1, SF2 in differing manners (e.g., one input handle 42 may vary its scaling factors SF1, SF2 based on the location of the input handle 42 and another input handle 42 may vary its scaling factors SF1, SF2 based on the direction of movement of the another input handle relative to the center “C”, each input handle 42 may vary its scaling factors SF1, SF2 using different linear or exponential formulas). It is also contemplated that an input handle 42 may vary one of the scaling factors SF1, SF2 and the other of the scaling factors SF1, SF2 may be constant. While points “F” to “J” are shown spaced evenly apart, it is contemplated that points “F” to “J” may be spaced apart different distances from one another. For example, be spaced closer to one another as the points get closer to the limit of movement.


Referring back to FIG. 1, the robotic surgical system 1 includes an imaging arm 52 that is controlled by the processing unit 30 and may be selectively controlled by the user interface 40. The imaging arm 52 includes an imaging device 56 disposed on a distal end thereof. The imaging device 56 is located over or within the surgical site “S” and is configured to capture the tools 20 acting within the surgical site “S” of the patient “P” and transmit the captured images to the display 44. The imaging device 56 may be a three-dimensional (3D) camera and the display 44 may be a 3D display enabling the clinician to view the surgical site “S” in three dimensions. The imaging device 56 is moveable relative to or within the surgical site “S” in six degrees of freedom and may be selectively moved by one of the input handles 42. The surgical system 1 may include a foot switch (not shown) that switches one of the input handles 42 between a run mode where the input handle 42 is associated with one of the arms 12 to move the tool 20 within the surgical site “S” and a camera mode where the input handle 42 is associated with the imaging arm 52 to move the imaging device 56 about the surgical site “S”. It is also contemplated that one of the input handles 42 may include a camera button (not shown) to operatively associate the input handle 42 with the imaging arm 52.


With reference to the method 400 shown in FIG. 5, the processing unit 30 may determine the location of the imaging device 56 relative to or within the surgical site “S”, at step 458, to determine the scaling factor SF, at step 459, used to associate the movement of the input handles 42 detected, at step 410, to the movement of the tools 20 within the surgical site “S”, at step 460. As detailed herein, the processing unit 30 determines the location of the imaging device 56 relative to the surgical site “S” along the “Z” axis to determine the scaling factor SF; however, the processing unit 30 may determine the scaling factor SF based on the location of the imaging device 56 relative to or within the surgical site “S” in each of the “X”, “Y”, and “Z” axes.


With continued reference to FIG. 1, when the imaging device 56 is in a first position “L” the scaling factor SF has a first value and when the imaging device 56 is in a second position “M” the scaling factor SF has a second value larger than the first value. For example, the first value may be about “1” when the imaging device 56 is in the first position “L” and as the imaging device 56 is moved closer to the surgical site “S” to the second position “M”, the second value may be about “2”, at step 459 of FIG. 5. Continued movement of the imaging device to a third location “N” may further increase the scaling factor SF to a third value (e.g., about 3), at step 459 of FIG. 5. It will be appreciated that by increasing the scaling factor SF as the imaging device 56 is moved closer to the surgical site “S”, the movement of the tools 20 within the surgical site “S” relative to the movement of the input handles 42 within the predefined workspace “W” is decreased such that the movement of the tools 20 as shown in a field of view on the display 44 is relatively constant for a movement of the input handles 42 within the predefined workspace “W”.


The processing unit 30 may be operatively associated with the imaging arm 52 such that as the scaling factor SF is increased or decreased the processing unit 30 zooms the imaging device 56 in and out from the surgical site “S” to match the movement of the input handles 42 within the predefined workspace “W” to the movement of the tools 20 within the surgical site “S” as viewed by the clinician on the display 44. The zooming in and out of the imaging device 56 may be accomplished by manipulating a lens assembly (not explicitly shown) of the imaging device 56 or by moving the imaging device 56 towards and away from the surgical site “S”. The processing unit 30 may zoom the imaging device 56 out when one of the input handles 42 approaches a limit or edge of the predefined workspace “W” to keep the tools 20 within the field of view of the imaging device 56. In addition, the processing unit 30 may reposition the imaging device 56 such that the center “C” (FIG. 2) of the input handles 42 are substantially centered on the display device 44.


It is contemplated that any of the methods of varying the scaling factor SF or moving of the imaging device 56 may be selectively activated or deactivated by a clinician operating the robotic surgical system 1 before or during a surgical procedure.


It will be appreciated that the scaling factor SF determined by the processing unit 30 based on the position of the imaging device 56 relative to the surgical site “S” may be varied as detailed above based on the movement of or location of the input handles 42 relative to the center “C” of the predefined workspace “W”.


While several embodiments of the disclosure have been shown in the drawings, it is not intended that the disclosure be limited thereto, as it is intended that the disclosure be as broad in scope as the art will allow and that the specification be read likewise. Any combination of the above embodiments is also envisioned and is within the scope of the appended claims. Therefore, the above description should not be construed as limiting, but merely as exemplifications of particular embodiments. Those skilled in the art will envision other modifications within the scope of the claims appended hereto.

Claims
  • 1. A method of operating a robotic surgical system, the method comprising: detecting a plurality of movements of an input handle of the robotic surgical system;determining a location of an imaging device, which is supported by a second linkage of the robotic surgical system, relative to a workspace, which relative location defines a field of view, wherein the field of view represents a movement range of a surgical tool of the robotic surgical system;varying a scaling factor for the movement range of the surgical tool within the field of view based on the location of the imaging device, wherein the scaling factor is one of linearly varied or exponentially varied based on the location and a limit of movement of the input handle; andscaling a detected movement of the input handle based on the scaling factor.
  • 2. The method of claim 1, further comprising capturing a three-dimensional image using the imaging device.
  • 3. The method of claim 1, further comprising: displaying the workspace in three-dimensions.
  • 4. The method of claim 1, wherein varying a scaling factor based on the location of the imaging device includes varying the scaling factor based on an anti-gravity direction with respect to the workspace.
  • 5. The method of claim 1, further comprising: switching between a run mode and a camera mode based on a user's selection.
  • 6. The method of claim 5, further comprising: moving the imaging device based on the scaled movement of the input handle in the camera mode.
  • 7. The method of claim 5, further comprising: moving the surgical tool based on the scaled movement of the input handle in the run mode.
  • 8. A robotic surgical system comprising: a first linkage moveably supporting a surgical tool relative to a base and an input handle moveable in a plurality of directions;a second linkage moveably supporting an imaging device relative to the base, the imaging device being configured to capture the surgical tool; anda processing unit in communication with the input handle operatively associated with the first linkage to move the surgical tool based on a scaled movement of the input handle, and, with the imaging device operatively associated with the second linkage, to move the imaging device,wherein the processing unit is configured to: determine a location of the imaging device relative to a workspace, which relative location defines a field of view, wherein the field of view represents a movement range of the surgical tool;vary a scaling factor for the movement range of the surgical tool within the field of view based on the location of the imaging device, wherein the scaling factor is one of linearly varied or exponentially varied based on the location and a limit of movement of the input handle; andscale a movement of the input handle based on the scaling factor.
  • 9. The robotic surgical system of claim 8, wherein the imaging device is a three-dimensional camera.
  • 10. The robotic surgical system of claim 9, further comprising: a display configured to display the workspace in three-dimensions.
  • 11. The robotic surgical system of claim 8, wherein the imaging device moves in six degrees of freedom.
  • 12. The robotic surgical system of claim 8, wherein the location is based on an anti-gravity direction with respect to the workspace.
  • 13. The robotic surgical system of claim 8, further comprising: a switch operably connected to the processing unit and configured to switch between a run mode and a camera mode.
  • 14. The robotic surgical system of claim 13, wherein the surgical tool is moved by the input handle in the run mode.
  • 15. The robotic surgical system of claim 13, wherein the imaging device is moved by the input handle in the camera mode.
  • 16. A robotic surgical system comprising: a first linkage moveably supporting a surgical tool relative to a base and an input handle moveable in a plurality of directions;a second linkage moveably supporting an imaging device relative to the base, the imaging device being configured to capture the surgical tool; anda processing unit in communication with the input handle operatively associated with the first linkage to move the surgical tool based on a scaled movement of the input handle, and with the imaging device operatively associated with the second linkage to move the imaging device,wherein the processing unit is configured to: determine a location of the imaging device relative to a workspace, which represents a movement range of the surgical tool;vary a scaling factor based on the location; andzoom the imaging device in or out from the workspace as the scaling factor is increased decreased.
CROSS-REFERENCE TO RELATED APPLICATION

This application is a Continuation application claiming the benefit of and priority to U.S. patent application Ser. No. 16/546,554, filed on Aug. 21, 2019 (now U.S. Pat. No. 11,331,158), which is a Continuation application claiming the benefit of and priority to U.S. patent application Ser. No. 15/548,439, filed on Aug. 3, 2017 (now U.S. Pat. No. 10,695,142), which is a U.S. National Stage Application, filed under 35 U.S.C. § 371(a), of International Patent Application Serial No. PCT/US2016/014031, filed on Jan. 20, 2016, which claims the benefit of and priority to U.S. Provisional Patent Application Ser. No. 62/118,123, filed on Feb. 19, 2015, the entire disclosure of each of which being incorporated by reference herein.

US Referenced Citations (377)
Number Name Date Kind
6132368 Cooper Oct 2000 A
6206903 Ramans Mar 2001 B1
6246200 Blumenkranz et al. Jun 2001 B1
6312435 Wallace et al. Nov 2001 B1
6331181 Tierney et al. Dec 2001 B1
6394998 Wallace et al. May 2002 B1
6424885 Niemeyer et al. Jul 2002 B1
6441577 Blumenkranz et al. Aug 2002 B2
6459926 Nowlin et al. Oct 2002 B1
6491691 Morley et al. Dec 2002 B1
6491701 Tierney et al. Dec 2002 B2
6493608 Niemeyer Dec 2002 B1
6565554 Niemeyer May 2003 B1
6645196 Nixon et al. Nov 2003 B1
6659939 Moll Dec 2003 B2
6671581 Niemeyer et al. Dec 2003 B2
6676684 Morley et al. Jan 2004 B1
6685698 Morley et al. Feb 2004 B2
6699235 Wallace et al. Mar 2004 B2
6714839 Salisbury, Jr. et al. Mar 2004 B2
6716233 Whitman Apr 2004 B1
6728599 Wang et al. Apr 2004 B2
6746443 Morley et al. Jun 2004 B1
6766204 Niemeyer et al. Jul 2004 B2
6770081 Cooper et al. Aug 2004 B1
6772053 Niemeyer Aug 2004 B2
6783524 Anderson et al. Aug 2004 B2
6793652 Whitman et al. Sep 2004 B1
6793653 Sanchez et al. Sep 2004 B2
6799065 Niemeyer Sep 2004 B1
6837883 Moll et al. Jan 2005 B2
6839612 Sanchez et al. Jan 2005 B2
6840938 Morley et al. Jan 2005 B1
6843403 Whitman Jan 2005 B2
6846309 Whitman et al. Jan 2005 B2
6866671 Tierney et al. Mar 2005 B2
6871117 Wang et al. Mar 2005 B2
6879880 Nowlin et al. Apr 2005 B2
6899705 Niemeyer May 2005 B2
6902560 Morley et al. Jun 2005 B1
6936042 Wallace et al. Aug 2005 B2
6951535 Ghodoussi et al. Oct 2005 B2
6974449 Niemeyer Dec 2005 B2
6991627 Madhani et al. Jan 2006 B2
6994708 Manzo Feb 2006 B2
7048745 Tierney et al. May 2006 B2
7066926 Wallace et al. Jun 2006 B2
7118582 Wang et al. Oct 2006 B1
7125403 Julian et al. Oct 2006 B2
7155315 Niemeyer et al. Dec 2006 B2
7239940 Wang et al. Jul 2007 B2
7306597 Manzo Dec 2007 B2
7357774 Cooper Apr 2008 B2
7373219 Nowlin et al. May 2008 B2
7379790 Toth et al. May 2008 B2
7386365 Nixon Jun 2008 B2
7391173 Schena Jun 2008 B2
7398707 Morley et al. Jul 2008 B2
7413565 Wang et al. Aug 2008 B2
7453227 Prisco et al. Nov 2008 B2
7524320 Tierney et al. Apr 2009 B2
7574250 Niemeyer Aug 2009 B2
7594912 Cooper et al. Sep 2009 B2
7607440 Coste-Maniere et al. Oct 2009 B2
7666191 Orban et al. Feb 2010 B2
7682357 Ghodoussi et al. Mar 2010 B2
7689320 Prisco et al. Mar 2010 B2
7695481 Wang et al. Apr 2010 B2
7695485 Whitman et al. Apr 2010 B2
7699855 Anderson et al. Apr 2010 B2
7713263 Niemeyer May 2010 B2
7725214 Diolaiti May 2010 B2
7727244 Orban, III et al. Jun 2010 B2
7741802 Prisco Jun 2010 B2
7756036 Druke et al. Jul 2010 B2
7757028 Druke et al. Jul 2010 B2
7762825 Burbank et al. Jul 2010 B2
7778733 Nowlin et al. Aug 2010 B2
7803151 Whitman Sep 2010 B2
7806891 Nowlin et al. Oct 2010 B2
7819859 Prisco et al. Oct 2010 B2
7819885 Cooper Oct 2010 B2
7824401 Manzo et al. Nov 2010 B2
7835823 Sillman et al. Nov 2010 B2
7843158 Prisco Nov 2010 B2
7865266 Moll et al. Jan 2011 B2
7865269 Prisco et al. Jan 2011 B2
7886743 Cooper et al. Feb 2011 B2
7899578 Prisco et al. Mar 2011 B2
7907166 Lamprecht et al. Mar 2011 B2
7935130 Williams May 2011 B2
7963913 Devengenzo et al. Jun 2011 B2
7983793 Toth et al. Jul 2011 B2
8002767 Sanchez Aug 2011 B2
8004229 Nowlin et al. Aug 2011 B2
8012170 Whitman et al. Sep 2011 B2
8054752 Druke et al. Nov 2011 B2
8062288 Cooper et al. Nov 2011 B2
8079950 Stern et al. Dec 2011 B2
8100133 Mintz et al. Jan 2012 B2
8108072 Zhao et al. Jan 2012 B2
8120301 Goldberg et al. Feb 2012 B2
8142447 Cooper et al. Mar 2012 B2
8147503 Zhao et al. Apr 2012 B2
8151661 Schena et al. Apr 2012 B2
8155479 Hoffman et al. Apr 2012 B2
8182469 Anderson et al. May 2012 B2
8202278 Orban, III et al. Jun 2012 B2
8206406 Orban, III Jun 2012 B2
8210413 Whitman et al. Jul 2012 B2
8216250 Orban, III et al. Jul 2012 B2
8220468 Cooper et al. Jul 2012 B2
8256319 Cooper et al. Sep 2012 B2
8285517 Sillman et al. Oct 2012 B2
8315720 Mohr et al. Nov 2012 B2
8335590 Costa et al. Dec 2012 B2
8347757 Duval Jan 2013 B2
8374723 Zhao et al. Feb 2013 B2
8418073 Mohr et al. Apr 2013 B2
8419717 Piolaiti et al. Apr 2013 B2
8423182 Robinson et al. Apr 2013 B2
8452447 Nixon May 2013 B2
8454585 Whitman Jun 2013 B2
8499992 Whitman et al. Aug 2013 B2
8508173 Goldberg et al. Aug 2013 B2
8528440 Morley et al. Sep 2013 B2
8529582 Devengenzo et al. Sep 2013 B2
8540748 Murphy et al. Sep 2013 B2
8551116 Julian et al. Oct 2013 B2
8562594 Cooper et al. Oct 2013 B2
8594841 Zhao et al. Nov 2013 B2
8597182 Stein et al. Dec 2013 B2
8597280 Cooper et al. Dec 2013 B2
8600551 Itkowitz et al. Dec 2013 B2
8608773 Tierney et al. Dec 2013 B2
8620473 Diolaiti et al. Dec 2013 B2
8624537 Nowlin et al. Jan 2014 B2
8634957 Toth et al. Jan 2014 B2
8638056 Goldberg et al. Jan 2014 B2
8638057 Goldberg et al. Jan 2014 B2
8644988 Prisco et al. Feb 2014 B2
8649554 Markovic Feb 2014 B2
8666544 Moll et al. Mar 2014 B2
8668638 Donhowe et al. Mar 2014 B2
8746252 McGrogan et al. Jun 2014 B2
8749189 Nowlin et al. Jun 2014 B2
8749190 Nowlin et al. Jun 2014 B2
8758352 Cooper et al. Jun 2014 B2
8761930 Nixon Jun 2014 B2
8768516 Diolaiti et al. Jul 2014 B2
8786241 Nowlin et al. Jul 2014 B2
8790243 Cooper et al. Jul 2014 B2
8808164 Hoffman et al. Aug 2014 B2
8816628 Nowlin et al. Aug 2014 B2
8821480 Burbank Sep 2014 B2
8823308 Nowlin et al. Sep 2014 B2
8827989 Niemeyer Sep 2014 B2
8838270 Druke et al. Sep 2014 B2
8852174 Burbank Oct 2014 B2
8858547 Brogna Oct 2014 B2
8862268 Robinson et al. Oct 2014 B2
8864751 Prisco et al. Oct 2014 B2
8864752 Diolaiti et al. Oct 2014 B2
8903546 Diolaiti et al. Dec 2014 B2
8903549 Itkowitz et al. Dec 2014 B2
8911428 Cooper et al. Dec 2014 B2
8912746 Reid et al. Dec 2014 B2
8944070 Guthart Feb 2015 B2
8989903 Weir et al. Mar 2015 B2
9002518 Manzo Apr 2015 B2
9014856 Manzo et al. Apr 2015 B2
9016540 Whitman et al. Apr 2015 B2
9019345 O'Grady et al. Apr 2015 B2
9043027 Durant et al. May 2015 B2
9050120 Swarup et al. Jun 2015 B2
9055961 Manzo et al. Jun 2015 B2
9068628 Solomon et al. Jun 2015 B2
9078684 Williams Jul 2015 B2
9084623 Gomez et al. Jul 2015 B2
9095362 Dachs, II et al. Aug 2015 B2
9096033 Holop et al. Aug 2015 B2
9101381 Burbank et al. Aug 2015 B2
9113877 Whitman et al. Aug 2015 B1
9138284 Krom et al. Sep 2015 B2
9144456 Rosa et al. Sep 2015 B2
9179979 Jinno Nov 2015 B2
9198730 Prisco et al. Dec 2015 B2
9204923 Manzo et al. Dec 2015 B2
9226648 Saadat et al. Jan 2016 B2
9226750 Weir et al. Jan 2016 B2
9226761 Burbank Jan 2016 B2
9232984 Guthart et al. Jan 2016 B2
9241766 Duque et al. Jan 2016 B2
9241767 Prisco et al. Jan 2016 B2
9241769 Larkin et al. Jan 2016 B2
9259275 Burbank Feb 2016 B2
9259277 Rogers et al. Feb 2016 B2
9259281 Griffiths et al. Feb 2016 B2
9259282 Azizian et al. Feb 2016 B2
9261172 Solomon et al. Feb 2016 B2
9265567 Orban et al. Feb 2016 B2
9265584 Itkowitz et al. Feb 2016 B2
9283049 Diolaiti et al. Mar 2016 B2
9301811 Goldberg et al. Apr 2016 B2
9314307 Richmond et al. Apr 2016 B2
9317651 Nixon Apr 2016 B2
9345546 Toth et al. May 2016 B2
9393017 Flanagan et al. Jul 2016 B2
9402689 Prisco et al. Aug 2016 B2
9417621 Diolaiti Aug 2016 B2
9424303 Hoffman et al. Aug 2016 B2
9433418 Whitman et al. Sep 2016 B2
9446517 Bums et al. Sep 2016 B2
9452020 Griffiths et al. Sep 2016 B2
9474569 Manzo et al. Oct 2016 B2
9480533 Devengenzo et al. Nov 2016 B2
9503713 Zhao et al. Nov 2016 B2
9550300 Danitz et al. Jan 2017 B2
9554859 Nowlin et al. Jan 2017 B2
9566124 Prisco et al. Feb 2017 B2
9579164 Itkowitz et al. Feb 2017 B2
9585641 Cooper et al. Mar 2017 B2
9615883 Schena et al. Apr 2017 B2
9623563 Nixon Apr 2017 B2
9623902 Griffiths et al. Apr 2017 B2
9629520 Diolaiti Apr 2017 B2
9662177 Weir et al. May 2017 B2
9664262 Donlon et al. May 2017 B2
9675354 Weir et al. Jun 2017 B2
9687312 Dachs, II et al. Jun 2017 B2
9700334 Hinman et al. Jul 2017 B2
9718190 Larkin et al. Aug 2017 B2
9730719 Brisson et al. Aug 2017 B2
9737199 Pistor et al. Aug 2017 B2
9795446 DiMaio et al. Oct 2017 B2
9797484 Solomon et al. Oct 2017 B2
9801690 Larkin et al. Oct 2017 B2
9814530 Weir et al. Nov 2017 B2
9814536 Goldberg et al. Nov 2017 B2
9814537 Itkowitz et al. Nov 2017 B2
9820823 Richmond et al. Nov 2017 B2
9827059 Robinson et al. Nov 2017 B2
9830371 Hoffman et al. Nov 2017 B2
9839481 Blumenkranz et al. Dec 2017 B2
9839487 Dachs, II Dec 2017 B2
9850994 Schena Dec 2017 B2
9855102 Blumenkranz Jan 2018 B2
9855107 Labonville et al. Jan 2018 B2
9872737 Nixon Jan 2018 B2
9877718 Weir et al. Jan 2018 B2
9883920 Blumenkranz Feb 2018 B2
9888974 Niemeyer Feb 2018 B2
9895813 Blumenkranz et al. Feb 2018 B2
9901408 Larkin Feb 2018 B2
9918800 Itkowitz et al. Mar 2018 B2
9943375 Blumenkranz et al. Apr 2018 B2
9948852 Lilagan et al. Apr 2018 B2
9949798 Weir Apr 2018 B2
9949802 Cooper Apr 2018 B2
9952107 Blumenkranz et al. Apr 2018 B2
9956044 Gomez et al. May 2018 B2
9980778 Ohline et al. May 2018 B2
10008017 Itkowitz et al. Jun 2018 B2
10028793 Griffiths et al. Jul 2018 B2
10033308 Chaghajerdi et al. Jul 2018 B2
10034719 Richmond et al. Jul 2018 B2
10052167 Au et al. Aug 2018 B2
10085811 Weir et al. Oct 2018 B2
10092344 Mohr et al. Oct 2018 B2
10123844 Nowlin Nov 2018 B2
10155315 Ogawa et al. Dec 2018 B2
10188471 Brisson Jan 2019 B2
10201390 Swarup et al. Feb 2019 B2
10213202 Flanagan et al. Feb 2019 B2
10258416 Mintz et al. Apr 2019 B2
10278782 Jarc et al. May 2019 B2
10278783 Itkowitz et al. May 2019 B2
10282881 Itkowitz et al. May 2019 B2
10335242 Devengenzo et al. Jul 2019 B2
10405934 Prisco et al. Sep 2019 B2
10433922 Itkowitz et al. Oct 2019 B2
10464219 Robinson et al. Nov 2019 B2
10485621 Morrissette et al. Nov 2019 B2
10500004 Hanuschik et al. Dec 2019 B2
10500005 Weir et al. Dec 2019 B2
10500007 Richmond et al. Dec 2019 B2
10507066 DiMaio et al. Dec 2019 B2
10510267 Jarc et al. Dec 2019 B2
10524871 Liao Jan 2020 B2
10548459 Itkowitz et al. Feb 2020 B2
10575909 Robinson et al. Mar 2020 B2
10592529 Hoffman et al. Mar 2020 B2
10595946 Nixon Mar 2020 B2
10603127 Hasser et al. Mar 2020 B2
10695142 Peine Jun 2020 B2
10881469 Robinson Jan 2021 B2
10881473 Itkowitz et al. Jan 2021 B2
10898188 Burbank Jan 2021 B2
10898189 Mcdonald, II Jan 2021 B2
10905506 Itkowitz et al. Feb 2021 B2
10912544 Brisson et al. Feb 2021 B2
10912619 Jarc et al. Feb 2021 B2
10918387 Duque et al. Feb 2021 B2
10918449 Solomon et al. Feb 2021 B2
10932873 Griffiths et al. Mar 2021 B2
10932877 Devengenzo et al. Mar 2021 B2
10939969 Swarup et al. Mar 2021 B2
10939973 DiMaio et al. Mar 2021 B2
10952801 Miller et al. Mar 2021 B2
10965933 Jarc Mar 2021 B2
10966742 Rosa et al. Apr 2021 B2
10973517 Wixey Apr 2021 B2
10973519 Weir et al. Apr 2021 B2
10984567 Itkowitz et al. Apr 2021 B2
10993773 Cooper et al. May 2021 B2
10993775 Cooper et al. May 2021 B2
11000331 Krom et al. May 2021 B2
11013567 Wu et al. May 2021 B2
11020138 Ragosta Jun 2021 B2
11020191 Diolaiti et al. Jun 2021 B2
11020193 Wixey et al. Jun 2021 B2
11026755 Weir et al. Jun 2021 B2
11026759 Donlon et al. Jun 2021 B2
11040189 Vaders et al. Jun 2021 B2
11045077 Stern et al. Jun 2021 B2
11045274 Dachs, II et al. Jun 2021 B2
11058501 Tokarchuk et al. Jul 2021 B2
11076925 DiMaio et al. Aug 2021 B2
11090119 Burbank Aug 2021 B2
11096687 Flanagan et al. Aug 2021 B2
11098803 Duque et al. Aug 2021 B2
11109925 Cooper et al. Sep 2021 B2
11116578 Hoffman et al. Sep 2021 B2
11129683 Steger et al. Sep 2021 B2
11135029 Suresh et al. Oct 2021 B2
11147552 Burbank et al. Oct 2021 B2
11147640 Jarc et al. Oct 2021 B2
11154373 Abbott et al. Oct 2021 B2
11154374 Hanuschik et al. Oct 2021 B2
11160622 Goldberg et al. Nov 2021 B2
11160625 Wixey et al. Nov 2021 B2
11161243 Rabindran et al. Nov 2021 B2
11166758 Mohr et al. Nov 2021 B2
11166770 DiMaio et al. Nov 2021 B2
11166773 Ragosta et al. Nov 2021 B2
11173597 Rabindran et al. Nov 2021 B2
11185378 Weir et al. Nov 2021 B2
11191596 Thompson et al. Dec 2021 B2
11197729 Thompson et al. Dec 2021 B2
11213360 Hourtash et al. Jan 2022 B2
11221863 Azizian et al. Jan 2022 B2
11234700 Ragosta et al. Feb 2022 B2
11241274 Vaders et al. Feb 2022 B2
11241290 Waterbury et al. Feb 2022 B2
11259870 DiMaio et al. Mar 2022 B2
11259884 Burbank Mar 2022 B2
11272993 Gomez et al. Mar 2022 B2
11272994 Saraliev et al. Mar 2022 B2
11291442 Wixey et al. Apr 2022 B2
11291513 Manzo et al. Apr 2022 B2
11331158 Peine May 2022 B2
20070021738 Hasser et al. Jan 2007 A1
20070083098 Stern Apr 2007 A1
20080215065 Wang Sep 2008 A1
20100332033 Diolaiti Dec 2010 A1
20110295268 Roelle et al. Dec 2011 A1
20110301616 Sanchez et al. Dec 2011 A1
20120116416 Neff et al. May 2012 A1
20130331644 Pandya et al. Dec 2013 A1
20140316434 Simaan Oct 2014 A1
20150025549 Kilroy Jan 2015 A1
20200289221 Denlinger Sep 2020 A1
20210000558 Penny Jan 2021 A1
20210038296 Sartor Feb 2021 A1
20220387128 Bail Dec 2022 A1
20230110248 Faraji Apr 2023 A1
20230126506 Wilson Apr 2023 A1
Foreign Referenced Citations (4)
Number Date Country
20120068597 Jun 2012 KR
20130015437 Feb 2013 KR
1020130015437 Feb 2013 KR
2015012241 Jan 2015 WO
Non-Patent Literature Citations (18)
Entry
Indian Office Action dated Jul. 6, 2020 corresponding to counterpart Patent Application IN 201717027545.
International Search Report dated Apr. 20, 2016, issued in PCT/US2016/014031.
European Search Report dated Oct. 15, 2018 cited in EP 16752764.
Australian Examination Report No. 1 dated Sep. 18, 2019 corresponding to counterpart Patent Application AU 2016220501.
Japanese Office Action dated Sep. 26, 2019, issued in JP Appln. No. 2017542842.
Chinese Third Office Action dated Dec. 4, 2020 corresponding to counterpart Patent Application CN 201680011105.4.
Australian Examination Report No. 3 for Application No. AU 2020201796 dated Aug. 12, 2020.
Australian Examination Report No. 1 dated Jun. 26, 2020 corresponding to counterpart Patent Application AU 2020201796.
Chinese First Office Action dated Oct. 29, 2019 corresponding to counterpart Patent Application CN 201680011105.4.
Japanese Office Action dated Jan. 9, 2020 corresponding to counterpart Patent Application JP 2017-542842.
Australian Examination Report for Application No. 2020201796 dated Jul. 17, 2020.
Chinese Second Office Action dated Jun. 10, 2020 corresponding to counterpart Patent Application CN 201680011105.4.
Chinese Fourth Office Action dated Apr. 27, 2021 corresponding to counterpart Patent Application CN 201680011105.4.
Japanese Office Action dated Mar. 31, 2021 corresponding to counterpart Patent Application JP 2020-083039.
Chinese Decision of Rejection dated Aug. 13, 2021 corresponding to counterpart Patent Application CN 201680011105.4.
Australian Examination Report No. 1 dated Sep. 1, 2021 corresponding to counterpart Patent Application AU 2020227038.
Canadian Office Action dated Sep. 16, 2022 corresponding to counterpart Patent Application CA 2,975,907.
Chinese First Office Action dated May 12, 2023 issued in Application No. CN 2021113610744 (6 pages).
Related Publications (1)
Number Date Country
20220273385 A1 Sep 2022 US
Provisional Applications (1)
Number Date Country
62118123 Feb 2015 US
Continuations (2)
Number Date Country
Parent 16546554 Aug 2019 US
Child 17745994 US
Parent 15548439 US
Child 16546554 US