Robotic surgical devices, systems, and related methods

Information

  • Patent Grant
  • 10806538
  • Patent Number
    10,806,538
  • Date Filed
    Wednesday, August 3, 2016
    8 years ago
  • Date Issued
    Tuesday, October 20, 2020
    4 years ago
Abstract
The various inventions relate to robotic surgical devices, consoles for operating such surgical devices, operating theaters in which the various devices can be used, insertion systems for inserting and using the surgical devices, and related methods.
Description
FIELD OF THE INVENTION

The embodiments disclosed herein relate to various medical devices and related components, including robotic and/or in vivo medical devices and related components. Certain embodiments include various robotic medical devices, including robotic devices that are disposed within a body cavity and positioned using a support component disposed through an orifice or opening in the body cavity. Other embodiments relate to various systems that have a robotic surgical device and a controller, wherein the device has one or more sensors and the controller has one or more motors such that the sensors transmit information that is used at the controller to actuate the motors to provide haptic feedback to a user.


BACKGROUND OF THE INVENTION

Invasive surgical procedures are essential for addressing various medical conditions. When possible, minimally invasive procedures such as laparoscopy are preferred.


However, known minimally invasive technologies such as laparoscopy are limited in scope and complexity due in part to 1) mobility restrictions resulting from using rigid tools inserted through access ports, and 2) limited visual feedback. Known robotic systems such as the da Vinci® Surgical System (available from Intuitive Surgical, Inc., located in Sunnyvale, Calif.) are also restricted by the access ports, as well as having the additional disadvantages of being very large, very expensive, unavailable in most hospitals, and having limited sensory and mobility capabilities.


There is a need in the art for improved surgical methods, systems, and devices.


BRIEF SUMMARY OF THE INVENTION

Discussed herein are various robotic surgical systems, including certain systems having camera lumens configured to receive various camera systems. Further embodiments relate to surgical insertion devices configured to be used to insert various surgical devices into a cavity of a patient while maintaining insufflations of the cavity.


A system of one or more computers can be configured to perform particular operations or actions by virtue of having software, firmware, hardware, or a combination of them installed on the system that in operation causes or cause the system to perform the actions. One or more computer programs can be configured to perform particular operations or actions by virtue of including instructions that, when executed by data processing apparatus, cause the apparatus to perform the actions.


In one Example, a robotic surgical system, including a robotic surgical device including a device body including a distal end; a proximal end, and a camera lumen defined within the device body, the camera lumen including (1) a proximal lumen opening in the proximal end of the device body; (2) a socket portion defined distally of the proximal lumen opening, the socket portion including a first diameter and a first coupling component; (3) an extended portion defined distally of the socket portion, the extended portion having a second, smaller diameter; and (4) a distal lumen opening in the distal end of the device body, the distal lumen opening defined at a distal end of the extended portion; first and second shoulder joints operably coupled to the distal end of the device body; a first robotic arm operably coupled to the first shoulder joint; and a second robotic arm operably coupled to the second shoulder joint; and a camera component, including a handle including a distal end configured to be positionable within the socket portion; a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket portion; an elongate tube operably coupled to the handle, where the elongate tube is configured and sized to be positionable through the extended portion, the elongate tube including a rigid section; an optical section; and a flexible section operably coupling the optical section to the rigid section, where the elongate tube has a length such that at least the optical section is configured to extend distally from the distal lumen opening when the camera component is positioned through the camera lumen. Other embodiments of this aspect include corresponding computer systems, apparatus, and computer programs recorded on one or more computer storage devices, each configured to perform the actions of the methods.


Implementations may include one or more of the following features. The robotic surgical system where the camera lumen further includes a seal portion defined distally of the socket portion and proximally of the extended portion. The robotic surgical system where the seal section is configured to receive a ring seal and a one-way seal. The robotic surgical system where the seal section is further configured to receive a retention component, where the ring seal is retained within the ring-seal retention component. The robotic surgical system where the ring-seal retention component includes at least one protrusion extending from an outer wall of the ring-seal retention component. The robotic surgical system where the socket portion further includes a channel defined in an inner wall of the socket portion, where the channel is configured to receive the at least one protrusion. The robotic surgical system where the handle includes a controller configured to operate the camera component. The robotic surgical system where the distal lumen opening is positioned between the first and second shoulder joints. The robotic surgical system where the optical section is configured to be tiltable at the flexible section in relation to the rigid section, where the optical section has a straight configuration and a tilted configuration. The robotic surgical system where the elongate tube is configured to be rotatable in relation to the handle. The robotic surgical system where the socket portion further includes an inner wall including a channel configured to receive an insertion device. The robotic surgical system where the camera lumen includes a proximal lumen opening in the proximal end of the device body; a socket portion defined distally of the proximal lumen opening, the socket portion including a first diameter and a first coupling component; an extended portion defined distally of the socket portion, the extended portion having a second, smaller diameter; and a distal lumen opening in the distal end of the device body, the distal lumen opening defined at a distal end of the extended portion. The robotic surgical system where the first robotic arm further includes a first arm upper arm; a first arm elbow joint; and a first arm lower arm, where the first arm upper arm is configured to be capable of roll, pitch and yaw relative to the first shoulder joint and the first arm lower arm is configured to be capable of yaw relative to the first arm upper arm by way of the first arm elbow joint. The surgical robotic system where the first robotic arm further includes at least one first arm actuator disposed within the first robotic arm. The robotic surgical system where the second robotic arm further includes a second arm upper arm; a second arm elbow joint; and a second arm lower arm, where the second arm upper arm is configured to be capable of roll, pitch and yaw relative to the second shoulder joint and the second arm lower arm is configured to be capable of yaw relative to the second arm upper arm by way of the second arm elbow joint. The surgical robotic system where the second robotic arm further includes at least one second arm actuator disposed within the second robotic arm. The surgical robotic system including a handle including a distal end configured to be positionable within the socket portion; and a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket portion. The surgical robotic system further including at least one PCB disposed within at least one of the first or second robotic arms and in operational communication with at least one of the first robotic arm and second robotic arm, where the PCB is configured to perform yaw and pitch functions. Implementations of the described techniques may include hardware, a method or process, or computer software on a computer-accessible medium.


In one Example, a robotic surgical system, including a robotic surgical device including a device body including a distal end; a proximal end, and a camera lumen defined within the device body; first and second shoulder joints operably coupled to the distal end of the device body; a first robotic arm operably coupled to the first shoulder joint; and a second robotic arm operably coupled to the second shoulder joint; and a camera component, including a handle including a distal end configured to be positionable within the socket portion; a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket portion; an elongate tube operably coupled to the handle, where the elongate tube is configured and sized to be positionable through the extended portion, the elongate tube including a rigid section; an optical section; and a flexible section operably coupling the optical section to the rigid section, where the elongate tube has a length such that at least the optical section is configured to extend distally from the distal lumen opening when the camera component is positioned through the camera lumen. Other embodiments of this aspect include corresponding computer systems, apparatus, and computer programs recorded on one or more computer storage devices, each configured to perform the actions of the methods.


Implementations may include one or more of the following features. The robotic surgical system where the camera lumen includes a proximal lumen opening in the proximal end of the device body; a socket portion defined distally of the proximal lumen opening, the socket portion including a first diameter and a first coupling component; an extended portion defined distally of the socket portion, the extended portion having a second, smaller diameter; and a distal lumen opening in the distal end of the device body, the distal lumen opening defined at a distal end of the extended portion. The robotic surgical system where the first robotic arm further includes a first arm upper arm; a first arm elbow joint; and a first arm lower arm, where the first arm upper arm is configured to be capable of roll, pitch and yaw relative to the first shoulder joint and the first arm lower arm is configured to be capable of yaw relative to the first arm upper arm by way of the first arm elbow joint. The surgical robotic system where the first robotic arm further includes at least one first arm actuator disposed within the first robotic arm. The robotic surgical system where the second robotic arm further includes a second arm upper arm; a second arm elbow joint; and a second arm lower arm, where the second arm upper arm is configured to be capable of roll, pitch and yaw relative to the second shoulder joint and the second arm lower arm is configured to be capable of yaw relative to the second arm upper arm by way of the second arm elbow joint. The surgical robotic system where the second robotic arm further includes at least one second arm actuator disposed within the second robotic arm. The surgical robotic system including a handle including a distal end configured to be positionable within the socket portion; and a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket portion. The surgical robotic system further including at least one PCB disposed within at least one of the first or second robotic arms and in operational communication with at least one of the first robotic arm and second robotic arm, where the PCB is configured to perform yaw and pitch functions. Implementations of the described techniques may include hardware, a method or process, or computer software on a computer-accessible medium.


In one Example, a robotic surgical system, including a robotic surgical device including a device body including a distal end; a proximal end, and a camera lumen defined within the device body, the camera lumen including (1) a proximal lumen opening in the proximal end of the device body; (2) a socket portion defined distally of the proximal lumen opening, the socket portion including a first diameter and a first coupling component; (3) an extended portion defined distally of the socket portion, the extended portion having a second, smaller diameter; and (4) a distal lumen opening in the distal end of the device body, the distal lumen opening defined at a distal end of the extended portion; first and second shoulder joints operably coupled to the distal end of the device body; a first robotic arm operably coupled to the first shoulder joint; and a second robotic arm operably coupled to the second shoulder joint; and a camera component, including an elongate tube operably coupled to the handle, where the elongate tube is configured and sized to be positionable through the extended portion, the elongate tube including a rigid section; an optical section; and a flexible section operably coupling the optical section to the rigid section, where the elongate tube has a length such that at least the optical section is configured to extend distally from the distal lumen opening when the camera component is positioned through the camera lumen. Other embodiments of this aspect include corresponding computer systems, apparatus, and computer programs recorded on one or more computer storage devices, each configured to perform the actions of the methods.


Implementations may include one or more of the following features. The surgical robotic system including a handle including a distal end configured to be positionable within the socket portion; and a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket portion. The surgical robotic system further including at least one PCB disposed within at least one of the first or second robotic arms and in operational communication with at least one of the first robotic arm and second robotic arm, where the PCB is configured to perform yaw and pitch functions. Implementations of the described techniques may include hardware, a method or process, or computer software on a computer-accessible medium.


While multiple embodiments are disclosed, still other embodiments of the present invention will become apparent to those skilled in the art from the following detailed description, which shows and describes illustrative embodiments of the invention. As will be realized, the invention is capable of modifications in various obvious aspects, all without departing from the spirit and scope of the present invention. Accordingly, the drawings and detailed description are to be regarded as illustrative in nature and not restrictive.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A is a front view of a robotic surgical device according to one embodiment.



FIG. 1B is perspective front view of the device of FIG. 1A.



FIG. 1C is a side view of the device of FIG. 1A.



FIG. 1D is an end view of the device of FIG. 1A.



FIG. 1E is a further front view of the device of FIG. 1A, without the camera component.



FIG. 1F is a further side view of the device of FIG. 1A, without the camera component.



FIG. 1G is a front view of the camera component, according to the embodiment of FIG. 1A.



FIG. 2A is a perspective view of the proximal end of a robotic surgical device according to one embodiment.



FIG. 2B is a rotated perspective view of the device of FIG. 2A.



FIG. 2C is a cutaway view of the proximal end of the device of FIG. 2A.



FIG. 3A is a perspective front view of a seal insertion component according to one embodiment.



FIG. 3B is a close-up view of the distal end of the insertion component of FIG. 3A without seals.



FIG. 3C is a close-up view of the distal end of the insertion component of FIG. 3A showing the O-ring carrier.



FIG. 3D is a perspective view of the insertion component of FIG. 3A above a robotic device body.



FIG. 3E is a perspective cutaway view of the embodiment of FIG. 3D.



FIG. 4A is an end view of a camera component according to one embodiment.



FIG. 4B is a side view of the embodiment of FIG. 4A, in a “down” configuration.



FIG. 4C is a side view of the embodiment of FIG. 4A, in an “up” configuration.



FIG. 4D is a three-quarters rotated view of the embodiment of FIG. 4C.



FIG. 4E is an end view of a camera component showing the internal components, according to one embodiment.



FIG. 4F is a front view of the embodiment of FIG. 4E, in an “up” configuration.



FIG. 4G is a side view of the embodiment of FIG. 4A, in an “up” configuration.



FIG. 4H is a three-quarters rotated view of the embodiment of FIG. 4G.



FIG. 5A is a cutaway side view of the proximal end of a camera component inserted into a robotic surgical device according to one embodiment.



FIG. 5B is a further close-up cutaway side view of the embodiment of FIG. 5A.



FIG. 5C is a side view of the internal components of a camera component according to one embodiment.



FIG. 5D is a further internal side view of the embodiment of FIG. 5C.



FIG. 5E is a perspective view of the embodiment of FIG. 50.



FIG. 5F is a perspective internal view of a camera component inserted into a robotic surgical device according to one embodiment.



FIG. 5G is a further perspective internal view of the embodiment of FIG. 5F.



FIG. 6A is a side view of the internal components of a camera component according to one embodiment.



FIG. 6B is a perspective internal view of the internal components of a camera component according to one embodiment.



FIG. 6C is a side internal view of the internal components of a camera component according to the embodiment of FIG. 6B.



FIG. 6D is a internal front view of the internal components of a camera component according to the embodiment of FIG. 6B.



FIG. 6E is a perspective internal view of a camera component inserted into a robotic surgical device according to one embodiment.



FIG. 7A is a perspective internal view of the distal end of the camera component according to one embodiment.



FIG. 7B is a perspective internal view of the distal end of the camera component according to another embodiment.



FIG. 7C is a schematic flow of camera information from a lens to a surgical console, according to one embodiment.



FIG. 8A is an internal front view of the device body without a housing, according to one embodiment.



FIG. 8B is a side view of the embodiment of FIG. 8A.



FIG. 8C is a perspective view of the embodiment of FIG. 8A.



FIG. 8D is an end view of the embodiment of FIG. 8A.



FIG. 8E is a rear three-quarters perspective view of the device body without a housing, according to one embodiment.



FIG. 8F is a side view of the embodiment of FIG. 8E.



FIG. 8G is a front three-quarters perspective view of the embodiment of FIG. 8E with the housing.



FIG. 9A is an internal front view of the device body showing the internal components without a housing or support structures, according to one embodiment.



FIG. 9B is a perspective view of certain yaw components of the embodiment of FIG. 9A.



FIG. 9C is a perspective view of certain pitch components of the embodiment of FIG. 9A.



FIG. 10 is a perspective view of a robotic arm having six degrees of freedom according to one embodiment.



FIG. 11A is a side view of an upper robotic arm without its housing according to one embodiment.



FIG. 11B is a rotated side view of the embodiment of FIG. 11A.



FIG. 11C is yet another rotated side view of the embodiment of FIG. 11A.



FIG. 11D is an end view of the embodiment of FIG. 11A.



FIG. 11E is a perspective view of an upper robotic arm according to one embodiment.



FIG. 11F is a rotated perspective view of the embodiment of FIG. 11E, without the housing.



FIG. 12A is a further rotated view of the embodiment of FIG. 11E.



FIG. 12B is another internal view of the components of an upper robotic arm according to one embodiment.



FIG. 12C is a perspective view of certain yaw components of the embodiment of FIG. 12B.



FIG. 12D is a perspective view of certain pitch components of the embodiment of FIG. 12B.



FIG. 13A is a perspective view of a lower robotic arm according to one embodiment.



FIG. 13B is a reverse perspective view of the embodiment of FIG. 13A, without the housing.



FIG. 13C is a side view of a lower robotic arm without its housing according to one embodiment.



FIG. 13D is a rotated side view of the embodiment of FIG. 13C.



FIG. 13E is yet another rotated side view of the embodiment of FIG. 13C.



FIG. 13F is an end view of the embodiment of FIG. 13C.



FIG. 13G is a perspective view of the embodiment of FIG. 13C.



FIG. 14A is another internal view of the components of a lower robotic arm according to one embodiment.



FIG. 14B is a perspective view of certain roll components of the embodiment of FIG. 14A.



FIG. 14C is a perspective view of certain end effector interaction coupling components of the embodiment of FIG. 14A.



FIG. 14D is a cross-sectional side view of a forearm, according to one embodiment.



FIG. 14E is a cutaway perspective side view of a forearm, according to one embodiment.



FIG. 14F is cross-sectional side view of a forearm, according to one embodiment.



FIG. 15A is a perspective view of an end effector, according to one embodiment.



FIG. 15B is a perspective view of an end effector, according to one embodiment.



FIG. 16A is a schematic view of a monopolar cautery connection, according to one embodiment.



FIG. 16B is a schematic view of a bipolar cauter connection, according to one embodiment.



FIG. 17A is top view of one implementation of the device within a sleeve, according to one embodiment.



FIG. 17B is a side perspective view of the arms of the device disposed within sleeves, according to one embodiment.



FIG. 18A is a perspective view of one implementation of the device within a sleeve, according to one embodiment.



FIG. 18B is an up-close view of the implementation of FIG. 18A.



FIG. 18C is a rotated up-close view of the implementation of FIG. 18A, without the end effectors shown.



FIG. 19A is a perspective view of one implementation of the device within a sleeve, according to another embodiment.



FIG. 19B is an up-close view of the implementation of FIG. 19A.



FIG. 19C is a rotated, cross-sectional up-close view of the implementation of FIG. 19A.



FIG. 20A is a perspective view of a device arm inside a sleeve and in an extended position, according to one embodiment.



FIG. 20B is a perspective view of the embodiment of FIG. 20A in a bent position.



FIG. 21A is a front view of a device arm having a semi-rigid slide guide, according to one embodiment.



FIG. 21B is a side view of the implementation of FIG. 21A in a bent position.



FIG. 21C is a side view of the implementation of FIG. 21A in a further bent position.



FIG. 22A is a front view of a sleeve having an “outer box” pleat, according to one implementation.



FIG. 22B is a front view of a sleeve having an “inner box” pleat, according to one implementation.



FIG. 22C is a front view of a “bent” sleeve, according to one implementation.



FIG. 23A is a perspective view of a disposable sleeve having an adhesive strip, according to one implementation.



FIG. 23B is a side view of a disposable sleeve having an adhesive strip, according to another implementation.



FIG. 23C is a further side view of a disposable sleeve, according to one implementation.



FIG. 24A is a perspective cutaway view of a side of the device and the device port prior to insertion of the device, according to one implementation.



FIG. 24B is a perspective cutaway view of the device of FIG. 24A immediately following insertion.



FIG. 24C is a further perspective cutaway view of the device of FIG. 24A following insertion, where the device has been tilted.



FIG. 24D is a further perspective cutaway view of the device of FIG. 24A following insertion, where the device has been rotated.



FIG. 25A is a side cross-sectional view of a device port, according to one implementation.



FIG. 25B is a top view of the port of FIG. 25A.



FIG. 25C is a front view of one device implementation attached to a robot support arm, according to one implementation.



FIG. 26A is a schematic view of one implementation of a robotic surgical device and operations system.



FIG. 26B is a front view of a surgical console, according to one implementation.



FIG. 26C is a side view of the surgical console of FIG. 26B.



FIG. 26D is a perspective view of the surgical console of FIG. 26B.



FIG. 26E is a top view of the foot controllers of the surgical console of FIG. 26B.



FIG. 27A is a screen view of a graphical user interface on the console, according to one implementation.



FIG. 27B is a screen view of another graphical user interface on the console, according to one implementation.



FIG. 27C is a screen view of yet another graphical user interface on the console, according to one implementation.



FIG. 28A is a schematic view of the workspace of one arm of a robotic device, according to one implementation.



FIG. 28B is a further schematic view of the workspace of one arm of a robotic device, according to one implementation.



FIG. 28C is yet a further schematic view of the workspace of one arm of a robotic device, according to one implementation.



FIG. 28D is schematic depiction of system and device operation, according to one implementation.



FIG. 29A is an end view of the hand controller limits in the haptic feedback system, according to on implementation.



FIG. 29B is a side view of the limits of FIG. 29A.



FIG. 29C is a top view of the limits of FIG. 29A.



FIG. 29D is a side view of the limits of FIG. 29A, showing the system disposed within those limits.



FIG. 30A is a perspective view of a hand controller, according to one implementation.



FIG. 30B is a cutaway view of the hand controller of FIG. 30A.





DETAILED DESCRIPTION

The various systems and devices disclosed herein relate to devices for use in medical procedures and systems. More specifically, various embodiments relate to various medical devices, including robotic devices and related methods and systems.


It is understood that the various embodiments of robotic devices and related methods and systems disclosed herein can be incorporated into or used with any other known medical devices, systems, and methods.


It is understood that the various embodiments of robotic devices and related methods and systems disclosed herein can be incorporated into or used with any other known medical devices, systems, and methods. For example, the various embodiments disclosed herein may be incorporated into or used with any of the medical devices and systems disclosed in U.S. Pat. No. 8,968,332 (issued on Mar. 3, 2015 and entitled “Magnetically Coupleable Robotic Devices and Related Methods”), U.S. Pat. No. 8,834,488 (issued on Sep. 16, 2014 and entitled “Magnetically Coupleable Surgical Robotic Devices and Related Methods”), U.S. patent application Ser. No. 14/617,232 (filed on Feb. 9, 2015 and entitled “Robotic Surgical Devices and Related Methods”), U.S. patent application Ser. No. 11/966,741 (filed on Dec. 28, 2007 and entitled “Methods, Systems, and Devices for Surgical Visualization and Device Manipulation”), U.S. Patent Application 61/030,588 (filed on Feb. 22, 2008), U.S. Pat. No. 8,343,171 (issued on Jan. 1, 2013 and entitled “Methods and Systems of Actuation in Robotic Devices”), U.S. Pat. No. 8,828,024 (issued on Sep. 9, 2014 and entitled “Methods and Systems of Actuation in Robotic Devices”), U.S. patent application Ser. No. 14/454,035 (filed Aug. 7, 2014 and entitled “Methods and Systems of Actuation in Robotic Devices”), U.S. patent application Ser. No. 12/192,663 (filed Aug. 15, 2008 and entitled Medical Inflation, Attachment, and Delivery Devices and Related Methods”), U.S. patent application Ser. No. 15/018,530 (filed Feb. 8, 2016 and entitled “Medical Inflation, Attachment, and Delivery Devices and Related Methods”), U.S. Pat. No. 8,974,440 (issued on Mar. 10, 2015 and entitled “Modular and Cooperative Medical Devices and Related Systems and Methods”), U.S. Pat. No. 8,679,096 (issued on Mar. 25, 2014 and entitled “Multifunctional Operational Component for Robotic Devices”), U.S. Pat. No. 9,179,981 (issued on Nov. 10, 2015 and entitled “Multifunctional Operational Component for Robotic Devices”), U.S. patent application Ser. No. 14/936,234 (filed on Nov. 9, 2015 and entitled “Multifunctional Operational Component for Robotic Devices”), U.S. Pat. No. 8,894,633 (issued on Nov. 25, 2014 and entitled “Modular and Cooperative Medical Devices and Related Systems and Methods”), U.S. Pat. No. 8,968,267 (issued on Mar. 3, 2015 and entitled “Methods and Systems for Handling or Delivering Materials for Natural Orifice Surgery”), U.S. Pat. No. 9,060,781 (issued on Jun. 23, 2015 and entitled “Methods, Systems, and Devices Relating to Surgical End Effectors”), U.S. patent application Ser. No. 14/745,487 (filed on Jun. 22, 2015 and entitled “Methods, Systems, and Devices Relating to Surgical End Effectors”), U.S. Pat. No. 9,089,353 (issued on Jul. 28, 2015 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), U.S. patent application Ser. No. 14/800,423 (filed on Jul. 15, 2015 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), U.S. patent application Ser. No. 13/573,849 (filed Oct. 9, 2012 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), U.S. patent application Ser. No. 13/738,706 (filed Jan. 10, 2013 and entitled “Methods, Systems, and Devices for Surgical Access and Insertion”), U.S. patent application Ser. No. 13/833,605 (filed Mar. 15, 2013 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), U.S. patent application Ser. No. 14/661,465 (filed Mar. 18, 2015 and entitled “Methods, Systems, and Devices for Surgical Access and Insertion”), Ser. No. 13/839,422 (filed Mar. 15, 2013 and entitled “Single Site Robotic Devices and Related Systems and Methods”), U.S. Pat. No. 9,010,214 (issued on Apr. 21, 2015 and entitled “Local Control Robotic Surgical Devices and Related Methods”), U.S. patent application Ser. No. 14/656,109 (filed on Mar. 12, 2015 and entitled “Local Control Robotic Surgical Devices and Related Methods”), U.S. patent application Ser. No. 14/208,515 (filed Mar. 13, 2014 and entitled “Methods, Systems, and Devices Relating to Robotic Surgical Devices, End Effectors, and Controllers”), U.S. patent application Ser. No. 14/210,934 (filed Mar. 14, 2014 and entitled “Methods, Systems, and Devices Relating to Force Control Surgical Systems), U.S. patent application Ser. No. 14/212,686 (filed Mar. 14, 2014 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), U.S. patent application Ser. No. 14/334,383 (filed Jul. 17, 2014 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), U.S. patent application Ser. No. 14/853,477 (filed Sep. 14, 2015 and entitled “Quick-Release End Effectors and Related Systems and Methods”), U.S. patent application Ser. No. 14/938,667 (filed Nov. 11, 2015 and entitled “Robotic Device with Compact Joint Design and Related Systems and Methods”), and U.S. Patent Application 62/338,375 (filed May 18, 2016 and entitled “Robotic Surgical Devices, Systems, and Related Methods”), and U.S. Pat. No. 7,492,116 (filed on Oct. 31, 2007 and entitled “Robot for Surgical Applications”), U.S. Pat. No. 7,772,796 (filed on Apr. 3, 2007 and entitled “Robot for Surgical Applications”), and U.S. Pat. No. 8,179,073 (issued May 15, 2011, and entitled “Robotic Devices with Agent Delivery Components and Related Methods”), all of which are hereby incorporated herein by reference in their entireties.


Certain device and system implementations disclosed in the applications listed above can be positioned within a body cavity of a patient in combination with a support component similar to those disclosed herein. An “in vivo device” as used herein means any device that can be positioned, operated, or controlled at least in part by a user while being positioned within a body cavity of a patient, including any device that is coupled to a support component such as a rod or other such component that is disposed through an opening or orifice of the body cavity, also including any device positioned substantially against or adjacent to a wall of a body cavity of a patient, further including any such device that is internally actuated (having no external source of motive force), and additionally including any device that may be used laparoscopically or endoscopically during a surgical procedure. As used herein, the terms “robot,” and “robotic device” shall refer to any device that can perform a task either automatically or in response to a command.


Certain embodiments provide for insertion of the present invention into the cavity while maintaining sufficient insufflation of the cavity. Further embodiments minimize the physical contact of the surgeon or surgical users with the present invention during the insertion process. Other implementations enhance the safety of the insertion process for the patient and the present invention. For example, some embodiments provide visualization of the present invention as it is being inserted into the patient's cavity to ensure that no damaging contact occurs between the system/device and the patient. In addition, certain embodiments allow for minimization of the incision size/length. Further implementations reduce the complexity of the access/insertion procedure and/or the steps required for the procedure. Other embodiments relate to devices that have minimal profiles, minimal size, or are generally minimal in function and appearance to enhance ease of handling and use.


Certain implementations disclosed herein relate to “combination” or “modular” medical devices that can be assembled in a variety of configurations. For purposes of this application, both “combination device” and “modular device” shall mean any medical device having modular or interchangeable components that can be arranged in a variety of different configurations. The modular components and combination devices disclosed herein also include segmented triangular or quadrangular-shaped combination devices. These devices, which are made up of modular components (also referred to herein as “segments”) that are connected to create the triangular or quadrangular configuration, can provide leverage and/or stability during use while also providing for substantial payload space within the device that can be used for larger components or more operational components. As with the various combination devices disclosed and discussed above, according to one embodiment these triangular or quadrangular devices can be positioned inside the body cavity of a patient in the same fashion as those devices discussed and disclosed above.


Certain embodiments disclosed or contemplated herein can be used for colon resection, a surgical procedure performed to treat patients with lower gastrointestinal diseases such as diverticulitis, Crohn's disease, inflammatory bowel disease and colon cancer. Approximately two-thirds of known colon resection procedures are performed via a completely open surgical procedure involving an 8- to 12-inch incision and up to six weeks of recovery time. Because of the complicated nature of the procedure, existing robot-assisted surgical devices are rarely used for colon resection surgeries, and manual laparoscopic approaches are only used in one-third of cases. In contrast, the various implementations disclosed herein can be used in a minimally invasive approach to a variety of procedures that are typically performed ‘open’ by known technologies, with the potential to improve clinical outcomes and health care costs. Further, the various implementations disclosed herein can be used for any laparoscopic surgical procedure in place of the known mainframe-like laparoscopic surgical robots that reach into the body from outside the patient. That is, the less-invasive robotic systems, methods, and devices disclosed herein feature small, self-contained surgical devices that are inserted in their entireties through a single incision in the patient's abdomen. Designed to utilize existing tools and techniques familiar to surgeons, the devices disclosed herein will not require a dedicated operating room or specialized infrastructure, and, because of their much smaller size, are expected to be significantly less expensive than existing robotic alternatives for laparoscopic surgery. Due to these technological advances, the various embodiments herein could enable a minimally invasive approach to procedures performed in open surgery today.


As shown in FIGS. 1A-1G, certain exemplary embodiments relate to a device 10 having a body 12 with two arms 14, 16 operably coupled thereto and a camera component 18 positionable therein. That is, device 10 has a first (or “right”) arm 14 and a second (or “left) arm 16, both of which are operably coupled to the body 12 as discussed in additional detail below. The body 12 as shown has a casing (also referred to as a “cover” or “enclosure”) 20. The body 12 is also referred to as a “device body” 20 and has two rotatable cylindrical components (also referred to as “housings” and “turrets”): a first (or “right”) housing 22 and a second (or “left”) housing 24. Each arm 14, 16 also has an upper arm (also referred to herein as an “inner arm,” “inner arm assembly,” “inner link,” “inner link assembly,” “upper arm assembly,” “first link,” or “first link assembly”) 14A, 16A, and a forearm (also referred to herein as an “outer arm,” “outer arm assembly,” “outer link,” “outer link assembly,” “forearm assembly,” “second link,” or “second link assembly”) 14B, 16B. The right upper arm 14A is operably coupled to the right housing 22 of the body 12 at the right shoulder joint 26 and the left upper arm 16A is operably coupled to the left housing 24 of the body 12 at the left shoulder joint 28. Further, for each arm 14, 16, the forearm 14B, 16B is rotatably coupled to the upper arm 14A, 16A at the elbow joint 14C, 16C.


In the exemplary implementation as shown, each of the arms 14, 16 also has an end effector 30, 32 operably coupled to the distal end of the forearm 14B, 16B. An end effector can also be referred to herein as an “operational component,” and various embodiments will be discussed herein below in further detail.


In one implementation, each of the arms 14, 16 has six degrees of freedom. That is, as explained in further detail below, each arm 14, 16 has three degrees of freedom at the shoulder joint 26, 28, one degree of freedom at the elbow joint 14C, 16C, and two degrees of freedom at the end effector 30, 32 (which can be, in certain embodiments, rotated—end effector roll—and opened/closed). As such, the six degrees of freedom of each arm 14, 16 are analogous to the degrees of freedom of a human arm, which also has three degrees of freedom at the shoulder and one at the elbow. One advantage of an arm having four degrees of freedom (with an end effector having two degrees of freedom) is that the end effector can have multiple orientations at the same Cartesian point. This added dexterity allows the surgeon or other user more freedom and a more intuitive sense of control while operating the device.


The camera component 18, as shown in FIG. 1G in accordance with one embodiment, is easily insertable into and removable from the body 12. As shown, the camera component 18 has a handle 40, a camera body or tube 42, a distal tube end 18A having a camera lens 48, and two shoulders 44, 46 defined at the distal end of the handle 40. That is, the first shoulder 44 has a first diameter and the second shoulder 46 has a second diameter that is larger than the first diameter.


According to one embodiment, FIGS. 2A, 2B, and 2C depict the proximal end of the device body 12 having sealed electrical connections or connectors 50, 52, a support rod 54, a latch 56 and a head 58 with an opening 60 defined in the body 12. As discussed in relation to FIGS. 25A-C herein, in these implementations, the support rod is configured to be attached to the support arm to dispose the device for use.


In various implementations, the electrical connectors 50, 52 can provide robot power and bus communications required for robot functionality, including power and communications connectors, bipolar cautery connectors and monopolar cautery connectors, such as LEMO® push-pull circular connectors. In certain implementations, three connectors can be provided. In the implementation of FIGS. 2A-C, the first electrical connector 50 is configured to send and receive robot power and bus communications and the second electrical connector 52 is configured for combined cautary mono- and bi-polar connectivity. Alternatively, the three connectors may be combined into a single integrate custom connector, In yet a further alternative, and as shown in FIGS. 3D-F, a single cable 53 integrated directly into the robot can be provided. It is understood that a sealed, strain relieved cable egress location would then exist in this location instead of the connectors 50, 52.


According to these implementations, the opening 60 is in fluid communication with a lumen 62 that is defined through the length of the body 12. The lumen 62 is configured to receive the camera component 18 and has a receiving portion (also referred to herein as a “socket portion” or “socket”) 62A, a seal portion 62B, and an extended portion 62C.


In certain implementations, the socket portion 62A is configured to be “tight fitting,” that is, it is configured to mate with the camera component 18 handle 40 to react or resist all loads or prevent all rotational and translational motion. In various implementations, the latch 56 is disposed within the socket portion 62A so as to be capable of coupling to the clasping portion 72 of the camera component 18.


In various implementations, a seal or seals 63A, 63B are provided in the seal portion 62B, so as to maintain a fluidic seal around the camera 18 as it is disposed in the lumen 62. The seal portion 62B is distal in relation to the receiving portion 62A and is configured to house a seal or seals 63A, 63B against the wall 68 of the lumen 62, as is described in relation to FIGS. 2C-3F.


In the implementation depicted in FIGS. 2C and 3A-F, the device utilizes a first seal 63A that is a one-way duckbill seal, though it is understood that various other one-way seals can be used in alternate embodiments. In these implementations, a second seal 63B—which can be an O-ring carrier seal—is also disposed proximally to the first seal 63A. As shown in FIGS. 3A and 3C-F, in various implementations, the O-ring carrier seal 63B comprises an O-ring 65 configured to urge the first seal 63A distally. It is understood that in various implementations, the O-ring can provide a seal against the camera component 18, while the O-ring carrier seal 63A can provide a seal against the lumen 62A against the escape of gasses or fluids as described herein.


As described below, in these implementations, when the seals are installed, the O-ring carrier seal 63B compresses on the lip 63A1 of the first seal 63A, thereby creating a seal against the inner wall of the housing (shown at 67). The use of first and second seals 63A, 63B in certain implementations provides certain advantages described herein. In situations when the camera component 18 is not present, the pressure from the abdominal cavity will cause the one-way duck bill seal 63A to close and prevent the loss of that pressure. In situations where the camera present, the camera and tube 42 will cause duck bill seal 63A to be open and allow passage into the lumen 62, while the O-ring 65 and O-ring carrier seal 63A will seal against the rigid camera tube 42 and lumen 62, respectively, to maintain cavity pressure. It is understood that further implementations are of course possible.


As shown in FIGS. 3A-F, in various implementations, an elongate insertion component 15 is provided, which allows the insertion and removal of the seal or seals 63A, 63B for replacement and/or cleaning. As is shown in FIGS. 3A-C, the insertion component 15 can have a seal coupling ring 13 with mounting projections 13A, 13B configured to mate to a seal such as the O-ring carrier seal 63B and maintain the rotational position of the seal while it is being disposed in the lumen 62. In various implementations, ridges 17 can also be provided to secure the seal in place. Returning to FIG. 2C, the distal end of the receiving portion 62A is defined by a shoulder 64 that is configured to receive the insertion component 15. At least one channel 66 is defined in a portion of the shoulder 64 and the lumen 62 as shown and is configured to receive a corresponding protrusion or protrusions 67A, 67B disposed on the O-ring carrier seal 63B such that the protrusion or protrusions 67A, 67B can be positioned in and slid along the channel 66, thereby allowing for the seals 63A, 63B to be place and locked into position in the lumen 62. The insertion component 15 can subsequently be removed, such that the seals 63A, 63B are contained within the lumen 62 for use.


More specifically, the channel 66 is defined in the lumen 62 with a longitudinal length 66A and a radial length 66B. In certain implementations, the channel 66 is tapered along the longitudinal length 66A. As such, a protrusion 67A is positioned in the longitudinal length 66A and the insertion component 15 is advanced distally until the protrusion 67A reaches the end of the longitudinal length 66A. At this point, the insertion component 15 can be rotated around its longitudinal axis such that the protrusion 67A is advanced along the radial length 66B. As shown in FIG. 3B, the mounting projections 13A, 13B can prevent the rotation of the O-ring seal 63B relative to the insertion component 15 during this rotation. Further the rigid O-ring 65 provides sufficient distal force against the first seal 63A such that it is fixed into place as a result of this rotation and locking. The resulting coupling of the seals 63A, 63B within the lumen is a mechanical coupling that is sufficiently strong for a user to pass the camera component 18 through the seals 63A, 63B for use without dislodging the seals 63A, 63B.



FIGS. 4A-H depict an exemplary implementation of the camera component 18. The camera component 18 in this specific implementation is configured to be removably incorporated into a robotic device body 12, as is shown in FIGS. 1A-B. More specifically, the camera component 18 is configured to be removably positioned through the lumen 62 defined in the device body 12 such that the camera component 18 is inserted through the proximal opening 60, receiving portion into the receiving portion 62A, through the seal portion 62B and seal or seals 63A, 63B, and into the extended portion 62C such that a distal portion of the camera component 18A and camera lens 48 protrudes from the distal opening 60A (as best shown in FIG. 1A).


As shown in FIGS. 4A-H, this camera component 18 embodiment has a controller (also referred to as a “handle”) 40 and an elongate component (also referred to herein as a “tube”) 42 operably coupled at its proximal end to the handle 40. As best shown in FIG. 4C, the tube 42 has a rigid section 42A, a flexible section 42B, and an optical section 42C. As such, in these implementations the camera component has two degrees of freedom: pan (left and right rotation) and tilt, meaning looking “up and down” in the surgical workspace. Further discussion of these degrees of freedom can be found in relation to FIG. 10.


In one embodiment, the handle 40 is configured to contain local electronics for video transmission, along with actuators and associated mechanisms (as are best shown in relation to FIG. 4H) for actuating pan and tilt functionality of the tube 42. It is understood that the local electronics, actuators, and associated mechanisms can be known, standard components. In a further implementation, the handle 40 can also contain a light engine. Alternatively, the light engine can be a separate component, and a light cable can operably couple the light engine to the handle.


According to one implementation, the rigid section 42A of the tube 42 is substantially rigid and contains appropriate wires and optical fibers as necessary to operably couple to the optical component in the optical section 42C to the handle 40. The substantial rigidity of the rigid section 42A allows for easy manipulation of the tube 42, including easy insertion into the lumen 62.


The flexible section 42B, in accordance with one embodiment, is configured to allow for movement of the optical section 42C between a straight configuration in FIG. 4B and a tilted configuration as shown in FIG. 4C, or any position in between. The optical section 42C is substantially rigid, much like the rigid section 42A, and contains the optical element, along with appropriate local electronics.


Accordingly, various implementations of the camera component 18 of this implementation have two mechanical degrees of freedom: pan (look left/right) and tilt (look up/down). In use, the camera component 18 has pan and tilt functionality powered and controlled by the actuators and electronics in the handle 40. In various implementations, the handle 40 further comprises a button 70 and camera clasp 72 configured to mate with the latch 56, as is shown in further detail in FIGS. 5A-D.


The tilt functionality relates to tilting the optical section 42C such that the camera 48 is oriented into the desired workspace, as is discussed further in relation to FIGS. 24A-D. This tilting can be accomplished via a cable that is operably coupled to the flexible section 42B or the optical section 42C such that actuation of the cable causes the optical section 42C to tilt by bending the flexible section 42B as shown for example in FIG. 4C and FIGS. 7A-B. Alternatively this tilt function can be achieved by any other known mechanism or method for bending the tube 42 at the flexible section 42B.


As shown in the implementations of FIGS. 5A-7B, the camera component 18 houses several internal electrical and mechanical components capable of operation and movement of the camera 48, tube 42 and other operative components. In various implementations, the camera component 18 has a presence sensing system configured to detect the insertion of the component 18 into the lumen 62. Further, to prevent damage, in certain embodiments, the camera component 18 is configured to provide a “mechanical lockout,” such that the camera component 18 cannot be removed unless the tube 42 is in the “straight” (tilt=0°) configuration.


As discussed above, FIGS. 5A-B depict an implementation of the device 10 where the camera component 18 has been inserted into the lumen 62 of the body 12. In these implementations, following the placement of the seals 63A, 63B in the seal portion 62B (as described in relation to FIGS. 3A-F), the camera component 18 can be inserted into the lumen 62 and the latch 56 engaged.


As shown in FIG. 5C, in certain implementations, the camera component 18 comprises a light guide post 74, an actuation connector 76 and/or a coax connector 78. In various implementations, the light guide post 74 facilitates the transference of light from an external light generator to the fiber optics in the camera. An actuation connector 76 can provide power and communications functions such as robotic and camera power and communications functions discussed below in relation to FIGS. 26-29D. In various implementations, the coax connector 78 can provide additional functionality, such as transmission of video signal, as is described herein in relation to FIGS. 7C and 26A. In various implementations, the fiber optic cable 80 is in operational and luminary communication with the light guide post and extends distally into the lumen (not shown). Further, a communications line 82 extends with the fiber optic cable 80 in these implementations, as is discussed further in relation to FIGS. 5E-F.


The implementation of FIGS. 5C-E depicts one implementation of the camera component 18 having a mechanical lockout. In certain of these implementations, depressing the handle button 70 can activate the re-orientation of the tube 42 into the straight orientation, so as to allow for removal of the camera component when tilt=0°. In this implementation, the camera clasp 72 is disposed on a clasping member 84 which also comprises the button 70, such that depressing the button 70 results in the pivoting of the clasping member around a clasp pivot 86 that is rotationally coupled to the camera housing (not shown), such that a plunger 88 disposed adjacent to the clasping member so as to be capable of being urged inward in response to the actuation of the button 70.


In various implementations, the plunger 88 end 88A is aligned with a slot 90A in the lead screw nut 90, which is linearly translated in response to camera tilt, as is described in further detail below. In these implementations, slot 90A and plunger 88 alignment only occurs when the camera tube 42 is in the “straight” orientation. In these implementations, the plunger is also fixedly attached to a trigger arm 92, such that when the arm is displaced—even slightly—the arm triggers a limit switch 94, initiating a “go-straight” subroutine, thereby straightening the camera. It is understood that the length of plunger 88 in these implementations is such that it is unable to enter the slot 90A when the camera is tilted (as described below in relation to FIGS. 5E-6D, so that that clasp 72 will not disengage from the robot when slot 90A is not aligned.


It is understood that in certain implementations, the “go-straight” subroutine is triggered in response to the actuation of the button 70, regardless of whether the plunger end 88A enters the slot 90A. In these implementations, and as best shown in FIG. 5E, the space between the non-slotted portions (shown generally at 91) of the lead screw nut 90 and the plunger end 88A is less than the distance of the overlap between the clasp 72A and latch edges 56A (shown in FIG. 5B), thereby preventing unclasping. In these implementations, the distance between the trigger arm 92 and limit switch 94 is also less than the distance between the space between the non-slotted portions (shown generally at 91) of the lead screw nut 90 and the plunger end 88A, such that the limit switch 94 will be actuated in response to the actuation of the button 70 whether or not the plunger end 88A enters the slot 90A. In certain implmentations, an actuator spring 96 is operationally coupled to the plunger 88 to urge the plunger outward, thereby keeping the clasp 72 and button 70 tensioned when not in use.


As best shown in the implementation of FIG. 5E-G, camera tilt is driven by a tilt actuator 100 disposed within the handle 40. The tilt actuator 100 in these implementations can be a 6 mm Maxon BLDC motor, or other such actuator. In these implementations, the tilt actuator 100 is capable of causing translational movement in the lead screw nut 90. In these implementations, the tilt actuator 100 drives a planetary gearhead and spur gear 102, which is coupled to a drive gear 104. In these implementations, the drive gear 104 is in rotational communication with the lead screw 106. In these implementations, rotation of the lead screw 106 within the lead screw nut 90 causes translational motion of the lead screw nut 90 and optionally a cable coupler assembly 108 fixedly attached to the lead screw nut 90 exterior. It is understood that in these implementations the lead screw nut 90 is rotationally coupled, but linearly decoupled, from the pan shaft 112, such that rotation of lead screw 106 causes linear translation of lead screw nut 90. In various implementations, an actuation cable (best shown at 120A, 120B in FIGS. 5F-G) is fixedly coupled to coupler assembly 108 such that translation of the lead screw nut 90 and cable coupler assembly 108 causes tilt actuation to occur.


As shown in FIGS. 5F-G, In various implementations, tilt functionality can be accomplished via the following configuration. In this embodiment, the flexible section 42B includes an elbow joint 124 and a pair of tilt cables 120A, 120B, wherein each of the tilt cables 120A, 120B is operably coupled at its distal end to the optical section 42C. In various implementations, and as shown, the cables 120A, 120B comprise a teflon sheathing 120C, 120D. In these implementations, the sheathings can remain static, while the cables 120A, 120B disposed within are able to slide relative to the sheathing as described generally herein.


The first tilt cable 120A is depicted in FIG. 5F-G is an active tilt cable 120A that is coupled on one side of the optical section 42C in relation to the joint 124 as shown such that urging the cable 120A proximally causes the optical section 42C to tilt upward on that side, as is designated by reference arrow C. The second tilt cable 120B is a passive tilt cable 120B that is coupled on the other side of the optical section 42C in relation to the joint 124 and the first title cable 120A. The second tilt cable 120B is not actuated by a user. Instead, the second tilt cable 120B is maintained at a predetermined level of tension by a passive spring 122 such that the cable 120B is continuously urged in the proximal direction, thereby urging the optical section 42C into a straight configuration such as that shown in FIG. 4B.


As such, in this implementation of FIGS. 5F-G, the default position of the optical section 42C will be the straight configuration. That is, the tensioned passive tilt cable 120B causes the optical section 42C to be in the straight configuration when no forces are being applied to the active tilt cable 120A by the cable coupler assembly 108. A user can cause proximal movement of the cable coupler assembly 108 through the lead screw nut, as described above, causing the active title cable 120A to be urged proximally to tilt the optical section 42C. In response to other activities, such as depressing the button 70, the cable 120A can be caused to allow the section 42C to return to the straight configuration by way of the spring 122 and return cable 120B. The straight configuration of FIG. 4B makes it easy to position the camera component 18 into the lumen 62 and further to remove the camera component 18 from the lumen 62 as well. In use, a user can urge the active cable 120A proximally to tilt the optical section 42C as desired/needed. In alternative embodiments, the system can have an actuation button (or other type of user interface) (not shown) that can be configured to actuate the system to move to the straight configuration, thereby facilitating easy insertion and/or removal.


The pan functionality is accomplished via rotation of the tube 42 around the longitudinal axis of the tube 42 as shown by arrow D in FIG. 5G-F. The rigid section 42A, the flexible section 42B, and the optical section 42C of the tube (not shown) are coupled together such that the sections 42A, 42B, 42C cannot rotate in relation to each other. In other words, the sections 42A, 42B, 42C rotate together as a single unit. The tube 42, however, is rotatably coupled to the handle 40 such that the tube 42 can rotate as shown by arrow D in relation to the handle 40. As a result, the panning functionality is provided by positioning the optical section 42C in a tilted configuration (such as the configuration of FIG. 5F) and rotating the tube 42 in relation to the handle 40. This results in the optical component in the optical section 42C being rotated around the tube 42 axis such that it can potentially capture images up to and including 360° around the camera component 18. In certain implementations, pan is limited to smaller ranges, such as 100°, such as by way of a hard stop 133, as shown in FIGS. 6A-B. In these implementations, the hard stop 33 rotates with the tube 42 and tube head 137 (or lead screw 106), while the tube housing 135 (or lead screw nut 90) maintains a fixed position, thereby limiting the range of tube 42 motion, as the hard stop 133 is unable to rotate through the tube housing 135. A hard stop opening 133A can also be provided in the tube head 137, as is shown in FIG. 6A. It is understood that in certain implementations, the limiting of pan range is done because of wire service loops.


As such, in the implementation of FIGS. 5F-6E, pan functionality is performed by way of a pan actuator 130 disposed within the handle 40, which is best shown in FIG. 6A. The pan actuator 130 in these implementations can be a 6 mm Maxon BLDC motor, or other such actuator. In these implementations, the pan actuator 130 is capable of causing rotational movement in the tube 42. In these implementations, the pan actuator 130 drives a planetary gearhead and spur gear 132, which is coupled to a drive gear 134. In these implementations, the drive gear 134 is in rotational communication with the pan shaft 112, which in turn is in rotational communication with the tube 42. It is understood that in these implementations the pan shaft 112 is rotationally coupled to the tube 42, such that rotation of the pan shaft 112 causes rotation of the entire tube 42, including the optical portion 42C, thus resulting in pan functionality. FIGS. 6B-D depict further implementations of the internal components of the camera handle 40, showing the pan actuator 130 and tilt actuator 100 disposed within the handle housing (not shown).


In these implementations, the pan assembly (generally at 128) has a ground slot 136 (which does not rotate) and a pan shaft slot 138 (which rotates), both being configured such that wires (not shown) may pass through the slots 136, 138 safely and not be damaged during pan actuation.


For example, as is shown in FIG. 6E, the image sensor power/communication lines 82 and the fiber optic illumination cable 80 are routed over a support 140 and pass through the slots 136, 138 in the to enter the camera tube 42 and extend to the lens 48. At least one handle rigid-flex PCB component, or “PCB” 142 is also provided to control various the camera handle functions, such as tilt and pan. It is understood that in certain implementations, a third degree of freedom is attainable with digital (software) based zoom.



FIGS. 7A and 7B depict various internal views of the flexible section 42B and distal camera components.


The implementation of FIG. 7A has a camera lens 48 which contains a stack 48A of lenses configured to optimize, among other parameters, field of view (such as approximately 90 degrees) and depth of field (approximately 1″ to 6″ focal range). A plurality of fiber optic lights 160 are also disposed in a lens housing 162. As is shown in FIGS. 7A-B, in various implementations, these fiber optics 160A, 160B, 160C can be disposed on opposite sides of the lens 48 (FIG. 7B) or radially around and/or “under” the lens 48 (FIG. 7A). These fiber optics 160A, 160B, 160C are in luminary communication with the fiber optic cable or cables 80A, 80B extending down from the handle, as discussed above, for example in relation to FIG. 6F.


In the implementation of FIGS. 7A-B, an image sensor 164 (such as an OmniVision 22720 IC, capable of 1080p @ 30 fps) is disposed behind the lens stack 48A on flex tape 166. In these implementations, the image sensor 164 or sensors outputs data in a MIPI format through the flex tape 166, which is in turn threaded through the flexible portion 42B. The flex tape 166 terminates at a termination point 166A into a rigid PCB 168 at the distal end of the camera tube (not shown). It is understood that the flexible tube portion 42B in the implementation of FIG. 7A comprises a plurality of articulating members 170A, 170B, 170C, as has been previously described, though other implementations are possible.


In various embodiments, and as shown generally in FIG. 7C, the sensor image signal (box 170) from the flex tape 166 is converted in the camera (box 171) from MIPI to LVDS by an FPGA (box 172) on the PCB 168. In an exemplary implementation, this LVDS signal is then transmitted through the internal camera harness, through the connector on the camera handle, through the 5 camera pigtail, through a connector pair, through a 20′ harness, through a panel mount connector on the surgeon console, through the surgeon console internal harness to a panel mount connector on the CCU (camera control unit—box 173), through the internal CCU harness, into the daughter card.


In the implementation of FIG. 7C, the CCU (box 173) translates the LVDS signal to parallel data (boxes 174 and 175), then to an HDMI output. The HDMI output is routed to the surgeon console computer (box 176) to an onboard video processing card (box 177). In various implementations, the video processing card (box 177) mixes the camera feed with GUI overlays (box 178), such that the mixed signal can be passed to the main monitor (box 179) on the surgeon console (box 176) where it is viewed. This signal is also mirrored on an HDMI output on the surgeon console connector panel, where it may be routed to an auxiliary monitor. It is understood that there are many different signaling protocals that may be used. In one example, the rigid PCB 168 at the distal end of the rigid tube 42 may convert the MIPI data to serial data instead and transmit the serialized signal along a coaxial cable back to the CCU. In another example, the video processing card (box 177) and GUI overlays (box 178) may be omitted, and the video signal may be routed directly from the CCU to the main display. In a further example, the video signal may be mirrored from the main display (box 179) instead of (or in addition to) the surgeon console connector panel.



FIGS. 8A-G and 9A-D according to one embodiment, depict the internal components of the body 12, which is shown in these figures without its casing 20. FIGS. 9B-C depict the right half of the body 12 and the internal components that control/actuate the right arm 14A. It is understood that the internal components in the left half (not shown) that operate/control/actuate the left arm 14B are substantially the same as those depicted and described herein and that the descriptions provided below apply equally to those components as well.



FIGS. 8A-G include the internal structural or support components of the body 12. In one implementation, the body 12 has an internal top cap 200 and an internal support shell 202 as shown. These components maintain the structure of the body 12 and provide structural support for the components disposed therein. FIG. 8D is an enlarged view of the distal end of the body 12.


In contrast to FIGS. 8A-D, FIGS. 9B-C depict the internal actuation and control components of the body 12 with the internal structural or support components hidden in order to better display the internal actuation and control components. These internal actuation and control components are configured to provide two degrees of freedom at the shoulder joint 26, 28.


In one embodiment, certain of the internal components depicted in FIGS. 9A-C are configured to actuate rotation at the shoulder joint 26, 28 around axis A (as best shown in FIG. 9A), which is parallel to the longitudinal axis of the body 12. This rotation around axis A is also referred to as “yaw” or “shoulder yaw.” The rotation, in one aspect, is created as follows. A yaw actuator 204 is provided that is, in this implementation, a yaw motor assembly. The yaw motor assembly 204 is operably coupled to the yaw motor gear 206, which is coupled to the driven gear 208 such that rotation of the yaw motor gear 206 causes rotation of the driven gear 208. The driven gear 208 is fixedly coupled to a transmission shaft 210, which has a transmission gear 212 at the opposite end of the shaft 210.


The transmission gear 212 is coupled to a driven gear 214, which is fixedly coupled to the shaft 216. A magnet holder 218 containing a magnet is also operably coupled to the transmission gear 212. The holder 218 and magnet are operably coupled to a magnetic encoder (not shown). It is understood that the magnet holder 218, magnet, and magnetic encoder (and those similar components as discussed elsewhere herein in relation to other joints) are components of an absolute position sensor that is the same as or substantially similar to one or more of the absolute position sensors disclosed in U.S. Provisional Application 61/680,809, filed on Aug. 8, 2012, which is hereby incorporated herein by reference in its entirety. The shaft 216 is fixedly coupled at its distal end to a rotatable pitch housing 220 (as best shown in FIGS. 9A-B) such that rotation of the driven gear 214 causes rotation of the shaft 216 and thus rotation of the housing 220 around axis A as shown in FIG. 8B and FIG. 9B (this is also shown in FIG. 10 at axis Z1).


According to one implementation, certain other internal components depicted in FIG. 9C are configured to actuate rotation of the shoulder joint 26, 28 around axis B (as best shown in FIGS. 8C and 9C), which is perpendicular to the longitudinal axis of the body 12. This rotation around axis B is also referred to as “pitch” or “shoulder pitch.” The rotation, in one embodiment, is created as follows. A pitch actuator 230 is provided that is, in this implementation, a pitch motor assembly 230. The pitch motor assembly 230 is operably coupled to a motor gear 232, which is coupled to the driven gear 234 such that rotation of the motor gear 232 causes rotation of the driven gear 234. The driven gear 234 is fixedly coupled to a transmission shaft 236, which has a transmission gear 238 at the opposite end of the shaft 236. The transmission gear 238 is coupled to a driven gear 240, which is fixedly coupled to the shaft 242. A magnet holder 244 containing a magnet is also operably coupled to the driven gear 240. The holder 244 and magnet are operably coupled to a magnetic encoder (not shown). As best shown in FIG. 9B-C, a portion of the shaft 242 is disposed within the lumen 216A of the shaft 216 described above and extends out of the distal end of the shaft 216 into the housing 220. As best shown in FIG. 9C, the distal end of the shaft 242 is coupled to a rotation gear 244 that is a bevel gear 244. The rotation gear 244 is operably coupled to link gear 246, which is also a bevel gear 246 according to one implementation. The link gear 246 is operably coupled to the shoulder link 16A (discussed above) such that rotation of the shaft 242 causes rotation of the rotation gear 244 and thereby the rotation of the link gear 246 and thus rotation of the link 16A around axis B as best shown in FIG. 9D, also shown in FIG. 10 at axis Z2.


In this embodiment, these two axes of rotation are coupled. That is, if solely rotation around axis A (pure yaw) is desired, then the “pitch drive train” (the pitch motor 230 and all coupled gears and components required to achieve rotation around axis B) must match the speed of the “yaw drive train” (the yaw motor 204 and all coupled gears and components required to achieve rotation around axis A) such that there is no relative angular displacement between the pitch housing 220 and the rotation gear 244. In contrast, if solely rotation around axis B (pure pitch) is desired, then the yaw drive train must hold position while the pitch drive train is actuated.


In one implementation as shown in FIG. 9A, the body 12 has a rigid-flex PCB 250 positioned in the body. The PCB 250 is operably coupled to and controls the motors 204, 230 and magnetic encoders (not shown). In one implementation, and as shown in FIGS. 8F, 9A and elsewhere the various actuators or motors described herein have at least one temperature sensor 248 disposed on the surface of the motor, for example by temperature-sensitive epoxy, such that the temperature sensors 248 can collect temperature information from each actuator for transmission to the control unit, as discussed below. In one embodiment, any of the motors discussed and depicted herein can be brush or brushless motors. Further, the motors can be, for example, 6 mm, 8 mm, or 10 mm diameter motors. Alternatively, any known size that can be integrated into a medical device can be used. In a further alternative, the actuators can be any known actuators used in medical devices to actuate movement or action of a component. Examples of motors that could be used for the motors described herein include the EC 10 BLDC+GP10A Planetary Gearhead, EC 8 BLDC+GP8A Planetary Gearhead, or EC 6 BLDC+GP6A Planetary Gearhead, all of which are commercially available from Maxon Motors, located in Fall River, Mass. There are many ways to actuate these motions, such as with DC motors, AC motors, permanent magnet DC motors, brushless motors, pneumatics, cables to remote motors, hydraulics, and the like.


As also described herein, each link (body, upper arm, and forearm) can also contain Printed Circuit Boards (PCBs) that have embedded sensor, amplification, and control electronics. For example, in certain implementations, identical PCBs 168, 250, 290, 320, 328 are used throughout where each one controls two motors. One PCB is in each forearm and upper arm and two PCBs are in the body. Each PCB also has a full 6 axis accelerometer-based Inertial Measurement Unit and temperature sensors that can be used to monitor the temperature of the motors. Each joint can also have either an absolute position sensor or an incremental position sensor or both. In certain implementations, the some joints contain both absolute position sensors (magnetic encoders) and incremental sensors (hall effect). Joints 5 & 6 only have incremental sensors. These sensors are used for motor control. The joints could also contain many other types of sensors. A more detailed description of one possible method is included here.



FIG. 10 shows the robot motions. As shown in relation to FIG. 10, the shoulder joint 26 connects the upper arm 14A to the body 12. Shoulder yaw (θ1 about Z1), shoulder pitch (θ2 about Z2) and shoulder roll (θ3 about Z3) may or may not have the three axes largely intersect so as to form a spherical-like joint. The elbow joint 14C (θ4 about Z4) connects the upper arm 14A to the forearm 14B. Then the tool can roll (θ5 about Z5). Finally, the tool itself (or end effector) has a motion that can be used to open and close the tool. The right arm 14 of this design is a mirror image of the left 16. FIGS. 11A-14C, according to one embodiment, depict the internal components of the right arm 14. It is understood that the internal components in the left arm 16 are substantially the same as those depicted and described herein and that the descriptions provided below apply equally to those components as well.



FIGS. 11A-F and 12A-D, according to one embodiment, depict the internal components of the right upper arm 14A, which is shown in FIGS. 11A-E and 12A-D without its housing 252. More specifically, these figures depict the right arm 14A and the internal components therein. FIGS. 12A-D depict the internal components of the right upper arm 14A, including actuators, drive components, and electronics, with the internal structural or support components hidden in order to better display the internal components. In contrast to FIGS. 12A-D, FIGS. 11A-F include both the internal actuator, drive, and electronics components, but also the internal structural or support components of the right upper arm 14A.


In one embodiment, certain of the internal components depicted in FIGS. 11A-F and 12A-D are configured to actuate rotation at the shoulder link 26 around Z3 as θ3 (as best shown in FIG. 10), which is parallel to the longitudinal axis of the right upper arm 14A. This rotation θ3 is also referred to as “shoulder roll.” The rotation, in one aspect, is created as follows. An actuator 260 is provided that is, in this implementation, a motor assembly 260. The motor assembly 260 is operably coupled to the motor gear 262. The motor gear 262 is supported by a bearing pair 264. The motor gear 262 is coupled to the driven gear 266 such that rotation of the motor gear 262 causes rotation of the driven gear 266. The driven gear 266 is fixedly coupled to the shoulder link (not shown) such that rotation of the driven gear 266 causes rotation of the upper arm 14A around axis Z3 as shown in FIG. 10. The driven gear 266 is supported by a bearing pair 268. A magnet holder 270 containing a magnet is also operably coupled to the driven gear 266. The holder 270 and magnet are operably coupled to a magnetic encoder, as has been previously described.


The rotation of the shoulder link 26 around axis Z3 causes the right upper arm 14A (and thus the forearm 14B) to rotate in relation to the body 12. According to one embodiment, this rotation adds an additional degree of freedom not provided in prior two-armed surgical devices.


According to one implementation, certain of the internal components depicted in FIGS. 11A-12D are configured to actuate rotation at the elbow link 14C around axis Z4 (as best shown in FIG. 3C), which is perpendicular to the longitudinal axis of the right upper arm 14A. This rotation around axis Z4 is also referred to as “elbow yaw.” The rotation, in one aspect, is created as follows. An actuator 272 is provided that is, in this implementation, a motor assembly 272. The motor assembly 272 is operably coupled to the motor gear 274, which is a beveled gear in this embodiment. The motor gear 274 is supported by a bearing 276. The motor gear 274 is coupled to the driven gear 278 such that rotation of the motor gear 274 causes rotation of the driven gear 278. The driven gear 278 is fixedly coupled to the elbow link 14C such that rotation of the driven gear 278 causes rotation of the elbow link 14C around axis Z4 as shown in FIG. 10. The driven gear 278 is supported by a bearing pair 280. A magnet holder containing a magnet is also operably coupled to the elbow link 14C. The holder and magnet are operably coupled to a magnetic encoder 282. According to one embodiment, the additional coupling of a link gear 284 and the elbow link 14C can provide certain advantages, including an additional external reduction (because the gear 284 has fewer gear teeth than the elbow link 14C) and shortening of the upper arm 14A (thereby improving the joint range of motion). The gear 284 is coupled to another gear which has the magnetic holder 282 on it. Additionally, this other gear (not shown) has a torsion spring attached to it, which functions as an anti-backlash device.


As shown in FIG. 12A-12B, the upper arm 14A can have at least one rigid-flex PCB 290 positioned therein. In one embodiment, the PCB 290 is operably coupled to and controls the motors 260, 272 and magnetic encoders (coupled to the holders 270). In these implementations, flex tapes 292 can be used to communicate with the PCB 290, motors 260, 272 and magnetic encoders, as would be appreciated by a skilled artisan. According to another embodiment, at least one connection component is associated with the upper arm 14A. More specifically, in this implementation, a power/communication line and the cautery power line enter through a port (not shown) at the proximal end of the upper arm 14A and exit through a port (not shown) at the distal end, as has been previously described.


As set forth below, each forearm 14B, 16B also has two electrically isolated cautery circuits, enabling both bipolar and monopolar cautery end effectors. Certain embodiments are configured to allow for easy removal and replacement of an end effector (a “quick change” configuration). Further embodiments contain sealing elements that help to prevent fluid ingress into the mechanism.



FIGS. 13A-G depict various embodiments of a right forearm 14B. FIGS. 13B-G show the forearm 14B without its housing 254. The various implementations disclosed and depicted herein include the actuators, drive components, and electronics that can be used to accomplish both tool roll and tool drive (open/close action), as will be described in further detail below. As set forth below, the forearm 14B also has two electrically isolated cautery circuits, enabling both bipolar and monopolar cautery end effectors. Certain embodiments are configured to allow for easy removal and replacement of an end effector 300 (a “quick change” configuration). Further embodiments contain sealing elements that help to prevent fluid ingress into the mechanism. As shown in FIG. 13B, a power and communications lumen 303 and cautery lumen 305 can be used to allow wires (not shown) to be routed from the body 12 to the forearm.


According to one implementation, certain of the internal components depicted in FIGS. 13A-G and 14A-F are configured to actuate rotation at the end effector 300 around axis Z5 (as best shown in FIG. 10), which is parallel to the longitudinal axis of the right forearm 14B. This rotation around axis Z5 is also referred to as “tool roll.”


The rotation, in one aspect, is created as follows. As best shown in FIG. 14B, an actuator 301 is provided that is, in this implementation, a motor assembly 301. The motor assembly 301 is operably coupled to the motor gear 302, which is a spur gear in this embodiment. The motor gear 302 is coupled to the driven gear 304 such that rotation of the motor gear 302 causes rotation of the driven gear 304. The driven gear 304 is fixedly coupled to the roll hub 306, which is supported by a bearing 308. The roll hub 306 is fixedly coupled to the tool base interface 310, which has an tool lumen 311 and external threads 310A which are threadably coupled to the end effector 300. Thus, rotation of the driven gear 304 causes rotation of the roll hub 306, which causes rotation of the tool base interface 310, which causes rotation of the end effector 300 around axis Z5 as shown in FIG. 10.


In one embodiment, certain of the internal components depicted in FIGS. 14A and 14C are configured to actuate the end effector to open and close. This rotation of the end effector arms such that the end effector opens and closes is also called “tool drive.” The actuation, in one aspect, is created as follows. An actuator 312 is provided that is, in this implementation, a motor assembly 312. The motor assembly 312 is operably coupled to the motor gear 314, which is a spur gear in this embodiment. The motor gear 314 is coupled to the driven gear 316 such that rotation of the motor gear 314 causes rotation of the driven gear 316. The driven gear 316 is fixedly coupled to a female tool spline 318, which is supported by bearing pair 320. The female tool spline 318 is configured to interface with a male tool spline feature on the end effector to open/close the tool as directed.


According to one implementation, the end effector 300 can be quickly and easily coupled to and uncoupled from the forearm 14B in the following fashion. With both the roll and drive axes fixed or held in position, the end effector 300 can be rotated, thereby coupling or uncoupling the threads 310A. That is, if the end effector 300 is rotated in one direction, the end effector 300 is coupled to the forearm 14B, and if it is rotated in the other direction, the end effector 300 is uncoupled from the forearm 14B.


Various implementations of the system 10 are also designed to deliver energy to the end effectors 300 so as to cut and coagulate tissue during surgery. This is sometimes called cautery and can come in many electrical forms as well as thermal energy, ultrasonic energy, and RF energy all of which are intended for this robot. Here electrosurgical cautery is described as an example.


In accordance with one embodiment, and as shown in FIGS. 14D-F, the forearm 14B has two independent cautery channels (referred to herein as “channel A” and “channel B”), which enable the use of either bipolar or monopolar cautery end effectors with this forearm 14B.


In these implementations, the channel A components are set forth in the forearm 14B as shown. A PCB 328 is electrically isolated from lead A 342 and/or lead B 344 a cautery power line (such as discussed below) that is coupled to an external power source. The PCB 328 is further electrically coupled to at least one flex tape 330A, 330B which is in electronic communication with the motors 301, 312. As such, energizing lead A in the cautery line 342 energizes channel A in the bipolar cautery end effector 300.


As is shown in FIGS. 14E-F, in certain implementations the end effector 300 is disposed within the forearm 14B in a rotor assembly 343A, 343B such that the rotor contacts 341A, 341B and stator contacts or hoops 345A, 345B are in electrical communication with the tool contacts 330, 332. In these implementations, the cautery wire enters through a lumen 305 in the back plate of the forearm (as shown in FIG. 13A). For a bipolar forearm (which uses a pair of conductors), conductor A is soldered to tab A 342 on the stator hoop A. Conductor B is soldered to tab B 344 on the stator hoop 345B. For the monopolar forearm, there is only 1 conductor, so conductor A 342 is soldered to tab A 342 on the stator hoop 345A and the other stator hoop 345B has no connection.


In various implementations, the stator assembly 347 contains the two stator hoops 345A, 345B. The assembly 347 is fixed to the forearm 14B and does not move. The rotor assembly 343 contains two rotor rings 341A, 341B. The rotor 343 is held concentric to the stator 347 through a bearing assembly (not shown) and is free to rotate within the stator 347. Each rotor ring 341A, 341B has a pair of leaf spring contacts (best shown in FIG. 14F at 349A, 349B) which maintain electrical contact to the stator rings 345A, 345B as would be the case for a slip ring.


In these implementations, the rotor rings 341A, 341B extend into the rotor assembly, and the end effectors have a corresponding pair of tool contacts 330, 332 disposed toward the proximal end. These tool contacts 330, 332 contacts can also have leaf spring protrusions.


In use, when the end effector 300 is properly seated within the rotor 343, the leaf spring protrusions of the end effector tool contacts 330, 332 press against the internal circumference of the rotor rings 341A, 341B, so as to form an electrical connection. Additionally, the rotor can have as “arrow shaped” protrusions along its internal surface, to create a lead in, so it is self aligning when you install the tool, while the end effector can have matching cut outs. In these implementations, when the end effector is inserted the protrusions and cut outs mate, such that they form a torque transfer feature between the end effector and the rotor assembly. In this way, when the rotor spins via the roll motor, the end effector spins with it. Thus there is no relative motion between the rotor assembly and the end effector 300.


In one implementation, as shown in FIGS. 15A-B the forearm 14B can be fitted with an insertable bi-polar cautery tool (300A in FIG. 15B), or an insertable mono-polar cautery tool (300B in FIG. 15A) designed for single or multiple use.


In these implementations, the end effector 300A, 300B has at least one fluidic seal interface that helps to prevent fluid ingress into the forearm 14B. One such mechanism is a single-piece housing 322 according to one embodiment. As best shown in FIG. 15A-B the housing 322 can have an O-ring 324 positioned in a groove defined in the housing 322.


In the specific embodiment of the bi-polar tool 300A of FIG. 15A, there are two bronze contacts 330, 332 at the proximal end of the tool 330A. When inserted, these contacts 330, 332 interface with wipers that make an electrical connection between the robot and the tool 300A. As has been previously described, for example in U.S. application Ser. No. 14/212,686, which has been incorporated by reference in its entirety, a wiper is a tensioned component that supported on one end by a mechanical strut. An insulating insert is positioned between the wiper and the mechanical strut. At its free end, the wiper is supported by a preloader. Based on this configuration, the wiper is loaded or urged (like a leaf spring) against tool base interface and thus is electrically coupled to the tool base interface. The tool base interface is mechanically coupled to the end effector 28A and electrically coupled to channel B of that end effector. In these implementations, the wipers and contacts 330, 332 are designed so that relative tool motion (tool roll or end effector roll) can occur while maintaining electrical contact between the wiper and contact. These two independent contacts 330, 332 are then connect to each of the jaws respectively, such as by solid copper wires 334. The tools are kept electrically isolated from one another using several techniques including a non-conductive divider 336. The electrical energy is then delivered to the tissue held between the two jaws 338A, 338B. In this implementation, a jaw guide 340 is also provided.


In the specific embodiment of the bi-polar tool 300B of FIG. 15B, there are two bronze contacts 330, 332 at the proximal end of the tool 330B. When inserted, these contacts 330, 332 interface with wipers that make an electrical connection between the robot and the tool 300B. Mono-polar energy from the generator (described in relation to FIGS. 16A-B) flows via one electrical connection to the tool 300B so that potential energy exists at the tool tip 340. The energy then returns to the generator through the surgical target via the return pad. The cables can contain connectors so as to simplify use and handling of the robot. This figure shows one additional feature in that the outgoing energy is transmitted through a shielded cable and the shield may or may not be connected to the return path. Having the shield connected to the return pad can be a safety feature in that it prevents energy leakage to the patient. Here leaked energy would be very likely to be collected by the shield and safely returned to the generator.


Various implementations of the system have a monopolar cautery power line 350 (as shown in FIG. 16A) and/or bipolar cauter power line 352 (as shown in FIG. 16B) in electrical communication with an at least one external cautery generator 354 and the respective monopolar 300B and bipolar 300B end effectors. In the implementation of FIG. 16A, the monopolar cautery line 352 is a single coaxial cable 352 which also in electrical communication with a return pad 356 for placement elsewhere on the patient's body. In these implementations, a shield 358A can be provided around the central conductor 358. In various implementations, the shield 358A can extend the length of the central conductor 358 from the generator 354 into the body 12 so as to terminate distally (shown at 358B) in the forearm 14B. In certain implementations, a shield tie 360 is provided, which electrically ties the shield 358 to the return pad 356 and/or electrical generator 354 to prevent radiation from escaping, as would be understood by the skilled artisan.


In the implementation of FIG. 16B, a bipolar power line 352 provides electrical communication between the bipolar cautery lines 352A, 352B and the external cautery generator 354. In various implementations, the monopolar 350 and/or bipolar lines 352A, 352B can connect directly to the body 12 or be connected by way of a “pigtail” 360A, 360B, 360C, as has been previously described.


As shown in FIG. 17A, another fluidic seal can be provided according to another embodiment in the form of a flexible membrane 400 (or “sleeve”) disposed on the exterior of the arms 14, 16. As shown in FIG. 17B, in certain implementations the membrane 400 is attached at the distal end end 402 to the forearm housing 14B, 16B. This membrane 400 serves to provide a fluidic seal for the internal components of the arms 14, 16 against any external fluids. In one implementation, the seal is maintained whether the end effector 300 is coupled to the forearm 14B, 16B or not.


It is understood that large or “bulky” membranes can interfere with the operation of the camera component 18, particularly for membranes 400 having a belt, as has been previously described. In various implementations, the presently disclosed membrane 400 addresses camera interference. As discussed herein in relation to FIGS. 18A-C through 22C, in certain implementations, the membrane 400 can be permanent, while in alternate implementations, and as shown in FIG. 23A-C, the membrane 400 can be disposable. Alternatively, the membrane 400 can be replaced with a metallic bellows, as has been previously described.


In various implementations, the sleeves 400 can be fabricated by cutting a pattern out of a thin film extrusion, such that a 2D pattern is cut out of a flat piece of plastic and the sleeve is then formed by bonding the 2D pieces together, such as by ultrasonic welding. Alternatively, thermal bonding or adhesives may be used to create the sleeve. In yet a further alternative, a molding process may be utilized to create these sleeve, as has been previously described. This can include dip molding, injection molding, or other known molding options. It is understood that the permanent sleeves can be made of thicker plastic or other material than disposable sleeves to enhance durability.


As is shown in FIGS. 18A-C, in certain implementations, a permanent membrane 400 is disposed over each of the arms 14, 16. In these implementations, the membrane has a rigid termination component 404 at the distal end 402. In certain implementations, and as shown in FIG. 18C, the termination component 404 can use an internal static seal 404A that clips to snap into place and seal with the forearm housing 14B, 16B. In alternate implementations, the membrane 400 can be bonded directly to the forearm housing 14B, 16B at the distal end 402 using UV cured bio-compatible epoxy. In yet further implementations, the distal end 402 can be attached to the forearm housing using mechanical capture between the forearm housing and forearm chassis sub-structure (as is described at the proximal end in relation to FIGS. 19A-C)


Turning to the implementations of FIGS. 19A-C, at the proximal end 406, an O-ring assembly 408 can be used to “pinch” the membrane 400 into a corresponding groove 410 in the body 12. As is shown in FIG. 19C, in these implementations, an outer body housing 412 can be provided over the attached membrane 400. In alternate implementations, the membrane 400 can be bonded directly at the proximal end 406 using UV cured bio-compatible epoxy.


In further implementations, and as shown in FIGS. 20A-20B, a flex-mesh component 414 can be used in conjunction with the membrane (not shown) to prevent “sleeve shear” in highly articulated elbow joints 14C. In these implementations, the mesh 414 ensures that the sleeve does not collapse into the pinch zones created by the arm joints, such as the elbow 14C and shoulder (generally 26, 28).


In the implementations of FIGS. 21A-C, a semi-rigid slide guide 420 can be used to prevent sleeve shear in the permanent membrane 400. In these implementations, the slide guide 420 extends the length of the membrane (not shown) so as to prevent the membrane from entering the space between the joints of the arm (described above). In various implementations, the semi-rigid guide 420 can be made of thin teflon, delrin, or other flexible, low friction polymers.


In certain implementations, and as shown in FIGS. 21A-21C, the semi-rigid guide 420 is fixed 421 at one location, either at the forearm 14B as shown, or on the upper arm (not shown), so as to allow the guard to move linearly relative to the arm if necessary. In the implementations of FIGS. 21A-21C, the semi-rigid guide 420 is disposed within a guide bushing 422 at the opposite end (here, the proximal end). It is understood that this allows the sliding of the guide as the robot articulates, and creates a moving barrier to prevent the sleeve from entering the pinch zones.


In the implementations of FIGS. 22A-C, sleeve geometry can be optimized in a number of ways. Optimization can be achieved by accounting for the required change in length of the sleeve as the arm goes from a straight to a bent configuration. Several different implementations can be used. As is shown in the implementations of FIGS. 22A-B, the sleeve 400 can be fabricated such that excess material—which is required when the arm is bent—is stored/managed when the arm is straight FIG. 22A depicts a sleeve 400 having an “outer box” pleat 430. FIG. 22B depicts an “inner box” pleat 432. FIG. 22C depicts a “bent” sleeve 434 configuration. Alternatively, as is shown in FIG. 22C, by fabricating the sleeve with a bent configuration 434 such that the bend corresponds to the robots elbow bent to the middle of its range of motion the sleeve was improved to reduce the overall parasitic torque, that is the torque applied to the robot by the sleeve during actuation. Additionally, these implementations can provide an improved “fit,” meaning having reduced bunching and/or stretching (as is shown, for example, in FIG. 23B), and can be easier to clean. Each of these optimizations can also be applied to disposable sleeves.



FIGS. 23A-C depict various implementations of disposable sleeves 436. In various implementations, the disposable sleeves 436 must be easy to install and form a barrier to bio-burden and fluids. In certain circumstances, the sleeves 436 may be attached using integrated O-rings that snap into O-Ring grooves, as has been previously described. The sleeves 436 may also be attached using integrated adhesive strips 437 which attach it to the device 10. As shown in FIG. 23B, excessive bunching 436A can occur if the sleeves are not properly sized or optimized. In various implementations, adhesive strips 437 may be optimally located rotationally around the proximal termination section (such as at the waist of robot) to minimize material buildup in critical zones, such as where the camera exits the lumen.


In use, FIGS. 24A-D depict the insertion and operation of the device 10, according to exemplary implementations. As has been previously described and discussed further in relation to FIGS. 27A-C, these steps can be accomplished while the device 10 is visualized, for example on a console, using a laparoscope 448 inserted into the abdominal cavity from another port.


As shown in FIG. 24A, during insertion, the device 10 is first held above a gel port 450 that allows for the abdominal cavity to remain insulated while the gel port 450 seals around the irregular shape of the device 10. As shown in FIG. 24B, the device 10 is then is inserted though the gel port 450. The elbows 14C, 16C can then be bent to accommodate further insertion. The device 10 can then be inserted further until the arms 14, 16 are substantially within the abdominal cavity, as best shown in FIG. 24C. This then allows the device 10 to be rotated and moved to the desired position for surgery, as shown in FIG. 24D.



FIGS. 25A-B depicts a gel port 450, according to one implementation. In the embodiment of FIG. 25A, the gel port 450 has first 452 and second 454 openings, or “slits.” In certain implementations, the passage of the device 10 into the gel port 450 causes splaying of the arms 14, 16, which can result in patient injury. In these implementations, the slits 452, 454 facilitate the retention of the device 10 in an upright orientation. In certain of these implementations, the gel port 450 has a pair of semi-rigid corals 456A, 456B configured to urge the arms 14, 16 centrally and prevent splaying during insertion.


As shown in FIG. 25C, In certain implementations the robotic device 10 is clamped to (or otherwise coupled to) the distal end of the robot support arm 470. The proximal end of the support arm 470 is clamped or otherwise coupled to a standard support strut on the operating table. In this embodiment, the support arm 470 has 6 degrees of freedom, which are manually released by a single knob. In use, the user can release the support arm 470 by loosening the knob, move the robotic device 10 to a suitable position, then tighten the knob, thereby rigidizing the arm 470 and fixing the robotic device 10 in place. One example of a commercially-available support arm 470 is the Iron Intern™, made by Automated Medical Products Corp.


In use, according to one embodiment as shown in FIG. 26A, the system 500 can operate in the following fashion. A user—typically a surgeon—positions herself at the surgeon console 502. As discussed in further detail below, the console 502 can have a visual display 510, a touch screen 514 and input components such as foot input devices (also called “foot controllers”) 512, and hand input devices (also called “hand controllers”) 518. The user can operate the system 500 by operating the hand controllers 518 with her hands, the foot controllers 512 with her feet, and the touch screen (also referred to herein as the “graphical user interface” or “GUI”) 514 with her hands, while using the visual display 510 to view feedback from the camera 18 relating to the robot 10 positioned in the target cavity of the patient. The console 502 is coupled to the robot 10 and its components in three different ways in this embodiment. That is, the console 502 is coupled directly to the robot 10 itself via the cable 502 that carries both power and communications between the console 502 and the robot 10. Further, the console 502 is also coupled to the camera 18 on the robot 10 via the cable 504 that carries both power and communications between the console 502 and the camera 18. In addition, the console 502 is also coupled to the cautery end effectors 300A, 300B on the robot 10 via the cable 506 that carries power and communications between the console 502 and the cautery end effectors 300A, 300B (as discussed above in relation to FIGS. 16A-B). In other implementations, the console 502 can be coupled to the robot 10 via other connection components and/or via other robot components.


According to one embodiment as best shown in FIGS. 26B-26E, the console 502 allows the user to control the robotic device 10 using the hand controllers 518 and/or foot controllers 512. The hand controllers 518 and the foot controllers 512 can be used to control the arms and other components and functions of the robotic device 10. In various implementations, the device 10 is controlled by using the hand controllers 518 and/or foot controllers 512 to cause the device 10 to move in the same way as the hand controllers 518 and/or foot controllers 512 are moved. More specifically, for example, the right hand controller 518 can be used to actuate the right arm of the robotic device 10 such that the movement of the hand controller 518 causes the right arm of the device 10 to replicate or simulate the same motion. For example, if the right hand controller 518 is extended outward away from the user, the right arm of the device 10 is actuated to extend outward away from device 10 in the same direction. The left hand controller 518 and left arm of the robotic device 10 can operate in a similar fashion. This virtual connection and interaction between the console 502 and the robotic device 10 can be called a tele-operation (or “tele-op”) mode.


One embodiment of an exemplary GUI 530 is depicted in FIGS. 27A-C. In this implementation, various buttons 520 are provided which can be used to control the insertion, retraction, and operation of the device 10. More specifically, as shown in FIG. 27B of this embodiment, the user can select the specific operational page to be displayed on the GUI 530. If the user selects the “Insert” button as shown in FIG. 27B, then the insertion page is displayed as shown in FIG. 27A. Thus, the GUI 530 can provide the user with the ability to control settings and functions of the robotic surgical system. In various implementations, the touch screen can include settings for motion scaling, camera position, and indicators that show robot modes (cautery state, GUI state, etc) and the like. Further, in the tele-op mode as shown in FIG. 27A, the display 530 can depict a real-time robot animation (generally at 10) that displays the current configuration of the device 10, including the specific positions of the arms of the device 10.


In certain embodiments, the virtual connection between the console 502 and device 10 as described above can be interrupted using a “clutch.” In one specific implementation, the clutch can be activated using a button 520 on the GUI 530. Alternatively, the user can activate the clutch by depressing one of the foot pedals 512. The clutch is activated to break the virtual connection described above, thereby disconnecting the device 10 from the console 502 such that the device 10 and its components enter a “frozen” or “paused” state in which the components of the device 10 remain in the last position the components were in when the clutch was activated and until the clutch is deactivated. This clutch feature can be utilized for several different reasons. For example, the clutch feature can be used in an emergency pausing situation in which the device 10 components are moving toward a position which one or more of the components might damage the internal tissues of the patient and the clutch activation prevents that. In another example, the clutch feature can be used to reset the virtual connection in the same way that a computer mouse is lifted off the mousepad to reset the connection between the mouse and the cursor on the computer screen. In other words, the clutch feature can be used to reposition the hand controllers to a more desirable position while pausing the device 10.


Certain system embodiments disclosed or contemplated herein can also have hand controllers (such as controllers 518 discussed above) that feature haptic feedback. That is, the hand controllers (such as controllers 518) have haptic input devices, which are made up of motors operably coupled to the hand controllers such that the motors can be actuated to apply force to the controllers (such as controllers 518), thereby applying force to the user's hands that are grasping the controllers. This force applied to the user's hands that is created by the haptic input devices is called haptic feedback and is intended to provide information to the user. For example, one use of haptic feedback is to indicate to the user a collision between the robotic arms. In another example, the haptic feedback is used to indicate to the user that the robotic device or one of its components (such as one of the arms) is approaching or has reached its reachable or dexterous workspace.



FIGS. 28A and 28B provide a schematic representation of haptic feedback relating to the dexterous workspace of a robotic arm according to a specific embodiment. More specifically, FIG. 28A represents a two-dimensional “slice” of the workspace 600 of one arm of a robotic device of any system embodiment herein. That is, the image is a representation of the range of motion 600 of the distal end of the robotic arm in two dimensions (the x and y directions) such that it depicts all of the area 600 that the end effector of the robotic arm can reach when the arm can move in those two dimensions and the device body is kept motionless. As is shown in FIG. 28C, this workspace 602 can be extended into three dimensions, as the device 10 is capable of operating in the z-direction as well).


In these implementations, and as best shown in FIG. 28A, the full reachable workspace 600 is made up of both the exterior portion 602 and the interior portion 604 of the workspace 600. The interior portion 604 is the operational workspace 604 of the robotic arm. That is, it is the workspace 604 in which the arm is functional and operates optimally. The outer portion 602 is the undesirable or non-optimal workspace 602 of the robotic arm.


In this specific embodiment as best shown in FIG. 28B, the system has been configured to provide haptic feedback as the end effector of the arm reaches the outer points of the workspace 600. More specifically, the system, or a software component thereof, defines the haptic boundary 604 as the operational workspace 604. When the user moves the robotic arm such that the end effector is inside the haptic boundary 604, the haptic input devices apply no force to the hand controllers, thereby indicating to the user that the robotic arm is in the operational workspace 604. If the end effector of the arm moves outside of the haptic boundary 604 and into the non-optimal workspace 602, the haptic input devices provide force that urges the hand controller, and thus the robotic arm, back toward the closest point on the haptic boundary. In one embodiment, the force applied at the hand controller is proportional to the distance from the haptic boundary such that it feels to the user like a virtual spring is pushing the user's hand (and thus the robotic arm) back inside the boundary 604. Alternatively, it is understood that other models for forces can be created other than proportional distance.


Once possible use of the system is shown in FIG. 28D. In this implementation, the user can operate the hand controllers 518 (as shown in FIGS. 226A-E) in translation (box 620) and/or orientation (box 622) modes through a variety of steps. In certain implementations, controllers have seven degrees of haptic feedback relating to the position of the device in the surgical theater. Here, translation mode refers to x-, y-, and z-haptics, while the orientation mode refers to roll, pitch and yaw feedback, and the trigger operation can account for a seventh degree. In certain implementations, it is desirable to lock certain of these feedback movements—such as orientation—while leaving others—such as translation—free to be moved relative to the console. This selective locking allows for gross re-positioning of the user's hands and controllers without causing a corresponding movement of the device 10 within the surgical theater.


For example, the user can enter tele-op mode (box 630) such that the haptic input devices (described in detail in relation to FIGS. 26A-27C) are aligned to their center (0, 0, 0) regardless of the position of the device (box 632), while the orientation vectors (box 634) are aligned with the orientation of the device arms 14, 16 with respect to yaw, pitch and roll.


In tele-op mode, these positions are set (boxes 632 and 634), such any movements of the controllers will directly correspond with the movement of the device 10, and any force applied to the device 10 will cause a corresponding force to be applied back to the user through the controllers. However, in certain situations, the user may desire to re-orient the hand controllers relative to the console without causing a corresponding change in the movement of the device.


When the system is paused (box 636) the system is “locked” (boxes 638 and 640), such that the hand controllers 518 are locked in place. No movement or commands to the device 10 are being sent, such that the device 10 holds position regardless of what the user does to the hand controllers, meaning that even if the user overpowers the haptic locks and moves the hand controllers, the robot will not move.


In further implementations, to move the controllers independently, the user can engage the clutch (box 642) so as to disengage the translation of the controllers only (box 644) while the device arms 14, 16 and controllers maintain a fixed orientation (box 646). When the clutch 512 is disengaged (box 648) the robot and hand controllers are then virtually re-connected, so as to again fix translation and orientation between the controllers and device.


In these implementations, the workspace can be defined (box 650) when the device 10 is positioned. As discussed above, the translational movement of the arms and controllers is limited by the workspace boundry (box 650), and the orientation movements are aligned with a valid vector (box 652) to ensure safety and precision.


In certain implementations, the haptic lock can be also interrupted by other functions such as “camera clutch” (box 654), where the two hand controllers can move together. In these implementations, it may be necessary to re-orient the hand controllers and/or device arms relative to the position and/or orientation of the camera. That is, as would be understood, because the camera is capable of pan and tilt functions, the camera has a specific frame of reference with regard to the workspace and device 10. In certain implementations, the console depicts this frame of reference, and the translation and/or orientation of the arms and controllers are fixed relative to the camera orientation. When the camera is moved, it may be necessary to re-orient the controllers and/or arms relative to the second camera frame of reference, which can be designated by a. Accordingly, it is possible to urge the hand controls in various directions (such as horizontally relative to the ground), but cause a corresponding vertical motion of the robot arms, in circumstances where the device and camera are pointed straight down. Other versions of this workflow are possible.



FIGS. 29A-D show another possible use of haptic feedback in the force dimension workspace 600. In these implementations, the motion—translation and/or orientation—of the hand controllers have certain limits. In certain embodiments, and as shown in FIGS. 29A-C, the haptic feedback system described above can be used to indicate to the user that the hand controllers have been moved to a limit 602 of their motion. Here another virtual spring could be implemented, or a visual alert, or audible alert, or vibratory alert could be provided.



FIGS. 30A-B show an operator detection system 700, which can be operationally integrated with any of the preceding embodiments as part of the user input device or controller 701. In these implementations, the operator detection system 700 is configured to detect the presence of the user 702 so as to prevent unintended motion of the device 10. One implementation is to use a mechanical switch 704 that is engaged as the user 702 inserts his/her hands 702 into contact with the user input device 701 and/or applies pressure to the controller sides 706, 708. Various implementations can also take the form of a capacitive sensor, a pressure sensor, and optical sensor, an optical beam break sensor, or many other forms. In various alternate implementations, the operator detection system 700 can utilize voice activation and/or a vision system.


The various embodiments are disclosed in additional detail in the attached figures, which include some written description therein.


Further, according to certain embodiments, a device as shown and described in the attached figures is inserted into the patient using the following procedure.


First, an incision is made through the abdominal wall using standard techniques. In this embodiment an incision of length 2.5″ is required to create a suitable orifice for the system to pass through.


Next, a retractor is placed in the incision. In this embodiment, an Applied Medical Alexis Wound Retractor (http://www.appliedmedical.com/Products/Alexis.aspx) is utilized. It consists of a thin walled (<0.005″) flexible tubular membrane with rigid ring shaped end caps. Once the distal ring is inserted into the patient, the proximal ring is rolled to take up the excess slack in tube and pull the wound open.


Then, a port is placed on the retractor. In this embodiment, a modified Applied Medical Gel port (http://www.appliedmedical.com/Products/Gelport.aspx) is utilized. The port is capable of maintain a pressure differential such that insufflation of the abdominal cavity may be achieved. The port is capable of having items (ie robot) plunged through it while maintaining this pressure differential/gas seal. This port consists of a rigid ring which mechanically clamps to the external rigid ring of the retractor. This clamp is capable of sealing to the ring, preserving insufflation pressure. The port further consists of a pair of circular gel membranes. Each membrane is ˜0.75″ thick. Each membrane has a slit through it. The slit has length of ˜50% of the membrane diameter. When assembled, the slit of membrane 1 is rotated 90 degrees with respect to the slit of membrane 2. Due to the gel/conforming nature of the membranes, a seal is maintained against oddly shaped objects as they pass through the slits of the membranes and into the abdominal cavity.


According to one alternative embodiment relating to the port, a lattice of non-elastic cords is embedded in the membranes, mitigating doming/blowout as a result of the internal pressure. In a further alternative, a thin film of a rigid/puncture resistant polymer was deposited at the interface of membrane 1 and 2. The purpose of this polymer is to prevent the end effectors of the robot from puncturing membrane 2 after it passes through the slit in membrane 1.


Once the retractor and gel port are in place, the robot may be inserted into the patient.


Next, a camera (a robot camera as disclosed in the attached figures or an auxiliary camera) is inserted through an auxiliary port to view the insertion.


Next, the insertion/extraction mode of the robot is activated from the GUI.


After that, the robot and/or system determines a path from its current state to its insertion pose (arms straight down), and the operator steps through this path to achieve the required pose.


Subsequently, the operator inserts the robot into the patient (through the gel port and through the retractor port) until the elbow joints of the robot clear the interior surface of the abdominal wall.


After that, the operator steps through the insertion path until the elbows reach their end point (90 degrees). The operator then further inserts the robot into the patient until the shoulder joints clear the interior surface of the abdominal wall. The operator continues to step through the insertion path until the robot achieves its “ready” pose (arms in a nominal operating position), at which point, the surgical procedure can proceed.


When the procedure is complete, device extraction follows the above sequence in reverse.


Although the present invention has been described with reference to preferred embodiments, persons skilled in the art will recognize that changes may be made in form and detail without departing from the spirit and scope of the invention.

Claims
  • 1. A robotic surgical system, comprising: a robotic surgical device comprising: an elongate device body having proximal and distal ends and comprising: a camera lumen defined within the device body, the camera lumen comprising: a proximal lumen opening; a socket defined distally of the proximal lumen opening, the socket comprising a first diameter and a first coupling component; an extended portion defined distally of the socket, the extended portion having a second, smaller diameter; and a distal lumen opening in the distal end of the device body, the distal lumen opening defined at a distal end of the extended portion; a seal structure defined distally of the socket and proximally of the extended portion, the seal structure comprising a ring seal and a one-way seal; first and second shoulder joints operably coupled to the distal end of the device body; a first segmented robotic arm operably coupled to the first shoulder joint; and a second segmented robotic arm operably coupled to the second shoulder joint; and a camera component, comprising: a handle comprising: a distal end configured to be positionable within the socket; a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket; and at least one actuator disposed within the handle; and an elongate tube operably coupled to the handle, wherein the elongate tube is configured and sized to be positionable through the camera lumen, the elongate tube comprising: a rigid section defining a lumen; an optical section; and a flexible section operably coupling the optical section to the rigid section; wherein the elongate tube has a length such that at least the optical section is configured to extend distally from the distal lumen opening when the camera component is positioned through the camera lumen, and the at least one actuator disposed within the handle is constructed and arranged to perform camera component pan and camera component tilt.
  • 2. The robotic surgical system of claim 1, wherein the seal structure comprises a ring-seal retention component, wherein the ring seal is retained within the ring-seal retention component.
  • 3. The robotic surgical system of claim 2, wherein the ring-seal retention component comprises at least one protrusion extending from an outer wall of the ring-seal retention component.
  • 4. The robotic surgical system of claim 3, wherein the socket further comprises a channel defined in an inner wall of the socket, wherein the channel is configured to receive the at least one protrusion.
  • 5. The robotic surgical system of claim 1, wherein the handle comprises a controller configured to operate the camera component.
  • 6. The robotic surgical system of claim 1, wherein the distal lumen opening is positioned between the first and second shoulder joints.
  • 7. The robotic surgical system of claim 1, wherein the optical section is configured to be tiltable at the flexible section in relation to the rigid section, wherein the optical section has a straight configuration and a tilted configuration.
  • 8. The robotic surgical system of claim 1, wherein the elongate tube is configured to be rotatable in relation to the handle.
  • 9. The robotic surgical system of claim 1, wherein the socket further comprises an inner wall comprising a channel configured to receive an insertion device.
  • 10. A robotic surgical system, comprising: (a) a robotic surgical device comprising: (i) a device body comprising: (A) a distal end;(B) a proximal end; and(C) a camera lumen defined within the device body, the camera lumen comprising: (1) a socket defined in the camera lumen, the socket comprising a first coupling component; and(2) a seal structure defined distally of the socket, wherein the seal structure comprises a ring seal and a one-way seal;(ii) first and second shoulder joints operably coupled to the distal end of the device body;(iii) a first robotic arm operably coupled to the first shoulder joint; and(iv) a second robotic arm operably coupled to the second shoulder joint;(b) a camera component, comprising: (i) a handle comprising: (A) a distal end configured to be positionable within the socket;(B) a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket; and(C) at least one actuator disposed within the handle; and(ii) an elongate tube operably coupled to the handle, wherein the elongate tube is configured and sized to be positionable through the camera lumen, the elongate tube comprising: (A) a rigid section;(B) an optical section; and(C) a flexible section operably coupling the optical section to the rigid section,wherein the elongate tube has a length such that at least the optical section is configured to extend distally from a distal lumen opening of the camera lumen when the camera component is positioned through the camera lumen, andthe at least one actuator disposed within the handle is constructed and arranged to perform camera component pan and camera component tilt.
  • 11. The robotic surgical system of claim 10, wherein the first robotic arm further comprises: (a) a first arm upper arm;(b) a first arm elbow joint; and(c) a first arm lower arm,wherein the first arm upper arm is configured to be capable of roll, pitch and yaw relative to the first shoulder joint and the first arm lower arm is configured to be capable of yaw relative to the first arm upper arm by way of the first arm elbow joint.
  • 12. The surgical robotic system of claim 11, wherein the first robotic arm further comprises at least one first arm actuator disposed within the first robotic arm.
  • 13. The robotic surgical system of claim 11, wherein the second robotic arm further comprises: (a) a second arm upper arm;(b) a second arm elbow joint; and(c) a second arm lower arm,wherein the second arm upper arm is configured to be capable of roll, pitch and yaw relative to the second shoulder joint and the second arm lower arm is configured to be capable of yaw relative to the second arm upper arm by way of the second arm elbow joint.
  • 14. The surgical robotic system of claim 13, wherein the second robotic arm further comprises at least one second arm actuator disposed within the second robotic arm.
  • 15. The robotic surgical system of claim 10, wherein the first shoulder joint comprises a first rotatable housing rotatably coupled to the distal end of the device body and the second shoulder joint comprises a second rotatable housing rotatably coupled to the distal end of the device body.
  • 16. A robotic surgical system comprising: (a) a robotic surgical device comprising: (i) an elongate device body having proximal and distal ends and comprising a camera lumen defined within the device body, the camera lumen comprising: (A) a proximal lumen opening;(B) a socket defined distally of the proximal lumen opening, the socket comprising a first diameter and a first coupling component;(C) an extended section defined distally of the socket, the extended section having a second, smaller diameter; and(D) a distal lumen opening in the distal end of the device body, the distal lumen opening defined at a distal end of the extended section;(ii) first and second shoulder joints operably coupled to the distal end of the device body, the first shoulder joint comprising a first rotatable housing rotatably coupled to the distal end of the device body and the second shoulder joint comprising a second rotatable housing rotatably coupled to the distal end of the device body;(iii) a first segmented robotic arm operably coupled to the first shoulder joint; and(iv) a second segmented robotic arm operably coupled to the second shoulder joint; and(v) a first drivetrain comprising: (A) a first body actuator disposed within the device body; and(B) a first transmission shaft operably coupled to the first body actuator, wherein the first transmission shaft is operably coupled to the first rotatable housing;(vi) a second drivetrain comprising: (A) a second body actuator disposed within the device body; and(B) a second transmission shaft operably coupled to the second body actuator, wherein the second transmission shaft is operably coupled to a first shoulder link;(vii) a third drivetrain comprising: (A) a third body actuator disposed within the device body; and(B) a third transmission shaft operably coupled to the third body actuator, wherein the third transmission shaft is operably coupled to the second rotatable housing; and(viii) a fourth drivetrain comprising: (A) a fourth body actuator disposed within the device body; and(B) a fourth transmission shaft operably coupled to the fourth body actuator, wherein the fourth transmission shaft is operably coupled to a second shoulder link;(b) a camera component, comprising: (i) a handle comprising: (A) a distal end configured to be positionable within the socket;(B) a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket; and(C) at least one actuator disposed within the handle; and(ii) an elongate tube operably coupled to the handle, wherein the elongate tube is configured and sized to be positionable through the camera lumen, the elongate tube comprising: (A) a rigid section defining a lumen;(B) an optical section; and(C) a flexible section operably coupling the optical section to the rigid section;wherein the elongate tube has a length such that at least the optical section is configured to extend distally from the distal lumen opening when the camera component is positioned through the camera lumen, and the at least one actuator disposed within the handle is constructed and arranged to perform camera component pan and camera component tilt.
  • 17. A robotic surgical system, comprising: (a) a robotic surgical device comprising: (i) a device body comprising: (A) a distal end;(B) a proximal end; and(C) a camera lumen defined within the device body, the camera lumen comprising a socket defined in the camera lumen, the socket comprising a first coupling component;(ii) first and second shoulder joints operably coupled to the distal end of the device body;(iii) a first robotic arm operably coupled to the first shoulder joint; and(iv) a second robotic arm operably coupled to the second shoulder joint;(v) a first drivetrain comprising: (A) a first body actuator disposed within the device body; and(B) a first transmission shaft operably coupled to the first body actuator, wherein the first transmission shaft is operably coupled to the first rotatable housing;(vi) a second drivetrain comprising: (A) a second body actuator disposed within the device body; and(B) a second transmission shaft operably coupled to the second body actuator, wherein the second transmission shaft is operably coupled to a first shoulder link;(vii) a third drivetrain comprising: (A) a third body actuator disposed within the device body; and(B) a third transmission shaft operably coupled to the third body actuator, wherein the third transmission shaft is operably coupled to the second rotatable housing; and(viii) a fourth drivetrain comprising: (A) a fourth body actuator disposed within the device body; and(B) a fourth transmission shaft operably coupled to the fourth body actuator, wherein the fourth transmission shaft is operably coupled to a second shoulder link;(b) a camera component, comprising: (i) a handle comprising: (A) a distal end configured to be positionable within the socket;(B) a second coupling component configured to releasably couple with the first coupling component, thereby releasably locking the handle into the socket portion; and(C) at least one actuator disposed within the handle; and(ii) an elongate tube operably coupled to the handle, wherein the elongate tube is configured and sized to be positionable through the camera lumen, the elongate tube comprising: (A) a rigid section;(B) an optical section; and(C) a flexible section operably coupling the optical section to the rigid section,wherein the elongate tube has a length such that at least the optical section is configured to extend distally from a distal lumen opening of the camera lumen when the camera component is positioned through the camera lumen, andthe at least one actuator disposed within the handle is constructed and arranged to perform camera component pan and camera component tilt.
CROSS-REFERENCE TO RELATED APPLICATION(S)

This application claims the benefit under 35 U.S.C. § 119(e) to U.S. Provisional Application 62/200,563, filed Aug. 3, 2015 and entitled “Robotic Surgical Devices, Systems, and Related Methods,” which is hereby incorporated herein by reference in its entirety.

US Referenced Citations (496)
Number Name Date Kind
3870264 Robinson Mar 1975 A
3989952 Timberlake et al. Nov 1976 A
4246661 Pinson Jan 1981 A
4258716 Sutherland Mar 1981 A
4278077 Mizumoto Jul 1981 A
4538594 Boebel et al. Sep 1985 A
4568311 Miyaki Feb 1986 A
4623183 Amori Nov 1986 A
4736645 Zimmer Apr 1988 A
4771652 Zimmer Sep 1988 A
4852391 Ruch et al. Aug 1989 A
4896015 Taboada et al. Jan 1990 A
4897014 Tietze Jan 1990 A
4922755 Oshiro et al. May 1990 A
4922782 Kawai May 1990 A
4990050 Tsuge et al. Feb 1991 A
5019968 Wang et al. May 1991 A
5108140 Bartholet Apr 1992 A
5172639 Wiesman et al. Dec 1992 A
5176649 Wakabayashi Jan 1993 A
5178032 Zona et al. Jan 1993 A
5187032 Sasaki et al. Feb 1993 A
5187796 Wang et al. Feb 1993 A
5195388 Zona et al. Mar 1993 A
5201325 McEwen et al. Apr 1993 A
5217003 Wilk Jun 1993 A
5263382 Brooks et al. Nov 1993 A
5271384 McEwen et al. Dec 1993 A
5284096 Pelrine et al. Feb 1994 A
5297443 Wentz Mar 1994 A
5297536 Wilk Mar 1994 A
5304899 Sasaki et al. Apr 1994 A
5307447 Asano et al. Apr 1994 A
5353807 DeMarco Oct 1994 A
5363935 Schempf et al. Nov 1994 A
5382885 Salcudean et al. Jan 1995 A
5388528 Pelrine et al. Feb 1995 A
5436542 Petelin et al. Jul 1995 A
5441494 Oritz Aug 1995 A
5458131 Wilk Oct 1995 A
5458583 McNeely et al. Oct 1995 A
5458598 Feinberg et al. Oct 1995 A
5471515 Fossum et al. Nov 1995 A
5515478 Wang May 1996 A
5524180 Wang et al. Jun 1996 A
5553198 Wang et al. Sep 1996 A
5562448 Mushabac Oct 1996 A
5588442 Scovil et al. Dec 1996 A
5620417 Jang et al. Apr 1997 A
5623582 Rosenberg Apr 1997 A
5624380 Shuichi et al. Apr 1997 A
5624398 Smith et al. Apr 1997 A
5632761 Smith et al. May 1997 A
5645520 Nakamura et al. Jul 1997 A
5657429 Wang et al. Aug 1997 A
5657584 Hamlin Aug 1997 A
5672168 de la Torre et al. Sep 1997 A
5674030 Sigel Oct 1997 A
5728599 Rosteker et al. Mar 1998 A
5736821 Suyaman et al. Apr 1998 A
5754741 Wang et al. May 1998 A
5762458 Wang et al. Jun 1998 A
5769640 Jacobus et al. Jun 1998 A
5791231 Cohn et al. Aug 1998 A
5792135 Madhani et al. Aug 1998 A
5797538 Heaton et al. Aug 1998 A
5797900 Madhani et al. Aug 1998 A
5807377 Madhani et al. Sep 1998 A
5808665 Green Sep 1998 A
5815640 Wang et al. Sep 1998 A
5825982 Wright et al. Oct 1998 A
5841950 Wang et al. Nov 1998 A
5845646 Lemelson Dec 1998 A
5855583 Wang et al. Jan 1999 A
5876325 Mizuno et al. Mar 1999 A
5878193 Wang et al. Mar 1999 A
5878783 Smart Mar 1999 A
5895417 Pomeranz et al. Apr 1999 A
5906591 Dario et al. May 1999 A
5907664 Wang et al. May 1999 A
5910129 Koblish et al. Jun 1999 A
5911036 Wright et al. Jun 1999 A
5971976 Wang et al. Oct 1999 A
5993467 Yoon Nov 1999 A
6001108 Wang et al. Dec 1999 A
6007550 Wang et al. Dec 1999 A
6030365 Laufer Feb 2000 A
6031371 Smart Feb 2000 A
6058323 Lemelson May 2000 A
6063095 Wang et al. May 2000 A
6066090 Yoon May 2000 A
6102850 Wang et al. Aug 2000 A
6106521 Blewett Aug 2000 A
6107795 Smart Aug 2000 A
6132368 Cooper Oct 2000 A
6132441 Grace Oct 2000 A
6139563 Cosgrove, III et al. Oct 2000 A
6156006 Brosens et al. Dec 2000 A
6159146 El Gazayerli Dec 2000 A
6162171 Ng et al. Dec 2000 A
D438617 Cooper et al. Mar 2001 S
6206903 Ramans Mar 2001 B1
D441076 Cooper et al. Apr 2001 S
6223100 Green Apr 2001 B1
D441862 Cooper et al. May 2001 S
6238415 Sepetka et al. May 2001 B1
6240312 Alfano et al. May 2001 B1
6241730 Alby Jun 2001 B1
6244809 Wang et al. Jun 2001 B1
6246200 Blumenkranz et al. Jun 2001 B1
D444555 Cooper et al. Jul 2001 S
6286514 Lemelson Sep 2001 B1
6292678 Hall et al. Sep 2001 B1
6293282 Lemelson Sep 2001 B1
6296635 Smith et al. Oct 2001 B1
6309397 Julian et al. Oct 2001 B1
6309403 Minoret et al. Oct 2001 B1
6312435 Wallace et al. Nov 2001 B1
6321106 Lemelson Nov 2001 B1
6327492 Lemelson Dec 2001 B1
6331181 Tierney et al. Dec 2001 B1
6346072 Cooper Feb 2002 B1
6352503 Matsui et al. Mar 2002 B1
6364888 Niemeyer et al. Apr 2002 B1
6371952 Madhani et al. Apr 2002 B1
6394998 Wallace et al. May 2002 B1
6398726 Ramans et al. Jun 2002 B1
6400980 Lemelson Jun 2002 B1
6408224 Lemelson Jun 2002 B1
6424885 Niemeyer et al. Jul 2002 B1
6432112 Brock et al. Aug 2002 B2
6436107 Wang et al. Aug 2002 B1
6441577 Blumenkranz et al. Aug 2002 B2
6450104 Grant et al. Sep 2002 B1
6450992 Cassidy, Jr. Sep 2002 B1
6451027 Cooper et al. Sep 2002 B1
6454758 Thompson et al. Sep 2002 B1
6459926 Nowlin et al. Oct 2002 B1
6463361 Wang et al. Oct 2002 B1
6468203 Belson Oct 2002 B2
6468265 Evans et al. Oct 2002 B1
6470236 Ohtsuki Oct 2002 B2
6491691 Morley et al. Dec 2002 B1
6491701 Nemeyer et al. Dec 2002 B2
6493608 Niemeyer et al. Dec 2002 B1
6496099 Wang et al. Dec 2002 B2
6508413 Bauer et al. Jan 2003 B2
6512345 Borenstein Jan 2003 B2
6522906 Salisbury, Jr. et al. Feb 2003 B1
6544276 Azizi Apr 2003 B1
6548982 Papanikolopoulos et al. Apr 2003 B1
6554790 Moll Apr 2003 B1
6565554 Niemeyer May 2003 B1
6574355 Green Jun 2003 B2
6587750 Gerbi et al. Jul 2003 B2
6591239 McCall et al. Jul 2003 B1
6594552 Nowlin et al. Jul 2003 B1
6610007 Belson et al. Aug 2003 B2
6620173 Gerbi et al. Sep 2003 B2
6642836 Wang et al. Nov 2003 B1
6645196 Nixon et al. Nov 2003 B1
6646541 Wang et al. Nov 2003 B1
6648814 Kim et al. Nov 2003 B2
6659939 Moll et al. Dec 2003 B2
6661571 Shioda et al. Dec 2003 B1
6671581 Niemeyer et al. Dec 2003 B2
6676684 Morley et al. Jan 2004 B1
6684129 Salisbury, Jr. et al. Jan 2004 B2
6685648 Flaherty et al. Feb 2004 B2
6685698 Morley et al. Feb 2004 B2
6687571 Byme et al. Feb 2004 B1
6692485 Brock et al. Feb 2004 B1
6699177 Wang et al. Mar 2004 B1
6699235 Wallace et al. Mar 2004 B2
6702734 Kim et al. Mar 2004 B2
6702805 Stuart Mar 2004 B1
6714839 Salisbury, Jr. et al. Mar 2004 B2
6714841 Wright et al. Mar 2004 B1
6719684 Kim et al. Apr 2004 B2
6720988 Gere et al. Apr 2004 B1
6726699 Wright et al. Apr 2004 B1
6728599 Wright et al. Apr 2004 B2
6730021 Vassiliades, Jr. et al. May 2004 B2
6731988 Green May 2004 B1
6746443 Morley et al. Jun 2004 B1
6764441 Chiel et al. Jul 2004 B2
6764445 Ramans et al. Jul 2004 B2
6766204 Niemeyer et al. Jul 2004 B2
6770081 Cooper et al. Aug 2004 B1
6774597 Borenstein Aug 2004 B1
6776165 Jin Aug 2004 B2
6780184 Tanrisever Aug 2004 B2
6783524 Anderson et al. Aug 2004 B2
6785593 Wang et al. Aug 2004 B2
6788018 Blumenkranz Sep 2004 B1
6792663 Krzyzanowski Sep 2004 B2
6793653 Sanchez et al. Sep 2004 B2
6799065 Niemeyer Sep 2004 B1
6799088 Wang et al. Sep 2004 B2
6801325 Farr et al. Oct 2004 B2
6804581 Wang et al. Oct 2004 B2
6810281 Brock et al. Oct 2004 B2
6817972 Snow Nov 2004 B2
6817974 Cooper et al. Nov 2004 B2
6817975 Farr et al. Nov 2004 B1
6820653 Schempf et al. Nov 2004 B1
6824508 Kim et al. Nov 2004 B2
6824510 Kim et al. Nov 2004 B2
6832988 Sprout Dec 2004 B2
6832996 Woloszko et al. Dec 2004 B2
6836703 Wang et al. Dec 2004 B2
6837846 Jaffe et al. Jan 2005 B2
6837883 Moll et al. Jan 2005 B2
6839612 Sanchez et al. Jan 2005 B2
6840938 Morley et al. Jan 2005 B1
6852107 Wang et al. Feb 2005 B2
6858003 Evans et al. Feb 2005 B2
6860346 Burt et al. Mar 2005 B2
6860877 Sanchez et al. Mar 2005 B1
6866671 Tierney et al. Mar 2005 B2
6870343 Borenstein et al. Mar 2005 B2
6871117 Wang et al. Mar 2005 B2
6871563 Choset et al. Mar 2005 B2
6879880 Nowlin et al. Apr 2005 B2
6892112 Wang et al. May 2005 B2
6899705 Niemeyer May 2005 B2
6902560 Morley et al. Jun 2005 B1
6905460 Wang et al. Jun 2005 B2
6905491 Wang et al. Jun 2005 B1
6911916 Wang et al. Jun 2005 B1
6917176 Schempf et al. Jul 2005 B2
6933695 Blumenkranz Aug 2005 B2
6936001 Snow Aug 2005 B1
6936003 Iddan Aug 2005 B2
6936042 Wallace et al. Aug 2005 B2
6943663 Wang et al. Sep 2005 B2
6949096 Davison et al. Sep 2005 B2
6951535 Ghodoussi et al. Oct 2005 B2
6965812 Wang et al. Nov 2005 B2
6974411 Belson Dec 2005 B2
6974449 Niemeyer Dec 2005 B2
6979423 Moll Dec 2005 B2
6984203 Tartaglia et al. Jan 2006 B2
6984205 Gazdzinski Jan 2006 B2
6991627 Madhani et al. Jan 2006 B2
6993413 Sunaoshi Jan 2006 B2
6994703 Wang et al. Feb 2006 B2
6994708 Manzo Feb 2006 B2
6997908 Carrillo, Jr. et al. Feb 2006 B2
7025064 Wang et al. Apr 2006 B2
7027892 Wang et al. Apr 2006 B2
7033344 Imran Apr 2006 B2
7039453 Mullick May 2006 B2
7042184 Oleynikov et al. May 2006 B2
7048745 Tierney et al. May 2006 B2
7053752 Wang et al. May 2006 B2
7063682 Whayne et al. Jun 2006 B1
7066879 Fowler et al. Jun 2006 B2
7066926 Wallace et al. Jun 2006 B2
7074179 Wang et al. Jul 2006 B2
7077446 Kameda et al. Jul 2006 B2
7083571 Wang et al. Aug 2006 B2
7083615 Peterson et al. Aug 2006 B2
7087049 Nowlin et al. Aug 2006 B2
7090683 Brock et al. Aug 2006 B2
7097640 Wang et al. Aug 2006 B2
7105000 McBrayer Sep 2006 B2
7107090 Salisbury, Jr. et al. Sep 2006 B2
7109678 Kraus et al. Sep 2006 B2
7118582 Wang et al. Oct 2006 B1
7121781 Sanchez et al. Oct 2006 B2
7125403 Julian et al. Oct 2006 B2
7126303 Farritor et al. Oct 2006 B2
7147650 Lee Dec 2006 B2
7155315 Niemeyer et al. Dec 2006 B2
7169141 Brock et al. Jan 2007 B2
7182025 Ghorbel et al. Feb 2007 B2
7182089 Ries Feb 2007 B2
7199545 Oleynikov et al. Apr 2007 B2
7206626 Quaid, III Apr 2007 B2
7206627 Abovitz et al. Apr 2007 B2
7210364 Ghorbel et al. May 2007 B2
7214230 Brock et al. May 2007 B2
7217240 Snow May 2007 B2
7239940 Wang et al. Jul 2007 B2
7250028 Julian et al. Jul 2007 B2
7259652 Wang et al. Aug 2007 B2
7273488 Nakamura et al. Sep 2007 B2
7311107 Harel et al. Dec 2007 B2
7339341 Oleynikov et al. Mar 2008 B2
7372229 Farritor et al. May 2008 B2
7447537 Funda et al. Nov 2008 B1
7492116 Oleynikov et al. Feb 2009 B2
7566300 Devierre et al. Jul 2009 B2
7574250 Niemeyer Aug 2009 B2
7637905 Saadat et al. Dec 2009 B2
7645230 Mikkaichi et al. Jan 2010 B2
7655004 Long Feb 2010 B2
7670329 Flaherty et al. Mar 2010 B2
7731727 Sauer Jun 2010 B2
7762825 Burbank et al. Jul 2010 B2
7772796 Farritor et al. Aug 2010 B2
7785251 Wilk Aug 2010 B2
7785333 Miyamoto et al. Aug 2010 B2
7789825 Nobis et al. Sep 2010 B2
7794494 Sahatjian et al. Sep 2010 B2
7865266 Moll et al. Jan 2011 B2
7960935 Farritor et al. Jun 2011 B2
8021358 Doyle et al. Sep 2011 B2
8179073 Farritor et al. May 2012 B2
8353897 Doyle et al. Jan 2013 B2
9089353 Farritor Jul 2015 B2
20010018591 Brock et al. Aug 2001 A1
20010049497 Kalloo et al. Dec 2001 A1
20020003173 Bauer et al. Jan 2002 A1
20020013601 Nobles et al. Jan 2002 A1
20020026186 Woloszka et al. Feb 2002 A1
20020038077 de la Torre et al. Mar 2002 A1
20020065507 Azizi May 2002 A1
20020091374 Cooper Jul 2002 A1
20020103417 Gazdzinski Aug 2002 A1
20020111535 Kim et al. Aug 2002 A1
20020120254 Julien et al. Aug 2002 A1
20020128552 Nowlin et al. Sep 2002 A1
20020140392 Borenstein et al. Oct 2002 A1
20020147487 Sundquist et al. Oct 2002 A1
20020151906 Demarais et al. Oct 2002 A1
20020156347 Kim et al. Oct 2002 A1
20020171385 Kim et al. Nov 2002 A1
20020173700 Kim et al. Nov 2002 A1
20020190682 Schempf et al. Dec 2002 A1
20030020810 Takizawa et al. Jan 2003 A1
20030045888 Brock et al. Mar 2003 A1
20030065250 Chiel et al. Apr 2003 A1
20030089267 Ghorbel et al. May 2003 A1
20030092964 Kim et al. May 2003 A1
20030097129 Davison et al. May 2003 A1
20030100817 Wang et al. May 2003 A1
20030114731 Cadeddu et al. Jun 2003 A1
20030135203 Wang et al. Jul 2003 A1
20030139742 Wampler et al. Jul 2003 A1
20030144656 Ocel et al. Jul 2003 A1
20030167000 Mullick Sep 2003 A1
20030172871 Scherer Sep 2003 A1
20030179308 Zamorano et al. Sep 2003 A1
20030181788 Yokoi et al. Sep 2003 A1
20030229268 Uchiyama et al. Dec 2003 A1
20030230372 Schmidt Dec 2003 A1
20040117032 Roth et al. Jan 2004 A1
20040024311 Quaid Feb 2004 A1
20040034282 Quaid Feb 2004 A1
20040034283 Quaid Feb 2004 A1
20040034302 Abovitz et al. Feb 2004 A1
20040050394 Jin Mar 2004 A1
20040070822 Shioda et al. Apr 2004 A1
20040099175 Perrot et al. May 2004 A1
20040102772 Baxter et al. May 2004 A1
20040106916 Quaid et al. Jun 2004 A1
20040111113 Nakamura et al. Jun 2004 A1
20040138525 Saadat Jul 2004 A1
20040138552 Harel et al. Jul 2004 A1
20040140786 Borenstein Jul 2004 A1
20040153057 Davison Aug 2004 A1
20040173116 Ghorbel et al. Sep 2004 A1
20040176664 Iddan Sep 2004 A1
20040215331 Chew et al. Oct 2004 A1
20040225229 Viola Nov 2004 A1
20040254680 Sunaoshi Dec 2004 A1
20040267326 Ocel et al. Dec 2004 A1
20050014994 Fowler et al. Jan 2005 A1
20050021069 Feuer et al. Jan 2005 A1
20050029978 Oleynikov et al. Feb 2005 A1
20050043583 Killmann et al. Feb 2005 A1
20050049462 Kanazawa Mar 2005 A1
20050054901 Yoshino Mar 2005 A1
20050054902 Konno Mar 2005 A1
20050064378 Toly Mar 2005 A1
20050065400 Banik et al. Mar 2005 A1
20050083460 Hattori et al. Apr 2005 A1
20050095650 Khalili et al. May 2005 A1
20050096502 Khalili May 2005 A1
20050143644 Gilad et al. Jun 2005 A1
20050154376 Riviere et al. Jul 2005 A1
20050165449 Cadeddu et al. Jul 2005 A1
20050177026 Hoeg Aug 2005 A1
20050283137 Doyle et al. Dec 2005 A1
20050288555 Binmoeller Dec 2005 A1
20050288665 Woloszko Dec 2005 A1
20060020272 Gildenberg Jan 2006 A1
20060046226 Bergler et al. Mar 2006 A1
20060119304 Farritor et al. Jun 2006 A1
20060149135 Paz Jul 2006 A1
20060152591 Lin Jul 2006 A1
20060155263 Lipow Jul 2006 A1
20060195015 Mullick et al. Aug 2006 A1
20060196301 Oleynikov et al. Sep 2006 A1
20060198619 Oleynikov et al. Sep 2006 A1
20060241570 Wilk Oct 2006 A1
20060241732 Denker et al. Oct 2006 A1
20060253109 Chu Nov 2006 A1
20060258954 Timberlake Nov 2006 A1
20070032701 Fowler et al. Feb 2007 A1
20070043397 Ocel et al. Feb 2007 A1
20070055342 Wu et al. Mar 2007 A1
20070080658 Farritor et al. Apr 2007 A1
20070106113 Ravo May 2007 A1
20070123748 Meglan May 2007 A1
20070142725 Hardin et al. Jun 2007 A1
20070156019 Larkin et al. Jul 2007 A1
20070156211 Ferren et al. Jul 2007 A1
20070167955 De La Menardiere et al. Jul 2007 A1
20070225633 Ferren et al. Sep 2007 A1
20070225634 Ferren et al. Sep 2007 A1
20070241714 Oleynikov et al. Oct 2007 A1
20070244520 Ferren et al. Oct 2007 A1
20070250064 Darois et al. Oct 2007 A1
20070255273 Fernandez et al. Nov 2007 A1
20080004634 Farritor et al. Jan 2008 A1
20080015565 Davison Jan 2008 A1
20080015566 Livneh Jan 2008 A1
20080033569 Ferren et al. Feb 2008 A1
20080045803 Williams Feb 2008 A1
20080058835 Farritor et al. Mar 2008 A1
20080058989 Oleynikov et al. Mar 2008 A1
20080071289 Cooper Mar 2008 A1
20080103440 Ferren et al. May 2008 A1
20080109014 Pena May 2008 A1
20080111513 Farritor et al. May 2008 A1
20080119870 Williams et al. May 2008 A1
20080132890 Woloszko et al. Jun 2008 A1
20080161804 Rioux et al. Jul 2008 A1
20080164079 Ferren et al. Jul 2008 A1
20080183033 Bern et al. Jul 2008 A1
20080221591 Farritor et al. Sep 2008 A1
20080269557 Marescaux et al. Oct 2008 A1
20080269562 Marescaux et al. Oct 2008 A1
20090020724 Paffrath Jan 2009 A1
20090024142 Ruiz Morales Jan 2009 A1
20090048612 Farritor et al. Feb 2009 A1
20090054909 Farritor et al. Feb 2009 A1
20090069821 Farritor et al. Mar 2009 A1
20090076536 Rentschler et al. Mar 2009 A1
20090137952 Ramamurthy et al. May 2009 A1
20090143787 De La Pena Jun 2009 A9
20090163929 Yeung et al. Jun 2009 A1
20090171373 Farritor et al. Jul 2009 A1
20090234369 Bax et al. Sep 2009 A1
20090236400 Cole et al. Sep 2009 A1
20090240246 Devill et al. Sep 2009 A1
20090247821 Rogers Oct 2009 A1
20090248038 Blumenkranz et al. Oct 2009 A1
20090281377 Newell et al. Nov 2009 A1
20090305210 Guru et al. Dec 2009 A1
20100010294 Conlon et al. Jan 2010 A1
20100016659 Weitzner et al. Jan 2010 A1
20100016853 Burbank Jan 2010 A1
20100042097 Newton et al. Feb 2010 A1
20100056863 Dejima et al. Mar 2010 A1
20100069710 Yamatani et al. Mar 2010 A1
20100069940 Miller et al. Mar 2010 A1
20100081875 Fowler Apr 2010 A1
20100139436 Kawashima et al. Jun 2010 A1
20100198231 Manzo et al. Aug 2010 A1
20100204713 Ruiz Aug 2010 A1
20100245549 Allen et al. Sep 2010 A1
20100262162 Omori Oct 2010 A1
20100292691 Brogna Nov 2010 A1
20100318059 Farritor et al. Dec 2010 A1
20110015569 Kirschenman et al. Jan 2011 A1
20110020779 Hannaford et al. Jan 2011 A1
20110071347 Rogers et al. Mar 2011 A1
20110071544 Steger et al. Mar 2011 A1
20110077478 Freeman et al. Mar 2011 A1
20110098529 Ostrovsky et al. Apr 2011 A1
20110152615 Schostek et al. Jun 2011 A1
20110224605 Farritor et al. Sep 2011 A1
20110230894 Simaan Sep 2011 A1
20110237890 Farritor et al. Sep 2011 A1
20110238080 Ranjit et al. Sep 2011 A1
20110264078 Lipow Oct 2011 A1
20110270443 Kamiya et al. Nov 2011 A1
20120035582 Nelson et al. Feb 2012 A1
20120109150 Quaid et al. May 2012 A1
20120116362 Kieturakis May 2012 A1
20120179168 Farritor Jul 2012 A1
20120253515 Coste-Maniere et al. Oct 2012 A1
20130041360 Farritor Feb 2013 A1
20130131695 Scarfogliero et al. May 2013 A1
20130345717 Scarfogliero et al. May 2013 A1
20140046340 Wilson et al. Feb 2014 A1
20140058205 Frederick et al. Feb 2014 A1
20140039515 Mondry et al. Jun 2014 A1
20140221749 Grant Aug 2014 A1
20140303434 Farritor et al. Oct 2014 A1
20150051446 Farritor et al. Feb 2015 A1
20170078583 Haggerty Mar 2017 A1
Foreign Referenced Citations (51)
Number Date Country
1082821918 Dec 2012 CN
102010040405 Mar 2012 DE
1354670 Oct 2003 EP
2286756 Feb 2011 EP
2286756 Feb 2011 EP
2329787 Aug 2011 EP
2563261 Mar 2013 EP
2004144533 May 1990 JP
5115425 May 1993 JP
200716235 Jun 1993 JP
2006507809 Sep 1994 JP
07 136173 May 1995 JP
7306155 Nov 1995 JP
08-224248 Sep 1996 JP
2001500510 Jan 2001 JP
2001505810 May 2001 JP
2003220065 Aug 2003 JP
2004322310 Jun 2004 JP
2004180781 Jul 2004 JP
2004329292 Nov 2004 JP
2006508049 Mar 2006 JP
2009-106606 May 2009 JP
2010-533045 Oct 2010 JP
2010-536436 Dec 2010 JP
2011-504794 Feb 2011 JP
2011-045500 Mar 2011 JP
2011-115591 Jun 2011 JP
WO 199221291 May 1991 WO
WO 0189405 Nov 2001 WO
WO 2002082979 Oct 2002 WO
WO 2002100256 Dec 2002 WO
WO 2005009211 Jul 2004 WO
WO 2005044095 May 2005 WO
WO 2006052927 Aug 2005 WO
WO 2006 005075 Jan 2006 WO
WO 2006079108 Jan 2006 WO
WO2006079108 Jul 2006 WO
WO 2007011654 Jan 2007 WO
WO 2007111571 Oct 2007 WO
WO 2007149559 Dec 2007 WO
WO 2009023851 Aug 2008 WO
WO 2009144729 Dec 2009 WO
WO2010042611 Apr 2010 WO
WO2010046823 Apr 2010 WO
WO201050771 May 2010 WO
WO 2011118646 Sep 2011 WO
WO 2011135503 Nov 2011 WO
WO 2011135503 Nov 2011 WO
WO 2011075693 Jul 2012 WO
WO 2013009887 Jan 2013 WO
WO 2014011238 Jan 2014 WO
Non-Patent Literature Citations (173)
Entry
Abbott et al., “Design of an Endoluminal Notes Robotic System,” from the Proceedings of the 2007 IEEE/RSJ Int'l Cont on Intelligent Robot Systems, San Diego, CA, Oct. 29-Nov. 2, 2007, pp. 410-416.
Allendorf et al., “Postoperative Immune Function Varies Inversely with the Degree of Surgical Trauma in a Murine Model,” Surgical Endoscopy 1997; 11:427-430.
Ang, “Active Tremor Compensation in Handheld Instrument for Microsurgery,” Doctoral Dissertation, tech report CMU-RI-TR-04-28, Robotics Institute, Carnegie Mellon Unviersity, May 2004, 167pp.
Applicant Amendment after Notice of Allowance under Rule 312, filed Aug. 25, 2008, in related U.S. Appl. No. 11/695,944, 6pp.
Applicant Response to Office Action dated Apr. 17, 2007, in related U.S. Appl. No. 11/552,379, filed Aug. 8, 2007, 7 pp.
Applicant Response to Office Action dated Aug. 18, 2006, in related U.S. Appl. No. 11/398,174, filed Nov. 7, 2006, 8pp.
Applicant Response to Office Action dated Aug. 21, 2006, in related U.S. Appl. No. 11/403,756, filed Nov. 21, 2006, 52pp.
Applicant Response to Office Action dated Oct. 29, 2007, in related U.S. Appl. No. 11/695,944, filed Jan. 22, 2008, 6pp.
Atmel 8005X2 Core, http://www.atmel.com, 2006, 186pp.
Bailey et al., “Complications of Laparoscopic Surgery,” Quality Medical Publishers, Inc., 1995, 25pp.
Ballantyne, “Robotic Surgery, Telerobotic Surgery, Telepresence, and Telementoring,” Surgical Endoscopy, 2002; 16: 1389-1402.
Bauer et al., “Case Report: Remote Percutaneous Renal Percutaneous Renal Access Using a New Automated Telesurgical Robotic System,” Telemedicine Journal and e-Health 2001; (4): 341-347.
Begos et al., “Laparoscopic Cholecystectomy: From Gimmick to Gold Standard,” J Clin Gastroenterol, 1994; 19(4): 325-330.
Berg et al., “Surgery with Cooperative Robots,” Medicine Meets Virtual Reality, Feb. 2007, 1 pg.
Breda et al., “Future developments and perspectives in laparoscopy,” Eur. Urology 2001; 40(1): 84-91.
Breedveld et al., “Design of Steerable Endoscopes to Improve the Visual Perception of Depth During Laparoscopic Surgery,” ASME, Jan. 2004; vol. 126, pp. 1-5.
Breedveld et al., “Locomotion through the Intestine by means of Rolling Stents,” Proceedings of the ASME Design Engineering Technical Conferences, 2004, pp. 1-7.
Calafiore et al., Multiple Arterial Conduits Without Cardiopulmonary Bypass: Early Angiographic Results,: Ann Thorac Surg, 1999; 67: 450-456.
Camarillo et al., “Robotic Technology in Surgery: Past, Present and Future,” The American Journal of Surgery, 2004; 188: 2S-15.
Cavusoglu et al., “Telesurgery and Surgical Simulation: Haptic Interfaces to Real and Virtual Surgical Environments,” in McLaughliin, M.L, Hespanha, J.P., and Sukhatme, G., editors. Touch in virtual environments, IMSC Series in Multimedia 2001, 28pp.
Cavusoglu et al., “Robotics for Telesurgery: Second Generation Berkeley/UCSF Laparoscopic Telesurgical Workstation and Looking Towards the Future Applications,” Industrial Robot: An International Journal, 2003; 30(1): 22-29.
Chanthasopeephan et al., (2003), “Measuring Forces in Liver Cutting: New Equipment and Experimenal Results,” Annals of Biomedical Engineering 31: 1372-1382.
Choi et al., “Flexure-based Manipulator for Active Handheld Microsurgical Instrument,” Proceedings of the 27th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS), Sep. 2005, 4pp.
Cuschieri, “Technology for Minimal Access Surgery,” BMJ, 1999, 319: 1-6.
Dakin et al., “Comparison of laparoscopic skills performance between standard instruments and two surgical robotic systems,” Surg Endosc., 2003; 17: 574-579.
Dumpert et al., “Improving in Vivo Robot Visioin Quality,” from the Proceedings of Medicine Meets Virtual Realtiy, Long Beach, CA, Jan. 26-29, 2005. 1 pg.
Dumpert et al., “Stereoscopic In Vivo Surgical Robots,” IEEE Sensors Special Issue on In Vivo Sensors for Medicine, Jan. 2007, 10 pp.
Examiner Interview Summary dated Aug. 6 and Aug. 12, 2008, in related U.S. Appl. No. 11/695,944, 1 pg.
Examiner Interview Summary dated May 9, 2008, in related U.S. Appl. No. 11/695,944, 1 pg.
Examiner Interview Summary dated Nov. 30, 2006, in related U.S. Appl. No. 11/398,174, 2pp.
Falcone et al., “Robotic Surgery,” Clin. Obstet. Gynecol. 2003, 46(1): 37-43.
Faraz et al., “Engineering Approaches to Mechanical and Robotic Design for Minimaly Invasive Surgery (MIS),” Kluwer Academic Publishers (Boston), 2000, 13pp.
Fearing et al., “Wing Transmission for a Micromechanical Flying Insect,” Proceedings of the 2000 IEEE International Conference to Robotics & Automation, Apr. 2000; 1509-1516.
Fireman et al., “Diagnosing small bowel Crohn's desease with wireless capsule endoscopy,” Gut 2003; 52: 390-392.
Flynn et al., “Tomorrow's Surgery: micromotors and microbots for minimally invasive procedures,” Minimally Invasive Surgery & Allied Technologies.
Franklin et al., “Prospective Comparison of Open vs. Laparoscopic Colon Surgery for Carcinoma: Five-Year Results,” Dis Colon Rectum, 1996; 39: S35-S46.
Fraulob et al., “Miniature assistance module for robot-assisted heart surgery,” Biomed. Tech. 2002, 47 Suppl. 1, Pt. 1: 12-15.
Fukuda et al., “Mechanism and Swimming Experiment of Micro Mobile Robot in Water,” Proceedings of the 1994 IEEE International Conference on Robotics and Automation, 1994: 814-819.
Fukuda et al., “Micro Active Catheter System with Multi Degrees of Freedom,” Proceedings of the IEEE International Conference on Robotics and Automation, May 1994, pp. 2290-2295.
Fuller et al., “Laparoscopic Trocar Injuries: A Report from a U.S. Food and Drug Administration (FDA) Center for Devices and Radiological Health (CDRH) Systematic Technology Assessment of Medical Products (STAMP) Committe,” U.S. Food and Drug Adminstration, available at http://www.fdaJ:?;ov, Finalized: Nov. 7, 2003; Updated: Jun. 24, 2005, 11 pp.
Grady, “Doctors Try New Surgery for Gallbladder Removal,” The New York Times, Apr. 20, 2007, 3 pp.
Guber et al., “Miniaturized Instrumetn Systems for Minimally Invasive Diagnosis and Therapy,” Biomedizinishe Technic. 2002, Band 47, Erganmngsband 1:.
Patronik et al., “Development of a Tethered Epicardial Crawler for Minimally Invasive Cardiac Therapies,” IEEE, pp. 239-240.
Patronik et al., “Crawling on the Heart: A Mobile Robotic Device for Minimally Invasive Cardiac Interventions,” MICCAI, 2004, pp. 9-16.
Patronik et al., “Preliminary evaluation of a mobile robotic device for navigation and intervention on the beating heart,” Computer Aided Surgery, 10(4): 225-232, Jul. 2005.
Peirs et al., “A miniature manipulator for integration in a self-propelling endoscope,” Sensors and Actuators A, 2001, 92: 343-349.
Peters, “Minimally Invasive Colectomy: Are the Potential Benefits Realized?” Dis Colon Rectum 1993; 36: 751-756.
Phee et al., “Analysis and Development of Locomotion Devices for the Gastrointestinal Tract,” IEEE Transaction on Biomedical Engineering, vol. 49, No. 6, Jun. 2002, pp. 613-616.
Phee et al., “Development of Microrobotic Devices for Locomotion in the Human Gastrointestinal Tract,” International Conference on Computational Intelligence, Robotics and Autonomous Systems (CIRAS 2001), Nov. 28-30, 2001, Singapore.
Platt et al., “In Vivo Robotic Cameras can Enhance Imaging Capability During Laparoscopic Surgery,” in the Proceedings of the Society of American Gastrointestinal Endoscopic Surgeons (SAGES) Scientific Conference, Ft. Lauderdale, FL, Apr. 13-16, 2005, I pg.
Preliminary Amendment filed Apr. 11, 2007, in related U.S. Appl. No. 11/403,756, 7 pp.
Preliminary Amendment filed Jul. 30, 2008, in related U.S. Appl. No. 12/171,413, 4 pp.
RCE and Amendment filed Jun. 13, 2007, in related U.S. Appl. No. 11/403,756, 8 pp.
Rentschler et al., “Mobile In Vivo Biopsy and Camera Robot,” Studies in Health and Infonnalics Medicine Meets Virtual Reality, vol. 119., pp. 449-454, IOS Press, Long Beach, CA, 2006e.
Rentschler et al., Mobile In Vivo Biopsy Robot, IEEE International Conference on Robotics and Automation, Orlando, Florida, May 2006, pp. 4155-4160.
Rentschler et al., “Miniature in vivo Robots for Remote and Harsh Environments,” IEEE Transactions on Information Technology in Biomedicine, Jan. 2006; 12(1): 66-75.
Rentschler et al., “An In Vivo Mobile Robot for Surgical Vision and Task Assistance,” Journal of Medical Devices, Mar. 2007, vol. 1: 23-29.
Rentschler et al., “In vivo Mobile Surgical Robotic Task Assistance,” 1 pg.
Rentschler et al., “In vivo Robotics during the NEEMO 9 Mission,” Medicine Meets Virtual Reality, Feb. 2007, I pg.
Rentschler et al., “In Vivo Robots for Laparoscopic Surgery,” Studies in Health Technology and Infonnatics—Medicine Meets Virtual Reality, ISO Press, Newport Beach, CA, 2004a, 98: 316-322.
Rentschler et al., “Mechanical Design of Robotic In Vivo Wheeled Mobility,” ASME Journal of Mechanical Design, 2006a, I-II.
Rentschler et al., “Mobile In Vivo Camera Robots Provide Sole Visual Feedback for Abdominal Exploration and Cholecystectomy,” Journal of Surgical Endoscopy, 20-I: 135-138, 2006b.
Rentschler et al., “Mobile In Vivo Robots Can Assist in Abdominal Exploration,” from the Proceedings of the Gastrointestinal Endoscopic Surgeons (SAGES) Scientific Conference, Ft. Lauderdale, FL, Apr. 13-16, 2005b.
Rentschler et al., “Modeling, Analysis, and Experimental Study of In Vivo Wheeled Robotic Mobility,” IEEE Transactions on Robotics, 22 (2): 308-321, 2005c.
Rentschler et al., “Natural Orifice Surgery with an Endoluminal Mobile Robot,” The Society of American Gastrointestinal Endoscopic Surgeons, Dallas, TX, Apr. 2006d, 14 pp.
Rentschler et al., “Theoretical and Experimental Analysis of In Vivo Wheeled Mobility,” ASME Design Engineering Technical Conferences: 28th Biennial Mechanisms and Robotics Conference, Salt Lake City, Utah, Sep. 28-Oct. 2, 2004, pp. 1-9.
Rentschler et al., “Toward In Vivo Mobility,” Studies in Health Technology and Infonnatics—Medicine Meets Virtual Reality, ISO Press, Long Beach, CA, 2005a, III: 397-403.
Response to Rule 312 Amendment in related U.S. Appl. No. 11/695,944, dated Jan. 12, 2009, 2 pp.
Riviere et al., “Toward Active Tremor Canceling in Handheld Microsurgical Instruments,” IEEE Transactions on Robotics and Automation, Oct. 2003, 19(5): 793-800.
Rosen et al., “Force Controlled and Teleoperated Endoscopic, Grasper for Minimally Invasive Surgery—Experimental Performance Evaluation,” IEEE Transactions of Biomedical Engineering, Oct. 1999; 46(10): 1212-1221.
Rosen et al., “Objective Laparoscopic Skills Assessments of Surgical Residents Using Hidden Markov Models Based on Haptic Information and Tool/Tissue Interactions,” Studies in Health Technology and Infonnatics—Medicine Meets Virtual Reality, Jan. 2001, 7 pp.
Rosen et al., “Spherical Mechanism Analysis of a Surgical Robot for Minimally Invasive Surgery—Analytical and Experimental Approaches,” Studies in Health Technology and Informatics-Medicine Meets Virtual Reality, pp. 442-448, Jan. 2005.
Rosen et al., “Task Decomposition of Laparoscopic Surgery for Objective Evaluation of Surgical Residents' Learning Curve Using Hidden Markov Model,” Computer Aided Surgery, vol. 7, pp. 49-61, 2002.
Rosen et al., “The Blue Dragon—A System of Measuring the Kinematics and the Dynamics of Minimally Invasive Surgical Tools In-Vivo,” Proc. of the 2002 IEEE International Conference on Robotics and Automation, Washington, DC, pp. 1876-1881, May 2002.
Ruurda et al., “Robot-Assisted surgical systems: a new era in laparoscopic surgery,” Ann R. Coll Surg Engl., 2002; 84: 223-226.
Ruurda et al., “Feasibility of Robot-Assisted Laparoscopic Surgery,” Surgical Laparoscopy, Endoscopy & Percutaneous Techniques, 2002; 12(1):41-45.
Sackier et al., “Robotically assisted laparoscopic surgery,” Surgical Endoscopy, 1994; 8: 63-66.
Salky, “What is the Penetration of Endoscopic Techniques into Surgical Practice?” Digestive Surgery, 2000; 17:422-426.
Satava, “Surgical Robotics: The Early Chronicles,” Surgical Laparoscopy, Endoscopy & Percutaneous Techniques, 2002; 12(1): 6-16.
Schippers et al., (1996) “Requirements and Possibilities of Computer-Assisted Endoscopic Surgery,” In: Computer Integrated Surgery: Technology and Clinical Applications, pp. 561-565.
Schurr et al., “Robotics and Telemanipulation Technologies for Endoscopic Surgery,” Surgical Endoscopy, 2000; 14: 375-381.
Schwartz, “In the Lab: Robots that Slink and Squirm,” The New York Times, Mar. 27, 2007, 4 pp.
Sharp LL-151-3D, http://www.sharp3d.com, 2006, 2 pp.
Slatkin et al., “The Development of a Robotic Endoscope,” Proceedings of the 1995 IEEE International Conference on Robotics and Automation, pp. 162-171, 1995.
Smart Pill “Fastastic Voyage: Smart Pill to Expand Testing,” http://www.smartpilldiagnostics.com, Apr. 13, 2005, 1 pg.
Southern Surgeons Club (1991), “A prospective analysis of 1518 laparoscopic cholecystectomies,” N. Eng. 1 Med. 324 (16): 1073-1078.
Stefanini et al., “Modeling and Experiments on a Legged Microrobot Locomoting in a Tubular Compliant and Slippery Environment,” Int. Journal of Robotics Research, vol. 25, No. 5-6, pp. 551-560, May-Jun. 2006.
Stiff et al.., “Long-term Pain: Less Common After Laparoscopic than Open Cholecystectomy,” British Journal of Surgery, 1994; 81: 1368-1370.
Strong, et al., “Efficacy of Novel Robotic Camera vs. a Standard Laproscopic Camera,” Surgical Innovation vol. 12, No. 4, Dec. 2005, Westminster Publications, Inc., pp. 315-318.
Suzumori et al., “Development of Flexible Microactuator and its Applications to Robotics Mechanisms,” Proceedings of the IEEE International Conference on Robotics and Automation, 1991: 1622-1627.
Taylor et al., “A Telerobotic Assistant for Laparoscopic Surgery,” IEEE Eng Med Biol, 1995; 279-287.
Tendick et al.. (1993), “Sensing and Manipulation Problems in Endoscopic Surgery: Experiment, Analysis, and Observation,” Presence 2( 1): 66-81.
Palm, William, “Rapid Prototyping Primer” May 1998 (revised Jul. 30, 2002) (http://www.me.psu.edu/lamancusa/rapidpro/primer/chapter2.htm).
Stoianovici et al., “Robotic Tools for Minimally Invasive Urologic Surgery”, Jan. 1, 2002, pp. 1-17.
Cleary et al., “State of the Art in Surgical Rootics: Clinical Applications and Technology Challenges”, “Computer Aided Surgery”, Jan. 1, 2002, pp. 312-328, vol. 6.
Green, “Telepresence Surgery”, Jan. 1, 1995, Publisher: IEEE Engineering in Medicine and Biology.
International Preliminary Report on Patentability from related case PCT/US2007/014567, dated Jan. 8, 2009, 11 pp.
International Search report and Written Opinion from international application No. PCT/US2012/41911, dated Mar. 13, 2013.
International Search Report and Written Opinion from international application No. PCT/US12/46274, dated Sep. 25, 2012.
International Search Report and Written Opinion from international application No. PCT/US2007/089191, dated Nov. 10, 2008, 20 pp.
“International Search Report and Written Opinion from international application No. PCT/US07/14567, dated Apr. 28, 2008, 19 pp.”
International Search Report and Written Opinion of international application No. PCT/US2008/069822, dated Aug. 5, 2009, 12 pp.
International Search Report and Written Opinion of international application No. PCT/US2008/073334, dated Jan. 12, 2009, 11 pp.
International Search Report and Written Opinion of international application No. PCT/US2008/073369, dated Nov. 12, 2008, 12 pp.
International Search Report and Written Opinion issued in PCT/US11/46809, dated Dec. 8, 2011.
Ishiyama et al., “Spiral-type Micro-machine for Medical Applications,” 2000 International Symposium on Micromechatronics and Human Science, 2000: 65-69.
Jagannath et al., “Peroral transgastric endoscopic ligation of fallopian tubes with long-term survival in a porcine model,” Gastrointestinal Endoscopy, 2005; 61(3): 449-453.
Kalloo et al., “Flexible transgastric peritoneoscopy: a novel approach to diagnostic and therapeutic interventions in the peritoneal cavity,” Gastrointestinal Endoscopy, 2004; 60(1): 114-117.
Kang et al., “Robotic Assistants Aid Surgeons During Minimally Invasive Procedures,” IEEE Engineering in Medicine and Biology, Jan.-Feb. 2001; pp. 94-104.
Kantsevoy et al., “Endoscopic gastrojejunostomy with survival in a porcine model,” Gastrointestinal Endoscopy, 2005; 62(2): 287-292.
Kantsevoy et al., “Transgastric endoscopic splenectomy,” Surgical Endoscopy, 2006; 20: 522-525.
Kazemier et al. (1998), “Vascular Injuries During Laparoscopy,” J. Am. Coli. Surg. 186(5): 604-5.
Kim, “Early Experience with Telemanipulative Robot-Assisted Laparoscopic Cholecystectomy Using da Vinci,” Surgical Laparoscopy, Endoscopy & Percutaneous Techniques, 2002; 12(1):33-40.
Ko et al., “Per-Oral transgastric abdominal surgery,” Chinese Journal of Digestive Diseases, 2006; 7: 67-70.
Lafullarde et al., “Laparoscopic Nissen Fundoplication: Five-year Results and Beyond,” Arch/Surg, Feb. 2001; 136:180-184.
Leggett et al. (2002), “Aortic injury during laparoscopic fundoplication,” Surg. Endoscopy 16(2): 362.
Li et al. (2000), “Microvascular Anastomoses Performed in Rats Using a Microsurgical Telemanipulator,” Comp. Aid. Surg. 5: 326-332.
Liem et al., “Comparison of Conventional Anterior Surgery and Laparoscopic Surgery for Inguinal-hernia Repair,” New England Journal of Medicine, 1997; 336 (22): 1541-1547.
Macfarlane et al., “Force-Feedback Grasper Helps Restore the Sense of Touch in Minimally Invasive Surgery,” Journal of Gastrointestinal Surgery, 1999; 3: 278-285.
Mack et al., “Present Role of Thoracoscopy in the Diagnosis and Treatment of Diseases of the Chest,” Ann Thorac Surgery, 1992; 54: 403-409.
Mack, “Minimally Invasive and Robotic Surgery,” JAMA, Feb. 2001; 285(5): 568-572.
Mei et al., “Wireless Drive and Control of a Swimming Microrobot,” Proceedings of the 2002 IEEE International conference on Robotics & Automation, May 2002: 1131-1136.
Melvin et al., “Computer-Enhanced vs. Standard Laparoscopic Antireflux Surgery,” J Gastrointest Surg 2002; 6: 11-16.
Menciassi et al., “Locomotion of a Leffed Capsule in the Gastrointestinal Tract: Theoretical Study and Preliminary Technological Results,” IEEE Int. Conf. on Engineering in Medicine and Biology, San Francisco, CA, pp. 2767-2770, Sep. 2004.
Menciassi et al., “Robotic Solutions and Mechanisms for a Semi-Autonomous Endoscope,” Proceedings of the 2002 IEEE/RSJ Intl. Conference on Intelligent Robots and Systems, Oct. 2002; 1379-1384.
Menciassi et al., “Shape memory alloy clamping devices of a capsule for monitoring tasks in the gastrointestinal tract,” J. Micromech. Microeng, 2005, 15: 2045-2055.
Meron, “The development of the swallowable video capsule (M2A),” Gastrointestinal Endoscopy 2000; 52 6: 817-819.
Micron, http://www.micron.com, 2006, I/4-inch VGA NTSC/PAL CMOS Digital Image Sensor, 98 pp.
Midday Jeff et al., “Material Handling System for Robotic natural Orifice Surgery”, Proceedings of the 2011 Design of medical Devices Conference, Apr. 12-14, 2011, Minneapolis, MN, 4 pages.
Miller, Ph.D., et al., “In-Vivo Stereoscopic Imaging System with 5 Degrees-of-Freedom for Minimal Access Surgery,” Dept. of Computer Science and Dept. of Surgery, Columbia University, New York, NY, 7 pp.
Munro (2002), “Laparoscopic access: complications, technologies, and techniques,” Curro Opin. Obstet. Gynecol., 14(4): 365-74.
Nio et al., “Efficiency of manual vs robotical (Zeus) assisted laparoscopic surgery in the performance of standardized tasks,” Surg Endosc, 2002; 16: 412-415.
Office Action dated Apr. 17, 2007, received in related U.S. Appl. No. 11/552,379, 5 pp.
Office Action dated Apr. 3, 2009, received in related U.S. Appl. No. 11/932,516, 43 pp.
Office Action dated Aug. 18, 2006, received in related U.S. Appl. No. 11/398,174, 6 pp.
Office Action dated Aug. 21, 2006, received in related U.S. Appl. No. 11/403,756, 6 pp.
Office Action dated Oct. 29, 2007, received in related U.S. Appl. No. 11/695,944, 6 pp.
Office Action dated Oct. 9, 2008, received in related U.S. Appl. No. 11/932,441, 4 pp.
Oleynikov et al., “In Vivo Camera Robots Provide Improved Vision for Laparoscopic Surgery,” Computer Assisted Radiology and Surgery (CARS), Chicago, IL, Jun. 23-26, 2004b.
Oleynikov et al., “In Vivo Robotic Laparoscopy,” Surgical Innovation, Jun. 2005, 12(2): 177-181.
Oleynikov et al., “Miniature Robots Can Assist in Laparoscopic Cholecystectomy,” Journal of Surgical Endoscopy, 19-4: 473-476, 2005.
O'Neill, “Surgeon takes new route to gallbladder,” The Oregonian, Jun. 2007, 2 pp.
Orlando et al., (2003), “Needle and Trocar Injuries in Diagnostic Laparoscopy under Local Anesthesia: What is the True Incidence of These Complications?” Journal of Laparoendoscopic & Advanced Surgical Techniques 13(3): 181-184.
Park et al., “Trocar-less Instrumentation for Laparoscopy: Magnetic Positioning of Intra-abdominal Camera and Retractor,” Ann Surg, Mar. 2007; 245(3): 379-384.
Park et al., “Experimental studies of transgastric gallbladder surgery: cholecystectomy and cholecystogastric anastomosis (videos),” Gastrointestinal Endoscopy, 2005; 61(4): 601-606.
Tendick et al., “Applications of Micromechatronics in Minimally Invasive Surgery,” IEEE/ASME Transactions on Mechatronics, 1998; 3(1): 34-42.
Thomann et al., “The Design of a new type of Micro Robot for the Intestinal Inspection,” Proceedings of the 2002 IEEE Intl. Conference on Intelligent Robots and Systems, Oct. 2002: 1385-1390.
U.S. Appl. No. 60/180,960, filed Feb. 2000.
U.S. Appl. No. 60/956,032, filed Aug. 15, 2007.
U.S. Appl. No. 60/983,445, filed Oct. 29, 2007.
U.S. Appl. No. 60/990,062, filed Nov. 26, 2007.
U.S. Appl. No. 60/990,076, filed Nov. 26, 2007.
U.S. Appl. No. 60/990,086, filed Nov. 26, 2007.
U.S. Appl. No. 60/990,106, filed Nov. 26, 2007.
U.S. Appl. No. 60/990,470, filed Nov. 27, 2007.
U.S. Appl. No. 61/025,346, filed Feb. 1, 2008.
U.S. Appl. No. 61/030,588, filed Feb. 22, 2008.
U.S. Appl. No. 61/030,617, filed Feb. 22, 2008.
Way et al., (editors), “Fundamentals of Laparoscopic Surgery,” Churchill Livingstone Inc., 1995, 14 pp.
Wolfe et al., “Endoscopic Cholecystectomy: An analysis of Complications,” Arch. Surg. Oct. 1991; 126: 1192-1196.
Worn et al., “Espirit Project No. 33915: Miniaturised Robot for Micro Manipulation (MINIMAN)”, Nov. 1998; http://www.ipr.ira.ujka.de/-microbot/miniman.
Yu et al., “Microrobotic Cell Injection,” Proceedings of the 2001 IEEE International Conference on Robotics and Automation, May 2001; 620-625.
Yu, Bsn, Rn, “M2ATM Capsule Endoscopy a Breakthrough Diagnostic Tool for Small Intestine Imagining,” vol. 25, No. 1, Gastroenterology Nursing, pp. 24-27.
International Search Report and Written Opinion of international application No. PCT/US2010/061137, dated Feb. 11, 2011, 10 pp.
Abbou et al., “Laparoscopic Radical Prostatectomy with a Remote Controlled Robot,” The Journal of Urology, Jun. 2001, 165: 1964-1966.
Glukhovsky et al., “The development and application of wireless capsule endoscopy,” Int. J. Med. Robot. Comput. Assist. Surgery, 2004; I (1): 114-123.
Gong et al., Wireless endoscopy, Gastrointestinal Endoscopy 2000; 51(6): 725-729.
Hanly et al., “Value of the SAGES Learning Center in introducing new technology,” Surgical Endoscopy, 2004; 19(4):477-483.
Hanly et al., “Robotic Abdominal Surgery,” The American Journal of Surgery 188 (Suppl.to Oct. 1994): 19S-26S, 2004.
Heikkinen et al., “Comparison of laparoscopic and open Nissen fundoplication two years after operation: A prospective randomized trial,” Surgical Endoscopy, 2000; 14: 1019-1023.
Hissink, “Olympus Medical develops capsule camera technology,” Dec. 2004, accessed Aug. 29, 2007, http://www.letsgodigital.org , 3 pp.
Horgan et al., “Technical Report: Robots in Laparoscopic Surgery,” Journal of Laparoendoscopic & Advanced Surgical Techniques, 2001; 11(6): 415-419.
Franzino, “The Laprotek Surgical System and the Next Generation of Robotics,” Surg Clin North Am, 2003 83(6): 1317-1320.
Related Publications (1)
Number Date Country
20170035526 A1 Feb 2017 US
Provisional Applications (1)
Number Date Country
62200563 Aug 2015 US