In robotically-assisted or telerobotic surgery, the surgeon typically operates a control device to control the motion of surgical instruments at the surgical site from a location that may be remote from the patient (e.g., across the operating room, in a different room or a completely different building from the patient) or immediately adjacent to the patient. The controller usually includes one or more manually-operated input devices, such as joysticks, exoskeletal gloves or the like, which are coupled (directly or indirectly) to the surgical instruments with servo motors for articulating the instruments at the surgical site. The servo motors are typically part of an electromechanical device or surgical manipulator that supports and controls the surgical instruments that have been introduced directly into an open surgical site or through trocar sleeves into a body cavity, such as the patient's abdomen. During the operation, the surgical manipulator provides mechanical articulation and control of a variety of surgical instruments, such as tissue graspers, needle drivers, electrosurgical cautery probes, etc., that each perform various functions for the surgeon, e.g., holding or driving a needle, grasping a blood vessel, or dissecting, cauterizing or coagulating tissue. A surgeon may employ a large number of different surgical instruments/tools during a procedure.
Robotic manipulators, particularly small ones of the type used in minimally-invasive surgery, generally need to transmit actuation forces and sensor information some distance through the structure of the robotic device, because the mass and size of presently-available actuators (e.g., motors) and sensors (e.g., position encoders) are too large to permit them to be directly located where the force and sensing is needed. In a mechanical device of a type used in minimally-invasive surgery, it is desirable to provide (a) significant force and (b) positioning precision at the tip of a small structure. However, the motors and sensors used are typically too large and heavy to be placed inside the patient's body. In addition, safety concerns make transmission of electrical power to actuators or sensors located inside the patient's body problematic. Therefore, in these robotic applications, it is desirable to transmit the actuation forces for multiple degrees of freedom over significant distances, and the desire to minimize the size of incisions places a premium on the cross-sectional area used for the structure and the transmission of the actuation forces. In addition, it is desirable to keep the transmission mechanism for these forces as stiff and friction-free as possible, so that the proximally sensed position of the proximal actuator may be used as a fair representation of the distally positioned joint.
There are many means known to transmit force over a distance. Some examples include: cables in tension; rods or tubes in compression; torsion; and hydraulic or pneumatic pressure. None of these force transmission means are ideal for minimally-invasive surgery, and therefore there are practical limits to the precision with which a distally mounted structure member can be moved, and the force it can provide. For example, cables may stretch, thereby leading to errors in the distal structure position relative to the position of the proximal driving mechanism. Compliant cables may also have significant friction with the supporting structure. Friction is notoriously difficult to model, may behave in a time-varying way and is dependent upon any number of variables, including the position (and past history of the position) of any intermediate joints in the structure, and when combined with compliant actuation can lead to stick-slip behavior. These errors limit the utility of the robot, limit the number of degrees of freedom that can be controlled remotely through a small structure, limit the use of the robot's state to provide haptic feedback, and are particularly severe for a flexible surgical robot which must accurately control many degrees of freedom of actuation at the end of a long, flexible structure.
Feedback regarding the actual position of a structure is a commonly applied mechanism for compensating for these errors. If the position of the distal part is accurately measured with minimal delays, control algorithms can compensate for many of the deficiencies of the force transmission mechanisms. While there are limits to the compensation that can be done with feedback, those limits cannot even be approached without accurate, timely information about the positions of the distally-mounted parts. The main restriction on the applicability of feedback in these types of robotic arms is the lack of an effective means of determining the position of parts mounted at the distal end of the structure.
Two methods of determining the position of a structure are proprioception and exteroception. Proprioception refers to the internal sensing of the position of the structure and exteroception refers to the external sensing of the position.
A proprioceptive system may monitor the motors actuating the movement of each joint in a robotic structure. By monitoring the degree of movement of each motor, the expected movement of each corresponding joint can be estimated. For example, the da Vinci Surgical System by Intuitive Surgical, Inc., of Sunnyvale, Calif. utilizes motor sensors (e.g., encoders) for position feedback. The encoders are co-located with the actuators, which are positioned outside of the patient body and drive the structure's joints through a cable mechanism. However, this method allows the position control loop to compensate only for error sources that occur between the actuator and the encoder. Error sources distal to the encoder are not observed by the encoder and thus cannot be compensated. This arrangement does not provide for compensation of errors introduced by any structures or drivetrain mechanisms that are distal to the encoders. In order to estimate the position of distal joints, the cable mechanism is modeled as infinitely stiff, so the motor position is assumed to be a reliable indication of the actual joint position. If the drivetrain between the actuator and the joint is not sufficiently stiff or other errors exist between the joint and the encoder, there will be a difference between actual joint orientation and the expected orientation based on the motor position. These joints are often mounted serially, so that the relative orientation of each of the links in the structure must be known and transformed to determine the position of the distal end of the structure. As a result, errors in the sensing of the orientations of intermediate links may compound along the way.
In other cases, position sensors may be positioned directly at the joints. This arrangement may be effective for larger structures, such as the Canadarm2 used on the International Space Station, and industrial robots. In robotically-assisted surgical systems having many degrees of freedom near the distal end, joint encoders do not offer much benefit as they are too large to be placed where the most problematic joints are located, and the connections required to transmit their position data back to the controller compete for space with the force-transmitting apparatus inside the arm.
Tachometers and accelerometers have been used, typically to estimate the robot's configuration at successive time steps and for force-control applications. These systems also do not provide complete position data and therefore cannot be used to compensate for some types of errors described above.
In systems which utilize exteroception, externally observed position feedback may be used to determine the position of a robotic structure.
For example, GPS-based location (especially differential GPS) has been used in outdoor environments, but typically can only be applied to large structures due to the relatively poor spatial resolution.
Field-based sensing can be accomplished using either AC or DC magnetic fields. In this approach, a small sensing element is placed at the tip of the structure to be monitored, and a magnetic field generated externally. While these systems can achieve good accuracy under ideal conditions, the accuracy rapidly degrades when metallic objects are nearby (as is the case in most surgical applications), and an external apparatus is needed to generate the field. In addition, these sensors do not directly encode the positions of the joints or links in a structure unless a sensor is provided on each link. Therefore, different configurations of the structure which result in the same position of the sensor are not distinguishable, which limits the use of this type of data for feedback compensation as described above.
Another approach is to utilize a dedicated set of unloaded cables, attached to the movable member at the distal end and to an encoder or position measurement device at the proximal end. Although unloaded, these cables are still subject to the same bending at the intermediate joints as are the actuation devices, and to provide feedback for many degrees of freedom, many cables must be used, requiring a larger structure.
Methods have been used involving a combination of sensing of the external environment, locating landmarks, and using this information to simultaneously construct a map of the environment and localize the robot. However, these systems are only applicable to larger and slower-moving systems, such as mobile robotic platforms. In the surgical applications considered here, these methods may be undesirable because the landmarks (e.g. patient anatomy) are not well defined and may change position as a result of disease states or during the surgical manipulations.
Techniques known as visual-servoing have been proposed, in which a camera mounted on the end of the robot arm is used in combination with joint encoders to control the position of the robot end-effector with respect to a tracked object in the field of view of an imaging device. This approach suffers from the need to provide joint position feedback and identifiable landmarks in the viewed scene, which are also problematic in surgical applications.
These techniques suffer from various deficiencies when utilized in robotic surgical applications and do not provide the type of information desired for implementing closed-loop control of such robotic mechanisms and compensating for the deficiencies in the force-transmission means described above. In addition, these techniques may be complex and expensive to implement.
Accordingly, it would be desirable to provide systems and methods for determining the position of a surgical instrument at a surgical site on a patient. In particular, it would be desirable for these systems and methods to provide real-time position feedback to the control system for a robotic surgical instrument.
In accordance with embodiments of the present invention, a surgical instrument is provided, comprising: at least one articulatable arm having a distal end, a proximal end, and at least one joint region disposed between the distal and proximal ends; an optical fiber bend sensor provided in the at least one joint region of the at least one articulatable arm; a detection system coupled to the optical fiber bend sensor, said detection system comprising a light source and a light detector for detecting light reflected by or transmitted through the optical fiber bend sensor to determine a position of at least one joint region of the at least one articulatable arm based on the detected light reflected by or transmitted through the optical fiber bend sensor; and a control system comprising a servo controller for effectuating movement of the arm.
In accordance with other embodiments of the present invention, a surgical instrument is provided, comprising: at least one elongate arm comprising a passively-bendable region and an actively-controlled bendable region including at least one joint region; a control system comprising a servo controller for effectuating movement of the at least one joint region; an optical fiber bend sensor provided in at least one of the passively-bendable region and the actively-controlled bendable region; and a detection system coupled to the optical fiber bend sensor, said detection system comprising a light source and a light detector for detecting light reflected by or transmitted through the optical fiber bend sensor.
Other features and aspects of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings which illustrate, by way of example, the features in accordance with embodiments of the invention. The summary is not intended to limit the scope of the invention, which is defined solely by the claims attached hereto.
In the following description, reference is made to the accompanying drawings which illustrate several embodiments of the present invention. It is understood that other embodiments may be utilized and mechanical, compositional, structural, electrical, and operational changes may be made without departing from the spirit and scope of the present disclosure. The following detailed description is not to be taken in a limiting sense, and the scope of the embodiments of the present invention is defined only by the claims of the issued patent.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. Spatially relative terms, such as “beneath”, “below”, “lower”, “above”, “upper” and the like, may be used herein for ease of description to describe one element or feature's relationship to another element(s) or feature(s) as illustrated in the figures. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if the device in the figures is turned over, elements described as “below” or “beneath” other elements or features would then be oriented “above” the other elements or features. Thus, the exemplary term “below” can encompass both an orientation of above and below. The device may be otherwise oriented (rotated 90 degrees or at other orientations) and the spatially relative descriptors used herein interpreted accordingly.
As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising” specify the presence of stated features, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, steps, operations, elements, components, and/or groups thereof.
Described herein are embodiments of a multi-component system, apparatus, and method for performing robotically-assisted surgical procedures on a patient, particularly including open surgical procedures, neurosurgical procedures, such as stereotaxy, and endoscopic procedures, such as laparoscopy, arthroscopy, thoracoscopy and the like. The system and method of the present invention may be particularly useful as part of a robotic surgical system that allows the surgeon to manipulate the surgical instruments through a servo controller from a location remote from the patient. To that end, the manipulator apparatus of the present invention may be driven by a controller to form a telepresence system. The systems may comprise a master-slave configuration in which the controller and manipulator apparatus are separate devices, or may comprise an augmented system in which the controller and manipulator are provided as part of a single device. A description of a suitable slave-master system can be found in U.S. patent application Ser. No. 08/517,053, filed Aug. 21, 1995, the complete disclosure of which is incorporated herein by reference for all purposes.
Referring to
Control assembly 6 may be located at a surgeon's console C which is usually located in the same room as operating table O so that the surgeon may speak to his/her assistant(s) A and directly monitor the operating procedure. However, it should be understood that the surgeon S can be located in a different room or a completely different building from the patient P. Control assembly 6 generally includes a support 8, a monitor 10 for displaying an image of the surgical site to the surgeon S, and one or more control device(s) 12 for controlling the manipulator assemblies 4. The control device(s) 12 may include a variety of input devices, such as joysticks, gloves, trigger-guns, hand-operated controllers, voice recognition devices or the like. Preferably, the control device(s) 12 will be provided with the same degrees of freedom as the associated surgical instruments 20 to provide the surgeon with telepresence, or the perception that the control device(s) 12 are integral with the instruments 20 so that the surgeon has a strong sense of directly controlling instruments 20. In some embodiments, the control devices 12 are manual input devices which move with six degrees of freedom, and which may also include an actuatable handle for actuating instruments (for example, for closing grasping saws, applying an electrical potential to an electrode, or the like).
Position, applied force, and tactile feedback sensors (not shown) may also be employed on instrument assemblies 20 to transmit information regarding position, applied force, and tactile sensations from the surgical instrument back to the surgeon's hands as he/she operates the robotic system. One suitable system and method for providing telepresence to the operator is described in U.S. patent application Ser. No. 08/517,053, filed Aug. 21, 1995, which is incorporated by reference herein in its entirety.
Monitor 10 may be operatively coupled to the viewing scope assembly 19 such that an image of the surgical site is provided adjacent the surgeon's hands on surgeon console C. Preferably, monitor 10 will display on a display 18 an image of the surgical site and surgical instruments. The display 18 and the master control devices 12 may be oriented such that the relative positions of the imaging device in the scope assembly and the surgical instruments are similar to the relative positions of the surgeon's eyes and hands so the operator can manipulate the end effector and the hand control as if viewing the workspace in substantially true presence. By true presence, it is meant that the presentation of an image is a true perspective image simulating the viewpoint of an operator that is physically manipulating the surgical instruments 20.
As shown in
Servo controller 16 may be separate from, or integral with manipulator assemblies 4. In some embodiments, the servo controller 16 and manipulator assemblies 4 are provided as part of a robotic arm cart positioned adjacent to the patient's body. The servo controller 16 transmits signals instructing the manipulator assemblies 4 to move instruments having shafts which extend into an internal surgical site within the patient body via openings in the body. Robotic surgery systems and methods are further described in U.S. Pat. No. 5,797,900, filed on May 16, 1997, issued on Aug. 25, 1998, U.S. Pat. No. 6,132,368, filed on Nov. 21, 1997, issued on Oct. 17, 2000, U.S. Pat. No. 6,331,181, filed on Oct. 15, 1999, issued on Dec. 18, 2001, U.S. Pat. No. 6,441,577, filed on Apr. 3, 2001, issued on Aug. 27, 2002, U.S. Pat. No. 6,902,560, filed on Jan. 6, 2004, issued on Jun. 7, 2005, U.S. Pat. No. 6,936,042, filed on Apr. 16, 2002, issued on Aug. 30, 2005, and U.S. Pat. No. 6,994,703, filed on Dec. 4, 2002, issued on Feb. 7, 2006, the full disclosures of which are incorporated herein by reference. A suitable robotic surgical system currently in use is the da Vinci S Surgical System by Intuitive Surgical, Inc.
Each of the manipulator assemblies 4 may support a surgical instrument 20 and may comprise a series of manually articulatable linkages, generally referred to as set-up joints, and a robotic manipulator. The manipulator assemblies 4 enable the instrument 20 to be rotated around a point in space, as more fully described in issued U.S. Pat. Nos. 6,331,181, and 5,817,084, the full disclosures of which are incorporated herein by reference. The robotic manipulators may pivot the instrument about a pitch axis, a yaw axis, and an insertion axis (which is aligned along a shaft of the instrument). The instrument has still further driven degrees of freedom as supported by the manipulator, including sliding motion of the instrument along the insertion axis.
The robotic manipulator assemblies 4 may be driven by a series of actuators (e.g., motors). These motors actively move the robotic manipulators in response to commands from the servo control 16. The motors are further coupled to the surgical instrument so as to rotate the surgical instrument about the insertion axis, and to articulate a wrist at the distal end of the instrument about at least one, and often two or more, degrees of freedom. Additionally, the motors can be used to actuate an articulatable end effector of the instrument for grasping tissues in the jaws of a forceps or the like. The motors may be coupled to at least some of the joints of the surgical instrument using cables, as more fully described in U.S. Pat. Nos. 6,331,181, and 5,792,135, the full disclosures of which are also incorporated herein by reference. As described in those references, the manipulators will often include flexible members for transferring motion from the drive components to the surgical instrument. Each actuator may effectuate movement of one or more joint members in the instrument. For endoscopic procedures, the manipulators may include a cannula, which supports the surgical instrument, allowing the surgical instrument to rotate and move axially through the central bore of the cannula.
In accordance with embodiments of the present invention, a surgical instrument is provided. This surgical instrument includes an elongate body having a positionable distal end and at least one bendable region, such as a joint region or a flexible region. An optical fiber bend sensor comprising one or more optical fibers is provided in the bendable region of the body. Each of these optical fibers includes a Fiber Bragg Grating, preferably a collinear array of Fiber Bragg Gratings. A strain sensor system comprising a light source and a light detector is used to measure strain in the optical fibers in order to determine a position and shape of the body. This shape and position information can be used to assist in controlling movement of the robotic manipulator and/or surgical instrument. This position information may include both translational and rotational position.
Fiber optic bend sensors for determining a shape of a structure have been used. For example, optical fibers including Fiber Bragg Gratings (FBG) have been used in a variety of applications for providing strain measurements in structures in one or more dimensions. Various systems and methods for monitoring the shape and relative position of an optical fiber in three dimensions are described in U.S. patent application publication no. 2006/0013523, filed on Jul. 13, 2005, U.S. provisional patent application Ser. No. 60/588,336, filed on Jul. 16, 2004, and U.S. Pat. No. 6,389,187, filed on Jun. 17, 1998, the disclosures of which are incorporated herein in their entireties.
In accordance with embodiments of the present invention, an optical fiber, such as the optical fibers described in U.S. patent application publication no. 2006/0013523, is utilized as optical fiber 220 for monitoring the shape and relative position of each body segment 212 in the instrument 200. This information, in turn, in can be used to determine other related variables, such as velocity and acceleration of the parts of a surgical instrument. By obtaining accurate measurements of one or more of these variables in real time, the controller can improve the accuracy of the robotic surgical system and compensate for errors introduced in driving the component parts. The sensing may be limited only to the degrees of freedom that are actuated by the robotic system, or may be applied to both passive (e.g., unactuated bending of the rigid members between joints) and active (e.g., actuated movement of the instrument) degrees of freedom.
In this embodiment, the optical fiber 220 comprises three cores contained within a single cladding. Each core may be single-mode with sufficient distance and cladding separating the cores such that the light in each core does not interact significantly with the light carried in other cores. In other embodiments, the number of cores may vary or each core may be contained in a separate optical fiber.
An array of Fiber Bragg Gratings is provided within each core. Each Fiber Bragg Grating comprises a series of modulations of the core's refractive index so as to generate a spatial periodicity in the refraction index. The spacing may be chosen so that the partial reflections from each index change add coherently for a narrow band of wavelengths, and therefore reflect only this narrow band of wavelengths while passing through a much broader band. During fabrication of the Fiber Bragg Gratings, the modulations are spaced by a known distance, thereby causing reflection of a known band of wavelengths. However, when a strain is induced on the fiber core, the spacing of the modulations will change, depending on the amount of strain in the core.
To measure strain, light is sent down the fiber, and the reflected wavelength is a function of the strain on the fiber and its temperature. This FBG technology is commercially available from a variety of sources, such as Smart Fibres Ltd. of Bracknell, England. When applied to a multicore fiber, bending of the optical fiber induces strain on the cores that can be measured by monitoring the wavelength shifts in each core. By having two or more cores disposed off-axis in the fiber, bending of the fiber induces different strains on each of the cores. These strains are a function of the local degree of bending of the fiber. Regions of the cores containing FBGs, if located at points where the fiber is bent, can thereby be used to determine the amount of bending at those points. These data, combined with the known spacings of the FBG regions, can be used to reconstruct the shape of the fiber. Such a system has been described by Luna Innovations. Inc. of Blacksburg, Va.
A control system 250 is also provided for detecting the position of the surgical instrument 200 and for utilizing that information to assist in surgical procedures. In this embodiment, the control system 250 comprises a detection system 260 coupled to an imaging system 270 and the servo controller 16. The detection system 260 is utilized for generating and detecting the light used for determining the position of the instrument 200. The imaging system 270 is used to provide the surgeon or other operator with real-time position information for use in the control of the instrument 200. The servo controller 16 may utilize the position information as feedback for positioning the instrument 200.
A bending of fiber 320 in one of the sensor regions 350 will lengthen at least one core 330a-330d with respect to the opposing core 330a-330d. Interrogation of this length differential enables the angle and radius of bending to be extracted. This interrogation may be performed using the detection system 260, as described below.
There are a variety of ways of multiplexing the Fiber Bragg Gratings so that a single fiber core can carry many sensors and the readings of each sensor can be distinguished. In some embodiments, Optical Frequency Domain Reflectometry (OFDR) may be used in which the Fiber Bragg Gratings, all with the same grating period, are placed along each of the cores, and each core is terminated at the proximal end with a partially reflecting mirror. The Fiber Bragg Gratings are placed in such a way that the distance from each grating to the reflector is unique, which causes the reflection spectrum of each Fiber Bragg Grating to be modulated with a unique modulation frequency, thereby allowing the individual reflection spectra to be determined. In addition, OFDR may be used to interrogate the array of Fiber Bragg Gratings with sufficiently low delays such that that the bending data can be used as a feedback signal in a real-time motion control loop.
In the embodiment shown in
For embodiments which detect position based on two degrees of freedom at each joint of the surgical instrument, at least three fiber cores 330 are utilized, with an interrogation system for each fiber core 330. In the embodiment illustrated in
As a result, the differences in strain between Fiber Bragg Gratings G1, G2 in the first pair of opposing cores 330a, 330c and the second pair of opposing cores 330b, 330d in each sensor region 350 can be used to determine the fiber's bending at each instrumented axial location corresponding to the sensor region 350. With four cores 330a-330d disposed equally around the fiber's neutral axis, the strains measured for each pair of opposing cores 330 are subtracted. The difference between strains of the opposing cores 330a and 330c is proportional to the degree of bending along the first axis 331, and the difference between strains of the opposing cores 330b and 330d is proportional to the degree of bending along the second axis 332.
In the embodiment shown in
In other embodiments, it may not be desirable or practical to position the fiber 220 on the neutral axis of the instrument 200. If the fiber 220 is positioned off-axis, then bending of the instrument 200 will change the length of the path of the fiber 220 through the channel 216 in the instrument 200. If the extent of bending is very small, then the fiber 220 may stretch or compress to accommodate the change in path length. Thus, the fiber 220 may have multiple fixed connection points with the instrument 200. This arrangement may be undesirable in some cases because increased axial loading of the fiber increases the common-mode signal in the fiber cores, thereby reducing the accuracy of the bending measurement. Considering a single pair of cores on opposite sides of the fiber's neutral axis, if there is no overall strain on the fiber, then bending at the sensor causes one FBG wavelength to shift down and the other to shift up by the same amount, thereby resulting in a zero common-mode signal. If, however, there is an overall strain on the fiber, both FBG wavelengths will shift in the same direction, in the absence of any bending at the sensor. Because each core is only a small distance (e.g., ˜75 μm) from the fiber's neutral axis, the strains on the individual cores caused by bending of the fiber are quite small Therefore, even a small axial strain on the fiber introduces a large common-mode signal, and gain errors between the two sensor paths may result in a false differential reading caused by the large common-mode signal, falsely indicating bending when the fiber is, in fact, straight.
If the extent of bending is large, then the change in path length can be greater than can be accommodated by the elasticity of the fiber 220. Thus, it may be desirable to have the fiber 220 translate axially relative to the instrument 200, causing the sensor regions 350 to also translate axially relative to the joint regions 214. One way to allow for this translation is to ensure that the length of each sensor region (e.g. 350a) exceeds the length of its associated joint region (e.g. 214a) by a sufficient amount, and that the joint regions are spaced sufficiently apart by rigid body segments 212, so that no matter how the joints are bent, the cumulative translation is not enough to shift the associated sensor region 350 away from the joint region, and also not enough to cause one sensor region to shift into the area of a neighboring joint region. Thus a one-to-one correspondence between joint regions and associated sensor regions is always maintained. This design will cause only a portion of the FBG to expand or contract with the joint bending, and thus the associated FBG reflectance spectrum will change shape with bending, rather than shifting up or down as they would if the entire FBG experienced the strain of bending.
Alternately, the known mechanical properties of the fiber 220 and the constraints on the shape of the fiber 220 when the instrument 200 is bent in various positions can be used to model the shifts in the Fiber Bragg Grating sensor readings. The joint positions can then be derived based on this model. This derivation can be accomplished by analyzing the shape of the reflection spectrum, which will change if a portion of the fiber in the FBG region is bent, but not the entire FBG region. Based on the observable effects on the reflection spectrum caused by bending confined to only a portion of the FBG region, and a forward model of joint angle to reflection spectrum can be derived. This forward model can then be inverted to obtain the joint angle estimate from the measured reflection spectrum.
In the simplified example shown in
Compensation for the movement of the sensor region 350 may be accomplished in a variety of ways. For example, the detection of the strain in the first sensor region 350 associated with the first joint region 214a can be used to determine the direction and extent of rotation of the first joint region 214a. Based on this information, the expected location of the second sensor region 350 relative to the second joint region 214b can be determined. This determination can be accomplished by calculating the axial displacement based on predetermined models, or by looking up the axial displacement in a pre-calibrated table.
This information, in turn, is used in conjunction with the detected signals reflected from the second sensor region 350 to determine the direction and extent of rotation of the second joint region 214b. The information regarding the axial shifting of the second sensor region 350 may be used to compensate for the change in reflected signal caused by the shifting. Based on the known position and orientation of the first body segment 212a, the calculated direction and extent of rotation of the first and second joint regions 214, and the known length of the third body segment 212c, the position of the distal end 218 of the instrument may then be determined. In embodiments having a greater number of body segments, a similar determination may be made for each body segment based on the information received from the preceding body segments.
A similar situation occurs when the bending of a joint region 214 is acute enough that the length of the bent region does not extend over the entirety of the sensor region 350, even though no significant axial translation of the fiber 220 relative to the joint region 214 occurs. Because the strained portion of the sensor region 350 does not encompass the entire length of the sensor region 350, a change in spectral shape, rather than a simple shift of the reflectance spectrum, will occur.
Alternately, the sensor regions 350 may be made short enough that axial translation of the fiber 220 will move them partially out of the joint regions 214. As with the case described above, the result of this bending applied to only part of the sensor regions 350 is that the shape of the FBG reflectance spectrum will shift, rather than a simple shift in the center wavelength of the spectrum. In this case, the known axial shift in the fiber 220 at each joint region 214 can be used to assist in determining the degree of bending of the fiber 220 at each successive joint region 214. If, for example, the fiber 220 is anchored at the proximal end of the structure, the axial position of the fiber 220 at the first joint region 214 is fixed, and therefore the effects of the bending of the first joint region 214 on the FBG reflectance spectrum (a shift in the spectrum, a shape change, or a combination of the two), is known, and a measurement of the reflectance spectrum shape and shift can be directly used to determine the bend angle of this first joint region 214. That bend angle, along with the known structure of the channel containing the sensor fiber 220, can be used to determine the degree of axial shift in the fiber 220 at the second joint region 214. Therefore, the effects of the bending of the second joint region 214 on the FBG reflectance spectrum of its associated sensor region 350 become known, and the measurement of that reflectance spectrum can be used to determine the bend angle of the second joint region 214. This procedure can be iterated for each joint region 214 in the structure serially.
Each of the plots 1000a-1000b shows the reflectance spectrum of the single Fiber Bragg Grating section in each core 1030a-1030b, respectively. The vertical axis corresponds to the intensity of the reflection, and the horizontal access corresponds to wavelength (with wavelength increasing from left to right). The first line 1001a-1001b corresponds to the reflectance spectrum intensity when the fiber is unbent, and therefore each Fiber Bragg Grating is at its nominal wavelength. The second line 1002a-1002b corresponds to the reflectance spectrum that would result if the entire length of the Fiber Bragg Grating region were bent. The third line 1003a-1003b shows the reflectance spectrum corresponding to the partial bending illustrated in
For fiber core 1030a, which is partially compressed by the bending, the reflectance spectrum of the bent section shifts to shorter wavelengths, but the partial reflections from this strained section Ra2 combine linearly with the partial reflections from the unstrained section Ra1. These reflections add coherently to produce the total reflection 1003a. Because part of the reflected light is shifted and part is not, the reflectance spectrum 1003a changes in both shape and center point position from the default spectrum 1001a. For the illustrated case, where the shift in the bent section is roughly equal to the spectral width of the reflection from the entire grating, the net result is a broadening of the spectrum, as illustrated by the third line 1003a. For more extreme shifts, the spectrum may split into two distinct peaks, and because the reflections add coherently and the detection system measures the intensity of the reflected light, the intensity of each peak will be one quarter of that of the unbent grating (where the grating is positioned so only half of it is within the joint region). For the second fiber core 1030b, which is partially extended by the bending, the same effect occurs, except that the partial reflection is shifted to longer wavelengths, as illustrated by the third line 1003b.
In cases where the fiber core is not uniformly stressed, the peak of the reflection spectrum is not directly proportional to the strain in the fiber core. Similar effects occur for other partially-bent sections of the fiber. Therefore, if the fiber cannot be placed on the neutral axis of the structure, or otherwise there is not a one-to-one mapping between the sensor regions and the joint regions, the changes in the shape can be used along with the shift in the peak wavelength to better determine the degree of strain in the bent portion of the fiber, and thereby the position of the joint region.
It is understood that the arrangement may vary in other embodiments. For example, the fiber 220 may be fixed to the distal end 218 of the instrument, rather than at the proximal end, as described above. Alternatively, the fiber 220 may be fixed at any known intermediate position. In yet other embodiments, the fiber 220 may be fixed at multiple intermediate and/or end positions. Overall changes in tension caused by stretching or contraction of the fiber 220 could be removed as common-mode noise, similar to a temperature disturbance. In embodiments in which the fiber is fixed at two points, it may be desirable to affix the fiber under tension in the neutral state so that any contraction caused by bending will result in a decrease of the pre-tension, rather than compression of the fiber. In addition, the number, shape, and size of the body segments 212 may vary. For example, fewer or greater joint regions 214 may be provided, thereby increasing the number of body segments 212. In addition, the body segments 212 need not be linear in shape and may include rigid curves or angles.
In accordance with other embodiments of the present invention, one or more of the body segments may comprise a flexible material. In this case, the bending of the instrument 200 is not limited to solely the joint regions 214. Thus, the constraints imposed on the fiber by the flexible structure may be used to model the system. This model may then be inverted either analytically or numerically to derive the instrument's state from the detected set of signals reflected from the Fiber Bragg Gratings. In other words, a forward model is generated that computes the strains expected at the FBG regions from an arbitrary but physically realizable bending of the structure. For example, the stiffness of the flexible section might force it to take the shape of a smooth spline curve when it experiences external forces at the ends and at points along its length, but not allow it to take a right-angle bend in the middle. In the forward model, the input would be the structure's degree of bending at each point, and the output is the predicted strains at the FBG regions. Inverting this model provides another model, this one taking as input the strains at the FBG regions, and providing as output the structure's degree of bending. In some cases this may be done analytically, but for complicated systems, the computation would be done numerically, in real time.
For example, the fixed position region 560 of the instrument 500 may be made of a compressible material, such as rubber, or include a compressible portion, such as a torsion spring. Thus, when the distal tip of the instrument 500 is used to apply a force in the axial direction onto a surface, the compressible fixed position region 560 will compress, thereby compressing the fiber 520. The sensor region 540 may be used to detect the axial loading of the fiber 520 in the fixed position region 560, thereby enabling determination of the force applied by the tip of the instrument 500. This determination of force may also include a determination of torque as well.
In other embodiments, the fixed position region 560 is substantially rigid so that no external load is applied to the portion of the fiber 520 in the fixed position region 560. In this case, changes in temperature of the fiber 520 and the rigid section 560 cause thermal expansion or contraction of the fiber 520 and the rigid section 560. Because the fiber 520 is fixed at anchor points 502-503, the net thermal expansion of this composite structure causes strain in the fiber 520. This strain combined with the temperature-induced change in the refractive index causes a shift in the Fiber Bragg Grating's reflection peak, which can then be detected as described above. The temperature may be determined by looking up a detected signal in a pre-calculated lookup table, thereby eliminating the need for comparison with the second sensor region.
In some embodiments, a second sensor region is provided adjacent to the fiber 520 in a second fiber that is not bounded by fixed anchor points. Thus, thermal expansion of the second fiber in this location causes axial movement of the second fiber and does not cause strain in the second fiber. The two fibers are placed close together, so they are at substantially the same temperature. The common-mode signal in each fiber is a function of both the temperature and the strain in each fiber, and since the second fiber is free to move axially, it experiences no strain and thereby provides an output that is a function of temperature only. The shift in the reflected signal from the second fiber can then be subtracted from the shift in the reflected signal from the first fiber, allowing determination of both the temperature and axial strain in region 540 independently.
In yet other embodiments, multiple sensor regions may be provided along the length of the fiber 520, thereby enabling the instrument 500 to be used for multiplexed temperature sensing. In yet other embodiments, the temperature may be measured in the sections of the fiber corresponding to joint regions. The shift in wavelength depends on both the strain in the fiber core, and on the temperature of the fiber core. In the bend-sensitive sections of the fiber, there is no axial strain (which causes a similar wavelength change in all the FBGs in a sensing region, and hence a common-mode signal). In these sections, the bending strain causes opposite shifts in the wavelength of the FBGs on opposite sides of the core, and hence a differential-mode signal. Adding the wavelength shifts for all the FBGs in a single sensitive region thus nulls the bending signal, and amplifies the temperature signal. Thus, the temperature measurement may be accomplished by using the sum of the oppositely disposed cores' readings rather than the difference, as long as there is no axial loading of the fiber.
In these temperature sensing embodiments, it may be desirable for the body segments 512 in the fixed position region 560 to be made of a thermally conductive material, such as a metal, or for at least a portion of the fiber 520 to be exposed to the exterior of the instrument 560 so as to be placed in contact with the target environment for which temperature detection is desired.
In the embodiment shown in
In accordance with other embodiments of the present invention, the fiber containing Fiber Bragg Gratings may also be used for other purposes, such as to provide illumination in a body cavity. For example, if the distal end of the fiber is exposed at the distal end of a surgical instrument, such as an endoscope, the fiber may be used to provide illumination for the endoscope. The Fiber Bragg Grating sensors typically operate in a small band in the infrared region of the light spectrum, e.g., 1.55 μm wavelength. In this case, the fiber may also be used to convey light for illumination to the end of the endoscope without interfering with the operation of the strain sensors, as long as the illuminating light is sufficiently removed in wavelength that the illuminating light can be filtered from the infrared light used to interrogate the Fiber Bragg Grating sensors. The illuminating light may be provided in the visible range to directly illuminate a scene.
The surgical instrument 600 includes a shaft portion 602, a control housing 603 provided at a proximal end of the shaft portion 602, and a working end 610 at a distal end 604 of the shaft portion 602. The working end 610 comprises three joint assemblies 620a-620c coupling three body segments 612a-612c to the distal end 604 of the shaft portion 602. An end effector 630 is coupled to the third body segment 612c via a wrist assembly 622 at the distal end of the instrument 600. The end effector 630 is the working part of the surgical instrument 600 and can include clamps, jaws, scissors, graspers, needle holders, micro-dissectors, staple appliers, tackers, suction irrigation tools, and clip appliers, and non-articulated tools, such as cutting blades, cautery probes, irrigators, catheters, and suction orifices.
In this embodiment, each of the joint assemblies 620a-620c provides one, two, three, or more degrees of freedom. An optical fiber, such as the fiber 220 described above, may be provided in a channel through the shaft portion 602, the joint assemblies 620a-620c, and the body segments 612a-612c, in order to provide position information regarding the working end 610 and, more particularly, the end effector 630 of the instrument 600. Because the shaft portion 602 is substantially rigid, the position and orientation of the distal end 604 can be determined based on the known position and orientation of the proximal end of the shaft, which is under the control of the manipulator assembly of the robotic system. Therefore, the optical fiber need not be used to determine the position of the shaft portion 602 and Fiber Bragg Gratings need not be provided along the portion of the fiber within the shaft portion 602.
Although it may be possible to estimate the position of the end effector 622 based on the control inputs to the joint assemblies 620a-620c, errors in the estimated position may be introduced due to stretching of the joint actuation cables or other conditions of the driving mechanism. Therefore, sensor regions of the fiber may be provided at each joint assembly 620 in order to determine the extent and orientation of any bending at that joint. This can, in turn, be used to determine the position and orientation of subsequent body segments 612 and ultimately the end effector 630. Roll movement of body segments may be accommodated by providing the sensing fiber in a channel so as to allow rotation of the fiber within the channel. The fiber is quite rigid in torsion, and rotates freely inside the channel, so that roll movement of the body segments would not cause rolling or twisting of the fiber. The extent of rolling may be measured using conventional actuator sensors. This rotational information may then be combined with the information obtained from the bend sensors to determine the orientation of the distal links in space, and to transform the two-dimensional bending determined by the fiber sensor into the degrees of freedom that are actuated at each joint.
In accordance with one aspect of the present invention, the information regarding joint bending obtained from the FBG bend sensors may be used as part of the feedback loop for the control system 250. As described above, actuator sensors may be used to measure the rotation of the motors used to actuate linear movement of the cables, which, in turn, actuates rotational movement of the actively controlled joints. Due to errors introduced by any structures or drivetrain mechanisms that are distal to the actuator sensors, the actual rotation of the joints may not correspond to the rotation measured by the actuator sensors. The bending detected by the bend sensors located in those joints can then be provided back to the servo controller 16 as part of a feedback loop to improve the control of the motors.
In accordance with another aspect of the present invention, FBG sensors may be used in a surgical instrument including both passive and actively-controlled joint regions.
Unlike the shaft portion 602 in
At the distal end 706 of the passive region P is a working end 710, which forms the active region A of the instrument 700. In this embodiment, the working end 710 comprises a pair of actively-controlled robotic arms 740 and an imaging device (e.g., a camera 750). In other embodiments, the working end 710 may have greater or fewer components. For example, the working end 710 may include only a single arm 740 or three or more arms 740, and the imaging device may be omitted.
Each of the robotic arms 740 may be similar in operation to the working end 610 described above with respect to
The instrument 700 may be inserted into a patient's body and controlled using a robotic surgical system similar to the robotic surgical system 2 shown in
As with the surgical system 2 described above, the movement of each joint assembly 720a-720c may be actuated by motors controlled by a servo controller (not shown). The motors may be coupled to the joint assemblies 720a-720c using, e.g., actuation cables. In instrument 600, the shaft 602 is rigid and linear, so the joint assemblies 620a-620c may be actuated using a combination of rigid shafts and cables made of a stiff material, such as, e.g., tungsten. In contrast, in instrument 700, the working end 710 is separated from the proximal end 705 of the instrument 700 by a flexible region. As a result, rigid shafts cannot be used to transmit the actuation force, and stiff cables may unacceptably limit the bending of the passive region P and experience high frictional forces when being linearly translated through a curved path. Thus, the actuation cables may comprise a less rigid material, such as a polymer or composite.
Although a polymer cable may not experience as much friction as a tungsten cable when passed through a channel in the passive region P, the polymer cable may experience significantly increased stretching, thereby increasing the difficulty of measuring the position of each joint assembly 720a-720c based on measurements of the actuator sensors. In addition, because the passive region P is not directly controlled by the operator, it is difficult for the operator to accurately determine the position and orientation of the distal end 706 inside of the patient's body. Thus, even if the positions of each joint assembly 720a-720c were capable of being accurately measured, the position and orientation of the end effectors 730 would only be known relative to the position and orientation of the distal end 706. Thus, the position and orientation of the end effectors 730 relative to an external coordinate system (e.g., defined by base 701 in a known location outside of the patient's body) will also be indeterminable.
In accordance with embodiments of the present invention, a multi-core optical fiber, such as the fiber 220 described above, may be provided in a channel through the passive region P and/or the active region A, in order to provide position information of the instrument 700, as described above. In one embodiment, the fiber may be provided only in passive region P, so that the position and orientation of the distal end 706 can be determined relative to the external coordinate system. The position and orientation of the end effectors 730 relative to the distal end 706 can then be determined based on the measurements by the actuator sensors.
In another embodiment, the fiber including FBG bend sensors may be provided only in the active region A. The position and orientation of the distal end 706 of the passive region P may be determined using any of the conventional methods described above. In this case, the position and orientation of each individual segment along the length of the passive region P may be irrelevant, so long as the position and orientation of the distal end 706 is known. Thus, it may be possible to use an external detection system to determine the position and orientation of the distal end 706. The FBG sensors in the fiber can then be used to determine the position and orientation of each joint assembly 720a-720c in the arms 740. This information can be combined with the externally detected position of the distal end 706 to provide accurate position and orientation information regarding the end effectors 730.
In yet another embodiment, the fiber including FBG bend sensors may be provided along the entire length of the instrument 700. Thus, the position and orientation along the entire length of the instrument 700 can be determined relative to an external coordinate system.
This position and orientation information can be used for a variety of purposes. For instance, the operator may utilize the determined position of the end effectors 730 in order to navigate to a site of interest in the patient's body. Alternatively, the control system may utilize this information as part of a feedback loop.
Because the load 806 is connected to the motor 804 through the drivetrain 805 which is necessarily compliant, the load 806 does not react immediately to changes in motor position. The time lag between motor rotation and load movement can make the load feedback too slow to close the velocity loop 803. In a simple case, one solution is to connect the bend sensor, which directly senses the load, to the slower position loop 802 (where accuracy is most important), and to connect the actuator sensor to the faster velocity loop 803 (where responsiveness is most important). In other cases, the dual loop feedback may utilize more complex algorithms for mixing the two sensor signals. In addition, unmodeled friction in the drivetrain can make the position and velocity of the load a poor indicator of the position and velocity of the motor, and thus inappropriate for use as an actuator sensor.
In accordance with other aspects of the present invention, FBG sensors may be used to locate a portion of a surgical instrument (e.g., the tip of the instrument) in a fixed external coordinate system in order to combine external data (such as image data from a magnetic resonance imaging (MRI) or ultrasound imaging system) with the positioning information to assist the operator in performing the surgical procedure.
For example, the detection system 260 may be coupled to the imaging system 270 to provide real-time position information for use in the control of the instrument 200. The imaging system 270 may include image data of the patient from an external source, such as a pre-operative MRI. The proximal end of the instrument 200 can then be supported by the surgical system 2 in a known position relative to the patient. Thus, one end of the sensing fiber may have a fixed position and angular orientation in a coordinate system external to the patient. For example, in a surgical robot, the proximal end of the fiber may be fixed to the frame holding the apparatus, which would be fixed in position relative to the patient during a period of use. As described above, the detection system 260 can then utilize the FBG sensors to determine the angles of all the joint regions in the instrument, including those that are not actively controlled. This information can then be combined with the known lengths of the structure's links to determine the position and orientation of each portion of the instrument.
A calibration step may be used, depending on whether the external data to be merged is available in the coordinate system of the frame that holds the fixed end of the fiber. In some cases (such as for pre-operative MRI images), the image data will be referenced to landmarks of patient anatomy. The image data can be aligned by identifying three patient landmarks that are visible in the external data with the robotic camera, and viewing these from two different angles. The landmarks may be internal or external anatomical features, or fiducials placed prior to imaging.
The sensor system, by providing in three dimensions unambiguous position and orientation information for each link in the instrument, enables this coordinate mapping for all parts of the instrument. This takes advantage of the ability of the OFDR interrogation system to multiplex many FBG sensors along a single fiber. In addition to the image registration described above, this allows the system to provide a richer set of features to the user. For example, the position of tools or other structure elements that are not currently in the field of view of a camera can be determined, and indicated to the user, and these features can be provided more easily across a wide range of minimally invasive surgery platforms with different architectures.
In some embodiments, the position information can be combined with the external image data to provide the operator with an image showing the location of the tip of the instrument on the MRI image in order to provide additional context for the surgical procedure.
In accordance with other embodiments, the external data may be combined with the image produced by a camera carried on the surgical instrument, such as, e.g., camera 750 in
In accordance with other embodiments of the present invention, FBG bend sensors in a surgical instrument may be used for calibration purposes in order to correct for manufacturing defects. These manufacturing defects may affect the dimensions and control of the various components of a surgical instrument, thereby causing the end effector and any intermediate components to be located in a position other than the assumed ideal. In order to calibrate the device, the control system will attempt to position the instrument in one or more known configurations. The FBG bend sensors in the instrument may be used to compare the measured position of the instrument to the expected position of the instrument. Any differences between the measured position and the expected position can be then be compensated for by the controller software.
The sensing of the position and orientation of the distal end of a surgical instrument enables the performance of a system identification on each device's actual kinematics, mapping out errors (e.g., differences from the ideal kinematics, such as those induced by manufacturing or material variation) automatically, and storing this error map in non-volatile memory on the device. Thus, the accuracy of the device can be improved. This mapping might occur once in the factory, or might occur each time the device is used. The latter would allow for correction of errors induced by aging or autoclave cycles, or any other time-dependent cause.
Embodiments of the present invention may provide various advantages not provided by prior art systems. For example, embodiments may enable more precise control of the elements located at the distal end of a robotically-controlled surgical instrument. In other embodiments, precise positioning and control may be achieved of a robotically-controlled surgical instrument having unactuated or flexible bends or joints.
In some embodiments, the cross-sectional area of the surgical instrument may be reduced because a single fiber can be used for multiple purposes, including sensing the position of all of the joints in the structure, sensing the forces applied to the structure, sensing temperature, and providing illumination for an imaging system. This may have particular applicability for endoscopes having a limited number of working ports.
In some embodiments, the angles of each of the joints or flexible portions of the surgical instrument may be mapped in real time. This can enable precise transformation of the location coordinate system at the instruments end effector to a global coordinate system. This transformation may enable the combination of data and images acquired by other systems with data acquired by the surgical system, such as, e.g., images obtained from endoscopic cameras mounted on the end effector.
In addition, by determining the position and orientation of the segments of the surgical instrument in a world reference frame, a flexible robotic surgery system can more completely mimic the operation of pre-existing minimally-invasive surgical systems that are designed to provide this information, such as, the da Vinci Surgical System, providing the ability to re-use user features and the code base developed for those systems. Thus a consistent surgeon experience can be provided, independent of the robot system architecture.
While the invention has been described in terms of particular embodiments and illustrative figures, those of ordinary skill in the art will recognize that the invention is not limited to the embodiments or figures described. For example, the position.
In many of the embodiments described above, the optical fiber comprises a single fiber having multiple cores. In other embodiments, the multiple single-core optical fibers may be used.
Embodiments described above include the use of actuator sensors for detecting movement of the actuators. In various embodiments, the actuator sensors may detect one or more of the position, velocity, acceleration, and applied force of the actuators. It will be understood that the detected applied force may be an applied torque, and the detected velocity and acceleration may be an angular velocity and angular acceleration. In addition, the sensed information provided by the actuator sensors to the control system may be provided in various forms and need not be limited to a direct measurement of position, velocity, acceleration, and applied force. For example, the actuation sensor may provide a logarithm of position to the control system, which, in turn, converts that value into a position.
In addition, the robotic surgical system described above may utilize cables in tension in order to actuate movement of the surgical instrument. In other embodiments, other actuation mechanisms, such as rods or tubes in compression and torsion, may be used to actuate movement. In addition, the FBG sensors may be used to detect the bending of passive joints as well. Some embodiments may have particular applicability for systems utilizing hydraulic or pneumatic mechanisms to actuate movement. Although hydraulic and pneumatic mechanisms are not subject to the frictional forces afflicting cables or tubes, these mechanisms introduce other control problems. For instance, in order to utilize hydraulics or pneumatics to transmit actuation forces through a flexible or multi-jointed structure having multiple degrees of freedom, very small cylinders that may not be completely sealed are used. The use of such small cylinders may result in a drift in the expected position of the actuator. Having real-time position feedback of the positions of the hydraulically or pneumatically actuated joints would ameliorate this problem.
In addition, the anchoring of the optical fiber cores relative to the instrument body may be accomplished in a variety of ways. For example, the optical fiber may be attached to an interior surface of the channel passing through the instrument. Alternatively, the optical fiber may be attached to a structure external to the robotic arm, but at a fixed position relative to the arm.
Therefore, it should be understood that the invention can be practiced with modification and alteration within the spirit and scope of the appended claims. The description is not intended to be exhaustive or to limit the invention to the precise form disclosed. It should be understood that the invention can be practiced with modification and alteration and that the invention be limited only by the claims and the equivalents thereof.
This application is a continuation of U.S. patent application Ser. No. 15/870,116, filed Jan. 12, 2018, which is a continuation of U.S. patent application Ser. No. 14/813,944, filed Jul. 30, 2015, and issued as U.S. Pat. No. 9,883,914 on Feb. 6, 2018, which is a continuation of U.S. patent application Ser. No. 13/049,003, filed Mar. 16, 2011, and issued as U.S. Pat. No. 9,125,679 on Sep. 8, 2015, which is a continuation of U.S. patent application Ser. No. 11/491,384, filed on Jul. 20, 2006, and issued as U.S. Pat. No. 7,930,065 on Apr. 19, 2011, which claims the benefit of and priority to U.S. provisional patent application Ser. No. 60/755,157, filed on Dec. 30, 2005, entitled “Modular Force Sensor,” all of which are incorporated by reference herein in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
3096962 | Meijs | Jul 1963 | A |
3546961 | Marton | Dec 1970 | A |
4437114 | LaRussa | Mar 1984 | A |
4673988 | Jansson et al. | Jun 1987 | A |
4721097 | D'Amelio | Jan 1988 | A |
4742817 | Kawashima et al. | May 1988 | A |
4792715 | Barsky et al. | Dec 1988 | A |
4809191 | Domeier et al. | Feb 1989 | A |
4949927 | Madocks et al. | Aug 1990 | A |
5114403 | Clarke et al. | May 1992 | A |
5174276 | Crockard | Dec 1992 | A |
5251611 | Zehel et al. | Oct 1993 | A |
5297536 | Wilk | Mar 1994 | A |
5307437 | Facq et al. | Apr 1994 | A |
5397323 | Taylor et al. | Mar 1995 | A |
5417210 | Funda et al. | May 1995 | A |
5429604 | Hammersmark et al. | Jul 1995 | A |
5487757 | Truckai et al. | Jan 1996 | A |
5617515 | Maclaren et al. | Apr 1997 | A |
5624380 | Takayama et al. | Apr 1997 | A |
5752112 | Paddock et al. | May 1998 | A |
5755713 | Bilof et al. | May 1998 | A |
5759151 | Sturges | Jun 1998 | A |
5792135 | Madhani et al. | Aug 1998 | A |
5797900 | Madhani et al. | Aug 1998 | A |
5817084 | Jensen et al. | Oct 1998 | A |
5855569 | Komi et al. | Jan 1999 | A |
5855583 | Wang et al. | Jan 1999 | A |
5868760 | McGuckin, Jr. | Feb 1999 | A |
5876325 | Mizuno et al. | Mar 1999 | A |
5892860 | Maron et al. | Apr 1999 | A |
5899425 | Corey, Jr. et al. | May 1999 | A |
5982791 | Sorin et al. | Nov 1999 | A |
5999662 | Burt et al. | Dec 1999 | A |
6030130 | Paddock et al. | Feb 2000 | A |
6063095 | Wang et al. | May 2000 | A |
6066090 | Yoon | May 2000 | A |
6120433 | Mizuno et al. | Sep 2000 | A |
6132368 | Cooper | Oct 2000 | A |
6171277 | Ponzi | Jan 2001 | B1 |
6191414 | Ogle et al. | Feb 2001 | B1 |
6246200 | Blumenkranz et al. | Jun 2001 | B1 |
6275628 | Jones et al. | Aug 2001 | B1 |
6331181 | Tierney et al. | Dec 2001 | B1 |
6332089 | Acker et al. | Dec 2001 | B1 |
6347892 | Paddock et al. | Feb 2002 | B1 |
6352503 | Matsui et al. | Mar 2002 | B1 |
6366722 | Murphy et al. | Apr 2002 | B1 |
6371952 | Madhani et al. | Apr 2002 | B1 |
6389187 | Greenaway et al. | May 2002 | B1 |
6396574 | Lee et al. | May 2002 | B1 |
6436107 | Wang et al. | Aug 2002 | B1 |
6441577 | Blumenkranz et al. | Aug 2002 | B2 |
6451027 | Cooper et al. | Sep 2002 | B1 |
6470205 | Bosselmann et al. | Oct 2002 | B2 |
6471710 | Bucholtz | Oct 2002 | B1 |
6478028 | Paolitto et al. | Nov 2002 | B1 |
6487352 | Sobiski et al. | Nov 2002 | B1 |
6537205 | Smith | Mar 2003 | B1 |
6554793 | Pauker et al. | Apr 2003 | B1 |
6569084 | Mizuno et al. | May 2003 | B1 |
6571639 | May et al. | Jun 2003 | B1 |
6572629 | Kalloo et al. | Jun 2003 | B2 |
6574355 | Green | Jun 2003 | B2 |
6575644 | Paddock et al. | Jun 2003 | B2 |
6578967 | Paddock et al. | Jun 2003 | B1 |
6645196 | Nixon et al. | Nov 2003 | B1 |
6648816 | Irion et al. | Nov 2003 | B2 |
6671055 | Wavering et al. | Dec 2003 | B1 |
6671581 | Niemeyer et al. | Dec 2003 | B2 |
6728599 | Wang et al. | Apr 2004 | B2 |
6743239 | Kuehn et al. | Jun 2004 | B1 |
6758843 | Jensen et al. | Jul 2004 | B2 |
6783491 | Saadat et al. | Aug 2004 | B2 |
6790173 | Saadat et al. | Sep 2004 | B2 |
6817974 | Cooper et al. | Nov 2004 | B2 |
6837847 | Ewers et al. | Jan 2005 | B2 |
6839612 | Sanchez et al. | Jan 2005 | B2 |
6866671 | Tierney et al. | Mar 2005 | B2 |
6902560 | Morley et al. | Jun 2005 | B1 |
6936042 | Wallace et al. | Aug 2005 | B2 |
6942613 | Ewers et al. | Sep 2005 | B2 |
6960162 | Saadat et al. | Nov 2005 | B2 |
6960163 | Ewers et al. | Nov 2005 | B2 |
6984203 | Tartaglia et al. | Jan 2006 | B2 |
6991602 | Nakazawa et al. | Jan 2006 | B2 |
6994703 | Wang et al. | Feb 2006 | B2 |
6997870 | Couvillon, Jr. | Feb 2006 | B2 |
6999852 | Green | Feb 2006 | B2 |
7041052 | Saadat et al. | May 2006 | B2 |
7042573 | Froggatt | May 2006 | B2 |
7090637 | Danitz et al. | Aug 2006 | B2 |
7090683 | Brock et al. | Aug 2006 | B2 |
7128708 | Saadat et al. | Oct 2006 | B2 |
7158860 | Wang et al. | Jan 2007 | B2 |
7194118 | Harris et al. | Mar 2007 | B1 |
7222000 | Wang et al. | May 2007 | B2 |
7252660 | Kunz | Aug 2007 | B2 |
7320700 | Cooper et al. | Jan 2008 | B2 |
7371028 | Gordon et al. | May 2008 | B2 |
7386365 | Nixon | Jun 2008 | B2 |
7594912 | Cooper et al. | Sep 2009 | B2 |
7725214 | Diolaiti | May 2010 | B2 |
7930065 | Larkin et al. | Apr 2011 | B2 |
8075498 | Leo et al. | Dec 2011 | B2 |
8281670 | Larkin et al. | Oct 2012 | B2 |
8419717 | Diolaiti et al. | Apr 2013 | B2 |
8784435 | Cooper et al. | Jul 2014 | B2 |
8864752 | Diolaiti et al. | Oct 2014 | B2 |
9039685 | Larkin et al. | May 2015 | B2 |
9060678 | Larkin et al. | Jun 2015 | B2 |
9060793 | Larkin et al. | Jun 2015 | B2 |
9066739 | Larkin et al. | Jun 2015 | B2 |
9084624 | Larkin et al. | Jul 2015 | B2 |
9101380 | Larkin et al. | Aug 2015 | B2 |
9125679 | Larkin et al. | Sep 2015 | B2 |
9241769 | Larkin et al. | Jan 2016 | B2 |
9283049 | Diolaiti et al. | Mar 2016 | B2 |
9526583 | Larkin et al. | Dec 2016 | B2 |
9757149 | Cooper et al. | Sep 2017 | B2 |
9883914 | Larkin et al. | Feb 2018 | B2 |
9962066 | Rogers et al. | May 2018 | B2 |
9980630 | Larkin et al. | May 2018 | B2 |
10398520 | Larkin et al. | Sep 2019 | B2 |
10448812 | Yoon et al. | Oct 2019 | B2 |
10456166 | Cooper et al. | Oct 2019 | B2 |
10588711 | DiCarlo et al. | Mar 2020 | B2 |
10799308 | Wang et al. | Oct 2020 | B2 |
10932877 | Devengenzo et al. | Mar 2021 | B2 |
10959607 | Rogers et al. | Mar 2021 | B2 |
11013567 | Wu et al. | May 2021 | B2 |
11135023 | Larkin et al. | Oct 2021 | B2 |
11166646 | Allenby | Nov 2021 | B2 |
11202680 | Donhowe | Dec 2021 | B2 |
11219490 | Barbagli | Jan 2022 | B2 |
11278354 | Duindam | Mar 2022 | B2 |
11278364 | Cooper et al. | Mar 2022 | B2 |
11298505 | Bailey | Apr 2022 | B2 |
11445988 | Chopra | Sep 2022 | B2 |
11464411 | Barbagli | Oct 2022 | B2 |
11478138 | Hazelton | Oct 2022 | B2 |
20010049509 | Sekine et al. | Dec 2001 | A1 |
20010055462 | Seibel | Dec 2001 | A1 |
20020007110 | Irion | Jan 2002 | A1 |
20020055795 | Niemeyer et al. | May 2002 | A1 |
20020058929 | Green | May 2002 | A1 |
20020087169 | Brock et al. | Jul 2002 | A1 |
20020099293 | Fontenot et al. | Jul 2002 | A1 |
20020120252 | Brock et al. | Aug 2002 | A1 |
20020143319 | Brock | Oct 2002 | A1 |
20020173786 | Kortenbach et al. | Nov 2002 | A1 |
20030050649 | Brock et al. | Mar 2003 | A1 |
20030167061 | Schlegel et al. | Sep 2003 | A1 |
20030228039 | Green | Dec 2003 | A1 |
20030236505 | Bonadio et al. | Dec 2003 | A1 |
20040054355 | Gerbi et al. | Mar 2004 | A1 |
20040083808 | Rambow et al. | May 2004 | A1 |
20040092794 | Chin et al. | May 2004 | A1 |
20040138525 | Saadat et al. | Jul 2004 | A1 |
20040138529 | Wiltshire et al. | Jul 2004 | A1 |
20040153191 | Grimm et al. | Aug 2004 | A1 |
20040186350 | Brenneman et al. | Sep 2004 | A1 |
20040193146 | Lee et al. | Sep 2004 | A1 |
20040202400 | Kochergin et al. | Oct 2004 | A1 |
20040249367 | Saadat et al. | Dec 2004 | A1 |
20050004431 | Kogasaka et al. | Jan 2005 | A1 |
20050043718 | Madhani et al. | Feb 2005 | A1 |
20050059960 | Simaan et al. | Mar 2005 | A1 |
20050065397 | Saadat et al. | Mar 2005 | A1 |
20050065398 | Adams | Mar 2005 | A1 |
20050075536 | Otsuka et al. | Apr 2005 | A1 |
20050096502 | Khalili | May 2005 | A1 |
20050096694 | Lee | May 2005 | A1 |
20050102062 | Green | May 2005 | A1 |
20050107663 | Saadat et al. | May 2005 | A1 |
20050131343 | Abrams et al. | Jun 2005 | A1 |
20050215983 | Brock | Sep 2005 | A1 |
20050222554 | Wallace et al. | Oct 2005 | A1 |
20050234293 | Yamamoto et al. | Oct 2005 | A1 |
20050250990 | Le et al. | Nov 2005 | A1 |
20050251176 | Swanstrom et al. | Nov 2005 | A1 |
20050272977 | Saadat et al. | Dec 2005 | A1 |
20050273084 | Hinman et al. | Dec 2005 | A1 |
20050284221 | Danisch et al. | Dec 2005 | A1 |
20060013523 | Childlers et al. | Jan 2006 | A1 |
20060025652 | Vargas | Feb 2006 | A1 |
20060095022 | Moll et al. | May 2006 | A1 |
20060149418 | Anvari | Jul 2006 | A1 |
20060161136 | Anderson et al. | Jul 2006 | A1 |
20060200049 | Leo et al. | Sep 2006 | A1 |
20060258938 | Hoffman et al. | Nov 2006 | A1 |
20070021738 | Hasser et al. | Jan 2007 | A1 |
20070043338 | Moll et al. | Feb 2007 | A1 |
20070055291 | Birkmeyer et al. | Mar 2007 | A1 |
20070065077 | Childers et al. | Mar 2007 | A1 |
20070089557 | Solomon et al. | Apr 2007 | A1 |
20070135803 | Belson | Jun 2007 | A1 |
20070137371 | Devengenzo et al. | Jun 2007 | A1 |
20070151391 | Larkin et al. | Jul 2007 | A1 |
20070156020 | Foley et al. | Jul 2007 | A1 |
20070197896 | Moll et al. | Aug 2007 | A1 |
20070225554 | Maseda et al. | Sep 2007 | A1 |
20070265503 | Schlesinger et al. | Nov 2007 | A1 |
20070283970 | Mohr et al. | Dec 2007 | A1 |
20070287884 | Schena | Dec 2007 | A1 |
20070287889 | Mohr | Dec 2007 | A1 |
20080058861 | Cooper et al. | Mar 2008 | A1 |
20080064921 | Larkin et al. | Mar 2008 | A1 |
20080064927 | Larkin et al. | Mar 2008 | A1 |
20080064931 | Schena et al. | Mar 2008 | A1 |
20080065097 | Duval et al. | Mar 2008 | A1 |
20080065098 | Larkin et al. | Mar 2008 | A1 |
20080065099 | Cooper et al. | Mar 2008 | A1 |
20080065100 | Larkin | Mar 2008 | A1 |
20080065101 | Larkin | Mar 2008 | A1 |
20080065102 | Cooper | Mar 2008 | A1 |
20080065103 | Cooper et al. | Mar 2008 | A1 |
20080065104 | Larkin et al. | Mar 2008 | A1 |
20080065106 | Larkin | Mar 2008 | A1 |
20080065107 | Larkin et al. | Mar 2008 | A1 |
20080065109 | Larkin | Mar 2008 | A1 |
20080065110 | Duval et al. | Mar 2008 | A1 |
20080071288 | Larkin et al. | Mar 2008 | A1 |
20080071289 | Cooper et al. | Mar 2008 | A1 |
20080071290 | Larkin et al. | Mar 2008 | A1 |
20080071291 | Duval et al. | Mar 2008 | A1 |
20080091170 | Vargas et al. | Apr 2008 | A1 |
20080119872 | Brock et al. | May 2008 | A1 |
20080147089 | Loh et al. | Jun 2008 | A1 |
20080151041 | Shafer et al. | Jun 2008 | A1 |
20080156971 | Ogisu et al. | Jul 2008 | A1 |
20080171908 | Okada et al. | Jul 2008 | A1 |
20080188890 | Weitzner et al. | Aug 2008 | A1 |
20080212082 | Froggatt et al. | Sep 2008 | A1 |
20080269559 | Miyamoto et al. | Oct 2008 | A1 |
20090054728 | Trusty et al. | Feb 2009 | A1 |
20090192524 | Itkowitz et al. | Jul 2009 | A1 |
20090248040 | Cooper et al. | Oct 2009 | A1 |
20090314131 | Bailey | Dec 2009 | A1 |
20090322001 | Luke et al. | Dec 2009 | A1 |
20090326322 | Diolaiti | Dec 2009 | A1 |
20090326552 | Diolaiti | Dec 2009 | A1 |
20090326553 | Mustufa et al. | Dec 2009 | A1 |
20100170519 | Romo et al. | Jul 2010 | A1 |
20100274087 | Diolaiti et al. | Oct 2010 | A1 |
20100332030 | Larkin et al. | Dec 2010 | A1 |
20100332033 | Diolaiti et al. | Dec 2010 | A1 |
20110040305 | Gomez et al. | Feb 2011 | A1 |
20110040404 | Diolaiti et al. | Feb 2011 | A1 |
20110082365 | McGrogan et al. | Apr 2011 | A1 |
20110152613 | Zubiate et al. | Jun 2011 | A1 |
20110152879 | Williams | Jun 2011 | A1 |
20110184241 | Zubiate et al. | Jul 2011 | A1 |
20110196419 | Cooper | Aug 2011 | A1 |
20110201883 | Cooper et al. | Aug 2011 | A1 |
20110202068 | Diolaiti et al. | Aug 2011 | A1 |
20110277576 | Cooper | Nov 2011 | A1 |
20110277579 | Anderson et al. | Nov 2011 | A1 |
20110277580 | Cooper et al. | Nov 2011 | A1 |
20110277775 | Holop et al. | Nov 2011 | A1 |
20110277776 | McGrogan et al. | Nov 2011 | A1 |
20110282356 | Solomon et al. | Nov 2011 | A1 |
20110282357 | Rogers et al. | Nov 2011 | A1 |
20110282358 | Gomez et al. | Nov 2011 | A1 |
20110282359 | Duval | Nov 2011 | A1 |
20120010628 | Cooper et al. | Jan 2012 | A1 |
20150265368 | Chopra et al. | Sep 2015 | A1 |
20160183841 | Duindam et al. | Jun 2016 | A1 |
20170265952 | Donhowe et al. | Sep 2017 | A1 |
20190206004 | Shelton, IV et al. | Jul 2019 | A1 |
20190365202 | Larkin et al. | Dec 2019 | A1 |
20190374293 | Larkin et al. | Dec 2019 | A1 |
20200106937 | Cooper et al. | Apr 2020 | A1 |
20200146754 | Row et al. | May 2020 | A1 |
20210228063 | Rogers et al. | Jul 2021 | A1 |
Number | Date | Country |
---|---|---|
1155833 | Jul 1997 | CN |
1486667 | Apr 2004 | CN |
0776738 | Jun 1997 | EP |
1334700 | Aug 2003 | EP |
S57190549 | Nov 1982 | JP |
H06285009 | Oct 1994 | JP |
H07504363 | May 1995 | JP |
2000093522 | Apr 2000 | JP |
2000166936 | Jun 2000 | JP |
2001327460 | Nov 2001 | JP |
2003275223 | Sep 2003 | JP |
WO-9313916 | Jul 1993 | WO |
WO-9729690 | Aug 1997 | WO |
WO-0051486 | Sep 2000 | WO |
WO-0217810 | Mar 2002 | WO |
WO-2004052171 | Jun 2004 | WO |
WO-2005087128 | Sep 2005 | WO |
WO-2005115226 | Dec 2005 | WO |
WO-2006039092 | Apr 2006 | WO |
WO-2007120329 | Oct 2007 | WO |
WO-2007146987 | Dec 2007 | WO |
WO-2008028149 | Mar 2008 | WO |
WO-2009002701 | Dec 2008 | WO |
Entry |
---|
Tanimoto et al., Micro force sensor for intravascular neurosurgery and in vivo experiment, 1998, IEEE, p. 504-509 (Year: 1998). |
Haga et al., Biomedical microsystems for minimally invasive diagnosis and treatment, 2004, IEEE, p. 98-114 (Year: 2004). |
Tanimoto et al., Micro force sensor for intravascular neurosurgery, 1997, IEEE, p. 1561-1566 (Year: 1997). |
Shapo et al., Displacement and strain imaging of coronary arteries with intraluminal ultrasound, 1996, IEEE, p. 234-246 (Year: 1996). |
Abbott, Daniel J. et al., “Design of an Endoluminal NOTES Robotic System,” Conference on Intelligent Robots and Systems, 2007, pp. 410-416. |
Anisfield, Nancy; “Ascension Technology Puts Spotlight on DC Field Magnetic Motion Tracking,” HP Chronicle, Aug. 2000, vol. 17, No. 9, 3 Pages. |
Ascari, Luca et al., “A New Active Microendoscope for Exploring the Sub-Arachnoid Space in the Spinal Cord,” Proc. IEEE International Conference on Robotics and Automation, 2003, pp. 2657-2662, vol. 2, IEEE. |
Barnes Industries, Inc., “How a Ball Screw Works,” 4 pages, Copyright 2007; Internet: http://www.barnesballscrew.com/ball.htm. |
Berthold III, John W., “Historical Review of Microbend Fiber-Optic Sensors,” Journal of Lightwave Technology, vol. 13, No. 7, Jul. 1995, pp. 1193-1199. |
Blue Road Research, “Overview of Fiber Optic Sensors,” 40 pages, first posted Dec. 8, 2004. Internet www.bluerr.com/papers/Overview_of_FOS2.pdf. |
Cao, Caroline G.L., “Designing Spatial Orientation in Endoscopic Environments,” Proceedings of the Human Factors and Ergonomics Society 45th Annual Meeting, 2001, pp. 1259-1263. |
Cao, Caroline G.L., “Disorientation in Minimal Access Surgery: a Case Study,” Proceedings of the IEA 2000/HFES 2000 Congress, pp. 4-169-4-172. |
Childers, Brooks A., et al., “Use of 3000 Bragg grating strain sensors distributed on four eight-meter optical fibers during static load tests of a composite structure,” SPIE 8th International Symposium on Smart Structures and Materials, Mar. 4-8, 2001, Newport Beach, California, 10 Pages. |
Choi, Dong-Geol et al., “Design of a Spring Backbone Micro Endoscope,” Conference on Intelligent Robots and Systems, 2007, pp. 1815-1821. |
Co-pending U.S. Appl. No. 11/762,185, inventor Thomas; G. Cooper, filed Jun. 13, 2007. |
Cowie, Barbara M., et al., “Distributive Tactile Sensing Using Fibre Bragg Grating Sensors for Biomedical Applications,” 1st IEEE / RAS-EMBS International Conference on Biomedical Robotics and Biomechatronics (BioRob 2006), Feb. 2006, pp. 312-317. |
Danisch, Lee et al., “Spatially continuous six degree of freedom position and orientation sensor,” Sensor Review, 1999, vol. 19, Issue 2, pp. 106-112. |
Dario, Paolo et al., “A Miniature Device for Medical Intracavitary Intervention,” Micro Electro Mechanical Systems '91 Proc IEEE ‘An Investigation of Micro Structures, Sensors, Actuators, Machines and Robots’, 1991, pp. 171-175, IEEE. |
Duncan, Roger, “Sensing Shape: Fiber-Bragg-grating sensor arrays monitor shape at a high resolution,” 2005, pp. 18-21, SPIE. |
Extended European Search Report for Application No. EP20070798487, dated Jan. 30, 2015, 8 pages. |
Gagarina, T. et al., “Modeling and experimental analysis of a new bellow type actuators for active catheter end-effector,” Proc. 10th IEEE International Workshop on Robot and Human Interactive Communication, 2001, pp. 612-617, IEEE. |
Gander, M.J. et al., “Bend measurement using Bragg gratings in multicore fibre,” Electronics Letter, Jan. 20, 2000, vol. 36, No. 2, 2 Pages. |
Hill, Kenneth O., “Fiber Bragg grating technology fundamentals and overview,” IEEE Journal of Lightwave Technology, vol. 15, Issue 8, Aug. 1997, pp. 1263-1276. |
Ikuta, Koji et al., “Development of remote microsurgery robot and new surgical procedure for deep and narrow space,” Proc. IEEE International Conference on Robotics & Automation, 2003, pp. 1103-1108, vol. 1, IEEE. |
Ikuta, Koji et al., “Shape memory alloy servo actuator system with electric resistance feedback and application for active endoscope,” Proc. IEEE International Conference on Robotics and Automation, 1988, pp. 427-430, vol. 1, IEEE. |
International Search Report for application No. PCT/US07/71085, dated Sep. 17, 2008, 2 pages. |
Jin, Long et al., “Two-dimensional bend sensing with a cantilever-mounted FBG [Fiber Bragg Grating],” Meas. Sci. Technol., 2006, pp. 168-172, vol. 17, Institute of Physics Publishing. |
Kreger, Stephen et al., “Optical Frequency Domain Refiectometry for High Density Multiplexing of Multi-Axis Fiber Bragg Gratings,” 16th International Conference on Optical Fiber Sensors (OFS-16), Oct. 2003, Nara, Japan, pp. 526-529. |
Lertpiriyasuwat, Vatchara et al., “Extended Kalman Filtering Applied to a Two-Axis Robotic Arm with Flexible Links,” International Journal of Robotics Research, 2000, vol. 19., No. 3, pp. 254-270. |
Lunwei Z., et al., “FBG Sensor Devices for Spatial Shape Detection of Intelligent Colonoscope,” IEEE International Conference on Robotics and Automation, Apr. 2004, New Orleans, Louisiana, pp. 835-840. |
Martinez, A. et al., “Vector Bending Sensors Based on Fibre Bragg Gratings Inscribed by Infrared Femtosecond Laser,” Electronics Letters, 2005, pp. 472-474, vol. 41—Issue 8. |
Measurand, “ShapeTape Overview,” Measurand ShapeTape Advantage, pp. 1-3, first posted Nov. 3, 2004. Internet www.measurand.com/products/ShapeTape_overview.html. |
Meltz, Gerald, “Overview of Fiber Grating-Based Sensors,” Proceedings of SPIE Distributed Multiplexed Fiber Optic Sensors VI, Nov. 27, 1996, Eds. Kersey et al.,vol. 2838, pp. 2-22. |
Office Action dated Jun. 17, 2014 for Japanese Application No. 20130179563 filed Aug. 30, 2013, 7 pages. |
Payne C.J., et.al., “A Hand-Held Flexible Mechatronic Device for Arthroscopy,”IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2015, pp. 817-823. |
PCT/US07/71085 Written Opinion, dated Sep. 17, 2008, 5 pages. |
PCT/US09/46446 International Search Report and Written Opinion of the International Searching Authority, dated Dec. 14, 2009, 21 pages. |
PCT/US09/46446 Partial International Search Report and Invitation to Pay Additional Fees, dated Sep. 18, 2009, 9 pages. |
PCT/US2011/035113 International Search Report and Written Opinion of the International Searching Authority, dated Aug. 4, 2011, 13 pages. |
Russo et.al., “A Novel Robotic Platform for Laser-Assisted Transurethral Surgery of the Prostate,” IEEE Transactions on Biomedical Engineering, Feb. 2015, vol. 62 (2), pp. 489-500. |
Sareh S., et.al., “A 7.5mm Steiner Chain Fibre-Optic System for Multi-Segment Flex Sensing,” IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2015, 2336-2341. |
Searle T.C., et.al., “An Optical Curvature Sensor for Flexible Manipulators,” IEEE International Conference on Robotics and Automation, May 2013, pp. 4415-4420. |
Stieber, Michael E. et al., “Vision-Based Sensing and Control for Space Robotics Applications,” IEEE Transactions on Instrumentation and Measurement, Aug. 1999, vol. 48, No. 4, pp. 807-812. |
Sturges, Robert H. et al., “A Flexible, Tendon-Controlled Device for Endoscopy,” the International Journal of Robotics Research, 1993, vol. 12(2), pp. 121-131. |
Szewczyk, Jerome et al., “An active tubular polyarticulated micro-system for flexible endoscope,” Lecture Notes in Control and Information Sciences, vol. 271, Experimental Robotics VII, 2000, pp. 179-188, Springer-Verlag. |
U.S. Appl. No. 60/813,028, inventor Thomas Cooper; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,029, inventor David Q. Larkin; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,030, inventor David Q. Larkin; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,075, inventor David Q. Larkin; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,125, inventor David Q. Larkin; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,126, inventor Thomas G. Cooper; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,129, inventor Thomas G. Cooper; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,131, inventor Eugene Duval; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,172, inventor Thomas G. Cooper; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,173, inventor David Larkin; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,198, inventor David Q. Larkin; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,207, inventor Nicola Diolaiti; Palo Alto, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 60/813,328, inventor Catherine Moore; Menlo Park, Ca, filed Jun. 13, 2006. |
U.S. Appl. No. 61/334,978, inventor McGrogan; Anthony, filed May 14, 2010. |
Vertut, Jean and Phillipe Coiffet, Robot Technology: Teleoperation and Robotics Evolution and Development, English translation, Prentice-Hall, Inc., Inglewood Cliffs, NJ, USA 1986, vol. 3A, 332 pages. |
Wang, Yi-Ping et al., “A novel long period fiber grating sensor measuring curvature and determining bend-direction simultaneously,” IEEE Sensors Journal, 2005, pp. 839-843, vol. 5—Issue: 5, IEEE. |
Webster, Robert J. III et al., “Toward Active Cannulas: Miniature Snake-Like Surgical Robots,” 2006, 7 pages. |
Wong, Allan C. L. et al., “Multiplexed fibre Fizeau interferometer and fibre Bragg grating sensor system for simultaneous measurement of quasi-static strain and temperature using discrete wavelet transform,” Measurement Science and Technology, 2006, pp. 384-392, vol. 17—Issue 2, Institute of Physics Publishing. |
Extended European Search Report for Application No. EP21201271, dated May 30, 2022, 9 pages. |
Bergamasco M., et al., “Advanced Interfaces for Teleoperated Biomedical Robots,” IEEE Engineering in Medicine & Biology Society 11th Annual Int'l Conf., pp. 912-913, vol. 3, IEEE, Jun. 1989. |
Chen M.D., et al., “A Robotics System for Stereotactic Neurosurgery and Its Clinical Application,” IEEE, May 1998, pp. 995-1000. |
Mack M.J., et al., “Minimally Invasive and Robotic Surgery,” Opportunities for Medical Research, Feb. 2001, vol. 285 (5), pp. 568-572. |
Number | Date | Country | |
---|---|---|---|
20220047341 A1 | Feb 2022 | US |
Number | Date | Country | |
---|---|---|---|
60755157 | Dec 2005 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15870116 | Jan 2018 | US |
Child | 17477846 | US | |
Parent | 14813944 | Jul 2015 | US |
Child | 15870116 | US | |
Parent | 13049003 | Mar 2011 | US |
Child | 14813944 | US | |
Parent | 11491384 | Jul 2006 | US |
Child | 13049003 | US |