COMMUNICATION OF DETECTED TISSUE CHARACTERIZATION IN A SURGICAL ROBOTIC PLATFORM

Information

  • Patent Application
  • 20210307861
  • Publication Number
    20210307861
  • Date Filed
    March 12, 2018
    6 years ago
  • Date Published
    October 07, 2021
    3 years ago
Abstract
A system and method for displaying tissue information on a display screen includes a camera capturing an image of an anatomical working site including a surgical instrument, and a display displaying the image. A tissue information detection device determines information about the health, pathology etc. of tissue in contact with a surgical instrument, or about an interaction between the surgical instrument and the tissue. The system renders overlays depicting the tissue information with the image on the display.
Description
FIELD OF THE INVENTION

The invention relates generally to communication of information about a surgical procedure to operating room personnel. More specifically, the invention relates to communication of information concerning tissues, objects or instruments within the operative working site to operating room personnel, including the surgeon conducting the procedure.


BACKGROUND

There are several methods and systems for using multi- or hyper-spectral imaging for in vivo tissue diagnosis. These allow users to intra-operatively distinguish between different types of tissue, whether different organs, vessels or even cancerous versus benign tissue. Many of these techniques and systems are used in the endoscopic field as an alternative to biopsies. Other technologies exist for obtaining information about tissues or object within the operative working site. For example, computer vision can be used to identify the boundaries of surgical instruments “seen” by the surgical scope within the operative site.


There are various types of surgical robotic systems on the market or under development. Some surgical robotic systems use a plurality of robotic arms. Each arm carries a surgical instrument, or the camera used to capture images from within the body for display on a monitor. Other surgical robotic systems use a single arm that carries a plurality of instruments and a camera that extend into the body via a single incision. These types of robotic systems use motors to position and orient the camera and instruments and, where applicable, to actuate the instruments. Input to the system is generated based on input from a surgeon positioned at master console, typically using input devices such as input handles and a foot pedal. Motion and actuation of the surgical instruments and the camera is controlled based on the user input. The image captured by the camera is shown on a display at the surgeon console. The console may be located patient-side, within the sterile field, or outside of the sterile field.


Some surgical imaging systems offer intra-operative tissue diagnostic capability using fluorescence. Often fluorescence imaging relies on the introduction of fluorescent agents (such as indocyanine green (ICG)) that are administered to tissue and illuminated with a certain wavelength of light. Some tissue types auto-fluoresce when exposed to light of particular wavelengths, allowing fluorescence imaging of those tissues. In systems offering fluorescence, the surgeon can view the fluorescence images by observing the system's camera display.


This application describes methods for presenting tissue information, such as the results of tissue diagnosis techniques or tissue-identifying steps, or instrument information, to surgeons or users of a surgical robotic system.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 shows a surgical robotic system which may use aspects of the disclosed invention.



FIGS. 2(a) and 2(b) show examples of a video display in which color overlays are used over a portion of an image of a surgical instrument captured using the endoscopic system.



FIG. 3 illustrates an alternative overly display format suitable for displaying surgical instrument information on a camera image display.



FIGS. 4(a) and 4(b) illustrate another embodiment of an overlay display format suitable for displaying tissue information on a camera image display.



FIG. 5 illustrates another embodiment of an overlay display format.





DETAILED DESCRIPTION

This application describes a system and method for presenting to the surgeon information obtained about the health, type, nature, location etc (referred to here as “tissue information”) of tissue that lies within the field shown on the video display being observed by the surgeon. The displayed tissue information may be information obtained using a tissue information detection system that can detect tissue density, inflammation, ischemia, oxygen levels, blood presence (e.g. via transmittance loss, hemoglobin detection or blood flow detection) or identifies types of tissue using fluorescent techniques or other methods, or that represents an interaction between a surgical instrument and tissue (e.g. squeezing force applied by forceps, or the existence/absence of contact between a part of an instrument and adjacent tissue). Other tissue information might be obtained information about the location of tissue structures or features such as tumors, blood vessels, ureters, lymph nodes, etc. In many of the described embodiments, the information is communicated by rendered overlays (also referred to as indicia) displayed on the endoscopic display.



FIG. 1 shows a system that may incorporate the inventions described herein. The system 10 comprises at least one robotic arm which acts under the control of a control console 12 managed by the surgeon who may be seated at the console. The system shown in FIG. 1 includes multiple robotic arms 14, 15, 16. Three such arms are shown but a larger or smaller number may be used. Each robotic arm can support and operate a surgical instrument 1, 2, 3 for use on a patient. One of the instruments 3 is preferably a camera which records the operating field inside the patient. These images are shown in a display 23 at the surgeon console 12.


The arms 14-16 are operated by an electronic control unit 30 which causes the arms to perform the movements entered via the console 12. The unit 30 will receive the high-level movement commands (for example, desired position and inclination of the tool supported by the robot) and will execute them, converting them into the corresponding sequences of signals to be sent to the individual motors of the robot arm articulations. The console includes input devices 17, 18 which can be gripped by the surgeon and moved so as to deliver instructions to the system as to the desired movement and operation of the instruments supported by the arms. The surgeon's movements are suitably reproduced by the surgical instruments by means of movement of the robotic arms. The input devices may be equipped to provide the surgeon with tactile feedback so that the surgeon can feel on the input devices 17, 18 the forces exerted by the instruments on the patient's tissues. The console may also include a keyboard 19 and/or touch screen and/or other command input devices. These other command devices might include a pedal device 20, and a button(s) on or in proximity to one or both handles of the input devices 17, 18. Other features are described in Applicant's U.S. Pat. No. 9,360,934 which is incorporated herein by reference.


Components of a system for communicating detected tissue information include one or more surgical instrument moveable at a surgical site, such as the instruments 1, 2, and means for determining the location of at least a part of the surgical instrument at the surgical site. For example, a computer vision system can be used to recognize the instrument pose in the image of the instruments captured by the endoscopic camera 3. In other embodiments, the pose can be determined by a processor of the robotic system based on kinematics.


The system further includes a (preferably real-time) tissue information detection feature which captures tissue information during the course of the procedure. This might be a fluorescence system, an optical tissue interrogation system such as that described in U.S. Regular application Ser. No.______ entitled Apparatus for Optical Tissue Interrogation Using a Surgical Instrument” (Attorney Docket: TRX-11910) filed on the same day as the present application, which is incorporated herein by reference, or other types of systems for collecting information about the health, type, nature, location etc of tissue.


Another form of tissue information might relate to the interaction of the tissue with instruments in use by the surgeon. For example, the robotic system might determine the amount of force being applied to tissue by the jaws that are grasping the tissue. This force information may come from any number of sources, including, but not limited to Fabry-Perot interferometry, Fiber-Bragg gratings, strain gages, load cells, force sensors inside the instrument itself, force sensors inside the robotic surgical system to which the instrument itself is attached, etc.


Another type of instrument for obtaining tissue information that may represent an interaction with the tissue will be discussed below in connection with FIG. 5 and is described in detail in U.S. Regular application Ser. No.______ entitled Apparatus for Optical Tissue Interrogation Using a Surgical Instrument” (Attorney Docket: TRX-11910). Briefly stated, this type of instrument may be used to obtain tissue information representing the existence or absence of contact between a part of the instrument and tissue, or an identification of the part of the instrument that is contacting tissue and the part of the instrument that is not contacting tissue.


The system further includes a communication device configured to generate a visual, auditory and/or tactile notification communicating the tissue information for tissue in contact with in proximity to, or in the expected path the surgical instrument or a particular part of the surgical instrument.


In a first embodiment, communicated tissue information takes the form of visual overlays or cues displayed on the camera display 23 along with the video information captured by the camera of the vision system. The tissue information may be displayed in real time as it is obtained from the tissue information detection system.


In some implementations, the overlays may be shown directly on the endoscopic view, even to the extent of being positioned on the image of surgical instrument tips themselves. FIGS. 2(a) and 2(b) show examples of a video display in which color overlays are used over a portion of an image of a surgical instrument captured using the endoscopic system. These figures illustrate use of color-changing or pattern-changing overlays to communicate to the surgeon tissue information about tissue in question. When a surgeon is grasping tissue using an instrument, the instrument is visible on the image shown on the camera display. This first embodiment is configured to overlay shading of a particular color or pattern over the image of the jaws, with the color or pattern representing a particular type of tissue information (e.g. tissue pathology or certain metrics of the tissue), or informing the user as to whether tissue information is available. In this example, as the user uses an instrument to grasp a first tissue T1 (FIG. 2(a)), the jaws of the instrument on the image display of the instrument are colored using an overlay G that is green. When the instrument is then used to grasp a second tissue T2 (FIG. 2(b)), the jaws of the instrument on the image display are colored using an overlay P that is pink. The system may change the jaw overlay for the instrument grasping tissue to a first color or pattern to a second color or pattern depending on the type of information to be conveyed using the overlays. One of the colors/patterns may represent one specific tissue pathology or metric, and the other color/pattern may represent a second tissue pathology/metric. Alternatively, one of the colors/patterns may serve to indicate that tissue pathology or other metrics are available for that tissue, and the other color/pattern may serve to indicate that tissue pathology/metrics are unavailable. Where information is available, the user may be prompted to give a command to the system using a manual, verbal, eye tracking, or other form of input to display the available information. Upon receiving the command, the system may display the information on the screen.


It can be appreciated that this embodiment can be adapted for instruments that do not have jaws. For example, the overlay may be rendered over the tip of an instrument probe that is in contact with the tissue in question.



FIG. 3 depicts an image of a surgical instrument grasping tissue as it might appear on the display 23, and it shows that the overlays may appear over parts of the instrument other than the jaws. In this example, green overlays G are positioned at the instrument tips, as well as on the shaft of the instrument. It will be appreciated that many variations are conceivable within the scope of the disclosed invention. Overlays may also be rendered on the endoscopic view between the jaws, at or near the jaw pivot, or any combination of the locations described here. In other implementations, the overlays may be made on rendered images of the surgical instruments on another portion of the screen


Where the tissue information to be communicated is a depiction of the force being applied to the tissue by an instruments jaws, overlays such as those shown in FIGS. 2(a)-3 might show continuous changes along a color spectrum as the force increases or decreases, or stepwise color changes (e.g. green to yellow to red) as the force moves between force thresholds. Alternatively, the system might render a graphic on the display that increases or decreases in size as the force increases, and that is resized in the opposite direction as the force decreases. Any combination of the described visual renderings might also be used.



FIGS. 4(a) and 4(b) also depict images of a surgical instrument as it might appear on the display using a different type of information overlay. FIG. 4(a) illustrates that a color or pattern overlay may be superimposed over grasped tissue, or a portion of the grasped tissue. Here grasped tissue T3 has a pink overlay C over it. This overlay may be within the boundaries of the grasped tissue if the boundaries are determined (e.g. using the computer vision system). As described above, different colors/patterns might represent different tissue pathologies/metrics, or indicate the presence/absence of pathology information or metrics, or simply serve to identify the boundaries of a particular type of tissue (blood vessel, tumor etc) to the surgeon. Where information is available, the user may be prompted to give a command to the system as noted above to display the available information.



FIG. 4(b) is similar to FIG. 4(a) but shows that the overlay can be used to indicate a difference in the pathology within the tissue that is being grasped, possibly alerting the surgeon to the location of a vessel or tumor material within the tissue being grasped. In this figure, a portion of the grasped tissue has a pink overlay C over it.



FIG. 5 depicts an image of a surgical instrument as it might appear on a display, together with alternate rendered overlays representing obtained information. In this case the obtained tissue information may represent an interaction of the tissue with the edge of the instrument, namely the existence or lack of existence of contact between a part of the instrument and tissue, or an identification of the part of the instrument that is contacting tissue and the part of the instrument that is not contacting tissue. In the specific overlays shown, triangles mark a portion of the jaw exterior that is in contact with tissue, and the circles mark a portion of the jaw exterior that is not in contact with tissue. Alternatively, these overlays might represent that tissue information is available or unavailable for tissue in contact with that part of the jaw, or they might represent a particular characteristic of that tissue. The detected information depicted by the overlay may be from a single point of measurement, multiple discrete measurement points, or a continuous measurement or feedback area.


The overlays in an embodiment such as that shown in FIG. 5 may be formed of various shapes, patterns icons, line types, or any combination thereof. Other information that might be communicated using this type of overlay includes a representation of the tissue contact force, optionally using some of the color changing principles (or other types of changes such as pattern changes, or changes in line weights or overlay size) discussed above for depicting changes in contact force.


The concepts described here may also be used to present to the surgeon information obtained about other structures present in the field of view of the video display being observed by the surgeon. Overlays of the type shown in FIGS. 2(a) through 3 and described in the accompanying text may be rendered to highlight to the surgeon where the tips of the jaws are and/or where the shaft of the grasper is. This information may be determined using computer vision/image processing to identify the edges of the tool, or identify markers or fiducials attached to or etched onto the instrument shaft or jaws. The overlay may be positioned by the system over the fiducials or another area of the instrument.


Supplemental embodiments may include the use of text captions or other means of highlighting important areas including outlines or enclosed volumes around specific points of interest.


Additional forms of communication to the user might include the use of sound. This might be used to help the surgeon search the work site for objects/tissue of particular importance. For example, the system might emit a sound that changes in frequency in one direction (i.e. higher or lower) as the surgeon moves the instrument closer to a tissue type or object that is of interest, and that changes frequency in the opposite direction as the instrument is moved away from the tissue type/object. This can be particularly useful when searching for tissues/objects that are below the surface of the tissue at the worksite. In other implementations, a certain tone may also occur when the distance between the instrument and the tissue crosses a defined distance threshold. In a variation of this embodiment, multiple thresholds with multiple tones may be used, each tone indicating that a different defined distance threshold has been crossed. These segmented thresholds may provide more clear assurance that a certain distance to/from an object or structure has been achieved. In another variation, This may also be accomplished with a series of clicks whose inter-click period changes as the distance changes (analogous to a Geiger counter). Audio feedback could also be combined with visual feedback to report tissue characteristics to the surgeon, or used alone to indicate when the user is touching tissue having characteristics of interest.


Additional embodiments may include haptic or tactile feedback—e.g. using linear resonating actuators or eccentric motors within the handles grasped by the user at the surgeon console—to generate haptic cues to communicate not only that the surgeon has grasped or touched tissue, but also the pathology of the tissue being grasped. As an example, the handles may vibrate during removal of cancerous tissue if the surgeon has started to remove adjacent viable healthy tissue. The frequency of the haptic/tactile feedback is selected to stimulate receptors in the user's hands, without inducing unintended input motion of the user's hands. Alternatively, vibrations may be induced by a haptic controller using the motors already included in the handles 17, 18 for providing the 3D haptic feedback that allows the surgeon to feel on the input devices 17, 18 the forces exerted by the instruments on the patient's tissues.


In some embodiments, the metrics or overlays may persist in the image, and the surgical system may keep track of the tissue metrics in 3-dimensional space. The position and orientation of the end effectors is well-known by the surgical robotic system along with their relation to the endoscope (either because the endoscope is also controlled by the surgical robotic system or via machine vision techniques). Thus, it is possible to build up a 3-dimensional map of tissue information. These overlays may persist over time, or be recorded to be recalled to the user interface at some time in the future.


Structural scanning and mapping and/or tissue deformation modeling may be used to keep the overlay positioning accurate as the soft tissue deforms.


The concepts disclosed in this application allow for real-time communication of pathology results to the surgeon through a robotic surgical console through any of the following means: visual overlays, outlines, or captions, audible descriptions and surgeon console handle vibrations.


Applications and patents referred to herein, including for purposes of priority, are incorporated herein by reference.

Claims
  • 1. A method for displaying tissue information on a display screen comprising: (a) during a surgical procedure, capturing an image of an anatomical working site using a scope device and displaying the image on a display visible to a surgeon;(b) during step (a), obtaining tissue information about tissue within the working site; and(c) displaying indicia depicting the tissue information with the image on the display.
  • 2. The method according to claim 1, wherein the image includes an image of a surgical instrument within the working site, the indicia includes graphical marking on the image of the surgical instrument.
  • 3. The method according to claim 2, where the graphical marking includes color shading or a pattern on at least a portion of the image of the surgical instrument.
  • 4. The method according to claim 3, wherein the surgical instrument is an instrument grasping tissue, and the graphical marking represents tissue information about tissue grasped by the instrument.
  • 5. The method according to claim 4, wherein the graphical marking is displayed on the image of the jaws of the instrument.
  • 6. The method according to claim 1, wherein the tissue information is tissue type information, tissue pathology information, tissue metrics information, information concerning an interaction between tissue and the instrument, or information indicating the presence or absence of any such information.
  • 7. The method according to claim 2, wherein the graphical marking indicates whether tissue information is available for tissue in contact with a portion of the surgical instrument.
  • 8. The method according to claim 2, wherein the graphical marking represents tissue type, tissue pathology, or tissue metrics for tissue in contact with a portion of the surgical instrument.
  • 9. A method for augmenting an image displaying to a user during surgery, comprising: (a) during a surgical procedure, capturing an image of an anatomical working site using a scope device and displaying the image on a display visible to a surgeon;(b) during step (a), obtaining instrument information about a surgical instrument within the working site; and(c) displaying indicia depicting the instrument information with the image on the display.
  • 10. The method according to claim 9, wherein step (b) includes using computer vision to detect edges of the surgical instrument, and step (c) displays indicia highlighting portions of the instrument on the display.
  • 11. A method for communicating information to a user operating a surgical instrument during surgery, comprising: (a) moving a surgical instrument within an anatomical working site;(b) during step (a), obtaining tissue information about tissue within the working site;(c) communicating to the user when the instrument is within a defined distance from a tissue of interest within the working site.
  • 12. The method of claim 11, wherein the communicating step includes emitting an auditory signal, and altering the frequency of the signal as the surgical instrument approaches and/or moves away from the tissue of interest.
  • 13. The method of claim 11, wherein moving the surgical instrument includes moving a user input device to cause movement of the surgical instrument, and wherein the communicating step includes generating tactile feedback at a handle of the user input device.
  • 14. A method for communicating information to a user operating a surgical instrument during surgery: (a) causing movement of a surgical instrument within an anatomical working site by moving a handle of a user input device to direct movement of the surgical instrument;(b) during step (a), obtaining tissue information about tissue within the working site;(c) communicating information to the user about tissue in the working site by generating tactile feedback at a handle of the user input device.
  • 15. The method of claim 14, wherein the tactile feedback represents contact or proximity between the surgical instrument and tissue having tissue characteristics of interest.
  • 16. The method of claim 14, wherein the tissue characteristics are tissue type, pathology, density, vascularization, absence/presence/rate of blood flow, oxygen levels, presence of hemoglobin, inflammation, ischemia, thickness or other metrics.
  • 17. The method of claim 6, wherein the interaction is tissue contact or tissue force.
  • 18. The method of claim 17, wherein the indicia marks areas of the surgical instrument in contact with tissue.
  • 19. The method of claim 18, wherein the indicia represents a degree of force applied to tissue by the surgical instrument.
Parent Case Info

This application claims the benefit of U.S. Provisional Application No. 62/470,120, filed Mar. 10, 2017.

Provisional Applications (1)
Number Date Country
62420120 Nov 2016 US