This application relates to the use of computer vision to recognize events or procedural steps within a surgical site. Surgical systems providing this type of feature can reduce the surgeon's cognitive load by recognizing and alerting the surgeon to the detected events within the surgical site, by performing tasks the surgeon would otherwise need to perform (such as creating documentation recording the detected events, or adjusting system parameters), by adjusting aspects of the visual augmentation of the surgical site display, and/or by providing contextually-aware information to the surgeon when appropriate, and (in the case of on-screen display) causing that information to disappear when it is no longer needed. Reducing the surgeon's cognitive load can shorten the surgical procedure, thereby reducing risk of complications, lowering procedure costs, and increasing surgeon productivity.
Co-pending and commonly owned U.S. Ser. No. 17/368,753, filed Jul. 6, 2021, and entitled “Providing Surgical Assistance via Automatic Tracking and Visual Feedback During Suture” describes examples of ways in which computer vision is used to give feedback and assistance to a user performing suturing within the field of vision of a camera capturing images of a surgical site.
This application describes a system that creates surgical records in response to computer recognition of steps or events of a surgical procedure based on computer vision analysis or kinematic data.
System
A system useful for performing the disclosed methods, as depicted in
The camera 10 is one suitable for capturing images of the surgical site within a body cavity. It may be a 3D or 2D endoscopic or laparoscopic camera. Where it is desirable to use image data to detect movement or positioning of instruments, or tissue in three dimensions, configurations allowing 3D data to be captured or derived are used. For example, a stereo/3D camera might be used, or a 2D camera with software and/or hardware configured to permit depth information to be determined or derived using structured light techniques. The terms “camera,” “endoscopic camera,” and “laparoscopic camera” may be used interchangeably in this description without limiting the scope of the invention.
The computing unit 12 is configured to receive the images/video from the camera and input from the user input device(s). If the system is to be used in conjunction with a robot-assisted surgical system in which surgical instruments are maneuvered within the surgical space using one or more robotic components the system may optionally be configured so that the computing unit also receives kinematic information from such robotic components 18 for use in recognizing procedural steps or events as described in this application.
An algorithm stored in memory accessible by the computing unit is executable to, depending on the particular application, use the image data to perform one or more of the functions described with respect to the first and second embodiments.
The system may include one or more user input devices 16. When included, a variety of different types of user input devices may be used alone or in combination. Examples include, but are not limited to, eye tracking devices, head tracking devices, touch screen displays, mouse-type devices, voice input devices, foot pedals, or switches. Eye-tracking, head-tracking, mouse devices or other devices that cause movement of a cursor on a display may be used by a user to move a cursor around the display to identify regions of interest, to position the cursor on icons displayed and representing options the user can select. Various movements of an input handle used to direct movement of a robotic component of a surgical robotic system may be received as input (e.g., handle manipulation, joystick, finger wheel or knob, touch surface, button press). Another form of input may include manual or robotic manipulation of a surgical instrument having a tip or other part that is tracked using image processing methods when the system is in an input-delivering mode, so that it may function as a mouse, pointer and/or stylus when moved in the imaging field, etc. Input devices of the types listed are often used in combination with a second, confirmatory, form of input device allowing the user to enter or confirm (e.g., a switch, voice input device, foot pedal, button, icon to press on a touch screen, etc., as non-limiting examples).
Some hospitals or surgical centers have operating room dashboards or notification systems 20 that provide updates to hospital personnel or waiting room occupants as to the expected remaining duration of the procedure. The one or more computing units may further be configured to interact with such a system to keep the relevant parties apprised of the expected time remaining for the surgical procedure.
The initial step or set of steps relate to receiving data from which the computing unit can determine the occurrence of a surgical tasks, subtasks, or events. Images are captured from the camera 10 as shown. The system may optionally also receive kinematic data for any robotic components 18 being used to perform the surgery, such as data based on input generated by sensors of the robotic components.
The system executes algorithms that analyze the image data and other data to carry out any one or combination of the following:
The system then recognizes/detects whether tool movement patterns, sequences of movements, detected changes etc. indicate that a predetermined surgical task or subtask is being, or has been, performed. For example, recognizing the motion of a tool being used to wrap a suture around another tool allows the system to detect that the surgeon is in the process of tying a suture knot. Where a needle-holder is known or recognized as being controlled by a user's right hand, and a grasper is known or recognized as being controlled by the left hand but is then detected as being replaced with scissors, the system might detect that the surgeon has just completed the process of tying a knot. A database associated with the computing unit may store sequences of tasks and/or sub-tasks for a given procedure. Some may be steps of standard procedures with accepted best practices. Others may be tasks/sub-tasks of a custom procedure developed by a surgeon or team and input or saved to the system by a user. Such customized procedures might be patient-specific procedure steps developed based on the co-morbidities, prior surgeries, complications, etc. of the patient who is to be treated.
A database associated with the one or more computing units preferably includes a tracker that is updated as each task/sub-task is detected during a surgical procedure. If the system determines that the detected tool movement patterns, sequences of movements, detected changes etc. indicate that a predetermined surgical task or subtask is being, or has been, performed, the tracker is updated. Examples of tasks that might be tracked include access to or exposure of a treatment site, dissection, and suturing. More significant steps or procedures that might be part of, or steps marking the completion of, the procedural plan might include tissue or organ removal steps (i.e. “-ectomy” procedures), tissue cutting or formation of openings (“-otomy” procedures) etc.
The procedures stored in the database might have designated “major” steps or tasks, which for the first embodiment are defined as steps identified to the system as ones for which photo documentation is to be captured upon their completion. If the detected surgical task or sub-task is not a “major” task, the process repeats for the next task of the procedure or planned sequence of tasks. If it is a “major” task, a snapshot (e.g. a frame grab from the video image) is captured. The capture may be automatically performed by the system, or the user may take an action (optionally in response to a prompt to do so) that inputs instructions to capture it (e.g. by giving a voice command, touching a graphical user interface or, if the surgeon is at a robotic surgery console, by a button press). The snapshot is displayed on the screen as an overlay of the displayed endoscopic image and/or stored in the database for use in creating records of the surgery. By storing the snapshot, it is meant that a still image is saved independently of any video of the procedure that may be stored. As one example discussed in connection with
In addition to capturing the snapshot, the system may prompt the user to record a verbal annotation in real time. The verbal annotation is stored in a database and is associated with the snapshot, so the two may become part of a record of the procedure and/or subsequently used to train a machine learning algorithm.
The system may additionally or alternatively be configured to perform other functions in response to the scene detected. For example, where robotic components are used, it may be beneficial to set different force thresholds or scaling factors (between the user input and the instrument motion) depending on the steps being performed or the instruments being used. As one example, if the system detects that a 3 mm needle holder has been introduced, but it does not detect the presence of a suture needle in the scene captured in the images, the maximum force threshold for robotic control of that instrument may be set to a relatively lower limit. On the other hand, if the system detects that a suture needle is introduced into the scene, the maximum force threshold is raised to allow so that the needle holder will be able to hold the needle with ample holding force.
In many cases the above-listed steps may utilize a machine learning algorithm such as, for example, one utilizing neural networks that analyzes the images. Such algorithms can be used to recognize tools, and/or anatomic features, and/or recognize when movement patterns, sequences of movements, etc. are those of a predetermined surgical task or subtask. Certain surgical procedure steps require particular motion patterns or sequence of motions. The machine learning algorithm may thus be one that recognizes such patterns of motion and identifies them as being those for a particular surgical task. Movements taken into consideration may be movements of a tool alone or relative to a second tool, or movements of a tool relative to an identified feature of the anatomy. Others might include the combined movement patterns of, or interactions between, more than one tool. In some cases, multiple movements, or steps in a sequence of movements, are recognized as a predetermined surgical task or subtask if performed within a given length of time. Tool exchanges may also be recognized as being part of a recognized surgical task or subtask. As discussed above, detection of tool types that would be used to recognize tool exchanges may use computer vision or information from robotic components, as described above.
An example of use of the first embodiment will next be described with respect to
The system, using computer vision, recognizes the cystic duct, and displays an overlay over it, as shown in
As discussed in connection with
In a variation of the
In this embodiment, the procedures stored in the database have estimated completion times for certain steps (labeled “major steps” in
The concepts described in this application aid allow the surgeon's cognitive load to be reduced in in a variety of ways. These include providing auto-documentation capability, automatically adjusting system settings, and providing contextually-aware information (such as procedural task lists relevant to the procedure currently underway) to the surgeon.
All patents and applications described herein, including for purposes of priority, are incorporated by reference.
Number | Date | Country | |
---|---|---|---|
63088393 | Oct 2020 | US |