The present disclosure generally relates to a surgical visualization system and, more particularly, to devices and methods utilizing detectable radiation in surgery.
Typically, during endoscopic surgery, a surgical field is cluttered with different anatomical structures and surgical implements as well as fluids that can obscure a surgeon's view of relevant anatomical structures and surgical implements. It is often difficult to see the position of surgical implements relative to different anatomical structures and to properly position surgical instruments in the surgical field. The disclosure provide for various systems and methods to improve the visualization of surgical implemented in surgery settings.
In various implementations, the disclosure provides for surgical implements that comprise a fluorescent agent. The fluorescent agents may be incorporated in surgical tools or implements to assist in distinguishing the implements, or portions of the implements, from their surroundings in a surgical field. In general, the fluorescent agents may be excited in response to receiving an excitation emission of radiation over a range of excitation wavelengths. In response to the excitation emission, the fluorescent agent emits a fluorescent emission of radiation in a known wavelength band that is detectable in image data captured by the surgical camera. In response to the detection of the fluorescent emission, the camera may respond in a number of ways to improve the visualization, detection, and/or identification of the surgical implement associated with the fluorescent agent. In some cases, the excitation emission and/or the fluorescent emission may correspond to wavelengths of light capable of penetrating biological tissue. In such cases, the fluorescent emission may be detected by the camera system to identify a position or presence of the surgical implement through the biological tissue. Once identified, a display controller of the camera system may overlay or provide a visual indication of the position of the fluorescent portion of the surgical implement in the image data for improved visualization during surgery. These and other features are described in the following detailed description.
These and other features, objects and advantages of the present disclosure will become apparent upon reading the following description thereof together with reference to the accompanying drawings.
The features, aspects and advantages of the present disclosure will become better understood with regard to the following description, appended claims and accompanying figures wherein:
In the following description of the preferred implementations, reference is made to the accompanying drawings, which show specific implementations that may be practiced. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts. It is to be understood that other implementations may be utilized and structural and functional changes may be made without departing from the scope of this disclosure.
Referring to
As previously discussed, the camera system 10 may be configured to capture image data associated with the visible light emission 38 as well as the fluorescent emission 32. Once captured, the system 10 may enhance the image data representing the visible light with one or more overlays or graphics to generate enhanced image data that emphasizes and/or identifies portions of a field of view 44 corresponding to the surgical implement 14. In order to provide the enhanced image data, a camera controller 46 may be configured to selectively control each of the first and second light sources 36, 40 as well as process image data received from a first image sensor 42a and a second image sensor 42b. In a standard operating mode, the camera controller 46 may activate the visible light emission 38 output from the second light source 40 to illuminate the surgical site 26 in wavelengths of light in a visible range (e.g., 400 nm-650 nm). Reflections from the visible light emission 38 may be captured by the second image sensor 42b, which may correspond to a visible light image sensor. Such operation may provide for illumination of the surgical site 26 in visible wavelengths of light, such that the camera controller 46 can output image data demonstrating visible characteristics of the surgical site 26 to the display controller 20. An example of the surgical implement 14 demonstrated illuminated by the visible light emission 38 and captured by the second image sensor 42b is shown in
In order to generate the enhanced image data, the camera controller 46 may activate the first light source 36 to output the excitation emission 34. In response to the excitation emission 34, the fluorescent agent of the fluorescent portion 22 may become excited and output the fluorescent emission 32. Concurrent with the activation of the first light source 36, the camera controller 46 may also activate the second light source 40 to illuminate the surgical site 26 in the visible light emission 38. As a result, the fluorescent emission 32 and the visible light emission 38 may be captured within the field of view 44 of each of the image sensors 42. While the second image sensor 42b may be configured to capture the reflected visible light emission 38, the first image sensor 42a may correspond to a near-infrared image sensor configured to capture wavelengths of light in a near-infrared range (e.g., 650 nm-900 nm). As shown, each of the image sensors 42 may comprise one or more light filters, exemplified as a first light filter 52a and a second light filter 52b. In operation, the light filters 52a, 52b may filter the combined wavelengths of the fluorescent emission 32 and the visible light emission 38 in the field of view 44 to improve the fidelity of the detection of the corresponding wavelengths detected by each of the image sensors 42a, 42b. In this way, the camera controller 46 may process image data recorded by each of the image sensors 42a, 42b to detect and discriminate between the fluorescent emission 32 and the visible light emission 38 in the field of view 44 representative of the surgical site 26.
Though generally described as light filters 52, the first filter 52a and the second filter 52b may correspond to one or more high pass, low pass, and/or bandpass filters configured to transmit light over a range associated with a corresponding detection range of the image sensors 42a, 42b. For example, the first light filter 52a may correspond to a bandpass filter configured to pass a range of near-infrared wavelengths from approximately 800 nm to 850 nm. In this configuration, the first light filter 52a may be selected to have a center frequency of approximately 825 nm, which may effectively pass wavelengths of light associated with the fluorescent emission 32 to the first image sensor 42a. In such cases, the fluorescent emission 32 may correspond to an emission from a fluorescent agent in the form of an indocyanine green (ICG) dye. Accordingly, the fluorescent emission 32 output from the fluorescent portion 22 may pass through the first light filter 52a within the bandpass range, such that the associated light from the fluorescent emission 32 is captured and identified by the camera controller 46. Similarly, the visible light emission 38 and the corresponding light reflected from the surgical site 26 may pass through a second light filter 52b, which may be configured to pass wavelengths of light in a visible range (e.g., 400 nm-650 nm). In this way, the camera system 10 may actively detect the fluorescent emission 32 and generate overlays, graphics, or other visual enhancements to augment the image data illuminated by the visible light emission 38 in the field of view 44.
In addition to the first and second light filters 52a, 52b, the camera system 10 may further comprise additional filters, which may include one or more dichroic filters or mirrors configured to separate the fluorescent emission 32 from the visible light emission 38. Such filters, generally referred to as light filters 52, may be incorporated in an endoscope or camera 60, which may comprise the image sensors 42, light sources 36,40, and camera controller 46, as well as the light filters 52 in a unified package. For example, the camera 60 may comprise each of the light sources 36, 40, image sensors 42, filters 52, and the camera controller 46 in a compact endoscope similar to that discussed later in reference to
With the image data associated with the visible light emission 38 detected independently of the fluorescent emission 32, the camera system 10 may provide for the enhancement of the fluorescent portions 22 in the image data. In this way, one or more colors, patterns, or other visual enhancements or overlays 62 may be superimposed or overlaid on the image data to generate enhanced image data for presentation on the display device 24. As shown in
Referring generally to
In various examples, the fluorescent agent may be coated or used as an integral portion (e.g., embedded in a material or structure) of a surgical implement 14. In some cases, the fluorescent agent may be incorporated in the fluorescent portion 22 of the surgical implement 14 during manufacture. For example, a plastic surgical implement may have a fluorescent dye mixed into the plastic during manufacture. Additionally, light blocking packaging may be used to protect the fluorescent dye from light until the surgical implement 14 is ready for use. The surgical implement 14, such as, for example and without limitation, a sponge, a suture, a pin, a screw, a plate, a surgical tool, or an implant may be painted with a fluorescent material. As used herein, the term “surgical tool”, may comprise, without limitation, a biter, grasper, retriever, pick, punch, hook, probe, elevator, retractor or scissors. The surgical implement 14 may have a fluorescent agent coated on a portion to indicate a location, position, depth, orientation, or other characteristic of the surgical implement. Accordingly, the fluorescent portion 22 of the surgical implement 14 may be readily identified or detected in the enhanced image data provided by the camera system 10.
As discussed later in specific reference to
In various implementations, the fluorescent agent in the surgical implement 14 may be excited using a light source that emits excitation light in the excitation wavelength range of the particular fluorescent agent. For example, when ICG is used as the fluorescent agent, ICG fluorescence may be excited using light in a wavelength range of from about 600 nm to about 900 nm and in some cases around 780 nm. In such cases, the light source 36 may be a light emitting diode or a laser emitting diode with a center frequency within or centrally within the excitation range of the ICG. The image sensors 42 may be, for example, a complementary metal-oxide-semiconductor (CMOS) sensor or a charge-coupled device (CCD) sensor. The camera 60 may also include optics, such as, for example, lenses, filters, mirrors, and prisms to direct and independently detect the wavelengths of light associated with the visible light source 40 and the fluorescent emission 32.
In some implementations, the camera 60 is implemented as an endoscopic camera, which may include the image sensors 42, light sources 36, 40, as well as the light filters 52. Accordingly, the camera 60 may include both the first light source 36 as an excitation light source for exciting the fluorescent agent and the second light source 40 in the form of a white light source for illuminating the surgical site 26 in the visible range of wavelengths. The camera 60 may further include a corresponding image sensor 42a or detector for detecting the fluorescent emission 32 and an image sensor 42b or detector for detecting and recording image data in the visible light range. In some cases, the camera 60 may have additional light sources for exciting multiple fluorescent agents or for detecting other non-visible attributes of a surgical field. An example of a camera system usable to detect fluorescent agents in surgical implements is the Arthrex Synergy IDTM camera system which has a camera head and a camera control unit. The Arthrex Synergy IDTM camera system has a light source for exciting fluorescence from ICG and is capable of detecting visible and near infra-red (NIR) light such as light emitted by ICG.
Referring again to
Additionally,
Referring now to
Though discussed in reference to the excitation emission 34 being output from the dedicated lighting device 80, either or both of the light sources 36, 40 may be implemented in the dedicated lighting device 80 to output light in various ranges of wavelengths. In some implementations, the lighting device 80 or the camera 60 may be configured to emit a beam of light with a diameter small enough for targeting items in the surgical field for further action by a surgeon. In an implementation, the beam diameter may be less than about 5 mm. In some cases, the beam diameter may be less than about 2 mm or than about 1 mm. In general, the lighting device 80 or camera 60 may be configured to emit a beam of light of sufficient brightness and density to be detected within a surgical field. For example, in some cases, high sensitivity sensors 42 have been measured to detect light at intensities of 10 nW/cm2 or less (e.g., a high sensitivity CMOS sensor). The light sources 36, 40 may be positioned proximal to a distal end of the light emitting device 80 or camera 60. Additionally, the light source 36, 40 may be positioned away from the distal end and light emitting device 80 or camera 60 from the light source communicated to the distal end such as by, for example, fiber optics. The light emitted by the light emitting device 80 and/or camera 60 may have a variable shape that may be adjusted, such as by using optics to allow a user to better illuminate a desired target.
In some implementations, one or both of the light sources 36, 40 may be incorporated into a surgical instrument 14 other than the endoscopic camera system 10, for example, in a probe, a shaver 14a, an ablation device, or other instrument. In some examples, an LED may be located at a distal end of the device or instrument. In some example, a probe or other device may be formed at least partially of a light pipe that may receive light from an LED, laser, or other light source external to the body and transmit the radiation to the distal end of the instrument. The light emitting device 80 may be powered by an isolated power source coupled to the light emitting device. Additionally, the light emitting device 80 may be battery powered. The battery powered light emitting device may be configured for a single use or may be configured with a rechargeable battery for multiple uses. The light emitting device 80 may be packaged in a sterile container for a single use. Additionally, the light emitting device 80 may be configured for sterilization and repeated use. The light emitting device 80 may be a rigid device or a flexible device. The light emitting device may be an articulatable device.
Additionally, the light emitting device 80 or light sources 36, 40 may be placed outside of a surgical field or site 26 and light directed through biological tissue for detection by the camera 60 positioned in the surgical field. Additionally, the light emitting device may direct light from a surgical field through tissue for detection by a device positioned outside of a surgical field. In some cases, the light emitting device 80 may be placed outside of a body and direct light through tissue for detection by the camera 60 positioned inside the body. Additionally, the light emitting device 80 may be placed inside of a body and direct light through tissue for detection by a camera (e.g., the camera 60) positioned outside of the body. Additionally, the light emitting device 80 may be placed in a first portion of a surgical site 26 and direct light through tissue for detection in a second portion of the surgical site 26.
As demonstrated in
Referring now to
In addition to the obstructions in the field of view 44 within the shoulder cavity 44, in some cases, an anchor (represented in
In some cases, the fluorescent emissions 32 output from the fluorescent portions 100a, 100b, 100c may vary in wavelengths due to different compositions or combinations of fluorescent agents incorporated therein. In other cases, a concentration of a common fluorescent agent (e.g., ICG dye) may be incorporated at different levels in each of the fluorescent portions 100a, 100b, 100c. Accordingly, in response to receiving the excitation emission 34, each of the fluorescent emissions 32 output from the fluorescent portions 100a, 100b, 100c may vary in wavelength or intensity based on the composition of fluorescent agents or concentration of fluorescent agents incorporated therein. Based on the variations in the intensity or wavelengths associated with the fluorescent emissions 32, the display controller 20 may be operable to distinguish among the different fluorescent portions 100a, 100b, 100c and overlay each of the fluorescent portions 100a, 100b, 100c with different characteristic colors 102. Accordingly, the camera system 10 may be configured to distinguish among a plurality of fluorescent portions 100a, 100b, 100c and assign different respective characteristic colors 102 or patterns, such that the enhanced image data demonstrated on the display device 24 clearly distinguishes the locations of each of the surgical implements 14 (e.g., 92a, 92b, and 94b).
To be clear, the sutures 92a, 92b and second anchor 94b demonstrated in
In some cases, the characteristic colors 102 or patterns associated with the enhanced image data may be customized or modified to suit the preferences of a specific user. For example, some users may prefer a wide range of colors to assist in distinguishing among the various surgical implements 14, while others may prefer subtle color differences that may not distract their view from other aspects within the surgical site 26. In some cases, the display controller 20 may adjust a color template or color configuration of the characteristic colors 102 or patterns based on the colors of the local environment demonstrated in the image data captured by the second image sensor 42 associated with the visible light emission 38. For example, if the image data illuminated by the visible light emission 38 is displayed primarily in warm hues (e.g., red, yellow, orange), the display controller 20 may assign a cool color template (e.g., blue, purple, green) to distinguish the fluorescent portions 100a, 100b, 100c from the remainder of the image data in the field of view 44. Similarly, if the image data is dark, light or contrasting hues or patterns may be automatically applied to contrast the image data. Accordingly, the camera system 10 may provide for a variety of formats and color templates associated with the enhanced image data to assist in the visualization of the surgical site 26.
Referring to
Referring to
In some cases, the longitudinal markings 110, lateral markings 120, or various additional fluorescent portions 22 incorporated on the surgical implements 14 may be disposed within a groove 124 or indentation formed in an exterior surface of the surgical implement 14. By including the fluorescent portions 22 in the grooves or indentations associated with the orientation or positional markings 110, 120; the resulting fluorescent emissions 32 output from the grooves 124 or indentations may be captured in the field of view 44 of the camera system 10 through an orientation aperture associated with an interior surface of each of the grooves 124 directed to or facing the corresponding image sensors 42a, 42b of the camera 60. In this configuration, the dimensional or orientational markings 110, 120 incorporated on the surgical implement 14 may be hidden from the field of view 44 of the camera 60 until a portion of the fluorescent emission 32 is output from the corresponding fluorescent portions 22 disposed in the grooves 124. The result of the fluorescent portions 22 disposed in the grooves 124 may be an improved accuracy achieved similar to a sight that only exposes the fluorescent emission 32 when an interior surface of each of the grooves 124 is visible through the corresponding orientation aperture. In this way, the dimensional and orientational features (e.g., 110, 120) of the surgical implements 14 may provide for improved accuracy in determining the relative positioning or orientation of the surgical implement 14.
Referring now to
Referring now to
In response to detecting the fluorescent emission 32 in the image data captured by the first image sensor 42a, the display controller 20 of the camera system 10 may overlay the pixels in the image data associated with the fluorescent emission 32 with the overlay 62 (e.g., characteristic colors 102 or patterns) to generate the enhanced image data. Accordingly, the camera system 10 may provide for the detection and tracking of the position of one or more surgical implements 14 through biological tissue 132 by detecting the fluorescent emission 32. Once detected, the display controller 20 may further overlay, mark, or enhance corresponding portions of the image data to demonstrate the surgical implements 14 that would otherwise be completely hidden from a conventional camera system.
Referring now to
In some examples, the excitation light source or first light source 36 may output the excitation emission 34 at an intensity sufficient to penetrate biological tissue as discussed herein. For example, the first light source 36 may output the excitation emission 34 at an intensity ranging from approximately 1 mW/cm2 to 1 W/cm2. In some cases, the light intensity may be higher or lower depending on the specific light emitter technology implemented and the application. Depending on the application and the duration over which the excitation emission 34 is to be activated, the intensity of the excitation emission 34 may be limited or pulsed to control excess heat generation and limit damage to the biological tissue. As previously discussed, the excitation emission 34 may comprise wavelengths of radiation ranging from approximately 650 nm to 900 nm in the near-infrared range. For reference, the visible light emission 38 associated with the second light source 40 may be output in wavelengths corresponding to visible colors of light associated with the acuity of a human eye ranging from 400 nm to approximately 650 nm. The penetration of the excitation emission 34 and/or the fluorescent emission 32 through biological tissue may extend approximately from a depth of 1 mm to depths or thicknesses of biological tissue exceeding 10 mm. Experimental results have demonstrated a loss of intensity of emissions similar to the excitation emission 34 and the fluorescent emission 32 in the near-infrared range at a rate of approximately 3%-10%/mm of biological tissue penetrated. Accordingly, the first image sensor 42a may detect the fluorescent emission 32 or the excitation emission 34 after the corresponding light energy has penetrated multiple millimeters of biological tissue. Therefore, the camera system 10 may identify the relative location or orientation of the various surgical implements 14 and demonstrate the locations in the enhanced image data in a variety of cases where the surgical implements 14 may be hidden behind layers of biological tissue having various thicknesses.
Referring now to
As previously discussed in reference to
Referring now to
In step 178, if an object associated with the fluorescent emission 32 is detected in the image data, the method 170 may continue to mark, overlay, or annotate the image data to emphasize the regions in the field of view 44 where the fluorescent emission 32 is detected (180). The marked or annotated image data generated in step 180 may correspond to the enhanced image data comprising one or more overlays 62 in the form of characteristic colors, patterns, or other indicating features that may assist a viewer in recognizing a location, orientation, dimensions, proportions, or other information related to the surgical implement 14 from which the fluorescent emission 32 was emitted and detected by the camera system 10. Examples of surgical implements may include a biter, grasper, retriever, pick, punch, hook, probe, elevator, retractor or scissors. In some cases, the surgical implements 14 may correspond to items configured to trigger an alert or notification of the camera system 10 to indicate the detection of their presence. For example, partial components of tools, implants, sponges, or other various surgical implements within the surgical site 26 may be detected by camera system 10 in response to the presence of the fluorescent emission 32. In response to such a detection, the method 170 may output an indication (e.g., an alert, instruction, notification, etc.) indicating the presence of a fluorescent portion 22 and alerting a surgeon or medical professional of the presence of the corresponding surgical implement 14 (182). In some cases, the programming of the camera system 10 may define specific surgical implements 14 that may be associated with the fluorescent emission 32. In such cases, the notification output in step 182 may indicate the specific type or category of the surgical implement 14 identified in the image data by the camera system 10. Following step 182, the detection routine may continue until it is deactivated by an operator, as demonstrated in step 184.
Referring now to
In some cases, the display controller 20 may identify different intensities of the fluorescent emission 32 over time, such that the characteristic colors or patterns associated with the overlay 62 of the enhanced image data may be maintained even in cases where the corresponding surgical implements 14 are not simultaneously presented in the image data. For example, the display controller 20 may be preconfigured to associate a lower intensity fluorescent emission 32 with a first color, a medium intensity fluorescent emission 32 with a second color, and a third intensity fluorescent emission 32 with a third color. The relative intensities may correspond to percentages or relative levels of luminance associated with each of the fluorescent emissions 32. For example, if three levels of luminance are detected, a maximum intensity may be associated with the third color. An intermediate intensity may be associated with the second color, and a minimum or lowest intensity may be associated with the first color. Once the enhanced image data is generated, it may further be selectively displayed on the display device 24 by controlling an interface of the display controller (202). Following step 202, the display routine may continue until deactivated (204).
Referring now to
The light sources 36, 40 may correspond various light emitters configured to generate light in the visible range and/or the near infrared range. In various implementations, the light sources 36, 40 may include light emitting diodes (LEDs), laser diodes, or other lighting technologies. As previously discussed, the first light source 36 may generally correspond to a laser emitter configured to output emissions in the near infrared range including wavelengths from approximately 650 nm to 900 nm. In some instances, the first light source 36 may output the excitation emission 34 ranging from 650 nm to 680 nm with a center frequency of approximately 670 nm. In some cases, the first light source 36 may output the excitation emission 34 in a range of wavelengths from approximately 740 nm to 780 nm. More generally, the wavelengths associated with the first light source 36 and the excitation emission 34 may be selected to effectively energize the fluorescent agent of the fluorescent portion 22. The second light source 40 may correspond to a white light source in the visible spectrum including wavelengths ranging from approximately 380 nm to 700 nm or from approximately 400 nm to 650 nm.
The image sensors 42a, 42b may correspond to various sensors and configurations comprising, for example, charge-coupled devices (CCD) sensors, complementary metal-oxide semiconductor (CMOS) sensors, or similar sensor technologies. As previously discussed, the system 10, particularly the display controller 20 may process or compare the image data captured by each of the image sensors 42 to identify the fluorescent emission 32 and apply the overlay 62 in the form of one or more colors (e.g., the characteristic colors 102), patterns, markers, graphics, messages, and/or annotations indicating the presence and/or location of the fluorescent emission 32 in the image data. In operation, the light filters 52a, 52b (e.g. bandpass filters) may filter and effectively separate the combined wavelengths of the fluorescent emission 32 and the visible light emission 38 in the field of view 44. Accordingly, the filtered light received by the first image sensor 42a may provide a map identifying locations of the fluorescent emission 32 and the corresponding locations of the fluorescent portions 22 of the surgical implements 14 in the image data.
The camera controller 46 may correspond to a control circuit configured to control the operation of image sensors 42a, 42b and the light sources 36, 40 to provide for the concurrent or simultaneous capture of the image data in the visible light spectrum as well as the near infrared spectrum or wavelength associated with the fluorescent emission 32. Additionally, the camera controller 46 may be in communication with a user interface 210, which may include one or more input devices, indicators, displays, etc. The user interface may provide for the control of the camera 60 including the activation of one or more routines as discussed herein. The camera controller 46 may be implemented by various forms of controller, microcontrollers, application-specific integrated controllers (ASICs), and/or various control circuits or combinations.
The display controller 20 may comprise a processor 212 and a memory 214. The processor 212 may include one or more digital processing devices including, for example, a central processing unit (CPU) with one or more processing cores, a graphics processing unit (GPU), digital signal processors (DSPs), field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs) and the like. In some configurations multiple processing devices are combined into a System on a Chip (SoC) configuration while in other configurations the processing devices may correspond to discrete components. In operation, the processor 212 executes program instructions stored in the memory 214 to perform the operations described herein.
The memory 214 may comprise one or more data storage devices including, for example, magnetic or solid state drives and random access memory (RAM) devices that store digital data. The memory 214 may include one or more stored program instructions, object detection templates, image processing algorithms, etc. As shown, the memory 214 may comprise a detection module 216 and an annotation module 218. The detection module 216 include instructions to process the image data identifying the fluorescent emission 32 from the first image sensor 42a and detect the locations in the field of view 44 from which the fluorescent portion 22 of the surgical implement 14 emitted the fluorescent emission 32. In some cases, the detection module 216 may include instructions to detect or identify a type or classification associated with the surgical implement 14 in the image data captured by the camera 60. For example, the processor 212 may access instructions in the detection module 216 to perform various processing tasks on the image data including preprocessing, filtering, masking, cropping and various enhancement techniques to improve detection capability and efficiency. Additionally, the detection module 216 may provide instructions to process various feature detection tasks including template matching, character recognition, feature identification or matching, etc. In some examples, the detection module 216 may also include various trained models for object detection and/or labeling surgical implements 14 or related objects. In some implementations, the detection of a surgical implement, either by identity, presence, or classification, may initiate an instruction to output an alert or notification on the display device 24, the control console 16, an external device or server 220, or various connected devices associated with the surgical camera system 10.
The annotation module 218 may comprise instructions indicating various marking or overlay options to generate the enhanced image data as well as corresponding display filters to superimpose or apply the overlays 62 to the image data. As previously discussed, the enhanced image data may also include one or more graphics, annotations, labels, markers, and/or identifiers that indicate the location, presence, identity, or other information related to a classification or identification of the surgical implement 14. The annotation module 218 may further provide instructions to generate, graphics, labels, overlays or other associated graphical information that may be applied to the image data captured by the second image sensor 42b (e.g., the visible light sensor) to generate the enhanced image data for display on the display device 24.
The display controller 20 may further comprise one of more formatting circuits 222, which may process the image data received from the camera 60, communicate with the processor 212, and output the enhanced image data to the display device 24. The formatting circuits 222 may include one or more a signal processing circuit, analog to digital converters, digital to analog converters, etc. The display controller may comprise a user interface 224, which may be in the form of an integrated interface (e.g., a touchscreen, input buttons, an electronic display, etc.) or may be implemented by one or more connected input devices (e.g., a tablet) or peripheral devices (e.g., keyboard, mouse, etc.). As shown, the controller 20 is also in communication with an external device or server 220, which may correspond to a network, local or cloud-based server, device hub, central controller, or various devices that may be in communication with the display controller 20 and more generally the camera system 10 via one or more wired (e.g., Ethernet) or wireless communication (e.g., WiFi, 802.11 b/g/n, etc.) protocols. For example, the display controller 20 may receive updates to the various modules and routines as well as communicate sample image data from the camera 60 to a remote server for improved operation, diagnostics, and updates to the system 10. The user interface 224, the external server 220, and/or the surgical control console 16 may be in communication with the controller 20 via one or more I/O circuits 226. The I/O circuits may support various communication protocols including but not limited to Ethernet/IP, TCP/IP, Universal Serial Bus, Profibus, Profinet, Modbus, serial communications, etc.
In various implementations, the disclosure provides for a surgical camera system configured to capture image data indicative of a surgical implement comprising a fluorescent agent. The surgical camera system comprises a camera comprising at least one sensor configured to capture image data comprising a first range of wavelengths and a second range of wavelengths. An excitation light source emits an excitation emission at an excitation wavelength. A controller is in communication with the at least one sensor of the camera. The controller is configured process the image data from the at least one sensor and detect at least one fluorescent portion of the image data in response to a fluorescent emission generated by the fluorescent agent in the second range of wavelengths. The controller is further configured to generate enhanced image data demonstrating the at least one fluorescent portion of the surgical implement in the image data.
In various implementations, the systems and methods described in the application may comprise one or more of the following features or steps alone or in combination:
In various implementations, the disclosure provides for method for displaying a surgical implement may comprise illuminating a fluorescent portion of the surgical implement in light comprising a first range of wavelengths corresponding to visible light and a second range of wavelengths comprising an excitation emission. The method may further include capturing first image data comprising the first range of wavelengths and capturing second image data comprising the second range of wavelengths demonstrating a fluorescent emission output from the fluorescent portion in response to the excitation emission. The method further includes generating enhanced image data demonstrating the first image data with at least one overlay or graphic demonstrating the fluorescent portion defined by the second image data overlaid on the first image data and communicating the enhanced image data for display on a display device.
In various implementations, the systems and methods described in the application may comprise one or more of the following features or steps alone or in combination:
In some implementations, the disclosure provides for a surgical camera system configured to capture image data indicative of a surgical implement comprising a fluorescent agent. The surgical camera system comprises camera comprising at least one sensor configured to capture image data comprising a first range of wavelengths and a second range of wavelengths. An excitation light source emits an excitation emission at an excitation wavelength. A controller is in communication with the sensor of the camera. The controller is configured to process image data from the at least one image sensor comprising the first range of wavelengths and the second range of wavelengths and identify a plurality of intensity levels of at least one fluorescent emission output from the at least one fluorescent portion generated by the fluorescent agent in the second range of wavelengths. The controller is further configured to assign a distinctive color or pattern to each of the plurality of intensity levels and generate enhanced image data demonstrating the plurality of intensity levels of the fluorescent emission with the a distinctive colors or patterns. In some implementations, the enhancement of the image data comprises overlaying the distinctive color or pattern over the fluorescent portion demonstrating each of the plurality of intensity levels in the enhanced image data.
In some implementations, a surgical implement may comprise a body forming an exterior surface comprising a proximal end portion and a distal end portion. A fluorescent portion may comprise a fluorescent agent disposed on the exterior surface. The fluorescent portion may comprises at least one marking extending over the exterior surface and the fluorescent portion is configured to emit a fluorescent emission in a near-infrared range in response to an excitation emission.
In various implementations, the systems and methods described in the application may comprise one or more of the following features or steps alone or in combination:
In some implementations, the surgical detection system may be configured to identify at least one surgical implement in an operating region. The system may comprise
a camera comprising at least one sensor configured to capture image data comprising a first range of wavelengths and a second range of wavelengths. An excitation light source emits an excitation emission at an excitation wavelength. A controller is in communication with the at least one sensor of the camera, the controller configured to process image data from the at least one sensor and identify the fluorescent emission in the image data output from at least one fluorescent portion of a surgical implement. The controller is further configured to detect a presence of the surgical implement in response to the presence of the fluorescent emission.
In various implementations, the systems and methods described in the application may comprise one or more of the following features or steps alone or in combination:
In some implementations, the surgical camera system may be configured to capture image data indicative of a surgical implement comprising a fluorescent agent. The surgical camera system may comprise an endoscopic camera comprising at least one sensor configured to capture image data in a field of view comprising a first range of wavelengths and a second range of wavelengths. An excitation light source emits an excitation emission at an excitation wavelength. A controller is in communication with the sensor of the camera. The controller is configured to process the image data from the at least one sensor in the field of view depicting a cavity and detect a fluorescent emission output from at least one fluorescent portion of a surgical implement in the image data. The fluorescent emission is transmitted through a biological tissue forming at least a portion of the cavity. In response to a fluorescent emission, the controller generates enhanced image data demonstrating the at least one fluorescent portion of the surgical implement overlaid on the biological tissue depicted in the image data.
In various implementations, the systems and methods described in the application may comprise one or more of the following features or steps alone or in combination:
There is disclosed in the above description and the drawings, a surgical camera system and method that fully and effectively overcomes the disadvantages associated with the prior art. However, it will be apparent that variations and modifications of the disclosed implementations may be made without departing from the principles described herein. The presentation of the implementations herein is offered by way of example only and not limitation, with a true scope and spirit being indicated by the following claims.
As used herein, words of approximation such as, without limitation, “approximately, “substantially,” or “about” refer to a condition that when so modified is understood to not necessarily be absolute or perfect but would be considered close enough to those of ordinary skill in the art to warrant designating the condition as being present. The extent to which the description may vary will depend on how great a change can be instituted and still have one of ordinary skill in the art recognize the modified feature as having the required characteristics or capabilities of the unmodified feature. In general, but subject to the preceding discussion, a numerical value herein that is modified by a word of approximation such as “approximately” may vary from the stated value by ±0.5%, ±1%, ±2%, ±3%, ±4%, ±5%, ±10%, ±12%, or ±15%.
Any element in a claim that does not explicitly state “means” for performing a specified function or “step” for performing a specified function, should not be interpreted as a “means” or “step” clause as specified in 35 U.S.C. § 112.
It will be understood that any described processes or steps within described processes may be combined with other disclosed processes or steps to form structures within the scope of the present device. The exemplary structures and processes disclosed herein are for illustrative purposes and are not to be construed as limiting.
It is also to be understood that variations and modifications can be made on the aforementioned structures and methods without departing from the concepts of the present device, and further it is to be understood that such concepts are intended to be covered by the following claims unless these claims by their language expressly state otherwise.
The above description is considered that of the illustrated embodiments only. Modifications of the device will occur to those skilled in the art and to those who make or use the device. Therefore, it is understood that the embodiments shown in the drawings and described above are merely for illustrative purposes and not intended to limit the scope of the device, which is defined by the following claims as interpreted according to the principles of patent law, including the Doctrine of Equivalents.
This application claims priority under 35 U.S.C. § 119(e) and the benefit of U.S. Provisional Application No. 63/174,966 entitled SYSTEM AND METHOD FOR USING DETECTABLE RADIATION IN SURGERY, filed on Apr. 14, 2021, by Bruce Laurence Kennedy et al., the entire disclosure of which is incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
63174966 | Apr 2021 | US |