The invention relates to a laparoscopic tool system for minimally invasive surgery, the laparoscopic tool system comprises a laparoscopic tool adapted for use during a minimally invasive surgery procedure and/or a minimally invasive examination procedure.
Minimally invasive surgery (MIS) and in particular laparoscopy has been used increasingly in recent years due to the benefits compared to conventional open surgery as it reduces the trauma to the patient tissue, leaves smaller scars, minimizes post-surgical pain and enables faster recovery of the patient.
There are different kinds of MIS such as laparoscopy, endoscopy, arthroscopy and thoracoscopy. Whereas many of the MIS procedures are mainly for examination within natural openings of mammals, laparoscopy has in recent years developed to be a preferred method of performing both examination and surgical procedures.
In laparoscopic surgery the surgeon accesses a body cavity, such as the abdominal or pelvic cavity, through a series of small incisions. A laparoscope is inserted through an incision, and conventionally connected to a monitor, thereby enabling the surgeon to see the inside of the abdominal or pelvic cavity. In order to perform the surgical procedure, surgical instruments are inserted through other incisions. In addition, the body cavity (surgery cavity) around the surgical site is inflated with a fluid, preferably gas e.g. carbon dioxide in order to create an ‘air’ space within the cavity to make space for the surgeon to view the surgical site and move the laparoscopic instruments. Minimally invasive surgery is generally performed through openings in a patient's skin—often relatively small openings—and the surgical site is visualized for the surgeon by inserting a laparoscope which comprises illumination means and a camera into the body cavity and displaying the images on a screen.
In order to improve the vision for the surgeon, in particular to make it easier for the surgeon to determine the sizes of various organs, tissues, and other structures in a surgical site, several in-situ surgical metrology methods have been provided in the prior art. Different types of optical systems have been applied to provide an improved vision of the surgical site. Some prior art systems are focused on providing 3D images of the MIS cavity and other prior art systems are focused on monitoring the position of a laparoscopic tool.
US 2013/0296712 describes an apparatus for determining endoscopic dimensional measurements, including a light source for projecting light patterns on a surgical sight including shapes with actual dimensional measurements and fiducials, and means for analyzing the projecting light patterns on the surgical site by comparing the actual dimensional measurements of the projected light patterns to the surgical site.
WO 2013/163391 describes at system for generating an image, which the surgeon may use for measuring the size of or distance between structures in the surgical field by using an invisible light for marking a pattern to the surgical field. The system comprises a first camera; a second camera; a light source producing light at a frequency invisible to the human eye; a dispersion unit projecting a predetermined pattern of light from the invisible light source; an instrument projecting the predetermined pattern of invisible light onto a target area; a band pass filter directing visible light to the first camera and the predetermined pattern of invisible light to the second camera; wherein the second camera images the target area and the predetermined pattern of invisible light, and computes a three-dimensional image.
US2008071140 discloses an endoscopic surgical navigation system which comprises a tracking subsystem to capture data representing positions and orientations of a flexible endoscope during an endoscopic procedure, to allow co-registration of live endoscopic video with intra-operative and/or pre-operative scan images. Positions and orientations of the endoscope are detected using one or more sensors and/or other signal-producing elements disposed on the endoscope.
US2010268067 discloses methods, systems, devices, and computer-readable media for image guided surgery for allowing a physician to use multiple instruments for a surgery and simultaneously provide image-guidance data for those instruments.
US2011069159 discloses a system for orientation assistance and display of an instrument that is inserted or present in the natural or artificially produced hollow cavity (human, animal, object), and that is equipped with one or more sensor units. Multiple measurements of the 3D position of the instrument equipped with one or more sensor units are performed by positioning a measuring system, so that a precise orientation and positioning of the instrument in the body may be computed. The 3D position data are used to compute a virtual image of the instrument synchronously. The virtual images are then either projected directly in exact position onto the body surface of a person or combined in a body surface image (real video camera image of the patient) onto a monitor or superimposed (virtual or augmented reality).
It has also been suggested to generate augmented reality vision of surgery cavities for providing an improved view of internal structures of the organs of a patient to determine the minimal distance to a cavity surface or organ of a patient. Such systems are described in the articles “Augmented reality in laparoscopic surgical oncology” by Stéphane Nicolau et al. Surgical Oncology 20 (2011) 189-201 and “An effective visualization technique for depth perception in augmented reality-based surgical navigation” by Choi Hyunseok et al. The international journal of medical robotics and computer assisted surgery, 2015 May 5. doi: 10.1002/rcs.1657.
The object of the present invention is to provide an alternative system for providing good real life visibility of at least a part of a body cavity during a laparoscopic minimally invasive surgery to provide accurate real life information to the operator of a laparoscopic tool about spatial shape and/or positions and/or surface area information to thereby make it simpler for the operator to manipulate the laparoscopic tool during a laparoscopic procedure and/or for simpler examination of a surface area within a surgery cavity. Thereby the laparoscopic procedure may be performed faster and/or with reduced risk of unintended damage to the patient.
This and other objects have been solved by the invention or embodiments thereof as defined in the claims and as described herein below.
It has been found that the invention or embodiments thereof have a number of additional advantages, which will be clear to the skilled person from the following description.
In a first aspect the laparoscopic tool system for minimally invasive surgery, comprises
In a second aspect the laparoscopic tool system for minimally invasive surgery comprises
In a third aspect the laparoscopic tool system for minimally invasive surgery is a combination of the laparoscopic tool system of the first aspect and the laparoscopic tool system of the second aspect.
The laparoscopic tool system has been found to ensure a very accurate real time distance and/or spatial information to the operator during a minimally invasive surgery procedure. Even where smoke, fog or similar which usually results in blurring up images obtained by an camera, such as a stereo camera is within the minimally invasive surgery cavity, it has been found that the laparoscopic tool system of the invention still may operate with a very high accuracy.
The terms “distal” and “proximal” should be interpreted in relation to the orientation of the laparoscopic tool, such as a laparoscopic instrument i.e. the distal end of the laparoscopic tool is the part of the laparoscopic tool furthest from a handle or collar portion which constitutes the proximal end of the laparoscopic tool.
The phrase “distal to” means “arranged at a position in distal direction to the laparoscopic tool, where the direction is determined as a straight line between a proximal end of the laparoscopic tool and the distal end of the laparoscopic tool. The phrase “distally arranged” means arranged distal to the distal end of the laparoscopic tool.
The phrase “real time” is herein used to mean the time it requires the computer to receive and process constantly changing data optionally in combination with other data, such as predetermined data, reference data, estimated data which may be non-real time data such as constant data or data changing with a frequency of above 1 minute to return the real time information to the operator. “Real time” may include a short delay, such as up to 5 seconds, preferably within 1 second, more preferably within 0.1 second of an occurrence.
The Term “operator” is used to designate a surgeon or a robotic surgeon i.e. a robot programmed to perform a laparoscopic procedure on a patient.
The term “access port” means a port through which a surgical instrument may be inserted. The access port may comprise a seal or an insulation, a lid and/or similar which fully or partly locks or fills out the access port when the laparoscopic instrument is not inserted in the access port. The seal, insulation and/or seal ensure that undesired amounts of gasses do not escape and deflate the body cavity. When a laparoscopic tool is not inserted in the access port, the seal or the insulation advantageously seals against undesired leakage of gas.
The term “rigid connection” means a connection which ensures that the relative position between rigidly connected elements is substantially constant during normal use.
The term “laparoscopic tool” means herein any tool adapted for being inserted into a small surgical incision in the mammal skin e.g. the abdominal wall. The laparoscopic tool includes a laparoscopic instrument, a laparoscope, a penetrator and a cannula.
Laparoscopic tools comprise dynamic laparoscopic tools and static laparoscopic tool. Static laparoscopic tools include laparoscopes (sometimes referred to as endoscopes), penetrators and cannulas. Static laparoscopic tools are generally held in stationary position during the actual minimally invasive surgery procedure or in the case of a laparoscopic tool in the form of a penetrator. A penetrator is usually applied for making the incision to provide an access port for a dynamic laparoscopic tool. A penetrator is often in the form of an obturator or a trocar. A Trocar comprises an obturator and a cannula. Upon incision by the obturator the cannula is positioned to form an access port to the MIS cavity. The obturator may have a channel for feeding gas to form or to expand the minimally invasive surgery cavity, where after the obturator is removed, leaving the cannula as an access port.
Dynamic laparoscopic tools includes laparoscopic instruments, Such laparoscopic instrument are used for dynamic movements within the minimally invasive surgery cavity for performing the actual minimally invasive surgery procedure
The term “laparoscopic instrument” means herein a laparoscopic tool comprising a surgical tool adapted for performing surgery onto the tissue within the minimally invasive surgery cavity e.g. a grasper, a suture grasper, a cutter, a sealer, a stapler, a clip applier, a dissector, scissors, shears, a suction instrument, a clamp instrument, an electrode, a coagulation device, a curette, ablators, scalpels, a needle holder, a needle driver, a spatula, forceps, a biopsy and retractor instrument or a combination thereof.
As explained above a dynamic laparoscopic tool is a laparoscopic instrument.
A laparoscopic instrument comprising a grasper as its surgical tool for performing surgery is referred to as a grasper instrument and etc. The surgical tool may in an embodiment be detachable from the remaining part of the laparoscopic instrument.
In an embodiment the laparoscopic instrument is selected from graspers, scissors or staplers, preferably the laparoscopic instrument is adapted for performing a surgical procedure on the surface area, e.g. comprising direct interaction with tissue at or in the vicinity of the tissue surface area.
The laparoscopic instrument is not an incision instrument such as a penetrator or an obturator. Advantageously the surgical tool if the laparoscopic instrument does not have any sharp edge for cutting through mammal skin.
Preferably the laparoscopic instrument in adapted for being inserted into the minimally invasive surgery cavity via a cannula access port and manipulated within the surgical field distal from the cannula access port or port of entry with the surgical cavity said surgical field.
The dynamic laparoscopic tool is advantageously configured for being dynamically moved within the minimally invasive surgery cavity at a position distal relative to the access port and/or the cannula.
The term “cannula” means herein a hollow tool adapted for being inserted into an incision to provide an access port as defined above.
The term “laparoscope” means herein a laparoscopic tool which is not a laparoscopic instrument or a cannula. A laparoscope will usually carry a light receiver, a camera, an emitter, an illuminator or any combinations thereof.
The term “substantially” should herein be taken to mean that ordinary product variances and tolerances are comprised.
The term “about” is generally used to include what is within measurement uncertainties. When used in ranges the term “about” should herein be taken to mean that what is within measurement uncertainties is included in the range.
It should be emphasized that the term “comprises/comprising” when used herein is to be interpreted as an open term, i.e. it should be taken to specify the presence of specifically stated feature(s), such as element(s), unit(s), integer(s), step(s) component(s) and combination(s) thereof, but does not preclude the presence or addition of one or more other stated features.
Throughout the description or claims, the singular encompasses the plural unless otherwise specified or required by the context.
The laparoscopic tool system of the first aspect is specifically focused on making it simpler for the operator to manipulate a laparoscopic instrument with high precision during a laparoscopic procedure by providing the operator with real life information about a distance between a surface area and the laparoscopic instrument. The surface area is advantageously a tissue surface area within the surgery cavity where the minimal invasive surgery is to be performed, e.g. a target area for the laparoscopic procedure.
Often the surface of the minimally invasive surgery cavity is very curved. The term ‘target area’ of the surface of the minimally invasive surgery cavity is herein used to designate the area which the operator has focus on, and the target area may advantageously comprise a surgery site and/or a tissue surface area which potentially could be in risk of damage during the laparoscopic procedure.
In an embodiment of the first aspect of the invention the time-of-flight light emitter and the time-of-flight receiver is fixed to the distal portion of the laparoscopic tool, the time-of-flight light emitter is adapted to emitting light and the time-of-flight receiver is adapted to sensing reflected light emitted by the time-of-flight emitter and to generating received light data when reflected from a surface area within a preselected distance from the time-of-flight distance sensor set.
The time-of-flight light emitter and the time-of-flight receiver are preferably rigidly fixed to the distal portion of the laparoscopic tool to ensure high accuracy. The time-of-flight light emitter and the time-of-flight receiver may in an embodiment have two or more settings of emitting/receiving directions. Such settings are preferably step-wise for simpler calibration of the time-of-flight distance sensor set.
In that way the operator may receive distance information from the laparoscopic instrument to different parts of a selected surface area within the surgery cavity.
The optional plurality of settings of the emitting/receiving directions is advantageously provided by an angular tilting movement in one or more directions by a step motor, which may be operated by a suitable controller arranged at a handle portion of the laparoscopic instrument or operated by a robotic operator.
The term “emitting direction” means the optical axis of the emitted light and receiving direction” means the optical axis of the received (collected) light.
The computer system is advantageously in data communication with the time-of-flight distance sensor set to receiving the received light data.
It should be understood herein that the term “in data communication” may include any type of data communication capable of transmitting/receiving digital or analog data signal. For simplification it is generally desired to use digital data communication or wire based communication. The data communication may thus be by wire or wireless, e.g. using Bluetooth or similar data transmission systems.
Where the operator is a robot it may be desired to use wire based communication and the computer system or at least a part thereof may advantageously be integrated with the robot.
The computer system is programmed to calculate a real time distance between at least a part of the distal portion and the surface area.
Thus since the operator is constantly informed about the distance between the laparoscopic instrument and the surface area—e.g. one or more points of the surface area—the risk of unintended damaging of tissue is significantly reduced and the operator may be able to perform the laparoscopic procedure with less movements of the laparoscopic tool and/or substantially faster than without the laparoscopic tool system of the first aspect of the invention.
In an embodiment the laparoscopic instrument is a dynamic instrument configured for performing a laparoscopic surgery procedure on the surface area within the minimally invasive surgery cavity. The laparoscopic instrument is adapted for being moved both in axial direction relative to the laparoscopic instrument but also in several tilting directions for performing a minimally invasive surgery procedure.
A laparoscopic surgery procedure means herein a minimally invasive surgery procedure where the laparoscopic instrument is inserted into via a previously formed access port into a minimally invasive surgery cavity to be in interaction with tissue within the minimally invasive surgery cavity by or in the vicinity of the tissue surface area to perform a surgical action on the tissue. A laparoscopic examination procedure Is a minimally invasive surgery procedure where the tissue within the minimally invasive surgery cavity is examined for diagnostic purpose, e.g. for planning and/or for providing information to the operator if and/or how a laparoscopic surgery procedure should be performed.
The laparoscopic tool system of the first aspect is preferably adapted for performing a laparoscopic surgery procedure.
The time-of-flight light emitter and the time-of-flight receiver of the time-of-flight distance sensor set are advantageously configured to be correlated to each other to ensure that the distance, angle and other position based parameters are adjusted to ensure low measurement error.
In an embodiment the time-of-flight light emitter and the time-of-flight receiver of the time-of-flight distance sensor set are integrated into a time-of-flight distance module. Thereby the time-of-flight light emitter and the time-of-flight receiver are spatially correlated. Advantageously the time-of-flight light emitter and the time-of-flight receiver are rigidly interconnected such that an optional movement from a first setting of the emitting/receiving directions to another will be an equal movement of both of the time-of-flight light emitter and the time-of-flight receiver, thereby providing a simpler calibration.
In an embodiment the time-of-flight distance sensor set is not a rangefinder.
In an embodiment the time-of-flight distance module comprises a rangefinder.
Advantageously the time-of-flight distance sensor set is configured for dynamic operation during the minimally invasive surgery procedure while the laparoscopic instrument is dynamically moved with movements including tilting in various directions, such that the operator receives real-time information about the distance between the distal portion of the laparoscopic instrument and the tissue within the minimally invasive surgery cavity.
In an embodiment the time-of-flight distance module is preferably a rangefinder, preferably in miniature format. Rangefinders are well known within the art of hunting and are used to determine the distance to a target animal. The principle used in such range finders may be applied in the time-of-flight distance module, preferably modified to measure distances of up to 1 m such as up to 0.5 m, e.g. from about 1 mm to about 25 cm. In an embodiment the time-of-flight distance module is a miniature distance sensors as marketed by Sensopart, Germany. In an embodiment the time-of-flight distance module comprises at least one distance sensor, such as a short range laser based sensor e.g. as provided by SICK AG, Germany.
Advantageously the computer system comprises at least one computer unit which is integrated with the time-of-flight receiver and/or the time-of-flight light emitter. The incorporated computer unit may for example be applied to control the operation of the time-of-flight receiver and/or the time-of-flight light emitter and to ensure an operative correlation between the time-of-flight receiver and the time-of-flight light emitter.
The time-of-flight light emitter is advantageously adapted for emitting modulated light.
The term “modulated light” means light that is timely variated in a preselected timely pattern, e.g. with a modulation frequency (or merely called frequency) which may be stationary or may vary according to a preselected variation pattern.
The modulated light may be pulsed or continuous-wave (CW) modulated light. The modulation frequency largely depends on the type of modulation. In an embodiment the modulation frequency is at least about 200 Hz, such as at least about 100 KHz, such as at least about 1 MHz, such as at least about 20 MHz, such as up to about 200 MHz or more.
To ensure a high resolution it is preferred that the modulation frequency is at least 200 Hz and advantageously at least about 1 MHz.
In an embodiment the emitted light is pulsed light and the modulation is the pulsing of the light. The pulsed light need not be further modulated, however if desired it may additionally be modulated by one or more other modulation types, such as pulse duration and/or any of the below mentioned modulation types.
In an embodiment the pulsed light is additionally modulated with respect to pulse rep rate e.g. by using a pulse dumper/pulse picker to dump out pulses in a preselected pattern.
In an embodiment the emitted light is CW light and the light is modulated by modular changes of at least one of amplitude (and thereby the intensity), phase, light frequency, pulse frequency, polarization and/or pulse length of the light, preferably the time-of-flight light emitter comprises a tunable modulator.
Generation of such types of light modulations is well known to a skilled person.
The time-of-flight light emitter of the time-of-flight distance sensor set may be coherent, partially coherent or non-coherent.
Coherence is strongly related to the ability of light to exhibit interference effects. A light field is called coherent when there is a fixed phase relationship between the electric field values at different locations or at different times. Partial coherence means that there is some (although not perfect) correlation between phase values. There are various ways of quantifying the degree of coherence, as described below.
There are two very different aspects of coherence:
Spatial coherence means a strong correlation (fixed phase relationship) between the electric fields at different locations across the beam profile. For example, within a cross-section of a beam from a laser with diffraction-limited beam quality, the electric fields at different positions oscillate in a totally correlated way, even if the temporal structure is complicated by a superposition of different frequency components. Spatial coherence is the essential prerequisite of the strong directionality of laser beams.
Temporal coherence means a strong correlation between the electric fields at one location but at different times. For example, the output of a single-frequency laser may exhibit a very high temporal coherence, as the electric field temporally evolves in a highly predictable fashion: it exhibits a clean sinusoidal oscillation over extended periods of time.
Advantageously the time-of-flight light emitter of the time-of-flight distance sensor set is adapted for emitting a spatially and/or temporally coherent light beam. Thereby it is simpler to measure the distance to a selected surface area such as a point at the surface area in particular where the surface area is relatively uneven and undulating.
In an embodiment the emitter preferably comprises a laser, such as a semiconductor laser (sometimes called a semiconductor laser diode or merely a laser diode LD).
Whereas it for distance measurement is desired to use spatially coherent light, spatially incoherent light may in an embodiment be applied. This embodiment is in particular suitable where the surface area is relatively even.
In an embodiment the time-of-flight light emitter comprises at least one of a laser or a light emitting diode (LED), optionally the light emitter comprises two or more equal or different lasers and/or LEDs, such as two or more semiconductor lasers.
Generally the semiconductor LEDs on the market today emit incoherent light. The incoherent light is for example generated by spontaneous emission which may provide the semiconductor LEDs to produce light waves that lack a fixed-phase relationship. LEDs are often preferred due to their relatively low cost.
Semiconductor laser diodes (LD) which are preferred for the emitter of the time-of-flight distance sensor set, preferably emits spatially and temporally coherent light.
Generally the LDs on the market today produce light waves with a fixed-phase relationship (both spatial and temporal) between points on the electromagnetic wave. Light waves having a fixed-phase relationship are referred to as temporally coherent light. Since semiconductor LDs emit more focused light than LEDs, they are preferred for use in the determination of the distance between at least a part of the distal portion and the surface area in particular where the surface area is very undulating and irregular. In an embodiment the time-of-flight light emitter comprises a vertical-cavity surface-emitting laser (VCSEL).
The time-of-flight light emitter is advantageously adapted for emitting a light with a relatively narrow bandwidth. Thereby it is simpler to control and/or adjust for optional penetration of certain wavelengths and also the risk of interference, cross-talk and/or noise due to other light beams may be held at a very low level.
In an embodiment the time-of-flight light emitter comprises a light source having a band width (full width at half maximum—FWHM) of up to about 50 nm, such as from about 0.5 nm to about 40 nm such as from about 1 to about 10 nm.
The time-of-flight light emitter may in principle be configured for emitting any wavelength(s). The emitting wavelength(s) is/are advantageously selected in dependence of optional other light in the surgery cavity (i.e. to be distinguishable from such other light), in dependence of cost, in dependence of optional heat generation (i.e. to avoid undesired heating) and/or in dependence of penetration and/or absorptions properties (i.e. to ensure a highly reliable reflection of the emitted light at the surface area).
In an embodiment the time-of-flight light emitter is advantageously adapted for emitting at least one electromagnetic wavelength within the UV range of from about 10 nm to about 400 nm, such as from about 200 to about 400 nm.
In an embodiment the time-of-flight light emitter is adapted for emitting at least one electromagnetic wavelength within the visible range of from about 400 nm to about 700 nm, such as from about 500 to about 600 nm.
In an embodiment the time-of-flight light emitter is adapted for emitting at least one electromagnetic wavelength within the IR range of from about 700 nm to about 1 mm, such as from about 800 to about 2500 nm.
The time-of-flight light emitter may advantageously comprise a wavelength tunable light source. The wavelength tunable light source may advantageously be tunable by the operator e.g. to switch between preselected wavelengths or wavelengths ranges.
In an embodiment the time-of-flight light emitter comprises two or more light sources having equal or different bandwidths which preferably have different penetration depth and/or are different with respect to absorption of at least one type of tissue, blood and/or water. The two or more pattern light sources are preferably operatively interconnected to ensure that there is no interference or cross-talk between them. In an embodiment the two or more pattern light sources are operatively interconnected to emit light in an alternating order.
The power of the time-of-flight light emitter may advantageously be selectable (tunable) by the operator to be selected for the respective laparoscopic procedure.
In an embodiment the time-of-flight light emitter is adapted for emitting light at a power of from about 1 mW to about 100 mW, such as from about 3 mW to about 50 mW.
Where the time-of-flight light emitter is emitting spatially coherent light the time-of-flight light emitter is preferably adapted for emitting light at a relatively low power e.g. less than 100 mW, whereas where the time-of-flight light emitter is emitting spatially incoherent light the time-of-flight light emitter may be arranged for emitting light at a relatively high power even up to about 100 W, such as up to about 50 W, such as up to 10 W, such as up to about 5 W.
The time-of-flight receiver may be any receiver capable of receiving and detecting light and demodulating the received light. Advantageously the time-of-flight receiver comprises a photodetector, such as an avalanche photodiode (APD), a photomultipliers or a metal-semiconductor-metal photodetector (MSM photodetector).
Such receivers are known from Rangefinders.
The time-of-flight receiver preferably comprises a demodulator for demodulating the received light to determine the time it has taken the light to pass from the time-of-flight light emitter to the time-of-flight receiver and thereby determine the distance between the laparoscopic instrument and the surface area where the light was reflected. The demodulator may advantageously form part of the computer system. In an embodiment the computer system comprises a demodulator positioned at a distance to the time-of-flight distance sensor set.
Advantageously the time-of-flight receiver comprises a band pass filter for suppressing back-ground light.
Preferably the time-of-flight receiver is operatively connected to the time-of-flight light emitter, preferably for timely adjusting the operation of the emitter and the receiver.
Where the emitter is adapted for emitting pulsed light the receiver is advantageously timely synchronized with the time-of-flight light emitter.
To increase the sensitivity of the receiver, the time-of-flight receiver advantageously comprises at least one aperture lens, such as a Fresnel lens for collecting reflected light.
In the second aspect of the invention the laparoscopic tool system is particularly focused on generating 3D imaging of at least a surface area of the surgery cavity to increase the visual perception of the operator. The laparoscopic tool system of the second aspect has been found not only to increase the visual perception but also to provide in real life image scanning with a high flexibility because the system is focused to generate real life 3D imaging on a surface area which is changing as the laparoscopic tool is moved so that a relevant surface area at any time automatically may be subjected to the 3D imaging.
The laparoscopic tool system of the second aspect is therefore highly suitable for performing laparoscopic examination procedures
This flexibility of the 3D imaging may result in an even faster laparoscopic procedure and with even higher accuracy.
In an embodiment of the laparoscopic tool system of the second aspect the time-of-flight image sensor is fixed to the distal portion of the laparoscopic tool, the spatial position tracking means is adapted for tracking the spatial position of the time-of-flight image sensor and generating real time spatial position data of the time-of-flight image sensor.
Advantageously the time-of-flight light emitter is adapted to emitting light and the time-of-flight image sensor is adapted to sensing light from the time-of-flight emitter and generating sensed light data when light is reflected from a surface area within a preselected distance from the time-of-flight image sensor.
The computer system is in data communication with the spatial position tracking means for receiving real time spatial position data of the time-of-flight image sensor. The computer system is also in data communication with the time-of-flight camera set to receiving the sensed light data timely corresponding to the real time spatial position data of the time-of-flight image sensor. The computer system is programmed to calculate the 3D image(s) based on the sensed light data and real time spatial position data of the time-of-flight image sensor.
It is preferred that the computer system is programmed to calculate real life 3D images based on the sensed light data and real time spatial position data of the time-of-flight image sensor.
In the second aspect the laparoscopic tool is a tool which is adapted to being moved during the minimally invasive surgery i.e. at least the distal portion of the laparoscopic tool is adapted to being moved, preferably at least by lateral movements. The laparoscopic tool is preferably selected from a cannula, a laparoscope and a laparoscopic instrument.
In an embodiment the laparoscopic tool is a cannula as described in DK PA 2015 70483 with the modification that the cannula comprises a time-of-flight image sensor at its distal portion.
The spatial position tracking means may be any means suitable for tracking the position of the distal portion of the time-of-flight image sensor and thereby the laparoscopic tool. In an embodiment the spatial position tracking means is configured for tracking the distal portion of the laparoscopic tool and based on this determine the spatial position of the time-of-flight image sensor.
The spatial position tracking means is advantageously adapted for tracking the spatial position of the time-of-flight image sensor relative to a fixed point, a preselected point, and/or relative to an X,Y,Z coordinate matrix.
In an embodiment the spatial position tracking means is advantageously adapted for tracking the spatial position of the time-of-flight image sensor relative to a selected spatial start position. The selected spatial start position may be selected by a user—e.g. via a user interface and/or it may be selected from a database comprising suitable spatial start positions. In an embodiment the operator may be guided to position the laparoscopic tool such that the time-of-flight image sensor is in a selected spatial start position.
The X,Y,Z coordinate matrix may be a spatial reference system, such as a Spatial Reference System Identifier (SRID) system such as those known from spatially enabled databases (such as IBM DB2, IBM Informix, Microsoft SQL Server, MySQL, Oracle, Teradata, PostgreSQL and SQL Anywhere).
In an embodiment the X,Y,Z coordinate matrix may be a local matrix e.g., generated from a number of spatially arranged sensors.
In an embodiment the spatial position tracking means is adapted to tracking the spatial position of the time-of-flight image sensor relative to the time-of-flight light emitter.
In an embodiment the spatial position tracking means comprises at least one sensor mounted to the distal portion of the laparoscopic tool.
In an embodiment the position tracking means is adapted for tracking positions and orientations of the time-of-flight image sensor by a using one or more sensors and/or other signal-producing elements disposed on the laparoscopic tool, preferably immediately adjacent to the time-of-flight image sensor or optionally integrated with the time-of-flight image sensor.
In an embodiment the spatial position tracking means comprises a motion sensor adapted for determining motions of the laparoscopic tool.
In an embodiment the position tracking means comprises an inertial navigation system. The inertial navigation system advantageously includes at least a computer (e.g. of the computer system) and a platform or module containing accelerometers, gyroscopes and/or other motion-sensing devices.
The inertial navigation system may initially be provided with its position and velocity from another source (such as a human operator, a GPS satellite receiver, etc.), and thereafter computes its own updated position and velocity by integrating information received from the motion sensors. The advantage of an inertial navigation system is that it requires no external references in order to determine its position, orientation, or velocity once it has been initialized.
In an embodiment the position tracking means comprising a sensor element adapted to be or being physically connected to or integrated with the distal portion of the laparoscopic tool.
In an embodiment the position tracking means comprises a magnetic motion capture system comprising at least one sensor adapted to be or being physically connected to or integrated with the time-of-flight image sensor to measure low-frequency magnetic fields generated by a transmitter source.
In an embodiment the position tracking means comprises a MEMS sensor magnetic motion capture system adapted to be or being physically connected to or integrated with the time-of-flight image sensor to measure return signal upon activation by a transponder.
In an embodiment the position tracking means comprises an acoustic sensor including at least one sensor mounted on or integrated with the time-of-flight image sensor for increase accuracy—e.g. for the determination of direction and/or orientation of the laparoscopic tool.
In an embodiment the position tracking means comprises at least one distance sensor, such as a short range laser based sensor e.g. as provided by SICK AG, Germany.
A short range laser based distance sensor is operating by projecting a light beam spot onto a measurement object, e.g. using a laser diode. By means of an optical receiver, the reflection is mapped onto a light sensitive element (such as CMOS). Based on the position of the mapped light spot and the 3D surface data, the distance to the surface area may be determined.
In an embodiment the spatial position tracking means comprises at least one local reference sensor, preferably for spatial reference. Advantageously the at least one local reference sensor is adapted to being positioned on a patient and/or on a support (surgery table) for a patient.
In an embodiment there are a plurality of reference sensors, the reference sensors are preferably configured to communicate to locate a position of each other to thereby define an X.Y.Z dimensional space—e.g. as described in US 2007/060098 or U.S. Pat. No. 6,631,271. In an embodiment the 3D surface sensor system is as described in “Spatial Data Estimation in Three Dimensional Distributed Wireless Sensor Networks”, by Karjee et al. Embedded Systems (ICES), 2014 International Conference 3-4 Jul. 2014, IEEE ISBN 978-1-4799-5025-6.
In an embodiment the laparoscopic tool forms part of or is mounted on a robot for motorized manoeuvring of the laparoscopic tool and the spatial position tracking means comprises a motion sensor wherein the motion sensor is adapted for determining motions of the laparoscopic tool at least partly based on the motorized manoeuvring of the laparoscopic tool.
The robot may be motorized by use of any suitable motor(s) or motor systems e.g. comprising one or more step motors and/or one or more actuators.
In an embodiment the spatial position tracking means comprises at least one sensor mounted to the distal portion of the laparoscopic tool.
Advantageously the spatial position tracking means comprises at least one of a global position sensor and/or a local position sensor.
In a preferred embodiment the spatial position tracking means is adapted for detecting changes in rotational attributes, preferably comprising pitch, roll and yaw and optionally the spatial position tracking means is adapted for detecting acceleration.
In an embodiment the spatial position tracking means comprises an inertial measurement system, such as an inertial measurement unit (IMU), the inertial measurement system preferably comprises at least one of an accelerometer, a gyroscope and/or a magnetometer. Advantageously the spatial position tracking means comprises an IMU enabled GPS device. In this embodiment it is desired that the spatial position tracking means is adapted for tracking the spatial position of the time-of-flight image sensor relative to a selected spatial start position as described above.
In an embodiment the spatial position tracking means comprises a structured light projecting unit mounted to the distal portion of the laparoscopic tool and adapted for projecting a structured light pattern, wherein the structured light pattern differs in wavelength(s) from one or more wavelengths emitted by the time-of-flight light emitter.
In an embodiment the spatial position tracking means comprises a depiction system for generating a real time correlated depiction of movements of a laparoscopic tool as described in DK PA 2015 70642.
In an embodiment the spatial position tracking means comprises visual identifier adapted for tracking positions and orientations of the time-of-flight image sensor. Such a visual identifier is for example in the form of a pico-lantern e.g. as described by Philip Edgcumbel et al in “Pico Lantern: A Pick-up Projector for Augmented Reality in Laparoscopic Surgery” Med Image Comput Comput Assist Interv. 2014; 17(Pt 1):432-9. PMID:25333147.
In principle the time-of-flight light emitter of the laparoscopic tool system in its second aspect may be positioned anywhere provided that at least a portion of the emitted light reaches and is reflected from the surface area of the surgery cavity.
In an embodiment the time-of-flight light emitter is adapted to be positioned at a distance to the time-of-flight image sensor, such as at a unit adapted to be held substantially immovable.
To obtain a very accurate resolution, it is desired that the time-of-flight light emitter is either held at a stationary position and/or that the position or changes of position relative to the time-of-flight image sensor and/or to a X,Y,Z coordinate matrix is/are known.
Usually the laparoscopic tool may be calibrated with the time-of-flight light emitter in one or more selected positions and the measurements are thereafter based on this or these calibrations, e.g. modified in view of the movements of the time-of-flight light emitter.
In an embodiment the time-of-flight light emitter is adapted to be positioned at a distance to the time-of-flight image sensor. The time-of-flight light emitter may for example be positioned on a distal portion of a further laparoscopic tool, such as a laparoscope. In an embodiment a further spatial position tracking means—e.g. as the position tracking means described above—is adapted to tracking the spatial position of the time-of-flight light emitter.
Thereby the time-of-flight image sensor and the time-of-flight light emitter may be moved independently of each other while 3D real life images are generated by the laparoscopic tool system. Since the position of both the time-of-flight image sensor and the time-of-flight light emitter are known, the computer system may account for such independent movements and calculate 3D data for the surface area.
In an embodiment the time-of-flight light emitter is adapted to be positioned closer to the surface to be analyzed than the time-of-flight image sensor. For example the time-of-flight image sensor is positioned on the distal portion of a cannula or an endoscope and the time-of-flight light emitter is positioned on a distal portion of a further laparoscopic tool in the form of a laparoscopic instrument.
Thereby the time-of-flight image sensor may be held relatively far from the surface while still highly accurate dynamic real life measurements may be obtained because of the time-of-flight light emitter which may be relatively close to the surface area. The risk of interference and/or cross-talk between pixels/point sensor units of the time-of-flight image sensor is reduced, and accordingly signals with reduced noise are obtained.
In an embodiment the time-of-flight light emitter is fixed to the distal portion of the laparoscopic tool. Thereby the time-of-flight image sensor and the time-of-flight light emitter may be spatially stationary relative to each other. Preferably the time-of-flight light emitter and the time-of-flight image sensor are integrated into a time-of-flight camera module.
In an embodiment the time-of-flight camera module is mounted to or incorporated in a laparoscope. Whereas the resulting 3D images—preferably in real life dynamic mode—do not directly inform the operator about the position of the distal portion of the laparoscope, this information may be obtained from the position tracking means and/or tit may be derived from the 3D image by the computer system.
In an embodiment the time-of-flight camera module is mounted to or incorporated in a cannula. Also in this embodiment the position of the distal portion of the cannula may be obtained from the position tracking means.
In an embodiment the time-of-flight camera module is mounted to or incorporated in a laparoscopic instrument and the time-of-flight image sensor and time-of-flight light emitter is positioned at the distal end of the laparoscopic instrument. This may allow both 3D surface scanning and information on instrument position directly to the operator.
The time-of-flight light emitter of the laparoscopic tool system of the second aspect is advantageously adapted for emitting modulated light e.g. as the time-of-flight light emitter of the laparoscopic tool system of the first aspect.
The modulated light may be as described above.
For high and fast resolution it is desired that the modulation frequency is at least about 200 Hz, such as at least about 100 KHz, such as at least about 1 MHz, such as at least about 20 MHz, such as up to about 200 MHz or more.
The light may be pulsed or CW as described above.
In an embodiment the time-of-flight light emitter is adapted for emitting light at a power of from about 1 mW to about 100 W.
Preferably the time-of-flight light emitter of the laparoscopic tool system of the second aspect is emitting spatially incoherent light the time-of-flight light emitter is arranged for emitting light at a power of from about 1 W to about 50 W.
In an embodiment the time-of-flight light emitter of the second aspect is adapted for emitting a spatially coherent light beam and the emitter is further adapted
Triangular determinations are well known in the art to determine spatial relations, sizes and/or other 3D related dimensions.
In the second aspect it is however preferred that the time-of-flight light emitter is adapted for emitting incoherent light, thereby providing a relatively large spot size onto the surface area.
Preferably the time-of-flight light emitter of the second aspect comprises at least one light emitting diode (LED), optionally the light emitter comprises two or more equal or different LEDs.
The selected wavelength(s), band width, tenability and power may be as described above.
The time-of-flight image sensor is preferably a camera comprising an array of light sensor units, such as a photo detector e.g. a photo-electric sensor unit converting light energy (photons) into electricity (electrons).
Preferably the camera comprises an array of pixel sensors each comprising a photodetector (such as an avalanche photodiode (APD), a photomultiplier or a metal-semiconductor-metal photodetector (MSM photodetector). Preferably the time-of-flight image sensor comprises active pixel sensors (APS), preferably each pixel comprises an amplifier, more preferably the time-of-flight image sensor comprises at least about 1 kilo pixels, such as at least about 1 Mega pixels.
Preferably the camera is a time-of-flight camera build using MEMS technology.
The time-of-flight camera preferably comprises a demodulator for demodulating the received light of each pixel sensor to determine the real life 3D shape.
The demodulator may advantageously form part of the computer system. In an embodiment the computer system comprises a demodulator positioned at a distance to the time-of-flight camera set.
Advantageously the time-of-flight camera comprises a band pass filter for suppressing back-ground light.
The camera preferably has a relatively short integrating time for fast operation and for high accuracy also when the camera is moved. Preferably the time-of-flight camera is operatively connected to the time-of-flight light emitter, preferably for timely adjusting the operation of the emitter and the receiver.
Where the emitter is adapted for emitting pulsed light the receiver is advantageously timely synchronized with the time-of-flight light emitter.
To increase the sensitivity of the camera, the time-of-flight camera advantageously comprises at least one aperture lens, such as a Fresnel lens for collecting reflected light.
Advantageously the time-of-flight image sensor is selected from a charge-coupled device (CDD) image sensor and a complementary metal-oxide-semiconductor (CMOS) image sensor.
When using CMOS or other integrating detectors the camera preferably has a selectable integrating time, comprising settings of relatively low integration time such as 10 ms or less, such as 1 ms or less. Longer integration time may be used for calibration—e.g. up to 100 ms or longer. Also for relatively curved and hilly surface areas a longer integration time may be used.
The laparoscopic tool system of the third aspect is a combination of the laparoscopic tool system of the first aspect and the laparoscopic tool system of the second aspect.
In the third aspect of the laparoscopic tool system, the time-of-flight distance sensor set of the first aspect is advantageously provided as the position tracking means of the laparoscopic tool system of the second aspect.
Thus, in a preferred embodiment the spatial position tracking means comprises a time-of-flight distance sensor set for determining a distance between the distal portion of the laparoscopic tool and a surface area. Preferably the time-of-flight light emitter of the time-of-flight distance sensor set and the time-of-flight light emitter of the time-of-flight camera set are integrated with each other to a combined time-of-flight light emitter for the time-of-flight distance sensor set and the time-of-flight camera set.
In an embodiment the combined time-of-flight light emitter for the time-of-flight distance sensor set and the time-of-flight camera set also comprises the time-of-flight image sensor and preferably the time-of-flight receiver is integrated with or comprised by the time-of-flight image sensor.
In some minimally invasive surgery procedure the laparoscopic instrument is operated to burn off tissue areas which thereby generates smoke.
In an embodiment comprising a laparoscopic tool system of any one of the first, the second and the third aspect, the laparoscopic tool system further comprises an image camera secured to the laparoscopic tool at its distal portion and configured for acquiring images and transmitting the acquired image to the computer system and the computer system is configured for correcting the images based on the light data received from said time-of-flight receiver and/or time-of-flight image sensor. The image camera may be a stereo camera.
Thereby, if the minimally invasive surgery cavity is filled with smoke, fog or similar which reduces the quality of the images obtained by the image camera, the image of the image camera is corrected to remove errors occurred due to the smoke, fog or similar.
All features of the inventions and embodiments of the invention as described herein including ranges and preferred ranges may be combined in various ways within the scope of the invention, unless there are specific reasons not to combine such features.
The above and/or additional objects, features and advantages of the present invention will be further elucidated by the following illustrative and non-limiting description of embodiments of the present invention, with reference to the appended drawings.
The figures are schematic and may be simplified for clarity. Throughout the same reference numerals are used for identical or corresponding parts.
Further scope of applicability of the present invention will become apparent from the description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the invention, are given by way of illustration only, since various changes and modifications within the spirit and scope of the invention will become apparent to those skilled in the art from this detailed description.
The laparoscopic tool of
The cannula shown in
The cannula shown in
The laparoscopic instrument 20 comprises a distal portion 21 with a not shown surgical tool. The time-of-flight distance sensor module 23 is fixed at the distal portion 21 of the laparoscopic instrument 20. The time-of-flight distance sensor module 23 comprises a time-of-flight light emitter and a time-of-flight receiver. The time-of-flight light emitter emits light Lsc that at least is spatially coherent. The emitted light may advantageously also be temporally coherent. The light Lsc is impinging on the surface area 25 and at least a part of the light is reflected by the surface area 25 and captured by the time-of-flight receiver of the time-of-flight distance sensor set 23. By determining the time it takes the emitted light to reach the time-of-flight receiver, the distance between the distal portion 21 of the laparoscopic instrument 20 and the surface area 25 may be determined. The computer system 24 is schematically illustrated and may comprise two or more separate computing elements in data connection with each other—e.g. one or more drivers, one or more modulators and demodulators, one or more data storing elements, such as one or more data bases, etc. The various elements of the computer system are adapted to be at least temporarily in data connection with each other for transferring the data required to operate the time-of-flight distance sensor set and performing the distance determination. In an embodiment one or more elements of the computer system is/are integrated in the time-of-flight distance sensor set, e.g. one or more drivers and/or one or more modulators and/or demodulators. The data transmission may be by wire or be wireless, e.g. using blue tooth.
By having two or more time-of-flight receivers 23a positioned at a distance from each other at the distal portion 21 of the laparoscopic instrument 20 of the laparoscopic tool system, the system may also determine an angular position of the laparoscopic instrument relative to the surface area 25.
The time-of-flight receiver 23d is positioned at a distance from the outermost end of the distal portion 21 of the laparoscopic instrument 20, i.e. slightly retracted relative to the time-of-flight light emitters 23c. Thereby the risk of wetting the time-of-flight receiver 23d with blocking or blurring liquids such a blood and/or wound fluid is reduced.
Spatially incoherent light will spread out as it propagates. For the first aspect of the laparoscopic tool system it is generally desired to use spatially incoherent light because it spreads out and covers a relatively large surface area whereas a spatially coherent light generally will have a relatively narrow spot size.
Temporal coherence describes the correlation between waves observed at different moments in time. Monochromatic light is usually temporally coherent whereas broad band light is temporally incoherent.
Spatially coherent light is usually laser light. In the laparoscopic tool system of the first aspect it is generally desired to use spatially coherent light.
The laparoscopic tool system of an embodiment of the second aspect as shown in
The laparoscopic tool 50 comprises a distal portion 51 optionally with a not shown surgical tool. The time-of-flight camera module 53 is fixed at the distal portion 21 of the laparoscopic tool 50. The time-of-flight camera module 53 comprises a time-of-flight light emitter and a time-of-flight image sensor in the form of a time-of-flight camera. The time-of-flight light emitter emits light Lsi that is spatially incoherent. The emitted light may advantageously be temporally coherent (monochromatic). The light Lsi is impinging on the surface area 55 and at least a part of the light is reflected by the surface area 55 and captured by the time-of-flight camera of the time-of-flight camera set 53.
The time-of-flight camera comprises an array of pixels for receiving the reflected light and based on the received pattern of light and on real time spatial position data obtained from said position tracking means 56, a real time 3D image may be determined.
The spatial position tracking means 56 may be as described above. The spatial position tracking means is positioned at a known distance to the time-of-flight camera module 53.
The computer system 54 is schematically illustrated and may comprise two or more separate computing elements in data connection with each other—e.g. one or more drivers, one or more modulators and demodulators, one or more data storing elements, such as one or more data bases, etc. The various elements of the computer system are adapted to be at least temporarily in data connection with each other for transferring the data required to operate the time-of-flight camera set and performing the 3D determination. In an embodiment one or more elements of the computer system is integrated in the time-of-flight camera set e.g. one or more driver and/or one or more modulators and/or demodulators. The data transmission may be by wire or be wireless, e.g. using blue tooth.
The computer system 54 is in data communication with the spatial position tracking means 56 for receiving real time spatial position data of the time-of-flight camera module 53. Since the relative position between the spatial position tracking means and the time-of-flight camera module 53 is known the spatial position tracking means 56 may track the real time position of the time-of-flight camera module. And thus the position from where the light is emitted and received is known and the 3D determination may be performed.
The computer system 54 is also in data communication with the time-of-flight camera module 53 to receive sensed light data timely corresponding to said real time spatial position data, and the computer system is programmed to calculate the real time 3D images based on the sensed light data and the real time spatial position data.
The laparoscopic tool system shown in
By having the time-of-flight light emitter 53b and the time-of-flight image sensor 53a positioned at a distance to each other such that their relative positions are known, the system becomes more flexible and it is simpler to integrate the time-of-flight camera set into the laparoscopic tool 50.
The camera set 133a also comprises a spatial position tracking means e.g. in the form of the time-of-flight distance sensor set 133 of
Thereby an associated computer system may calculate real time 3D images of the surface area 135.
Number | Date | Country | Kind |
---|---|---|---|
PA 2015 70642 | Oct 2015 | DK | national |
PA 2016 70428 | Jun 2016 | DK | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/DK2016/050327 | 10/7/2016 | WO | 00 |