The present disclosure relates to a system and method for dynamic validation and correction of registration, and recovery of lost reference, for surgical navigation during operations.
During a navigated surgical procedure a surgeon typically needs to correlate the position of previously acquired imaging data (in three dimensions), that have been obtained from an imaging device or system (for example, ultrasound, CT or MRI data), with the physical position of the patient who is to be operated on. In some systems, for a navigation procedure a handheld instrument may be tracked using a tracking system, and a representation of the instrument's position and orientation may be displayed as an overlay on the three-dimensional imaging data from a scan of the patient's anatomy. To achieve this, a registration is obtained between the coordinate frame of the tracking system for the handheld instrument, the physical location of the patient in physical space, and the coordinate frame of the corresponding image data of the patient. Ensuring that the registration is aligned with and corresponds to the actual physical reality of the procedure is desirable and necessary for maintaining surgeon confidence in the information being presented and in ensuring the navigated procedure is accurately executed.
However, it tends to be difficult to measure and quantify registration accuracy. In the prior art, this accuracy has been reported to a surgeon as a confidence or tolerance number at the time that registration is computed. This was also described by the paper [The Silent Loss of Navigation Accuracy; Research-Human-Clinical Studies; Vol. 72, No. 5, May 2013, pages 796-807]. This number is not indicative of the complexity of registration accuracy, and, more significantly, is not indicative of the fact that accuracy can vary in different parts of the surgical field. Further, this number is used as a one-time accept/reject criterion for the registration—once the registration is accepted typically it is assumed to be correct for the duration of the procedure, or until the surgeon notices that something is significantly misaligned.
With the present state of the art misalignment of the navigation system is difficult to identify as a typical system only presents a virtual representation of the OR procedure, and as such it cannot be readily contrasted to the actual physical state of the OR at a given time. Currently, for a surgeon to measure registration accuracy during a procedure he or she typically locates the tool relative to an identifiable location on the actual patient anatomy while noting the degree to which the location of the virtual tool is displaced from the same location relative to the virtualized patient anatomy, where such virtual tool is displayed as an overlay on the three-dimensional imaging data from a scan of the patient's anatomy. Furthermore, once a registration misalignment is noticed, correcting for the error tends to be difficult, and often not achievable. Additionally, non-uniform displacement of tissue during a procedure also tends to mean that global corrections are not possible.
This application describes a system and method for validating registrations, and detecting and correcting registration.
An embodiment disclosed herein provides a method of confirmation of correct registration between one or more previously registered virtual objects in a coordinate frame of a surgical navigation system (which is located in an operating room in which a medical procedure is to be performed) and intra-operatively acquired imaging during the medical procedure in the coordinate frame of the surgical navigation system. Wherein a previously registered virtual object may be a computed tracked instrument visualization, or other computed tracked real object visualization. The surgical navigation system includes a tracking mechanism. The method include displaying intra-operatively acquired imaging of a surgical field containing one or more real objects corresponding to the one or more virtual objects, the surgical field containing a pre-selected number of landmarks in fixed and known locations with respect to the one or more real objects, with the landmarks being tracked by the tracking mechanism. The method includes overlaying a virtual image (as generated by a virtual camera) containing the one or more virtual objects previously registered onto the intra-operatively acquired imaging and detecting for any misalignment or non-concordance between any one of the one or more previously registered virtual objects contained in the virtual image and its corresponding real object contained in the intra-operatively acquired imaging, wherein a presence of misalignment or any non-concordance is indicative of a registration error.
An embodiment disclosed herein is a system for confirmation of correct registration between one or more previously registered virtual objects and intra-operatively acquired imaging during a medical procedure. The systems comprises a surgical navigation system having a coordinate frame of reference and including a tracking mechanism. The system further comprises a computer control system which includes a computer processor, and a storage device and a visual display both connected to the computer processor. The storage device has stored therein a visualization of one or more previously registered virtual objects in the coordinate frame of reference of the surgical navigation system. The system includes at least one sensor for acquiring intra-operative imaging of a surgical field during the medical procedure in which the surgical field contains one or more real objects corresponding to the one or more virtual objects and a pre-selected number of landmarks in known locations with respect to the one or more real objects. The landmarks and the at least one sensor are tracked by the tracking mechanism in the coordinate frame of the surgical navigation system. The computer processor is programmed with instructions to receive and display the intra-operatively acquired imaging of the surgical field containing the one or more real objects and to overlay the virtual image containing the one or more virtual objects previously registered onto the intra-operatively acquired imaging and wherein any misalignment or non-concordance between any one of the one or more previously registered virtual objects contained in the virtual image and its corresponding real object contained in the intra-operatively acquired imaging is indicative of a registration error between the virtual object and its corresponding real object in said coordinate frame of said surgical navigation system.
In an embodiment, the described system and method can also provide corrections based on the difference between local tissue characteristics and virtual instrument representations at the location that the surgeon is focusing on and a live video stream of the surgical field to immediately visualize (and optionally automatically or manually correct for) any difference between the expected (as calculated through a registration) and actual positions of tracked instruments and imaged patient tissue, which tends to achieve local, immediate, corrections of registration in an intuitive way.
A further understanding of the functional and advantageous aspects of the invention can be realized by reference to the following detailed description and drawings.
In an embodiment a method of confirmation of registration between one or more previously registered virtual objects in a common coordinate frame of a surgical navigation system and an operating room in which a medical procedure is to be performed, and intra-operatively acquired imaging during said medical procedure in said common coordinate frame, said surgical navigation system including a tracking system, the method comprising: a) displaying intra-operatively acquired imaging of a surgical field containing one or more real objects corresponding to said one or more previously registered virtual objects, the real objects being tracked by the tracking system; b) overlaying a virtual image containing the one or more previously registered virtual objects onto the intra-operatively acquired imaging, from the point of view of the intra-operatively acquired imaging; and c) detecting for any misalignment between any one of the one or more previously registered virtual objects contained in the virtual image and its corresponding real object contained in the intra-operatively acquired imaging, wherein a presence of misalignment is indicative of registration error between the virtual object and its corresponding real object.
In yet another embodiment a system for confirmation of correct registration between one or more previously registered virtual objects and intra-operatively acquired imaging during a medical procedure, comprising: a) a surgical navigation system having a coordinate frame and including a tracking mechanism; b) a computer control system including a computer processor, a storage device and a visual display both connected to said computer processor, said storage device having stored therein a computed tracked instrument visualization of one or more previously registered virtual objects in said coordinate frame of reference of said surgical navigation system; c) at least one sensor for acquiring intra-operative imaging of a surgical field during the medical procedure, said surgical field containing one or more real objects corresponding to said one or more virtual objects; and d) said computer processor being programmed with instructions to receive and display said intra-operatively acquired imaging of the surgical field and to overlay a virtual image from the point of view of the virtual camera onto the intra-operatively acquired imaging, wherein any misalignment between any one of the one or more previously registered virtual objects contained in the virtual image and its corresponding real object contained in the intra-operatively acquired imaging is indicative of a registration error between the virtual object and its corresponding real object.
Embodiments disclosed herein will be more fully understood from the following detailed description thereof taken in connection with the accompanying drawings, which form a part of this application, and in which:
Various apparatuses or processes will be described below to provide examples of embodiments of the invention. No embodiment described below limits any claimed invention and any claimed invention may cover processes or apparatuses that differ from those described below. The claimed inventions are not limited to apparatuses or processes having all of the features of any one apparatus or process described below or to features common to multiple or all of the apparatuses or processes described below. It is possible that an apparatus or process described below is not an embodiment of any claimed invention.
Furthermore, numerous specific details are set forth in order to provide a thorough understanding of the embodiments described herein. However, it will be understood by those of ordinary skill in the art that the embodiments described herein may be practiced without these specific details. In other instances, well-known methods, procedures and components have not been described in detail so as not to obscure the embodiments described herein. Also, the description is not to be considered as limiting the scope of the embodiments described herein.
Furthermore, in the following passages, different aspects of the embodiments are defined in more detail. In particular, any feature indicated as being preferred or advantageous may be combined with at least one other feature or features indicated as being preferred or advantageous.
As used herein, the phrase “intra-operatively acquired imaging” refers to images of a medical procedure being performed on an anatomical part. The imaging procedure may include using a sensor to acquire a continuous intra-operatively acquired image stream (i.e. obtained for example by a video camera corresponding to real-time imaging) or an intra-operatively acquired image taken at one or more specific times during the procedure using a sensor other than a video camera, for example a sensor (such as, but not limited to a camera) which is configured to record specific types of images one by one. The present disclosure includes both modalities.
In an embodiment, there is provided a continuously available and real-time confirmation of registration, with an intuitive interface for verification and correction (if necessary). In the embodiment, an overlay of computed tracked instrument visualization and patient imaging information on a video image of the surgical field is provided during a procedure. In
Persons skilled in the art will appreciate that being able to visualize a medical instrument when it is within a patient will aid in the improvement of the accuracy of the procedure. This can be seen in
Referring to
A virtual camera is the point and orientation in virtual 3D space that is used as a vantage point from which the entire virtual 3D scene is rendered. The virtual markers are objects within this virtual 3D scene. It is understood that those familiar in the art will recognize this methodology as a standard computer graphics technique for generating virtual 3D scenes.
The present system may be used with any compatible surgical navigation system. A non-limiting example of such a surgical navigation system is outlined in the co-pending U.S. patent application entitled “SYSTEMS AND METHODS FOR NAVIGATION AND SIMULATION OF MINIMALLY INVASIVE THERAPY”, U.S. Pat. Publication US20150351860 based on U.S. patent application Ser. No. 14/655,814, which claims the priority benefit of U.S. Provisional Patent Application Ser. Nos. 61/800,155 and 61/924,993, and wherein for the purposes of this present United States Patent Application, the Detailed Description, and Figures of U.S. Pat. Publication US20150351860 are incorporated herein by reference.
In an embodiment the validation system is used in a port based surgery the phases of which are depicted in
Once the plan has been imported into the navigation system (step 1505), the anatomical part of the patient is affixed into position using a head or body holding mechanism. The patient position is also confirmed with the patient plan using the navigation software (step 1510). In this embodiment, the plan is reviewed and the patient positioning is confirmed to be consistent with craniotomy needs. Furthermore, a procedure trajectory may be selected from a list of planned trajectories produced in the planning procedure.
Returning to
Those skilled in the art will appreciate that there are numerous registration techniques available and one or more of them may be used in the present application such as the one described in this disclosure and shown in
Registration can also be completed by conducting a surface scan procedure shown generally at (1690) in
As described herein the overlay of registered virtual and corresponding real objects in the surgical suite displayed by the navigation system allows for the identification of mis-registration arising between a virtual object and its corresponding real object in the surgical suite. As shown in
The subsequent steps after initial registration (1515) and confirmation of registration (1520) in a port-based procedure are further outlined in
Initial registration typically has occurred before the patent is draped (1530). Draping typically involves covering the patient and surrounding areas with a sterile barrier to create and maintain a sterile field during the surgical procedure. The purpose of draping is to eliminate the passage of microorganisms (i.e., bacteria) between non-sterile and sterile areas.
Upon completion of draping (step 1530), the next step is to confirm patient engagement points (step 1535) and then prepare and plan the craniotomy (step 1540).
Upon completion of the prep and planning of the craniotomy step (step 1540), the next step is to cut craniotomy (step 1545) where a bone flap may be removed from the skull to access the brain. The above steps of draping, and performing craniotomy, are known in the art to add to registration inaccuracy.
The next step is to confirm the engagement point and the motion range of the port (step 1555), and once this is confirmed the procedure typically advances to the next step of cutting the dura at the engagement point and identifying the sulcus (step 1560).
Thereafter, the cannulation process may be initiated (step 1562). Cannulation involves inserting a port into the brain, typically along a sulcus path as identified in step 1560 using an obturator (introducer). Cannulation may be an iterative process that involves repeating the steps of aligning the port, setting the planned trajectory (step 1580), and then cannulating to a target depth (step 1585), until the complete trajectory plan is executed (step 1562).
The surgery then proceeds (step 1565) by removing the obturator (introducer) from the port allowing access to the surgical site of interest. The surgeon then perform treatment at the distal end of the port, which may involve resection (step 1570) to remove part of the brain and/or tumor of interest. Lastly, the surgeon typically remove the port, close the dura and close the craniotomy (step 1575).
Active or passive fiduciary markers (1020) may be placed on the port (1950) and/or imaging sensor (same as 720 in
Markers (1020) may be arranged statically on the target on the outside of the patient's body or connected thereto. Tracking data of the markers acquired by the stereo camera are then logged and tracked by the tracking system. An advantageous feature is the selection of markers that can be segmented easily by the tracking system against background signals. For example, infrared (IR)-reflecting markers (1020) and an IR light source from the direction of the stereo camera can be used. Such tracking system is known, for example, such as the “Polaris” system available from Northern Digital Inc.
In a preferred embodiment, the navigation system may utilize reflective sphere markers in combination with a stereo camera system, to determine spatial positioning and pose of the medical instruments and other objects within the operating theater. Differentiation of the types of medical instruments and other objects and their corresponding virtual geometric volumes could be determined by the specific orientation of the reflective spheres relative to one another giving each virtual object an individual identity within the navigation system. This allows the navigation system to identify the medical instrument or other object and its corresponding virtual overlay representation (i.e. the correct overlay volume) as seen as (310) in
Other types of markers that could be used would be RF, EM, LED (pulsed and un-pulsed), glass spheres, reflective stickers, unique structures and patterns, where the RF and EM would have specific signatures for the specific tools they would be attached to. The reflective stickers, structures and patterns, glass spheres, LEDs could all be detected using optical detectors, while RF and EM could be picked up using antennas. Advantages to using EM and RF tags would include removal of the line-of-sight condition during the operation, whereas using an optical-based tracking system removes the additional noise and distortion from environmental influences inherent to electrical emission and detection systems.
In a further embodiment, 3-D design markers could be used for detection by an auxiliary camera and/or optical imaging system. Such markers could also be used as a calibration pattern to provide distance information (3D) to the optical detector. These identification markers may include designs such as concentric circles with different ring spacing, and/or different types of bar codes. Furthermore, in addition to using markers, the contours of known objects (i.e., side of the port) could be made recognizable by the optical imaging devices through the tracking system.
In another further embodiment, the medical instrument may be made or configured with an additional protrusion or feature that would not normally be obscured during a procedure, so that such protrusion or feature would typically be visible to the optical sensor during the procedure. Having such feature or protrusion on a tool would enable a verification of registration despite the fact that other portions of the tool may be obscured by patient anatomy, or other objects. As such, in such an embodiment it would be possible to verify registration without having to remove the tool from the patient.
Referring to
In general, each time a point is identified, the virtual and actual coordinate frames, (730) and (740) respectively, become more accurately aligned. For example, as the tip of the pointer (748) in
As shown in
In an embodiment, any tracked medical instrument(s) and 3D MR image data is computed for display as an overlay in the live video image feed, positioned relative to the registration transform, (for example a data overlay corresponding to an anatomical part (510), and the anatomical part). This would show alignment of the computed display with the video image of both the instrument(s) and the contours of the anatomical data if registration is correct, as shown in the bottom image (530) of
In an embodiment, a surface rendering of the MR, CT, ultrasound or other medical imaging data can be generated and displayed in a way to match the viewing position and optical properties (e.g. such as zoom, field of view, etc.) of the viewing camera. As this rendering is dependent on the computed registration between the image (or MR, CT, ultrasound or other medical imaging) dataset and the physical camera position, any mis-registration will tend to be instantly visible as a misalignment in the overlay display, and can be used to dynamically validate and ensure confidence in the current registration. An example of a misalignment in the overlay display can be seen in
Further, if a mis-registration is detected, a registration correction can be applied by manipulating the rendering of the MR, CT, ultrasound or other medical imaging data on the screen (for example, by rotation, translation, scaling, and any combination thereof) until it matches the overlaid video, or expressed more generally, until the virtual objects in the rendering are aligned with the real objects in the intra-operative imaging. In addition to rotation, translation, and scaling corrections, above, the rendering may also be skewed or manipulated non-linearly (such as optical flow) to generate an alignment with the real objects. Examples of linear translation, and rotation, are shown in the diagram in
An example embodiment of an automatic misalignment detection and correction process is shown in
The first step (2000) in the process is to identify and locate the Tracking Reference Marker of the Object of interest (TRMO) (for example, (1010) shown in
(xo,yo,zo,αo,βo,γo)
wherein the subscript “o” denotes the coordinates of the virtual volume of the object of interest as determined by the tracking device. The following step (2015) is to identify and locate the Tracking Reference Marker of the Imaging Sensor (TRMS) using the tracking device. Step (2020) is to register the TRMS location in the common coordinate frame by assigning it a coordinate value describing its exact location and pose in the common coordinate frame as assigned below for example:
(xs,ys,zs,αs,βs,γs)
wherein the subscript “s” denotes the coordinates of the imaging sensor in the common coordinate frame. The next step (2025) is to obtain the imaging feed acquired from the imaging sensor using the navigation system. The next step (2030) is to align the virtual imaging sensor with the imaging sensor in the common coordinate frame using the TRMS (i.e. so that the views of the two cameras are aligned as shown in the bottom right frame (705) of
(xt,yt,zt,αt,βt,γt)
wherein the subscript “t” denotes the coordinates of the virtual volume of the object of interest as determined by the imaging sensor. Step (2055) is to subtract the coordinates of the VVOT and VVOR as shown below for example:
(xq,yq,zq,αq,βq,γq)=(xo,yo,zo,αo,βo,γo)−(xt,yt,zt,αt,βt,γt)
wherein the subscript “q” denotes the deviation in location and orientation (spatial positioning and pose, respectively) of the overlaid and real objects in the imaging feed (for example (320) and (310) as shown in
xq>xQ;
yq>yQ;
zq>zQ;
αq>αQ;
βq>βQ; or
γq>γQ
wherein the subscript “Q” denotes the coordinates of a threshold metric used to determine if the virtual and real objects of interest are misaligned outside of a given tolerance, termed the “threshold coordinate”. The next step (2065), if the test coordinate is greater than the threshold coordinate, is to convert the test coordinate obtained in step (2055) into a translation transform and apply it to the VVOT to assign it a new location relative to the TRMO in the common coordinate frame, as shown below for example:
(xoa,yoa,zoa,αoa,βoa,γoa)=(xo,yo,zo,αo,βo,γo)−(xq,yq,zq,αq,βq,γq)
wherein the subscript “oa” denotes the coordinates of the overlaid virtual volume (VVOR) correction. This step (2055) also entails then setting the newly obtained VVOR coordinate to complete the correction, as shown below for example:
(xo,yo,zo,αo,βo,γo)=(xoa,yoa,zoa,αoa,βoa,γoa).
step (2060), if the test coordinate is less than the threshold coordinate,
or if step (2055) is completed, is to return to step (2000) and restart the loop.
In a further embodiment, the system can add registration using video overlay-match for MR, CT, ultrasound and any other imaging modality, with the addition of annotations of features on the image, (for example which may include solid outlines covering the port opening contour). These overlays can be fixed while the underlying medical image representation is manipulated (such as for a registration correction). A registration is achieved by manipulating the underlying MR to match these overlay positions, such as in the dimensions of the image data set or in three-dimensions. For example, three-dimension data points from tracked instrument(s), and patient features (such as tip of nose, corner of eyes, edge of ears, positions of bony protrusions, vessel bifurcations, etc.) may be overlaid, or the system can utilize landmarks such as a drawing of a surface of the patient or tracing structure (e.g. sulci, ears, exposed vessels) through a tool and the tracking system. An example of this may be seen depicted in the upper picture frame (810) in
In one embodiment, tracking of tools may be employed to improve the rendering of the optical images. A first improvement may be obtained by masking out the upper (relative to the bottom of the access port) portion of the inserted tools/surgical instruments. Often the tools are out of focus in the optical field at the top portion of their location into the access port. Here the image often experiences glare or out of focus issues. Since the system can track the tools and register the tools with a video image, a portion of the tool may be masked out. Masking may be performed, for example, based on known geometrical models of the tools, and/or based on real-time image segmentation as described herein and in the paper noted above [Monocular Model-Based 3D Tracking of Rigid Objects: A Survey]. For example, the upper portion of the tools, or another pre-defined portion of the tools, may be masked or otherwise modified. Accordingly, image artifacts may be reduced, and the ability to utilize the entire dynamic range of the system may be improved or enabled.
Additionally, in a related embodiment, the system may be employed to replace the selected region of the tracked tool with a rendered version of the tool that follows the three-dimensional profile of the tool, optionally including rendered features such as a shadowed rendering that indicates and/or emphasizes the change in the diameter of the tool, as it is further distal in the access port. This provides an opportunity to enhance three-dimensional understanding of tool locations. The tool would then be represented with a partial real-time video view of the actual tip, and a computer rendered view of the upper portion of the tool.
By focusing the camera's gaze on the surgical area of interest, a registration update can be manipulated to ensure the best match for that region, while ignoring any non-uniform tissue deformation areas affecting areas outside the surgical field of interest. By way of example, by focusing the imaging sensor (720) on the surgical area of interest, a re-registration can be configured to ensure the best match for that particular region (as shown as (850) in the lower picture frame in
Additionally, by matching overlay representations of tissue with an actual view of the tissue of interest, the particular tissue representation can be matched to the video image, and thus ensuring registration of the tissue of interest (850). For example, the embodiment can:
In other embodiments, multiple cameras (or a single camera moved to multiple positions) can be used and overlaid with tracked instrument(s) views, and thus allowing multiple views of the imaging data and their corresponding overlays to be presented. An example of this may be seen in the diagrams in
In an embodiment, recovery of loss of registration may also be provided. As described above, during a navigation procedure a handheld instrument is tracked using a tracking system, and a representation of the instrument's position and orientation may be provided and displayed as an overlay on a previously acquired or current image (such as a three-dimensional scan) of a patient's anatomy obtained with an imaging device or system (such as ultrasound, CT or MRI). To achieve this, a registration is needed between the coordinate frame of a tracking system, the physical location of the patient in space, and coordinate frame of the corresponding image of the patient. In an embodiment, a registration would be needed between the physical location of the patient, as well as the corresponding image of the patient and the tracking device (1110).
In an embodiment, and as shown in
In an embodiment, there is provided a system and method for the recovery of lost registration, while avoiding the need to perform a full re-registration. In the embodiment, provided are mechanisms for establishing backup reference positions that can be returned to in the event of a loss of registration. In the embodiment, this is provided by one or more secondary reference marker(s) being provided for navigation registration.
The one or more secondary reference marker(s) (for example as shown in
A secondary reference tool(s) is attached to the fixture (or fixtures) and a registration transformation to the secondary reference tool(s) is recorded at any time after initial registration, by transforming the primary registration to the secondary marker's (or markers') position and orientation. Alternately stated, a secondary tracking reference tool(s) (shown in
Once the registration is recorded, this secondary reference tool can optionally be removed. Since the secondary registration marker(s) does not need to be in position during the surgical procedure, the secondary reference marker(s) can be placed so that it is near the patient and occluding the surgical field. Generally, the secondary reference tool may be removed after registration, and need only be returned to the known position of a registration in order to provide recovery of lost registration of a primary (or other secondary) reference.
However, if one or more of the secondary reference(s) are maintained in position during a procedure they can also be used, which tends to improve reference position accuracy by using an average of all reference positions at all times to improve noise sensitivity in recording reference position; and/or providing a warning (such as a visual/audible) upon detection that the relative positions of the references has significantly changed. This may be used to provide an indication that one reference or both have moved and that the registration has been compromised (and so is in need for correction).
(xα,yα,zα)
Where the subscript “α” denotes that the coordinate belongs to RM1. Steps (1310) and (1315) are analogous to the two previous mentioned steps, only instead of applying the steps to RM1 the steps are applied to the second reference marker (“RM2”). Example coordinates for RM2 being
(xβ,yβ,zβ)
Where the subscript “β” denotes that the coordinate belong to RM2. The next step (1320) is to begin registration of an object of interest with a pre-selected number of registration points in fixed and known locations with respect to it. The object of interest is to be overlaid with its virtual counterpart during the surgical procedure. Wherein the registration is completed using the touch-point method depicted in
(rxαn,ryαn,rzαn)=(xα+xrαn,yα+yrαn,zα+zrαn)
or equivalently
(rxαn,ryαn,rzαn)=(Xxα,yα,zα)+(xrαn,yrαn,zrαn))
and relative to RM2 in the common coordinate frame as shown in the example below
(rxβn,ryβn,rzβn)=(xβ+xrβn,yβ+yrβn,zβ+zrβn)
or equivalently
(rxβn,ryβn,rzβn)=(xβ,yβ,zβ)+(xrβn,yrβn,zrβn)
Wherein the equality
(xα+xrαn,yα+yrαn,zα+zrαn)=(xβ+xrβn,yβ+yrβn,zβ+zrβn)
or equivalently
(rxαn,ryαn,rzαn)=(rxβn,ryβn,rzβn)
is inherently satisfied at the time of the execution of the touch-point registration method (shown in
(Rxn,Ryn,Rzn)=[(xα+xrαn,yα+yrαn,zα+zrαn)+(xβ+xrβn,yβ+yrβn,zβ+zrβn)]/2
or equivalently
(Rxn,Ryn,Rzn)=(([(xα+xrαn)+(xβ+xrβn)]/2),
([(yα+yrαn)+(yβ+yrβn)]/2),([(zα+zrαn)+(zβ+zrβn)]/2))
or equivalently
(Rxn,Ryn,Rzn)=[(rxβn,ryβn,rzβn)+(rxαn,ryαn,rzαn)]/2
Where the prefix R denotes the n registration point(s) average coordinates based on their relative location of two reference markers (RM1 and RM2). Step (1340) is to use the calculated point(s) (Rxn,Ryn,Rzn) (1 to n) to register the real object in the common coordinate frame so it can be overlaid with its virtual counterpart. It should be noted that if the coordinates of RM1 and RM2 are constant throughout the procedure the equality
(xα+xrαn,yα+yrαn,zα+zrαn)=(xβ+xrβn,yβyrβn,zβ+zrβn)
will be satisfied, resulting in the following
(Rxn,Ryn,Rzn)=(rxβn,ryβn,rzβn)=(rxαn,ryαn,rzαn)
This implies the relative position of the registration points (1 to n) will remain in an unchanged location (i.e. have constant coordinates) in the common coordinate frame equivalent to the initial touch-point registration coordinates. However if the points RM1 and/or RM2 change then the equality is broken and an averaged position located at the midpoints of the registration point sets relative to both RM1 (i.e. (rxαn,ryαn,rzαn)) and RM2 (i.e. (rxβn,ryβn, rzβn)) are calculated and used to determine the location of the virtual object overlay in the common coordinate frame.
The next three (3) steps in the process involve identifying any potential shifts of the reference marker locations to the point where the registration becomes inaccurate (as defined by a threshold value). Step (1345) indicates that the instantaneous (last updated) registration points relative to RM1 and RM2 must be subtracted and their absolute value calculated and defined as the total deviation and denoted with a prefix “t” as in the following example
(txn,tyn,tzn)=|(rxβn,ryβn,rzβn)−(rxαn,ryαn,rzαn)|
Once calculated step (1350) indicates the total deviation of the instantaneous (last updated) registration points relative to RM1 and RM2 will be compared to a threshold value defined by the user as shown in the following example
If
txn<Txn
or
tyn<Tyn,
or
tzn<Tzn
where the prefix “T” indicates the coordinate threshold values, then the process continues a loop by initiating step (1360) which is to update of the assigned location of RM1 and RM2
(i.e. (xα,yα,zα) and (xβ,yβ,zβ) respectively)
in the common coordinate frame returning followed by returning to step (1335). However
if (txn,tyn,tzn)>(Txn,Tyn,Tzn),
then the process moves to step (1355) and indicates to the surgeon that the registration of the object is inaccurate.
In an embodiment having multiple reference tools, it is possible to infer which reference tool has moved, by determining the one that has shifted position most significantly relative to a camera, and so this reference tool can automatically be dropped from calculation and the procedure can continue without interruption. In the embodiment, once a moved reference position has stabilized, a new position for the reference can be recorded and it can automatically be returned to function as a fixed reference in its new position, again, without interruption.
In use, if at any time during a procedure, a primary reference marker (1010) or instrument is moved and registration is lost (or, in effect considered no longer reliable), one or more secondary reference marker(s) or instruments(s) (that were previously registered can be affixed to the fixture and the secondary registration can be re-established using the secondary marker(s) or instruments(s). The procedure can continue using the secondary reference marker, or the secondary reference marker(s) or instruments(s) can be used as a fixed point to compute an updated registration to the (now) moved position of the primary reference, so that the procedure can continue using the updated primary reference registration. At that point the secondary reference marker(s) can optionally be removed from the fixation device, since registration using the primary reference marker (1010) would then be updated to reflect the moved position of the primary reference marker or instruments(s).
In an embodiment, a secondary reference marker or tool can be a separate tracking tool, or it can be a sterile, tracked surgical hand-piece or pointer, each of which would have been registered to a fixation fixture, which can be (re)attached temporarily to the fixation fixtures to re-establish a lost registration.
In some embodiments, multiple fixation fixtures can be provided and placed at any convenient position, each capable of holding a secondary reference marker(s) and/or having a handpiece or pointer initialized locations. In the embodiment, at any time during the procedure, any one of these fixtures can be used to re-establish or refine a registration by re-attaching the marker(s). In an example, secondary markers can be inclusive of both the tracking reference marker (1010) and the tracking sensor (1140).
In an embodiment, the fixation fixture may be a disposable clip that adheres, straps or screws into place. As noted above,
In an embodiment, a reference position may be provided, which allows a tracking camera to be moved dynamically during a procedure to ensure a good line-of-sight with the tracked instruments. Another method of recovering registration is based on the ability to temporarily lock the tracking camera in a fixed position relative to the patient. An example of an embodiment is shown in
In this embodiment, if the camera position needs adjustment, then the system may be placed into a “reference fixed” position, which establishes a registration relative to the position of the reference, and thereafter allows the camera to be moved to a new position without affecting the registration. The embodiment can also subsequently return to the “camera fixed” mode to again provide independence from the registration marker. An example embodiment of the system in “reference fixed” mode is shown in
However, in typical use during a procedure, after initial registration typically it will be advantageous that neither the tracking system sensor nor the reference marker be moved relative to the patient. A registration that averages a camera-fixed and reference-fixed computed transformation can be used to provide a more accurate registration in some embodiments. An algorithm for averaging a registration based upon two reference markers is depicted in
The system can also permit for “camera fixed” and “reference fixed” modes (or any combination of “tool X fixed” modes) to be enabled at the same time. In an embodiment, the restriction would be that a system state of “camera movable” and “reference movable” cannot be enabled together, or stated more generally, there must always be at least one tool remaining in “fixed mode” to act as a fixed patient reference. In an embodiment, the system can be configured to respond to an automatic signal sent when the camera mount is unlocked for movement, thus switching out of camera locked mode automatically when the camera is being moved and shifting back when the camera position is again locked.
The tracking of movement of a reference marker or tool can be monitored during a procedure, and a warning can be displayed or sounded if the reference has been moved relative to any other fixed reference.
In another embodiment, the use of secondary reference marker(s) and a tracking camera (1140) can be combined to provide a registration validation, where no fixed reference marker (1010) is required. In the embodiment, a tracking tool may be affixed to a fixation fixture and the registration is established relative to the tracking tool prior to moving the tracking sensor. After the tracking sensor is positioned and the registration is returned to be relative to the tracking sensor position, the reference tracking tool can then be removed. In the embodiment, the camera may be considered as a virtual secondary reference tool at the origin of the camera reference frame, which camera tends to serve the same function as the secondary reference marker or tool as described above in other embodiments
At least some of the elements of the systems described herein may be implemented by software, or a combination of software and hardware. Elements of the system that are implemented via software may be written in a high-level procedural language such as object oriented programming or a scripting language. Accordingly, the program code may be written in C, C++, C#, SQL or any other suitable programming language and may comprise modules or classes, as is known to those skilled in object oriented programming. At least some of the elements of the system 10 that are implemented via software may be written in assembly language, machine language or firmware as needed. In either case, the program code can be stored on a storage media or on a computer readable medium that is readable by a general or special purpose programmable computing device having a processor, an operating system and the associated hardware and software that is necessary to implement the functionality of at least one of the embodiments described herein. The program code, when read by the computing device, configures the computing device to operate in a new, specific and predefined manner in order to perform at least one of the methods described herein.
Furthermore, at least some of the methods described herein are capable of being distributed in a computer program product comprising a computer readable medium that bears computer usable instructions for execution by one or more processors, to perform aspects of the methods described. The medium may be provided in various forms such as, but not limited to, one or more diskettes, compact disks, tapes, chips, USB keys, external hard drives, wire-line transmissions, satellite transmissions, internet transmissions or downloads, magnetic and electronic storage media, digital and analog signals, and the like. The computer useable instructions may also be in various forms, including compiled and non-compiled code.
While the applicant's teachings described herein are in conjunction with various embodiments for illustrative purposes, it is not intended that the applicant's teachings be limited to such embodiments. On the contrary, the applicant's teachings described and illustrated herein encompass various alternatives, modifications, and equivalents, without departing from the embodiments, the general scope of which is defined in the appended claims. Except to the extent necessary or inherent in the processes themselves, no particular order to steps or stages of methods or processes described in this disclosure is intended or implied. In many cases the order of process steps may be varied without changing the purpose, effect, or import of the methods described.
This application is a National Phase application claiming the benefit of the international PCT Patent Application No. PCT/CA2014/050266, filed on Mar. 14, 2014, in English, which claims priority to U.S. Provisional Application No. 61/799,735, titled “SYSTEM AND METHOD FOR DYNAMIC VALIDATION AND CORRECTION OF REGISTRATION, AND RECOVERY OF LOST REFERENCE, FOR SURGICAL NAVIGATION” and filed on Mar. 15, 2013, the entire contents of which are incorporated herein by reference. This application claims priority to U.S. Provisional Application No. 61/801,530, titled “SYSTEMS, DEVICES AND METHODS FOR PLANNING, IMAGING, AND GUIDANCE OF MINIMALLY INVASIVE SURGICAL PROCEDURES” and filed on Mar. 15, 2013, the entire contents of which are incorporated herein by reference. This application claims priority to U.S. Provisional Application No. 61/818,280, titled “SYSTEMS, DEVICES AND METHODS FOR PLANNING, IMAGING, AND GUIDANCE OF MINIMALLY INVASIVE SURGICAL PROCEDURES” and filed on May 1, 2013, the entire contents of which are incorporated herein by reference. This application claims priority to U.S. Provisional Application No. 61/800,155, titled “PLANNING, NAVIGATION AND SIMULATION SYSTEMS AND METHODS FOR MINIMALLY INVASIVE THERAPY” and filed on Mar. 15, 2013, the entire contents of which are incorporated herein by reference. This application claims priority to U.S. Provisional Application No. 61/924,993, titled “PLANNING, NAVIGATION AND SIMULATION SYSTEMS AND METHODS FOR MINIMALLY INVASIVE THERAPY” and filed on Jan. 8, 2014, the entire contents of which are incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CA2014/050266 | 3/14/2014 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2014/139019 | 9/18/2014 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5792147 | Evans et al. | Aug 1998 | A |
6135946 | Konen et al. | Oct 2000 | A |
6724922 | Vilsmeier | Apr 2004 | B1 |
6785572 | Yanof | Aug 2004 | B2 |
6919867 | Sauer | Jul 2005 | B2 |
7097357 | Johnson et al. | Aug 2006 | B2 |
7491198 | Kockro | Feb 2009 | B2 |
7892165 | Nakamura | Feb 2011 | B2 |
8010181 | Smith et al. | Aug 2011 | B2 |
8073528 | Zhao et al. | Dec 2011 | B2 |
8211020 | Stetten et al. | Jul 2012 | B2 |
8251893 | Yamamoto et al. | Aug 2012 | B2 |
8398541 | DiMaio et al. | Mar 2013 | B2 |
20010039421 | Heilbrun | Nov 2001 | A1 |
20050215879 | Chuanggui | Sep 2005 | A1 |
20060142657 | Quaid | Jun 2006 | A1 |
20060173356 | Feilkas | Aug 2006 | A1 |
20060293557 | Chuanggui | Dec 2006 | A1 |
20070219639 | Otto | Sep 2007 | A1 |
20070270685 | Kang | Nov 2007 | A1 |
20080123910 | Zhu | May 2008 | A1 |
20080201016 | Finlay | Aug 2008 | A1 |
20080243142 | Gildenberg | Oct 2008 | A1 |
20080262297 | Gilboa et al. | Oct 2008 | A1 |
20110178394 | Fitzpatrick | Jul 2011 | A1 |
20110306986 | Lee | Dec 2011 | A1 |
20110320153 | Lightcap | Dec 2011 | A1 |
20120109150 | Quaid | May 2012 | A1 |
20130094742 | Feilkas | Apr 2013 | A1 |
20130204287 | Mark et al. | Aug 2013 | A1 |
20140171873 | Mark | Jun 2014 | A1 |
Number | Date | Country |
---|---|---|
103083089 | May 2013 | CN |
1708139 | Oct 2006 | EP |
1599148 | Apr 2011 | EP |
2641561 | Sep 2013 | EP |
2472066 | Jan 2011 | GB |
2006149560 | Jun 2006 | JP |
2013031660 | Jul 2012 | JP |
2013252452 | Dec 2013 | JP |
100956762 | May 2010 | KR |
2005074303 | Aug 2005 | WO |
2010140074 | Dec 2010 | WO |
Entry |
---|
Liao, Hognen, et al. “Real-time 3D-image-guided navigation system based on integral videography.” International Symposium on Biomedical Optics. International Society for Optics and Photonics, 2002. |
International Search Report (PCT/CA 2014/050266) dated Jul. 16, 2014. |
Written Opinion (PCT/CA 2014/050266 ) dated Jul. 16, 2014. |
The Silent Loss of Navigation Accuracy; Research-Human-Clinical Studies; vol. 72, No. 5, May 2013, pp. 796-807. |
European Search Report from EP2967297 dated Dec. 19, 2016. |
Number | Date | Country | |
---|---|---|---|
20160000515 A1 | Jan 2016 | US |
Number | Date | Country | |
---|---|---|---|
61799735 | Mar 2013 | US | |
61801530 | Mar 2013 | US | |
61800155 | Mar 2013 | US | |
61818280 | May 2013 | US | |
61924993 | Jan 2014 | US |