The present disclosure relates to acquisition of image data of a subject, and particularly to acquisition and display of image data collected from a tracked imaging system.
This section provides background information related to the present disclosure which is not necessarily prior art.
An imaging system can be used to image various portions of a subject. The subject can include a patient, such as a human patient. The portions selected to be imaged can be internal portions that are covered by skin or other tissue. A location of imaged portions of the subject may be selected to be known within the imaging system. The locations can be defined or established relative to instruments placed in the subject (e.g., a location of a heart wall relative to a catheter) or a location of the imaged portion relative to the instrument acquiring the image data.
An imaging modality may be used to generate image data that are used to render images of an interior of the subject. For example, x-rays may be used to generate image data of an interior of a subject. The images may be displayed for viewing by a user.
This section provides a general summary of the disclosure, and is not a comprehensive disclosure of its full scope or all of its features.
The subject disclosure relates to illustrating a position, including a pose of an instrument relative to a patient. The patient may be a human or non-human patient. Further, the patient or subject may be a non-living object where the location of an instrument within the subject is selected to be tracked.
A representation of the patient may be fit or morphed to illustrate or represent a size of the patient. Therefore, relative portions of the patient, such as relative portions of the anatomy, are determined or known relative to one another based upon the patient that is the current patient during a procedure. An avatar or indicia of the patient may be generated or altered to a specific patient substantially in real time during a procedure.
An instrument can be tracked relative to the patient. A representation of the instrument may be superimposed on the avatar of the patient. The avatar of the patient may provide a visual illustration of various portions of the anatomy of the patient without requiring image data being acquired of the patient. The display being viewed by the user may be generated without requiring image data of the patient, such as fluoroscopy image data.
Further areas of applicability will become apparent from the description provided herein. The description and specific examples in this summary are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.
The drawings described herein are for illustrative purposes only of selected embodiments, not all possible implementations, and are not intended to limit the scope of the present disclosure.
Corresponding reference numerals indicate corresponding parts throughout the several views of the drawings.
Example embodiments will now be described more fully with reference to the accompanying drawings.
The navigation system 10 can interface with or integrally include an imaging system 12 that is used to acquire pre-operative, intra-operative, or post-operative, or real-time image data of the patient 14. For example, the imaging system 12 can be an ultrasound imaging system (as discussed further herein) that has a tracking device 22 attached thereto. The imaging system 12 may also include any suitable CT system, x-ray system, C-arm, O-arm® imaging system, magnetic resonance imaging system, etc. The tracking device 22 may be tracked with the tracking system to determine a pose of the imaging system 12. The pose may include an orientation (e.g., three or more degrees of orientation (e.g., yaw, pitch, and roll) and/or a position (e.g., three degrees of freedom in physical space (e.g., x-axis, y-axis, and z-axis). The tracking system 10 may further determine appropriate pose information regarding the tracking device 22. The pose of the imaging system 12 can then be determined based on the tracked pose of the tracking device 22. The imaging system 12 may be used to generate image data to provide images for viewing with a selected display device 26, which may any appropriate display device including an augmented and/or virtual reality display device worn/used by the user 18 such as those disclosed in U.S. Pat. App. Pub. No. US2018/0078316A1, incorporated herein by reference.
It will be understood any appropriate subject can be imaged and any appropriate procedure may be performed relative to the subject. The navigation system 10 can be used to track various tracking devices, as discussed herein, to determine locations of the patient 14. The tracked poses of the patient 14 can be used to determine or select images for display to be used with the navigation system 10. The initial discussion, however, is directed to the navigation system 10 and the exemplary imaging system 12.
In the example shown, the imaging system 12 includes an ultrasound (US) imaging system with an US housing 16 that is held by a user 18 while collecting image data of the subject 14. It will be understood, however, that the US housing 16 can also be held by a stand or robotic system while collecting image data. The US housing 16 and included transducer can be any appropriate US imaging system 12, such as the M-TURBO® sold by SonoSite, Inc. having a place of business at Bothell, Washington. Associated with, such as attached directly to or molded into, the US housing 16 or the US transducer housed within the housing 16 is at least one imaging system tracking device 22. The tracking device 22 may be any appropriate tracking device such as an electromagnetic (EM) tracking device and/or an optical tracking device. The tracking devices can be used together (e.g., to provide redundant tracking information) or separately. Also, only one of the two tracking devices may be present. It will also be understood that various other tracking devices can be associated with the US housing 16, as discussed herein, including acoustic, ultrasound, radar, electrical impedance, and other tracking devices. Also, the tracking device 22 can include linkages or a robotic portion that can determine a location relative to a reference frame. Any suitable tracking device of any suitable tracking system may be used, such as, tracking devices and systems as understood by one skilled in the art.
In various embodiments, a secondary imaging system may also be presented and/or used to generated image data of the patient 14. The second imaging system may include an O-arm® imaging device sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado, USA, a C-arm imaging system, or other appropriate imaging system. The second imaging system can include those disclosed in U.S. Pat. Nos. 7,188,998; 7,108,421; 7,106,825; 7,001,045; and 6,940,941; all of which are incorporated herein by reference. The second imaging system, however, need not be present and the imaging system 12, which may be only the US transducer, may be the only imaging system to generate images to be displayed. As discussed herein, various other image data may also be displayed such as an avatar 120 of the patient 14.
The imaging system 12 may be tracked, as discussed above. Various tracking systems may include and/or require calibration of the imaging system. Thus, the pose of the tracking device 22 relative to a plane of the US imaging system may be determined and/or known. Various systems and methods are disclosed in U.S. Pat. Nos. 6,379,302; 6,669,635; 6,968,224; 7,085,400; 7,831,082; 8,320,653; 8,811,662; and 9,138,204 all of which are incorporated herein by reference.
As illustrated in
The navigation system 10 includes at least one tracking system. The tracking system can include at least one localizer. In one example, the tracking system can include an electromagnetic (EM) localizer 50. The tracking system can be used to track instruments relative to the patient 14 or within a navigation space. The navigation system 10 can use image data from the imaging system 12 and information from the tracking system to illustrate locations of the tracked instruments, as discussed herein. The tracking system can also include a plurality of types of tracking systems including an optical localizer 52 in addition to and/or in place of the EM localizer 50. When the EM localizer 50 is used, the EM localizer can communicate with or through a localizer communication 54 that may be wired or wireless. The EM localizer 50 may also include and/or take the form of an alternative pad or flat EM localizer 55.
The optical localizer 52 and the EM localizer 50 can be used together to track multiple instruments or used together to redundantly track the same instrument. Various tracking devices, including those discussed further herein, can be tracked and the information can be used by the navigation system 10 to allow for an output system to output, such as a display device to display, a position of an item. Briefly, tracking devices can include a patient or reference tracking device 56, also known as a dynamic reference frame, to track the patient 14, an instrument tracking device 60 to track an instrument 62, and/or other appropriate tracking devices for one or more portions. Patient or reference tracking device 56 and instrument tracking device 60 may include those disclosed in U.S. Pat. Nos. 8,060,185 and 8,644,907, both incorporated herein by reference. The tracking devices allow selected portions of the operating theater to be tracked relative to one another with the appropriate tracking system, including the optical localizer 52 and/or the EM localizer 50. The reference tracking device 56 can also or alternatively be positioned on an instrument and positioned within the patient 14, such as within a heart 15 of the patient 14.
It will be understood that any of the tracking devices 22, 56, 60 can be optical or EM tracking devices, or both, depending upon the tracking localizer used to track the respective tracking devices. It will be further understood that any appropriate tracking system can be used with the navigation system 10. Alternative tracking systems can include radar tracking systems, acoustic tracking systems, ultrasound tracking systems, electrical impedance tracking systems, radio frequency beacon, and the like. Exemplary tracking systems include those disclosed in U.S. Pat. Nos. 7,676,268; 8,532,734; and 8,494,608, all incorporated herein by reference. Each of the different tracking systems can be respective different tracking devices and localizers operable with the respective tracking modalities. Also, the different tracking modalities can be used simultaneously as long as they do not interfere with each other (e.g., an opaque member blocks a camera view of the optical localizer 52).
An exemplary EM tracking system can include the STEALTHSTATION® AXIEM™ Navigation System, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado. Exemplary tracking systems are also disclosed in U.S. U.S. Pat. No. 7,751,865, issued Jul. 6, 2010 and entitled “METHOD AND APPARATUS FOR SURGICAL NAVIGATION”; U.S. Pat. No. 5,913,820, titled “Position Location System,” issued Jun. 22, 1999 and U.S. Pat. No. 5,592,939, titled “Method and System for Navigating a Catheter Probe,” issued Jan. 14, 1997, all herein incorporated by reference.
Further, for EM tracking systems it may be necessary to provide shielding or distortion compensation systems to shield or compensate for distortions in the EM field generated by the EM localizer 50. Exemplary shielding systems include those in U.S. Pat. No. 7,797,032, issued on Sep. 14, 2010 and U.S. Pat. No. 6,747,539, issued on Jun. 8, 2004; distortion compensation systems can include those disclosed in U.S. patent Ser. No. 10/649,214, filed on Jan. 9, 2004, published as U.S. Pat. App. Pub. No. 2004/0116803, all of which are incorporated herein by reference.
With an EM tracking system, the EM localizer 50 and the various tracking devices can communicate through an EM controller. The EM controller can include various amplifiers, filters, electrical isolation, and other systems. The EM controller can also control the coils of the EM localizer 50 to either emit or receive an EM field for tracking. A wireless communications channel, however, such as that disclosed in U.S. Pat. No. 6,474,341, entitled “Surgical Communication Power System,” issued Nov. 5, 2002, herein incorporated by reference, can be used as opposed to being coupled directly to the EM controller. The EM controller can be incorporated into a navigation processing system 70.
It will be understood that the tracking system may also be or include any appropriate tracking system, including a STEALTHSTATION® TRIA®, TREON®, and/or S7™ Navigation System having an optical localizer, similar to the optical localizer 52, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado. Further, alternative tracking systems are disclosed in U.S. Pat. No. 5,983,126, to Wittkampf et al. titled “Catheter Location System and Method,” issued Nov. 9, 1999, which is hereby incorporated by reference. Other tracking systems include an acoustic, radiation, radar, electrical impedance, etc. tracking or navigation systems.
The navigation system 10 can include a navigation processing unit or module 74 that can communicate or include a navigation memory 76, which may be included in the navigation processing system 70. The navigation processing system 70 may further include a display device 77. The navigation processing unit 74 can include a processor and memory (e.g., microprocessor, a central processing unit, etc.). In various embodiments, the navigation processing unit may execute instructions to determine one or more poses of the tracking devices based on signals from the tracking devices. The navigation processing unit 74 can receive information, including image data, from the imaging system 12 and tracking information from the tracking systems, including the respective tracking devices and/or the localizers 50, 54. Image data can be displayed as an image 78 on the display device 26. The display device may be separate from and/or integrated into the navigation system, thus, the display device 26 may include the display device 77. The navigation system 70 can include appropriate input devices, such as a keyboard 84. It will be understood that other appropriate input devices can be included, such as a mouse, a foot pedal 88 or the like which can be used separately or in combination. Also, all of the disclosed processing units or systems can be a single processor module (e.g., a single central processing chip) that can execute different instructions to perform different tasks.
An image processing unit or module can process image data from the imaging system 12 and a separate first image processor (not illustrated) can be provided to process or pre-process image data from the imaging system 12. The image data from the image processor can then be transmitted to the navigation processor 74. It will be understood, however, that the imaging systems need not perform any image processing and the image data can be transmitted directly to the navigation processing unit 74. Accordingly, the navigation system 10 may include or operate with a single or multiple processing centers or units that can access single or multiple memory systems based upon system design.
In various embodiments, the imaging system 12 can generate image data that may be used to compose the image 78 and define an image space that can be registered to a patient space or navigation space that is defined by and/or relative to the patient 14. In various embodiments, the position of the patient 14 relative to the imaging system 12 can be determined by the navigation system 10 with the patient tracking device 56 and the imaging system tracking device(s) 22 to assist in and/or maintain registration. Accordingly, the position of the patient 14 relative to the imaging system 12 can be determined.
Manual or automatic registration of the image space to the subject space can occur. In various embodiments, the registration can occur by matching fiducial points in image data with fiducial points on the patient 14. Registration of image space to patient space allows for the generation of a translation map between the patient space and the image space. According to various embodiments, registration can occur by determining points that are substantially identical in the image space and the patient space. The identical points can include anatomical fiducial points or implanted fiducial points. Exemplary registration techniques are disclosed in U.S. patent application Ser. No. 12/400,273, filed on Mar. 9, 2009, now published as U.S. Pat. App. Pub. No. 2010/0228117 and in U.S. Pat. No. 9,737,235, issued Aug. 22, 2017, both incorporated herein by reference.
According to various embodiments, the imaging system 12 can be used with an un-navigated or navigated procedure. In a navigated procedure, a localizer and/or digitizer, including either or both of an optical localizer 52 and/or an electromagnetic localizer 50, 55 can be used to generate a field and/or receive and/or send a signal within a navigation domain relative to the patient 14. The navigated space or navigational domain relative to the patient 14 can be registered to the image 78. Correlation, as understood in the art, is to allow registration of a navigation space defined within the navigational domain and an image space defined by the image 78. The patient tracker or dynamic reference frame 56 can be connected to the patient 14 to allow for a dynamic registration and maintenance of registration of the patient 14 to the image 78.
Once registered, the navigation system 10 with or including the imaging system 12, can be used to perform selected procedures. Selected procedures can use the image data generated or acquired with the imaging system 12. Further, the imaging system 12 can be used to acquire image data at different times relative to a procedure. As discussed herein, image data can be acquired of the patient 14 prior to the procedure for collection of automatically registered image data or cine loop image data. Also, the imaging system 12 can be used to acquire images for confirmation of a portion of the procedure. Thus, image data may be acquired at any appropriate time and may be registered to the patient 14.
Registration and navigated procedures are discussed in U.S. Pat. No. 8,238,631, incorporated herein by reference. Upon registration and tracking of the instrument 62, a graphic representation 90 (e.g., an icon, indicium, animation or other or visual representation) may be displayed relative to, including overlaid (e.g., superimposed) on, the image 78. The image 78 may be an appropriate image and may include one or more 2D images, such as 2D images that are acquired at different planes. Images may also be a 3D image, or any appropriate image as discussed herein.
In addition to registering the subject space to the image space, however, the imaging plane of the US imaging system 12 can also be determined. By registering the image plane of the US imaging system 12, imaged portions can be located within the patient 14. For example, when the image plane is calibrated to the tracking device(s) 22 associated with the US housing 16 then a position of an imaged portion of the heart 15, or other imaged portion, can also be tracked as disclosed in U.S. Pat. Nos. 6,379,302; 6,669,635; 6,968,224; 7,085,400; 7,831,082; 8,320,653; 8,811,662; and 9,138,204 all of which are incorporated herein by reference.
Once the patient 14 is in condition for a procedure, the patient 14 may be positioned on the table 40, as noted above. The display 26 and/or the display 77 may display various information regarding the patient 14 and/or other information selected by the user 18. As noted above, the display 26 may illustrate the image 78 that may be acquired with the imaging system 12. The image 78 may also be displayed on the display 77. Also as noted above, the instrument 62 may be tracked and the graphic representation 90 of the instrument 62 may be displayed on the display 26, such as relative to the image 78. In addition and/or alternatively, the graphic representation 90 may also be displayed relative to a patient avatar 120.
The patient avatar 120 may be illustrated on a selected display, as illustrated in
The patient avatar 120 may generally illustrate a portion of the patient 14 and may be sized relative to the patient 14. For example, the patient avatar may illustrate an inferior portion of a neck 122, a pectoral region 124, and other appropriate regions of the patient 14. The patient avatar 120 may be displayed on the display device 26 to illustrate a general representation of the patient 14 without requiring an image acquisition of the same portion of the patient 14. Therefore, the patient avatar 120 may provide a general roadmap or indication of the position of various tracked portions, as discussed further herein. The avatar 120 may however, as discussed herein, be morphed to match a relative size of the patient 14 to provide an indication of relative anatomical portions of the patient 14 and a pose of the tracked portions, such as the instrument 62, relative thereto.
With continuing reference to
In various embodiments, for example, the user 18 may provide or include information regarding the patient 14. As noted above, the navigation processing system 70 may include various inputs that allow the user 18 to input information. Therefore, as illustrated in
In addition or alternatively thereto, the user 18 may also identify or size the avatar 120 to the patient 14 with a tracked indicator or tracked member 130. For example, the tracked indicator 130, which may be an indicator or probe usable to touch a surface of the patient 14, may be moved by the user 18 to various predetermined positions on the patient 14 as indicated on the avatar 120. The tracked indicator 130 may be tracked with any appropriate tracking system, including those discussed herein according to various embodiments such as the localizer system of U.S. Pat. No. 5,983,126, hereby incorporated by reference. The user 18 may move the tracked indicator 130 to various points, such as the exterior shoulder portions 132′, 134′ as illustrated by the icons 132 and 134. Additionally, the user 18 may move the tracked indicator 130 to other portions such as a suprasternal notch 136 and a sternum 140 of the patient 14 or outer boundary positions. It is understood that the tracked indicator 130 may be moved to any appropriate position and the above-noted positions are merely exemplary. Nevertheless, based upon the tracked position of the tracked indicator 130, the avatar 120 may be sized to morph the identified points to a representation of the patient 14 for display on the display device 26. Also, the points on the avatar 120 may be registered to the patient 14 by identifying them in the patient space relative to the patient 14 and identifying them on the avatar 120.
The avatar 120 may also be morphed to actual measurements taken of the patient 14. For example, a distance between the shoulder points 132′, 134′ on that the patient 14 may be measured, such as with a tape measure or other measuring device. The measured distance may be input within the navigation processing system at 77. The avatar 120 may be sized based upon the measurements.
Regardless of the technique, the avatar 120 may be matched or morphed to substantially mimic the shape of the patient 14. The avatar 120, as discussed further herein, may then be used as an illustration of a pose of the tracked instrument, such as the instrument 62, during a procedure on the patient 14. Thus, the avatar 120 may provide a representation of a position of the tracked instrument 62 without requiring an image acquisition of the patient 14, such as with a fluoroscopic imaging system. The avatar 120, as it is morphed to the specific patient may or will be displayed as a patient specific avatar. Thus, the avatar 120 may include general information, but is displayed and registered to the patient 14 as a patient specific avatar 120 including relevant locations and relative positions of various portions of the patient 14. The specific patient avatar illustrates the relative positions of the anatomy of the patient 14 that is the current patient. For example, the position of the heart 15 relative to the suprasternal notch.
The avatar 120 may be registered to the patient 14 in any appropriate manner, including those discussed above. In various embodiments, for example, when the tracked indicator 130 is tracked to determine positions of the patient 14 for marking or changing the avatar 120. The patient 14 after the various points are determined in both the patient 14 and the avatar 120 may be registered thereto. Additionally, and/or alternatively, the patient tracker 56 may be positioned on the patient 14 at an appropriate position. The position of the patient tracker 56 may be at the suprasternal notch 136′ which relates to a point on the avatar 120, such as the suprasternal notch 136. The position of the suprasternal notch 136 in the avatar 120 may be determined or known to be the position of the patient tracker 56 to allow for registration of the avatar 120 relative to the patient 14. Also, the patient tracker 56 may maintain registration even when the patient 14 moves relative to the localizer(s). It is understood that any appropriate number of the patient trackers 56 may be utilized, including each may be used for registration and/or more than one of the patient trackers 56 may be used.
The localizers, including the EM localizer 58, may also or alternatively be fixed or positioned relative to the patient 14 in an appropriate manner. For example, the localizer 58 may be fixed relative to the table 14 such that it is immovable during the procedure. Thus, the field generated with the EM localizer 58 can be at a known position relative to the patient 14 and allow for a registration of the avatar 120 to the patient 14 based upon the known position of the EM localizer 58 relative to the patient 14. The localizer 55 may also be positioned under the patient 14, such as between the patient 14 and the bed 40. Localizer 55 may include field generating portions, such as coils, that generate a field relative to the patient 14. The flat or panel localizer 55 may be positioned at a known position relative to the patient 14, such as aligning the shoulder edges 132′, 134′ to known or predetermined portions of the panel localizer 55. As the panel localizer 55 is at a known position relative to the patient 14, and the avatar 120 is morphed to the patient and 14, the avatar 120 may be registered to the patient 14 due to the position of the panel localizer 55.
Regardless, the avatar 120 may be registered to the patient 14. Registration of the avatar 120 relative to the patient 14 allows the avatar 120 to be used to assist navigation and guiding of the instrument 62, or any appropriate instrument relative to the patient 14. As the tracked position of the instrument 62 is determined with the navigation system 10, the pose of the instrument 62 may be illustrated with a graphical representation, such as the graphic representation 90 relative to the image 78 and/or the avatar 120.
As illustrated in
In various embodiments, for example, image data may be acquired of the patient 14. As noted above, the imaging system 12 may acquire image data of the patient 14. The imaging system 12 may generally be a non-ionizing imaging system, such as an ultrasound imaging system. The ultrasound image may generally be collected at a single plane relative to the housing 16. The US imaging system 12 with the tracker 22 may be tracked relative to the patient 14 so that the image may also be registered to the patient 14. Thus, the image 78 may also be displayed superimposed on the avatar 120, such as the image superimposition 78i. The image may also be displayed separately from the avatar, as also illustrated in
As the image 78 may be registered to the patient 14, a single image acquisition may occur and the image 78 may be displayed for the procedure. Further, as the image 78 is registered to the patient the graphic representation 90 of the instrument 62 may be displayed at the tracked pose on the image 78. Thus, the US imaging system 12 may be used at one point in time to acquire the image data, but it may be used to illustrate the selected portion of the patient 14, such as the heart, for an entire procedure. A constant imaging of the patient 14 is not required to illustrate or know a position of the instrument 62 as the instrument 62 is tracked and the image 78 is registered to the patient 14 due to the US tracker 22.
In a procedure, once the avatar 120 and/or the image 78 are registered to the patient, the user 18 may perform the procedure on the patient 14 with the tracked instrument(s). The procedure may be any appropriate procedure, such as placing a pacing lead and/or a catheter to assist in placing the lead. Other appropriate procedures may also be performed.
In various embodiments, the user 18 may form an incision, such as an incision 150 on the patient 14, and move the instrument 62 into the patient 14 through the incision 150. The tracked portion 60 on the instrument 62 may be tracked so that a position of the instrument 62 may be displayed relative to the avatar 120 and/or the image 78, as noted above. The instrument 62 may then be moved relative to the patient 14. In various embodiments, for example, the instrument 62 may be a catheter and/or lead for an implant, or a stylet or a delivery system for an implant such as a leadless pacemaker or a stent. As is understood by one skilled in the art, an implant, including an electrode, may be positioned relative to one or more portions of the heart 15 of the patient 14.
With continuing reference to
The avatar 120 may also assist the user 18 in identifying or determining an orientation of the instrument 62 as being in a selected or in a not-selected pose relative to the patient 14. For example, as illustrated in
Thus, the user 18 may view and understand the position of the instrument 62 during the placement of an implant, which may include, for example, a stimulation lead for a selected system. The lead may include a pacing electrode be for a cardiac stimulation system, such as the AZURE®, ADAPTA®, or other appropriate heart pacemaker or defibrillation systems sold by Medtronic, Inc. having a place for business in Minnesota, USA.
Turning reference to
Information regarding the patient, such as the specific patient, may be received in Block 218. The information regarding the patient may be any appropriate information, including that discussed above. Therefore, the information may include general dimensions of the current patient, including a height, weight, sex, or the like. Information received in Block 218 may also include specific tracked points of predetermined portions (e.g., shoulder edges) of the patient 14. Further the received information may include measurements of the patient. Regardless, the received information of the patient may be patient specific information regarding the patient for a current or real-time procedure.
A patient specific avatar is then generated in Block 230. The generation of a patient specific avatar may include morphing or fitting the standard avatar based upon the received information. In morphing the standard avatar, the method may include altering or scaling dimensions of the standard avatar based upon a ratio given the information received regarding the current patient. The ratio may be based upon the known dimensions of the current patient based upon known or average dimensions of various anatomical portions relative to one another based upon the dimensions of the specific patient. This may also include having predetermined average positions and relative positions of various anatomical features. In addition and/or alternatively thereto, the tracked position of the instrument may identify specific anatomical locations. The standard avatar may then be morphed based upon the tracked specific locations for the specific patient. Thus, the avatar may be morphed based upon selected instructions including an algorithm to generate the patient specific avatar in Block 230.
After generating the patient specific avatar in Block 230, the patient specific avatar may be displayed in Block 234. As discussed above, the display 26 may display the patient specific avatar for selected procedures. The display of the patient specific avatar may allow the user 18 to identify and understand relative positions or locations of various portions of the current patient 14.
An instrument may then be tracked, optionally and according to various embodiments, in Block 238. Although the tracking of the instrument is an optional part of the method 200, such tracking may allow for information regarding a pose of an instrument to be determined. Regardless of whether the instrument is tracked, the method 200 may include receiving tracking information in Block 240.
Receiving tracking information in Block 240 may include tracking one or more tracking devices to determine a pose of various elements or portions, such as the pose of the instrument 62 via the instrument tracker 60 and/or the patient 14 via the patient tracker 56. The patient specific avatar may be registered to the patient based upon the known location of one or more specific portion(s) of the patient, such as the positioning of the patient tracker 56 in the suprasternal notch of the patient 14. The suprasternal notch of the avatar 120 may then be registered for determination of the pose of the instrument 62 relative to the patient tracker 56. Therefore, the pose of the instrument 62 may be displayed on the display device 26 such as the superimposed graphic representation 90 on the avatar 120.
Thus, the pose of the instrument may be superimposed on the avatar 120 in Block 244. Again, this allows a user 18 to view and understand a pose of the tracked instrument 62 on a representation of the patient 14 without requiring image data to be acquired to the patient 14. While image data of the patient 14 may be acquired, such as with the US imaging system 12, such image information is not required. The image data information from the US imaging system 12 may assist in finalizing or determining a specific location of the instrument 62 relative to a selected portion, such as a specific portion of the heart 15 of the patient 14. According, only a selected portion or area of the patient may be imaged. However, the position of the instrument 62 may be determined without the image data. Thus, the user may understand a position of the instrument 62, including a movement of the instrument 62 toward a selected position, such as the RV septum of the patient 14.
The method 200 may then end in Block 250. And ending the method 250 it is understood that the method 200 may be repeated as often as necessary or selected to allow for up to date and/or real-time tracking and illustration of a pose of the instrument 62 relative to the avatar 120. Thus, the method 200 may be repeated at a selected rate and for a selected period, such as selected by the user 18.
As noted above, moving the instrument 62 relative to the patient 14, such as relative to the heart 15, may be accomplished while tracking a pose of the instrument 62. As noted above, tracking a pose of the instrument 62 can involve attempting to determine a physical location and/or orientation of the instrument 62 in the physical space. The tracked posed may include a location which includes X, Y, and Z axis positions and orientation which may include yaw, pitch, and roll orientations in space. The physical space may also be referred to as navigation space and is generally defined relative to the patient 14 and/or the heart 15. The patient tracker 56 allows the pose of the patient 14, including the heart 15, to be tracked. Tracking a pose relative to a portion of the patient 14, including the heart 15, may include systems such as those disclosed in U.S. Pat. No. 7,697,972 issued Apr. 13, 2010; U.S. Pat. No. 8,046,052 issued Oct. 25, 2011; U.S. Pat. No. 8,180,428 issued May 15, 2012; and U.S. Pat. No. 8,663,120 issued Mar. 4, 2014, all of which are herein incorporated by reference.
With continuing reference to
With reference to
Multiple images similar to the image 78, e.g., multiple similar images from a population of patients other than the patient 14, with similar identified portion(s) or location(s) may be combined to build the atlas 78m to illustrate the identified portion(s) or location(s) on a model of the imaged anatomy or heart, which model represents a typical or average patient anatomy. Alternatively, the atlas 78m may be built only from images of the anatomy of the patient 14, and the portion(s) or location(s) of interest applied to the atlas based on identification within the images of the patient's own anatomy, and/or based on a compilation of images from other patients' anatomy and an estimated and/or average position of the portion(s) or location(s) of interest within the compilation of images. The atlas 78m may also be morphed to actual measurements taken of the patient 14.
Based upon the user's 18 identification in the image 78 of the various portions/locations, which may include all of or selected number of portions/locations, a processing module or unit, such as the processor 74, may execute selected instructions to register or identify various portions in the image 78 and/or on the atlas 78m for viewing by the user 18. For example, in the atlas 78m the right ventricle septum may be identified as a selected point including a target 330. The image 78 may also, therefore, have the target identified thereon 330x to illustrate a proposed or selected position of the right ventricle septum (RV septum), e.g., to serve as a target for delivery of a catheter, stylet and/or pacing lead, or a leadless pacemaker. The predetermined portions in the model 78m, therefore, may be used to assist in identifying portions of the image 78 based on one or more portions identified by the user 18 and/or appropriate image identification systems. Therefore, the model 78m may be used to assist in identifying portions within the image 78 to assist in the procedure. The user may view the image 78 and identify or evaluate the target 330x prior to and/or during the procedure.
The image 78, as discussed above, may also include a plurality of images that are generated with the imaging system 12. The US imaging system 12 may acquire images of the patient 14, including the heart 15, in a plane. As the US housing 16 is moved, the plane of the image relative to the patient 14 may also change. Thus, the image 78 may include a plurality of images that are illustrated in one or more planes. For example, with reference to
In addition to providing more than one view of the heart 15 of the patient 14, the two images 78ap and 78ml may also assist in registration to the model 78m. Thus, the identification of anatomy in the images 78 and/or one or both of the images and 78ap and 78ml, may be assisted by the registration of more than one image relative to the model or atlas 78m. As discussed above, the ultrasound imaging system 12 is registered to the patient 14, thus the images acquired with of the imaging system 12 may also be registered to the patient 14. The registration of the images to the patient may be based upon the tracking of the imaging system 12, such as with the imaging system tracking device 22 and the tracking of the patient 14, such as with the patient tracking device 56. Once the image 78 is registered to the patient 14, the instrument 62 may be tracked relative to the patient 14 and its pose may be displayed relative to the image 78. As discussed above, the pose of the instrument 62 may also be displayed relative to the avatar 120. The graphic representation 90 allows the user 18 to view the pose of the instrument 62 relative to both of the image 78 and the avatar 120.
Further, the instrument 62 may be tracked in real time. Therefore, the instrument may be displayed relative to the avatar 120 and/or the image 78 to provide a real time graphical representation of the instrument 62 relative to the heart 15, based upon a graphical representation displayed relative to the image 78. In addition to viewing the tracked pose of the instrument 62, the graphical representation may show the pose of the instrument 62 over time. Therefore, the graphic representation 90 may illustrate or provide a sense of an entire path of the instrument 62 as it moves from the incision 150 in the patient 14 to the current location within the patient 14. Further, the graphic representation 90 may provide a real-time feedback to the user 18 of a movement of the instrument 62.
In various embodiments, the images 78 may be acquired or based upon image data acquired at a selected time. For example, prior to the incision 150, image data may be acquired with the imaging system 12 to generate the image 78. Image 78 may thus be registered to the patient 14 and displayed with a display device 26 either alone and/or superimposed on the avatar 120. Nevertheless, the image 78 may not be a real time or current image and may be a static image. The instrument 62, however, may be tracked in real time and provide a current determination of the pose of the instrument 62. The graphic representation 90 may, therefore, be updated in real time and at a selected rate to illustrate a current position of the instrument 62 and/or a current shape or condition of the patient anatomy. The display of the graphical representation may also be updated in a selected rate to provide a perception of motion of the instrument 62. For example, a refresh of the pose of the instrument 62 and of the graphic representation 90 may be 10 times a second, 30 times a second, or any appropriate rate such as about one time a second to about 120 times per second. Thus, the pose of the instrument 62 may be illustrated to show a motion within the display device 26.
With reference to
The graphic representation 90t2 may be displayed or have a selected position that is displaced a certain distance of 350 relative to the first graphic representation 90t1. The distance 350 may be any appropriate distance and may be based upon the various physical features, such as movement of a beating heart. The heart 15 of the patient 14 may beat (i.e., move due to the heartbeat), as is understood by one skilled in the art. Therefore, the graphic representation 90t2, including a terminal end 90x, may be shown in different positions based upon movements of the instrument 62. For example, as one skilled in the art will understand, when the terminal end of the instrument 62 contacts the RV septum 316, the terminal end of the instrument 62 may move due to the beating of the heart 15. Therefore, even if the user 18 is not moving the instrument 62, the instrument 62 may appear to move in real time. Real-time tracking of the instrument 62 with the tracking device and 60 may allow for a representation of the movement of the instrument 62 to be displayed on the display device 26. As discussed above, the movement of the instrument may be updated and displayed at any appropriate rate to provide a perception of movement to the user 18. Thus, the user 18 may be provided with additional information regarding a position of the instrument 62, such as being in contact with of the RV septum 316.
With reference to
After determining a pose of the instrument at a first time, an output for display of the determined pose of the instrument may be made in Block 382. The output of the determined pose may be based upon the registration of the image 78 and/or the avatar 120 to the physical space defined by the patient 14, as discussed above. Thus, the output may display the graphic representation 90 at a first position, for example of the first pose 90t1, as illustrated in
After determining and outputting the first pose, a wait or pause for a selected time may occur in Block 384. The wait for a selected time may be any appropriate time, such as about one second to about 120 seconds. The wait time may be any appropriate time to assist in illustrating a motion to the user 18, as discussed above. After the wait time, the second pose may be determined in Block 388. The second pose may be determined in the same manner as the first pose was determined in Block 384. An output for display of the second pose may be made in Block 392 to display the graphical representation of the instrument 62 at a second time, such as the graphic representation 90t2. As noted above, therefore, the pose of the instrument 62 may be displayed over a series of the events and/or time to show movement to the user 18.
A determination can be made to continue tracking at Block 394. If tracking is to be continued, a YES path 398 may be followed to again determine a first or subsequent pose in Block 378. The process 370 may provide an illustration to the user 18 for an appropriate period of time. If a tracking is not to be continued, a NO path 400 may be followed to end the process in Block 404. The process 370 may be used to illustrate and determine a pose of the instrument 62 over time for illustration relative to the image 78, as illustrated in
The display 26 may display the image of the patient 14 that is acquired with the imaging system 12. As discussed above, the image 78 may be an image that is acquired in a plane from the ultrasound housing 16. Therefore, rotation of the image may change a perspective of the image 78. With reference to
Further, the distance of the instrument 62 that is within the plane of the image may also be illustrated. For example, a first and second distance icon 410, 412 may be included or illustrated superimposed on both of the images 78e, 78f. The distance icons may be used to illustrate the distance that the instrument 62 is in the plane of the image. Thus, the user 18 may also understand that the portion of the instrument that is within the plane of the image and that other portions are not within the plane of the image.
The distance icon 410, 412 may be sized depending upon the length or portion of the instrument 62 that is within the plane of the image 78. Therefore, the first distance icon 410 may have a first distance or size 420. The second distance icon 412 may have a second distance or size 424. The two dimensions 420, 424 may be different or the same depending upon the amount of the instrument or length of the instrument that is within the plane of the image 78. If the instrument has a greater length within the plane, the distance icons 410, 412 may have greater dimensions. If any dimension of the instrument is smaller in the plane, then the dimension icons may be smaller. Regardless, the dimension icons 410, 412 may be displayed, such as superimposed on, the image 78 to represent the distance of the instrument within the image plane. This may allow the user 18 to understand or determine the distance of the instrument 62 within a plane of the imaging system and/or assist in moving of the imaging system to acquire a different perspective than in the image 78 relative to the instrument 62.
According to various embodiments, the display 26 may display the image of the patient 14 that is acquired with the imaging system 12 at more than one location. In other words, more than one image may be displayed for each position of the imaging system 12 at which image data is acquired. Moreover, one or more image data portions may be merged and/or used in a reconstruction of a displayed image. As discussed above and herein, all of the images and/or the reconstructed image may be registered to the patient 14 due at least to the tracking of the imaging system 12.
With reference to
Again, the edge image may allow for a viewing and understanding of a perspective of the instrument 62, as illustrated by the graphic representation 90, relative to a plane of the image. In various embodiments, the image acquired with the imaging system 12 may be a static image that is acquired with the imaging system 12. It is understood, however, that a live or current image may also be acquired and displayed. Nevertheless, the position of the instrument 62 relative to the plane of the image may be illustrated by rotating the image, such as showing it from an edge as illustrated in the edge image.
The instrument 62 may be tracked with any appropriate tracking system. As noted above, tracking and/or navigation systems include those disclosed above and those disclosed in U.S. Pat. No. 7,697,972 issued Apr. 13, 2010; U.S. Pat. No. 8,046,052 issued Oct. 25, 2011; U.S. Pat. No. 8,180,428 issued May 15, 2012; and U.S. Pat. No. 8,663,120 issued Mar. 4, 2014, all of which are herein incorporated by reference. The navigation system may allow for a determination of a pose of the instrument 62 in a physical location and/or orientation of the instrument 62 in the physical space. The tracked pose may include a location which includes X, Y, and Z axis positions and/or orientation which may include yaw, pitch, and roll orientations in space. A tracking and navigation system may also include one or more permanent magnets and/or magnetometers.
With reference to
Shown emanating from the probe housing icon 16′ is a field of view 24 of the imaging system 12. The portion of the anatomy within the field of view 24 captured by the imaging system 12 is shown as the ultrasound image 78. In the example illustrated, a heart image or model 110 is within the field of view 24. The heart image would be an actual ultrasound image captured by the imaging system 12. A 3D model would be a model created from segmented ultrasound images or from a registered CT or MRI scan. The location of the field of view 24 on the display 26 is determined based on the pose of the housing 16, as determined by tracking the position of the tracking device 22. Showing the field of view 24 on the display 26 assists the operator of the imaging system 12 with positioning the housing 16 to capture a desired portion of the anatomy. The user can see on the avatar 120 the portion of the anatomy within the field of view 24, which is displayed as the US image 78. If an area of interest is not within the field of view, the user can use the avatar 120 as a guide for maneuvering the housing 16 until the area of interest is within the field of view 24 and displayed in the ultrasound image 78. For example, the US image 78 of
The navigation processing system 70 can also be configured to generate the display 26 of
With renewed reference to
For some users, it may be difficult to identify areas of the heart that have been captured and areas of the heart that have yet to be captured, whether to ensure a capture of an entirety of the heart or of all of the areas necessary to generate the appropriate atlas or model. The navigation processing system 70 may be configured to generate the display 26 of
Thus, during the scanning process, the navigation system 70 tracks movement of the housing 16 relative to the heart, or other portions of the anatomy being scanned, to identify portions of the heart that have been scanned versus portions that have not been scanned. The navigation system 70 continuously updates the images of
The imaging system 12, including the ultrasound probe, may be moved relative to the subject to acquire an appropriate amount of image data, as discussed above. As illustrated in
With reference to
As discussed above and understood by one skilled in the art, the EM localizers 50, 55, or 58 may emit a field that induces a current in a tracking device, such as the tracking device 22. The display 26 also includes a range icon 162, which identifies the outer boundaries of the range of the particular localizer displayed. The range icon 162 identifies that extent (e.g., area or volume), in which the field emitted by the EM localizer is effective for navigation. Similarly, a field of view of an optical localizer may be displayed as the range icon 162. The range icon may be displayed relative to the subject aviator 120.
Displaying the sensor icon 160 and the range icon 162 facilitates positioning of the subject 14 and/or the localizer 50, 55, or 58 to capture an area of the body of interest. That is, the range icon 162 illustrates that extent to which the EM localizer 50, 55, or 58 is effective. As the avatar 120 illustrates the subject to be imaged and/or have tracking devices tracked relative thereto, the range illustrates the area in which a tracking device may be effectively tracked. Thus, the user may view the range icon 12 and decide if the volume included is effective for the procedure or not.
The localizer is registered in any suitable manner. For example, registration may be carried out with a reference tracker on the subject's body at a predefined location, such as the reference tracker 56. That is, the localizers 50, 55, or 58 may sense a reference marker placed on the subject at a known anatomical location and the localizer may be used to determine the pose of the localizer relative thereto. Further, the localizer range 162 may be relative to a second localizer that is tracked with a first localizer.
Example embodiments are provided so that this disclosure will be thorough and will fully convey the scope to those who are skilled in the art. Numerous specific details are set forth such as examples of specific components, devices, and methods, to provide a thorough understanding of embodiments of the present disclosure. It will be apparent to those skilled in the art that specific details need not be employed, that example embodiments may be embodied in many different forms and that neither should be construed to limit the scope of the disclosure. In some example embodiments, well-known processes, well-known device structures, and well-known technologies are not described in detail.
The foregoing description of the embodiments has been provided for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention. Individual elements or features of a particular embodiment are generally not limited to that particular embodiment, but, where applicable, are interchangeable and can be used in a selected embodiment, even if not specifically shown or described. The same may also be varied in many ways. Such variations are not to be regarded as a departure from the invention, and all such modifications are intended to be included within the scope of the invention.
It should be understood that various aspects disclosed herein may be combined in different combinations than the combinations specifically presented in the description and accompanying drawings. It should also be understood that, depending on the example, certain acts or events of any of the processes or methods described herein may be performed in a different sequence, may be added, merged, or left out altogether (e.g., all described acts or events may not be necessary to carry out the techniques). In addition, while certain aspects of this disclosure are described as being performed by a single module or unit for purposes of clarity, it should be understood that the techniques of this disclosure may be performed by a combination of units or modules associated with, for example, a medical device.
In one or more examples, the described techniques may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored as one or more instructions or code on a computer-readable medium and executed by a hardware-based processing unit that may also be referred to as a processor. Computer-readable media may include non-transitory computer-readable media, which corresponds to a tangible medium such as data storage media (e.g., RAM, ROM, EEPROM, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer).
Instructions may be executed by one or more processors or processor modules, such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor” or “processor module” as used herein may refer to any of the foregoing structure or any other physical structure suitable for implementation of the described techniques. Also, the techniques could be fully implemented in one or more circuits or logic elements.
This application claims the benefit of and priority to U.S. Provisional Patent Application 63/495,951 filed on Apr. 13, 2023. The subject application includes subject matter related to the subject matter disclosed in U.S. Provisional Patent Applications 63/420,274 filed on Oct. 28, 2022; 63/420,315 filed on Oct. 28, 2022; and 63/420,328 filed on Oct. 28, 2022, all of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
63495951 | Apr 2023 | US |