A transesophageal echocardiography (TEE) ultrasound probe is commonly used in cardiac monitoring and navigation. Currently available multi-plane imaging modes for a TEE ultrasound probe include X-plane and full three-dimensional (3D) volume.
Ultrasound tracking technology estimates the position of a passive ultrasound sensor (e.g., PZT, PVDF, copolymer or other piezoelectric material) in the field of view (FOV) of a diagnostic ultrasound B-mode image by analyzing the signal received by the passive ultrasound sensor as the imaging beams of the ultrasound probe sweep the field of view. Time-of-flight measurements provide the axial/radial distance of the passive ultrasound sensor from the imaging array, while amplitude measurements and knowledge of the beam firing sequence provide the lateral/angular position of the passive ultrasound sensor.
According to an aspect of the present disclosure, a controller for controlling tracking of an interventional medical device in a patient includes a memory that stores instructions, and a processor that executes the instructions. When executed by the processor, the instructions cause the controller to execute a process that includes controlling an imaging probe. The imaging probe is controlled to activate imaging elements to emit imaging signals to generate three or more imaging planes including a first imaging plane, a second imaging plane, and a third imaging plane perpendicular to the second imaging plane, to simultaneously capture an interventional device and anatomy targeted by the interventional device. The imaging probe is also controlled to simultaneously capture both the interventional device and the anatomy targeted by the interventional device. The imaging probe is controlled to capture at least one of the interventional device and the anatomy targeted by the interventional device in at least two of the three or more imaging planes, and to capture the other of the interventional device and the anatomy targeted by the interventional device in at least one of the three or more imaging planes.
According to another aspect of the present disclosure, a method for tracking an interventional medical device in a patient includes emitting, by activated imaging elements controlled by an imaging probe, imaging signals to generate three or more imaging planes including a first imaging plane, a second imaging plane, and a third imaging plane perpendicular to the second imaging plane, to simultaneously capture an interventional device and anatomy targeted by the interventional device. The method also includes simultaneously capturing the interventional device and the anatomy targeted by the interventional device. The imaging probe is controlled to capture at least one of the interventional device and the anatomy targeted by the interventional device in at least two of the three or more imaging planes, and to capture the other of the interventional device and the anatomy targeted by the interventional device in at least one of the three or more imaging planes.
According to yet another aspect of the present disclosure, a system for tracking an interventional medical device in a patient includes an imaging probe and a controller. The imaging probe is configured to activate imaging elements to emit imaging signals to generate three or more imaging planes including a first imaging plane, a second imaging plane, and a third imaging plane perpendicular to the second imaging plane, to simultaneously capture an interventional device and anatomy targeted by the interventional device. The controller controls the imaging probe to simultaneously capture both the interventional device and the anatomy targeted by the interventional device. The imaging probe is controlled to capture at least one of the interventional device and the anatomy targeted by the interventional device in at least two of the three or more imaging planes, and to capture the other of the interventional device and the anatomy targeted by the interventional device in at least one of the three or more imaging planes. The controller includes a signal processor that processes image signals that simultaneously capture at least one of the interventional device and the anatomy targeted by the interventional device in at least two of the three or more imaging planes and the other of the interventional device and the anatomy targeted by the interventional device in at least one of the three or more imaging planes.
The example embodiments are best understood from the following detailed description when read with the accompanying drawing figures. It is emphasized that the various features are not necessarily drawn to scale. In fact, the dimensions may be arbitrarily increased or decreased for clarity of discussion. Wherever applicable and practical, like reference numerals refer to like elements.
In the following detailed description, for purposes of explanation and not limitation, representative embodiments disclosing specific details are set forth in order to provide a thorough understanding of an embodiment according to the present teachings. Descriptions of known systems, devices, materials, methods of operation and methods of manufacture may be omitted so as to avoid obscuring the description of the representative embodiments. Nonetheless, systems, devices, materials and methods that are within the purview of one of ordinary skill in the art are within the scope of the present teachings and may be used in accordance with the representative embodiments. It is to be understood that the terminology used herein is for purposes of describing particular embodiments only, and is not intended to be limiting. The defined terms are in addition to the technical and scientific meanings of the defined terms as commonly understood and accepted in the technical field of the present teachings.
It will be understood that, although the terms first, second, third etc. may be used herein to describe various elements or components, these elements or components should not be limited by these terms. These terms are only used to distinguish one element or component from another element or component. Thus, a first element or component discussed below could be termed a second element or component without departing from the teachings of the inventive concept.
The terminology used herein is for purposes of describing particular embodiments only, and is not intended to be limiting. As used in the specification and appended claims, the singular forms of terms ‘a’, ‘an’ and ‘the’ are intended to include both singular and plural forms, unless the context clearly dictates otherwise. Additionally, the terms “comprises”, and/or “comprising,” and/or similar terms when used in this specification, specify the presence of stated features, elements, and/or components, but do not preclude the presence or addition of one or more other features, elements, components, and/or groups thereof. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
Unless otherwise noted, when an element or component is said to be “connected to”, “coupled to”, or “adjacent to” another element or component, it will be understood that the element or component can be directly connected or coupled to the other element or component, or intervening elements or components may be present. That is, these and similar terms encompass cases where one or more intermediate elements or components may be employed to connect two elements or components. However, when an element or component is said to be “directly connected” to another element or component, this encompasses only cases where the two elements or components are connected to each other without any intermediate or intervening elements or components.
In view of the foregoing, the present disclosure, through one or more of its various aspects, embodiments and/or specific features or sub-components, is thus intended to bring out one or more of the advantages as specifically noted below. For purposes of explanation and not limitation, example embodiments disclosing specific details are set forth in order to provide a thorough understanding of an embodiment according to the present teachings. However, other embodiments consistent with the present disclosure that depart from specific details disclosed herein remain within the scope of the appended claims. Moreover, descriptions of well-known apparatuses and methods may be omitted so as to not obscure the description of the example embodiments. Such methods and apparatuses are within the scope of the present disclosure.
As introduced above, use of an X-plane can provide a high frame rate, but only 2 adjustable imaging planes. On the other hand, use of a full three-dimensional (3D) volume can provide control over slicing, but a low frame rate. The present disclosure provides an ability to simultaneously visualize both an interventional medical device and anatomy targeted by the interventional medical device using, for example, the same ultrasound imaging probe by emitting imaging signals in three or more imaging planes. To be clear from the start, the simultaneous emission and capture by the ultrasound imaging probe may involve emitting and capturing the interventional medical device and targeted anatomy when the interventional medical device and targeted anatomy are physically separated in a three-dimensional space.
As described for embodiments below, tissue around a device can be visualized with other quantitative navigation metrics, without losing sight of desired anatomy. Device tracking output can be bootstrapped to an imaging plane selection algorithm, via an automatic feedback/control loop that links device location to control of imaging plane selection. An example of an automatic feedback/control loop is a remote control link (RCL), which tracks an identified device through imaging planes as the device is moved. By linking the interventional device tracking output and the imaging plane selection, multiple different embodiments described herein provide varying capabilities. In other words, the interventional device tracking can be used as part of a feedback loop to ensure that the ability to track the interventional device continues, so that one or more imaging planes can be tied or dedicated to the interventional device. Thus, device tracking can be used to automatically visually follow a device with the imaging planes, in order to continue tracking the interventional device.
The computer system 200 can be implemented as or incorporated into various devices, such as a stationary computer, a mobile computer, a personal computer (PC), a laptop computer, a tablet computer, an ultrasound system, an ultrasound probe, or any other machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. The computer system 200 can be incorporated as or in a device that in turn is in an integrated system that includes additional devices. In an embodiment, the computer system 200 can be implemented using electronic devices that provide voice, video or data communication. Further, while the computer system 200 is illustrated as a single system, the term “system” shall also be taken to include any collection of systems or sub-systems that individually or jointly execute a set, or multiple sets, of instructions to perform one or more computer functions.
As illustrated in
Moreover, the computer system 200 includes a main memory 220 and a static memory 230 that can communicate with each other via a bus 208. Memories described herein are tangible storage mediums that can store data and executable instructions, and are non-transitory during the time instructions are stored therein. As used herein, the term “non-transitory” is to be interpreted not as an eternal characteristic of a state, but as a characteristic of a state that will last for a period. The term “non-transitory” specifically disavows fleeting characteristics such as characteristics of a carrier wave or signal or other forms that exist only transitorily in any place at any time. A memory described herein is an article of manufacture and/or machine component. Memories described herein are computer-readable mediums from which data and executable instructions can be read by a computer. Memories as described herein may be random access memory (RAM), read only memory (ROM), flash memory, electrically programmable read only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), registers, a hard disk, a removable disk, tape, compact disk read only memory (CD-ROM), digital versatile disk (DVD), floppy disk, blu-ray disk, or any other form of storage medium known in the art. Memories may be volatile or non-volatile, secure and/or encrypted, unsecure and/or unencrypted.
As shown, the computer system 200 may further include a video display unit 250, such as a liquid crystal display (LCD), an organic light emitting diode (OLED), a flat panel display, a solid-state display, or a cathode ray tube (CRT). Additionally, the computer system 200 may include an input device 260, such as a keyboard/virtual keyboard or touch-sensitive input screen or speech input with speech recognition, and a cursor control device 270, such as a mouse or touch-sensitive input screen or pad. The computer system 200 can also include a disk drive unit 280, a signal generation device 290, such as a speaker or remote control, and a network interface device 240.
In an embodiment, as depicted in
In an alternative embodiment, dedicated hardware implementations, such as application-specific integrated circuits (ASICs), programmable logic arrays and other hardware components, can be constructed to implement one or more of the methods described herein. One or more embodiments described herein may implement functions using two or more specific interconnected hardware modules or devices with related control and data signals that can be communicated between and through the modules. Accordingly, the present disclosure encompasses software, firmware, and hardware implementations. Nothing in the present application should be interpreted as being implemented or implementable solely with software and not hardware such as a tangible non-transitory processor and/or memory.
In accordance with various embodiments of the present disclosure, the methods described herein may be implemented using a hardware computer system that executes software programs. Further, in an exemplary, non-limited embodiment, implementations can include distributed processing, component/object distributed processing, and parallel processing. Virtual computer system processing can be constructed to implement one or more of the methods or functionality as described herein, and a processor described herein may be used to support a virtual processing environment.
The present disclosure contemplates a computer-readable medium 282 that includes instructions 284 or receives and executes instructions 184 responsive to a propagated signal; so that a device connected to a network 101 can communicate voice, video or data over the network 201. Further, the instructions 284 may be transmitted or received over the network 201 via the network interface device 240.
In
At S320, a mode is determined. A mode may consist of a set of one or more selecting settings such as three or four imaging planes, rotations of planes about an axis, which planes are dedicated to an interventional device, and which planes are dedicated to anatomy targeted by an interventional device. The term “dedicated” as used herein may refer to an assignment of planes to a specific target, which for the purposes of the present disclosure is either an interventional device, or anatomy targeted by the interventional device. The interventional device may be targeted by dedicated planes that track the interventional device in two dimensions or three dimensions as the interventional device moves in the body of the patient.
The anatomy targeted by the interventional device may be designated by a user instruction, such as by using a mouse and cursor or a touch screen. The anatomy may be a specific position on the surface of an organ such as a heart or lung, and may be targeted by the interventional device in the sense that the interventional device is moved towards the anatomy targeted by the interventional device. The interventional device may also be designated by a user, but may alternatively be automatically identified and tracked, such as with the use of a sensor made of a specific material that is readily identified in ultrasound.
At S330, an ultrasound probe is controlled to emit imaging signals in three or more imaging planes, based on the mode, to simultaneously capture both the interventional device and the anatomy targeted by the interventional device. In known ultrasound, X-planes use 2 imaging planes, such as 2 perpendicular planes, and capture only one of an interventional device or anatomy targeted by the interventional device. However, at S330, three or more imaging planes are used, and between the three or more imaging planes, the interventional device and the anatomy targeted by the interventional device are both simultaneously captured. For example, each of the three or more imaging planes may specifically intersect with one or both of the interventional device and/or the anatomy targeted by the interventional device.
At S340, the ultrasound probe is controlled to capture both the interventional device and the anatomy targeted by the interventional device, based on the emitted imaging signals in three or more planes. One of the interventional device and the anatomy targeted by the interventional device is captured in at least two of the three or more imaging planes, and the other of the interventional device and the anatomy targeted by the interventional device is simultaneously captured in least one of the three or more imaging planes. In embodiments, both of the interventional device and the anatomy targeted by the interventional device are simultaneously captured in two of the imaging planes, albeit not necessarily the same two imaging planes. In other embodiments, one or the other of the interventional device and the anatomy targeted by the interventional medical device are captured in one and only one of the imagine planes.
At S350, positions of the interventional device and the anatomy targeted by the interventional device are identified, based, for example, on the capture of reflected/returned imaging signals. Alternately, positions of the interventional device can be tracked from signals of a passive ultrasound sensor, or by other methods and mechanisms. Positions may be identified in a predetermined coordinate system, such as in a three-dimensional cartesian coordinate system with dimensions for width (X), height (Y) and depth (Z). A center of the coordinate system may be set at a fixed point in the space (volume) in or around the patient body.
In an embodiment, multiple different medical imaging systems may be registered to one another, so as to reflect commonality in viewpoints. Registration in this manner may involve setting coordinate systems of the different medical systems to reflect a common origin and common directionality dimensions.
At S360, a distance between the interventional device and anatomy targeted by the interventional device is determined and displayed. The distance may be determined in two dimensions, such as width (X)/height (Y), or may be determined in three dimensions such as width (X)/height (Y)/depth (Z).
At S370, a display is controlled to simultaneously display, in real-time, the interventional device and the anatomy targeted by the interventional device. A display may be or may include a screen on a television or on an electronic device such as a monitor. The monitor may be a monitor specifically provided with an ultrasound system, and may have settings specifically appropriate for visualizing imagery captured by the ultrasound system as well as related information such as information related to the captured imagery.
Specifically, in
However, the device plane #1 (vertical) 591 and the device plane #2 (horizontal) 592 are shown to be perpendicular, and this characteristic is accurately reflective of how these planes are best used to capture a targeted interventional device or anatomy targeted by an interventional device. Similarly, the anatomy plane #1 (vertical) 596 and anatomy plane #2 (horizontal) 597 are also shown to be perpendicular, and this characteristic is also accurately reflective of how these planes are best used to capture a targeted interventional device or anatomy targeted by an interventional device. Nevertheless, perpendicular planes do not have to be perfectly perpendicular, and may be substantially perpendicular while still working in their intended manner. Examples of substantially perpendicular planes may be intersecting planes with a smaller angle therebetween greater than 67.5 degrees, greater than 75 degrees, or greater than 85 degrees.
Specifically, in
Accordingly, in the embodiment of
Specifically, in
In an alternative embodiment, the one anatomy plane #1 (horizontal) 797 can be a short-axis imaging plane rather than a long-axis imaging plane. In still another alternative to the embodiment shown in
Specifically, in
Here, the transverse plane of the interventional device X-plane tied to the interventional device position is adjusted to image the region of tissue “just ahead” of the current device position. The adjusted transverse plane thereby shows which tissue the interventional device will encounter if the interventional device is pushed ahead further in the current direction. Current direction can be determined from the recent history of device positions.
An example of projecting for the embodiments of
In
At any time during a procedure, a “distance to target” can be calculated from the current device location and the desired anatomical target, and shown to the user in real-time. This is shown in
Accordingly, interventional medical device tracking enables selective use of different numbers of imaging planes in order to simultaneously capture both an interventional device and anatomy targeted by the interventional device. This provides visualization of tissue around a device and other quantitative navigation metrics, without losing sight of targeted anatomy.
Although interventional medical device tracking has been described with reference to several exemplary embodiments, it is understood that the words that have been used are words of description and illustration, rather than words of limitation. Changes may be made within the purview of the appended claims, as presently stated and as amended, without departing from the scope and spirit of interventional medical device tracking in its aspects. Although interventional medical device tracking has been described with reference to particular means, materials and embodiments, interventional medical device tracking is not intended to be limited to the particulars disclosed; rather interventional medical device tracking extends to all functionally equivalent structures, methods, and uses such as are within the scope of the appended claims.
For example,
Although the present specification describes components and functions that may be implemented in particular embodiments with reference to particular standards and protocols, the disclosure is not limited to such standards and protocols.
The illustrations of the embodiments described herein are intended to provide a general understanding of the structure of the various embodiments. The illustrations are not intended to serve as a complete description of all of the elements and features of the disclosure described herein. Many other embodiments may be apparent to those of skill in the art upon reviewing the disclosure. Other embodiments may be utilized and derived from the disclosure, such that structural and logical substitutions and changes may be made without departing from the scope of the disclosure. Additionally, the illustrations are merely representational and may not be drawn to scale. Certain proportions within the illustrations may be exaggerated, while other proportions may be minimized. Accordingly, the disclosure and the figures are to be regarded as illustrative rather than restrictive.
One or more embodiments of the disclosure may be referred to herein, individually and/or collectively, by the term “invention” merely for convenience and without intending to voluntarily limit the scope of this application to any particular invention or inventive concept. Moreover, although specific embodiments have been illustrated and described herein, it should be appreciated that any subsequent arrangement designed to achieve the same or similar purpose may be substituted for the specific embodiments shown. This disclosure is intended to cover any and all subsequent adaptations or variations of various embodiments. Combinations of the above embodiments, and other embodiments not specifically described herein, will be apparent to those of skill in the art upon reviewing the description.
The Abstract of the Disclosure is provided to comply with 37 C.F.R. § 1.72(b) and is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, various features may be grouped together or described in a single embodiment for the purpose of streamlining the disclosure. This disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter may be directed to less than all of the features of any of the disclosed embodiments. Thus, the following claims are incorporated into the Detailed Description, with each claim standing on its own as defining separately claimed subject matter.
The preceding description of the disclosed embodiments is provided to enable any person skilled in the art to practice the concepts described in the present disclosure. As such, the above disclosed subject matter is to be considered illustrative, and not restrictive, and the appended claims are intended to cover all such modifications, enhancements, and other embodiments which fall within the true spirit and scope of the present disclosure. Thus, to the maximum extent allowed by law, the scope of the present disclosure is to be determined by the broadest permissible interpretation of the following claims and their equivalents, and shall not be restricted or limited by the foregoing detailed description.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2019/054399 | 2/22/2019 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62633788 | Feb 2018 | US |