The following generally relates to ultrasound and more particularly to ultrasound super resolution imaging.
Super resolution imaging is a technique to enhance a resolution of an image. Ultrasound super resolution imaging relies on tracking microbubbles of a microbubble based contrast agent to visualize microvasculature. However, tissue motion, e.g., from breathing, the heart beating, muscle contracting, etc., that is larger than a size of microvasculature limits the resolution of the super resolution image. For example, motion in the range of 100-200 μm in both the axial and lateral directions would limit the resolution of the ultrasound super resolution image to 100-200 μm, and vessels smaller than 100-200 μm such as microvasculature (e.g., 10 μm), would not be visible.
One motion correction scheme excludes frames with higher motion. (See J. Foiret et al., “Ultrasound localization microscopy to image and assess microvasculature in a rat kidney,” Scientific Reports, vol. 7, no. 1, pp. 13 662:1-12, 2017). Another scheme combines ultrasound microscopy and dual-frequency imaging, but the B-mode frame rate is only sufficient to capture motion from breathing and more rapid movements (e.g., a beating heart) are discarded. (See T. M. Kierski et al., “Superharmonic ultrasound for motion-independent localization microscopy: Applications to microvascular imaging from low to high flow rates,” IEEE Trans. Ultrason., Ferroelec., Freq. Contr., vol. 67, no. 5, pp. 957-967, 2020).
Unfortunately, the above and other approaches do not address spatial and temporal differences between the motion estimates from the B-mode frames and the contrast frames. As such, there is an unresolved need for an improved approach to super resolution imaging.
Aspects of the application address the above matters, and others.
In one aspect, an apparatus included a processor and a display. The processor includes a combiner configured to combine sets of contrast data acquired with a same sub-aperture, for each of a plurality of sub-apertures, to create a contrast frame for each of the sub-apertures. The processor further includes a microbubble detector configured to determine positions of microbubbles in the contrast frames. The processor further includes a motion estimator configured to estimate a motion field based on frames of B-mode data for each of the plurality of sub-apertures. The processor further includes a motion corrector configured to motion correct the positions of the microbubbles in the contrast frames based on the motion field and time delays between emissions for the sets of contrast data and the corresponding emission for B-mode data, for each of the plurality of sub-apertures, to produce motion corrected contrast frames. The display is configured to visually display the motion corrected contrast frames, wherein the motion corrected contrast frames include super resolution images.
In another aspect, a method includes combining sets of contrast data acquired with a same sub-aperture, for each of a plurality of sub-apertures, to create a contrast frame for each of the sub-apertures. The method further includes determining positions of microbubbles in the contrast frames. The method further includes estimating a motion field based on frames of B-mode data for each of the plurality of sub-apertures. The method further includes motion correcting the positions of the microbubbles in the contrast frames based on the motion field and time delays between emissions for the sets of contrast data and the corresponding emission for B-mode data, for each of the plurality of sub-apertures, to produce motion corrected contrast frames. The method further includes displaying the motion corrected contrast frames, wherein the motion corrected contrast frames include super resolution images.
In yet another aspect, a computer-readable storage medium storing instructions that when executed by a computer cause the computer to: combine sets of contrast data acquired with a same sub-aperture, for each of a plurality of sub-apertures, to create a contrast frame for each of the sub-apertures, determine positions of microbubbles in the contrast frames, estimate a motion field based on frames of B-mode data for each of the plurality of sub-apertures, motion correcting the positions of the microbubbles in the contrast frames based on the motion field and time delays between emissions for the sets of contrast data and the corresponding emission for B-mode data, for each of the plurality of sub-apertures, to produce motion corrected contrast frames, and display the motion corrected contrast frames, wherein the motion corrected contrast frames include super resolution images.
Those skilled in the art will recognize still other aspects of the present application upon reading and understanding the attached description.
The application is illustrated by way of example and not limited by the figures of the accompanying drawings, in which like references indicate similar elements and in which:
The probe 104 includes a transducer array 114 with one or more transducer elements 116. The transducer array 114 includes a one or two dimensional (1 or 2-D), linear, curved and/or otherwise shaped, fully populated or sparse, etc. array. The elements 116 are configured to convert excitation electrical pulses into an ultrasound pressure field and a received ultrasound pressure field (echo) into an electrical (e.g., a radio frequency (RF)) signal. Generally, the received pressure field is in response to the transmitted pressure field interacting with matter, e.g., contrast agent microbubbles, tissue, etc.
The console 106 includes transmit circuitry (TX) 118 configured to generate the excitation electrical pulses based on an algorithm(s) 120. For B-mode imaging, a suitable acquisition results in data that can be used to detect the linear signals from tissue (i.e. tissue data). For contrast-enhanced imaging using a contrast agent with microbubbles (which is administered to a subject prior to and/or during the scan), a suitable acquisition results in data that can be used to detect non-linear signals from the microbubbles (i.e. contrast data) and suppress the linear signals.
Briefly turning to
For contrast data, a first sub-aperture emits three times, one with full amplitude and two with half amplitude. This is shown in
Returning to
The console 106 further includes a switch (SW) 124 configured to switch between the TX 118 and RX 122 for transmit and receive operations, e.g., by electrically connecting and electrically disconnecting the TX 118 and RX 122. In a variation, separate switches are utilized to switch between the TX 118 and RX 122.
The console 106 includes further an RF processor 126. The RF processor 126 is configured to process the contrast and B-mode data to create ultrasound super resolution images. As described in greater detail below, creating contrast frames, detecting microbubbles in the contrast frames, estimating a local motion field from the B-mode frames, motion correcting the position of the microbubbles based on the local motion estimation, and registering the microbubbles across the contrast frames.
Generally, this approach utilizes a full motion field that is estimated as a function of space and time from all B-mode frames to co-register estimated tracks of microbubbles to a reference frame. In one instance, this ensures that no frame is skipped. In addition, this approach can reduce microbubble localization uncertainty relative to a configuration that does not utilize this approach.
The console 106 further includes a scan converter 128 and a display 130. The scan converter 128 is configured to scan convert the motion corrected contrast frames and/or B-mode frames for display, e.g., by converting the motion corrected contrast frames and/or the B-mode frames to the coordinate system of the display 130. The display 130 can display a motion corrected contrast frame and/or a B-mode frame alternately or concurrently, e.g., next to each other and/or on combined.
The console 106 further includes a user interface 132, which includes one or more input devices (e.g., a button, a touch pad, a touch screen, etc.) and one or more output devices (e.g., a display screen, a speaker, etc.). The console 106 further includes a controller 134 configured to control one or more of the transmit circuitry 118, the receive circuitry 122, the switch 124, the RF processor 126, the scan converter 128, the display 130, and/or the user interface 132.
It is to be appreciated that at least the RF processor 126 can be implemented by a hardware processor (e.g., a central processing unit (CPU), graphics processing unit (GPU), a microprocessor, etc.) executing computer readable instructions encoded or embedded on computer readable storage medium, which excludes transitory medium.
The RF illustrated processor includes a combiner 302. The combiner 302 is configured to combine the contrast data from the three emissions for each a sub-aperture to generate a contrast frame for the sub-aperture, for each of the sub-apertures. In one instance, this includes subtracting the data for the half amplitude acquisitions from the data for the full amplitude acquisition. Other approaches are also contemplated herein.
The RF processor further includes a microbubble detector 304. The microbubble detector 304 is configured to determine a centroid (or other region) of each of the microbubbles in each of the contrast frames. In one instance, this includes signal-to-noise (SNR) enhancement followed by microbubble localization. For SNR Enhancement, suitable approaches include thresholding, spatial filtering (e.g. Gaussian, Laplacian of Gaussian, etc.), spatio-temporal filtering, model fitting, and/or other approaches. For localization, suitable approaches include peak detection, centroid estimation (e.g. weighted centroid, etc.), learning-based, and/or other approaches.
The RF processor further includes a motion estimator 306. The motion estimator 306 is configured to process the B-mode frames to estimate tissue motion. In one instance, this includes using speckle tracking on the envelope data. An example of speckle tracking can be found in G. E. Trahey et al.,” IEEE Trans. Biomed. Eng., vol. BME-34, no. 12, pp. 965-967, 1987. For local motion estimation, each B-mode frame is divided into a plurality of partially overlapping sub-regions, with one of the frames being identified as a reference frame. Examples of suitable sub-region sizes include, but are not limited to, 1×1 to 10×10 millimeters square (mm2), larger, smaller, non-square, etc. Examples of a suitable B-mode reference frame includes a B-mode frame near the middle B-mode frame and/or other B-mode frame. In another instance, more than one B-mode reference frame is utilized.
The motion estimator 306 cross-correlates each sub-region in the B-mode reference frame with the corresponding sub-region in the other B-mode frames, and motion in the axial and lateral directions is estimated. The estimated motion for each sub-region is assigned to a center (and/or other location) of the corresponding sub-region, and the collection of motion estimates for a frame provides a discrete motion field through that frame. The estimated displacements vary spatially and temporally. Motion can be estimated at any point in any frame (i.e. space and/or time) using interpolation such a 3-D spline, etc. on the 3-D motion field.
The RF processor further includes a motion corrector 308. The motion corrector 308 is configured to apply the local motion estimates on the detected microbubble locations. By way of non-limiting example, for a detected microbubble in the kth frame at position {right arrow over ({circumflex over (r)})}=(z0, x0), the correct position {right arrow over ({circumflex over (r)})} of the microbubble can be approximated as shown in Equation 1:
Equation 1:
{right arrow over ({circumflex over (r)})}≈{right arrow over (r)}−{right arrow over (M)}(t, {right arrow over (r)}),
where {right arrow over (M)}(t, {right arrow over (r)}) represents the motion field,
where k represents the frame position, fr represents the imaging framerate frequency, and td(n, m) represents a time difference between an nth line of a contrast frame and an mth line of the corresponding B-mode frame. In one instance, e.g., in
where n, m ∈ {1, 2, . . . , Nlines}, Nlines represents a number of lines in the frame, and fprf represents the pulse repetition frequency.
The position of the detected microbubble at {right arrow over (r)}=(z0, x0) in the kth frame can be expressed as a function of the corrected microbubble position and value of the motion field at that position and time as shown in Equation 3:
where {right arrow over ({circumflex over (r)})}=({circumflex over (z)}0, {circumflex over (x)}0) and {circumflex over (n)} are the unknown correct position and its relative image line in the contrast image, respectively. In one instance, {right arrow over ({circumflex over (r)})} is computed using an optimization techniques such as a least-squares optimization as shown in Equation 4:
where {right arrow over (p)} and q are arguments.
An example algorithm for solving Equation 4 is provided next.
In general, the motion corrector 308 iteratively assigns motion estimates by calculating a time delay between each an image line of a contrast frame and an image line of the B-mode frame for a sub-aperture (the three emissions of the contrast sequence corresponded to the same image line as one emission in the B-mode sequence), updating a position of an estimated bubble in the contrast frame based on the motion field and the time delay, updating the image line in the contrast frame, and repeating the above until stopping criteria is satisfied or the maximum number of iterations is reached.
The RF processor further includes a microbubble tracker 310. The microbubble tracker 310 is configured to generate tracks that link microbubbles from frame to frame. Suitable approaches include, but are not limited to, nearest-neighbor, multi-frame data structure, dynamic programming, combinatorial, multi hypothesis, explicit motion models (e.g. Kalman filtering), learning-based, and/or other techniques. Provided herein is an example approach based on Kalman filtering.
This approach can be modeled as {right arrow over (r)}(t)={right arrow over (r)}(t−1)+d{right arrow over (r)}(t)+ϵ(t), where {right arrow over (r)}(t)=(rz(t), rx(t)) represents the position of a microbubble at time t, d{right arrow over (r)}(t)=(drz(t), drx(t)) represents the displacement of the microbubble, and ϵ(t) represents an uncertainty in the displacement. This model can be formulated, e.g., within the Kalman framework as shown below:
When entering the blood stream, the microbubbles have uncontrolled concentrations as well as different velocities and flow dynamics in different parts of the kidney. To track these different scenarios, a hierarchical structure of Kalman filters is utilized. Table 1 shows examples parameters for a hierarchical structure of Kalman filter.
At each level in Table 1, a Kalman filter is applied to the remaining untracked microbubbles from the previous levels. The tracked microbubbles at deeper levels have higher velocities. A maximum linking distance is a maximum acceptable distance between the estimated microbubble position by the Kalman filter and the nearest motion-corrected microbubble position in the next frame. Linking of multiple microbubbles can be performed using Hungarian assignment. If a microbubble is miss-detected over a couple of frames, the Kalman filter can predict its trajectory based on its previous detected positions. The gap closing parameter is a number of frames a microbubble can be miss-detected.
The motion-corrected and linked microbubbles provide ultrasound super resolution intensity image. In the illustrated embodiment, the RF processor 126 includes a velocity estimator 312 configured to process track locations to generate a super resolution velocity image. By way of non-limiting example, the velocity estimator 312 can be configured to determine a time derivative of track locations, which yields both the axial and lateral velocities. An axial and a lateral velocity image can then be constructed by drawing anti-aliased velocity weighted lines at the track positions. A mean velocity at each pixel in the velocity images can be found by dividing with a weighting image if it was different from zero. A Vector Flow Image (VFI) of the microbubble can be generated by combining the axial and lateral velocity images. In another instance, the velocity estimator 312 is omitted.
It is to be appreciated that the hierarchical tracking approach described herein provides a better visualization of opposite flows and improved flow quantification, relative to a configuration that does not employ the hierarchical tracking approach. The ultrasound super resolution intensity and/or velocity images can be displayed via the display 130.
The ordering of the following acts is for explanatory purposes and is not limiting. As such, one or more of the acts can be performed in a different order, including, but not limited to, concurrently. Furthermore, one or more of the acts may be omitted and/or one or more other acts may be added.
At 402, contrast and B-mode data are acquired, as described herein and/or otherwise.
At 404, microbubbles are detected in the contrast data, as described herein and/or otherwise.
At 406, motion is estimated from the B-mode data, as described herein and/or otherwise.
Although act 404 is described before act 406, it is to be understood that act 404 can be performed before act 406 or in parallel with act 406.
At 408, the microbubbles in the contrast data are motion corrected based on the estimated motion, as described herein and/or otherwise.
At 410, the microbubble are linked between frames of the contrast data, creating a super resolution intensity image, as described herein and/or otherwise.
At 412, the super resolution intensity image is displayed, as described herein and/or otherwise.
At 414, a vector velocity is determined, as described herein and/or otherwise.
At 416, a super resolution vector velocity image is displayed, as described herein and/or otherwise.
The above may be implemented by way of computer readable instructions, encoded or embedded on computer readable storage medium (which excludes transitory medium), which, when executed by a computer processor(s) (e.g., central processing unit (CPU), microprocessor, etc.), cause the processor(s) to carry out acts described herein. Additionally, or alternatively, at least one of the computer readable instructions is carried by a signal, carrier wave or other transitory medium (which is not computer readable storage medium).
The application has been described with reference to various embodiments. Modifications and alterations will occur to others upon reading the application. It is intended that the invention be construed as including all such modifications and alterations, including insofar as they come within the scope of the appended claims and the equivalents thereof.
Number | Date | Country | |
---|---|---|---|
63043839 | Jun 2020 | US |