The present invention relates generally to estimating blood flow velocity. More particularly, the invention relates to a method of using deep learning neural networks to provide a full three-dimensional characterization of blood flow dynamics in vivo.
Conventional blood flow velocity estimation techniques can only measure blood flow along the transmit ultrasound beam directions. These techniques cannot measure the flow that is perpendicular to the beam directions. This produces one-dimensional (1D) velocity estimation that is inaccurate for the measurement of blood flow velocity. It also limits the application of Doppler ultrasound to blood vessels that are along with the ultrasound beam directions, or have a small angle (typically within 30 degrees) deviation from the transmit beams. In addition, due to these limitations, these techniques cannot provide a full characterization of the hemodynamics in vivo.
Efforts in obtaining two-dimensional (2D) velocity measurement of blood flow velocity involves complicated transmit or receive sequencing or beamforming, computationally intensive estimators (e.g. least-squares multiple-angle estimation) or speckle tracking. These techniques are computationally intensive and may require complex sequencing or beamforming. Efforts to alleviate these limitations may result in a compromise in image quality or equipment cost.
The present invention advances the art to provide techniques to estimate 2D blood flow velocity using deep neural networks.
The present invention provides a method and system to characterize blood flow dynamics in vivo. A Doppler acquisition is obtained in vivo with two or more transmit-receive event pairs (step a). Each transmit event in the two or more transmit-receive event pairs at a different transmit angle, which could be two or more angles (step b). In one example the different transmit angles are three to five angles. The Doppler acquisition is processed to determine three-dimensional angle-resolved RF data. In one example, the three-dimensional angle-resolved RF data includes axial data, lateral data, and transmit angle data.
The steps (a) and (b) are then repeated to obtain four-dimensional dimensional angle-resolved RF data. In one example, the four-dimensional angle-resolved RF data includes axial data, lateral data, transmit angle data, and slow time.
The four-dimensional angle-resolved RF data is then provided as input into a trained neural network. This trained neural network is adaptive to the dimensions of the angle-resolved RF data, and outputs spatially resolved flow velocities in multiple blood flow dimensions. In one example, the multiple blood flow dimensions include at least an axial dimension and a lateral dimension.
The method and system could further include deriving from the outputs of the trained neural network and displaying temporal hemodynamic flow velocity profiles in the multiple blood flow dimensions at a given spatial location.
In one variation to plane wave transmits at different angles, the teachings of the invention could also be varied to include embodiments with transmit source and element allowing for diverging or focused transmits.
Embodiments of the invention estimate a 2D blood flow velocity using deep learning neural networks. This technique is fast and provides an accurate estimation of 2D blood flow velocity in all directions, including the flow that is perpendicular to transmit beam directions. Using 2D or other multidimensional ultrasound transducers, the technique can be improved to provide full 3D characterization of blood flow dynamics in vivo. The invention has been characterized in computer simulations, flow phantom experiments, and in vivo human liver vasculature studies.
Using existing deep learning frameworks and GPU hardware, the technique can be implemented and integrated into medical ultrasound scanners at low cost without major modification of hardware. No customized hardware is needed at all.
The invention has advantages for application that involves the measurement of blood flow velocity. For example, cardiac blood flow measurements to detect abnormal flow patterns, placental blood flow velocity measurements, tumor angiogenesis characterization, and kidney blood flow measurement, among others.
In one embodiment, a computer processor implements the training, validation, testing, and deployment of the technique.
In one embodiment the invention is a method of processing signals for input to a neural network structure to output three-dimensional characterization of blood flow dynamics in vivo (
From multiple Doppler frame acquisition data, step ((iii) could be repeated to produce videos of Doppler vector flow velocity, including magnitude and angle and producing measurement of temporal profiles of axial velocity, lateral velocity, magnitude of velocity and angle of velocity at any location in the field of view.
Embodiment of the invention are a method and system with signal processing steps as shown in
Step 1. The Doppler signal acquisition is done using the ultrasound scanner, where the output data dimensions include 4-dimensional (4D) data including:
The term “slow time” is derived from Doppler radar terminology, referring to the Doppler sampling dimension. Data sampling is performed at a fixed pulse-repetition-frequency. In the invention, each sample has a 3D data cube with axial, receive channel, and transmit event dimensions. Each sample in this dimension may be referred to as one ensemble. The number of samples in this dimension may be referred to as “packet size” or “ensemble length”.
For simplicity, “transmit events” and “transmits with different angles” are used interchangeably herein. In practice, any synthetic transmit aperture technique can be used in the transmit events, including using different diverging or focused waves, or different transmit apertures. “Angle compounding” in the following text means coherently summing data across the transmit event dimension in one Doppler acquisition. “Angle resolved” data refers to data that is not angle (or element/source) compounded.
Step 2. The Beamforming (dynamic receive beamforming) has output data dimensions: 4D data including:
Step 3. Clutter filtering can be embodied using the method of Singular Value Decomposition-based (SVD) filtering to reshape the 4D data into 2D. The reshaping vectorizes the axial, azimuth, and transmit event dimensions into one dimension, and the slow time dimension is kept as the second dimension. The resulted 2D matrix has these two dimensions:
After that, SVD decomposition, energy- or velocity-based thresholding on the singular values, SVD reconstruction are performed to remove unwanted tissue clutter and other types of noise, leaving only blood signals. The last step is reshaping the 2D filtered data back to 4D:
Note that in the first step, the reshaping method from 4D data to 2D data is through vectorization of 3 of the 4 dimensions of data, including axial, azimuth, and transmit event dimensions, into one-dimension (1D), while preserving the fourth original dimension, that is the slow time dimension. According to further embodiments of the invention, other filters, including Butterworth filter or eigen-based filters, can be used in a similar manner as well.
In one aspect, a power Doppler image is optionally produced to show the location of blood vessel. A threshold is applied to the power Doppler image to produce a binary mask, where all pixels in blood vessel region have values of 1, and all pixels outside the blood vessel have values of 0. The mask is then multiplied to all RF data to provide additional suppression of noise outside the blood vessel.
Step 4. The vector velocity estimation using neural networks includes two methods: Method 1 vector velocity estimation from angle-resolved beam-summed data using neural networks, and Method 2 vector velocity estimation from angle-compounded beam-summed data using neural networks (
Method 1
The neural network used has two major components: one component is the input layer that is tailored to extract features from the angle-resolved data, and the other major component is for vector flow velocity estimation from the features. The input layer for feature extraction is one convolutional layer which input dimension is dependent upon the number of plane waves utilized in the data acquisition. For an n-angle input data, the first layer of the network needs to be tailored as a convolution layer with n input channels and 16 output channels (with kernel size being 3, stride being 1). In theory, n can be any positive integer.
Method 2
The neural network structure used in Method 2 differs from Method 1 only in the input layer. Because angle-compounded data is used as input, the input layer has n=1 input channel instead of n>1 input channels. Alternatively, the input layer can also have n>1 channels and the angle-compounded data can be copied for n times before used as the input to the network. The rest of the structure used in Method 2 is the same as Method 1. The specific weights in the network, however, may be changed adaptively in the training process. This training does not alter the structure.
Turning to Method 1, the input includes angle-resolved, beamformed beam-summed radio-frequency data acquired with two Doppler events, where the data dimensions are 4D data comprising two frames of 3D data,
The beam-summation is performed in the receive element dimension, and is not performed in the transmit events dimension. The output includes vector flow estimation from the flow velocity field in 2D space. For each spatial location (x, z), there are two velocity estimation values corresponding to axial velocity (vz) and azimuth velocity (vx). The data dimensions are 3D, with the size of one dimension being two. The other two dimensions are azimuth and axial locations in space.
For Method 2, an additional step is performed that includes summing the 4D angle-resolved, beamformed beam-summed radio-frequency data in the transmit event dimension, and reducing the data to 3D (azimuth, axial, slow time). The data here is referred to as angle compounded beam-summed data. The input includes angle compounded beam-summed data acquired with two Doppler events, where the data dimensions include 3-dimensional data has two frames of 3D data, 1) axial, 2) azimuth. The output is a vector flow estimation, and a flow velocity field in 2D space. For each spatial location (x, z), there are two velocity estimation values corresponding to axial velocity (vz) and azimuth velocity (vx). The data dimensions include 3D, with the size of one dimension being two. The other two dimensions are azimuth and axial locations in space.
Regarding the performance difference between Method 1 and Method 2, Method 1 requires a larger data size to be used as input, because angle-resolved data is needed. The estimation is more accurate with lower variance, especially for flow in which the direction is perpendicular to transmit beam direction or ultrasound transducer axis.
Method 2 uses data with small sizes, because the angle compounded data has a smaller size than the angle resolved data from the same Doppler acquisition by a factor equal to the number of transmit events (e.g. transmit angles). It produces accurate flow estimation for flow in which the directions are not perpendicular to transmit beam direction or ultrasound transducer axis. For perpendicular flow, the variance produced using Method 2 is high than Method 1.
Returning to the signal processing steps, note that vector velocity estimation is produced from two consecutive Doppler frames (i.e. with an ensemble length of 2). For Doppler acquisitions with ensemble length greater than two, repeat Step 4 for the entire Doppler acquisition by processing all consecutive pairs of Doppler frames. The output would be the vector velocity estimation for all Doppler acquisitions. If multiple Doppler frames (more than 2 frames, preferably more than 100 for high video quality) are acquired and processed using Step 5, videos of vector flow velocity magnitudes and angles can be produced as shown in APPENDIX A of U.S. Provisional Patent Application 62/911,602 filed Oct. 7, 2019 to which this application claims the benefit and which is incorporated herein by reference. In addition, at any specific locations of the field of view (including in the blood vessels), temporal profiles of axial velocity, azimuth velocity, magnitude of velocity can be directly measured, as shown in the same APPENDIX A. The temporal profile of the angle of velocity can be calculated as the arctangent of the ratio of the axial velocity and azimuth velocity. The magnitude of the vector velocity can also be calculated. In the same APPENDIX A, the calculated angles are shown using the arrows.
Conversion from axial and azimuth velocities to the magnitude and angle of velocities. In this equation, νx is the lateral velocity, νz is the axial velocity, |ν| is the velocity magnitude, and θ is the flow angle.
Step 5 is optional, that includes performing post processing to improve the vector flow velocity estimation results. Typical methods include low-pass filtering or median filtering in spatial and/or slow time dimensions. The velocity information can be visualized as maps of flow velocities in the entire field of view. Further quantitative analysis of hemodynamics can be performed using the velocity estimation. Example of such quantitative analysis include finding the maximum or minimum of velocities, analyzing the pulse or the pulsing pattern, calculating the resistance of blood flow, etc.
Training of the neural network training includes:
λi (i=1, 2, 3, 4, 5) are the weights for the 5 different losses l1 to l5, respectively, and can be adjusted in the training process as hyperparameters. l1=∥predicted velocities−true velocities∥l1, in which ∥·∥l1 represents the L−1 norm. This loss calculates the L−1 norm of the difference between predicted flow velocities and the ground truth in the entire field of view.
This loss calculates the ratio between the number of false 0 velocity pixels and the number of true 0 velocity pixels.
n which ∥·∥l2 represents the L−2 norm. This loss calculates the L−2 norm of the difference between predicted flow velocities and the ground truth in the vessel regions.
l4=TV(predicted axial velocities), where TV(ν) represents the total variation, defined as TV(ν)=Σall pixels|ν(z+1)−ν(z)|. It calculates the summation of the absolute values of the differences between the axial flow velocity estimations at all adjacent pixels. Minimizing this loss function is equivalent to the introduction of prior knowledge that human blood vessels are smooth and that the axial flow velocity does not have abrupt and non-continuous changes, which is based on human physiology.
l5=TV (predicted azimuth velocities), which is the total variation of the flow estimation in the azimuth dimension. TV(ν) is defined as in l4. Minimizing this loss function is equivalent to the introduction of prior knowledge that human blood vessels are smooth and that the lateral flow velocity does not have abrupt and non-continuous changes.
Turning now to using input data collected with multiple plane wave transmits at different angles, the Doppler flow dataset differences from the optical flow data sets are first presented, where the data in the optical flow imaging case describes motion of relatively rigid objects, including chairs, trees, cars, etc. In the videos, the objects are rigid and do have significant local deformation. The shapes are preserved relatively well from frame to frame (i.e. in slow time dimension). In addition, the three-color channels, in most cases, show objects that have highly similar shapes and describe the same motion.
The data in Doppler flow imaging is different in these two major aspects. First, the blood flow signal does not describe rigid object motion without local deformation. Instead, the signals deform in slow time dimension, and have variable rates of change in appearance that is dependent on location. For example, the signals in the center of a blood vessel move faster and change their appearances faster in slow time, compared to signals near the edge of the same blood vessel. This phenomenon is well documented and rigorously proven in the literature, and is different from the videos in the optical flow data sets. Second, the angle resolved beamformed RF data acquired using different transmit angles show different images of the same blood vessel. In other words, the appearances of the blood signals acquired using plane waves at different angles, in general, show different patterns with different textures. The difference can be measured as a decrease of the correlation between them, which has been theoretically described and experimentally demonstrated.
Turning now to the increase in the number of angles in the angle-resolved method, where first presented is the impact. The increase in the number of angles in the angle-resolved method, in general and assuming all other conditions being the same, improves the estimation accuracy by reducing the variance of estimation. The trade-off is the longer training and inference computation time, as well as a lower limit on the Doppler pulse-repetition-frequency.
According to one embodiment of the method of the invention, for an n-angle input data, the first layer of the network needs to be tailored as a convolution layer with n input channels and 16 output channels (with kernel size being 3, stride being 1).
In theory, n can be any positive integer. Specifically, the first layer can be represented as:
For the angle number of 5, as an example, the feature extraction pyramid network structure can be significantly different from a conventional neural network structure. The entire feature extraction pyramid network structure with angle number 5 is:
The training with the network with this feature pyramid achieves fast convergence within 2 epochs.
This application claims priority from U.S. Provisional Patent Application 62/911,602 filed Oct. 7, 2019, which is incorporated herein by reference.
This invention was made with Government support under contract HD086252 awarded by the National Institutes of Health. The Government has certain rights in the invention.
Number | Date | Country | |
---|---|---|---|
62911602 | Oct 2019 | US |