The present invention relates to medical image analysis in general, and in particular, to ultrasound image analysis. The invention further teaches the application of multiple modes of machine learning and deep learning for image analysis.
In the field of medical imaging, image analysis has moved from the speciality of trained technicians to specialized machine learning algorithms. Trained machine learning algorithms can reliably and accurately solve general image recognition or pattern recognition problems. In complex medical situations, such as with multi-mode imaging diagnostics, image analysis becomes more complex because multiple types of images need to be evaluated.
Ultrasound medical imaging systems consist of numerous different imaging modes. The present invention teaches automatization and analysis of two in particular—brightness mode (B-mode) and spectral Doppler mode (pulsed wave (PW))—which are commonly used alone or in conjunction for diagnostic and procedural medical applications. B-mode ultrasound provides two-dimensional images of the magnitude of the reflections in tissues and represents structural information. Meanwhile, spectral Doppler imaging mode is used for representation of fluid flow (such as blood flow) information from a defined location, called Doppler gates.
It is standard in the field of vein cannulation to use ultrasound to guide catheter insertion, among other applications of ultrasound imaging. A standard procedure of cannulation requires a technician to scan for blood vessels, manually identify blood vessels, and then the technician can switch ultrasound modes of operations and manually adjust the imaging parameters. Ultimately, ultrasound scanning provides some information to the technician about the state of the vessels and their character, and while numerous known solutions are provided to determine the character of a certain type of vessel or under specific conditions, reliable means of identification of an artery or vein in general are not found in the art. Most commonly, ultrasound scanners used for vein cannulation guidance are capable of only B-mode ultrasound scanning, and therefore, lack the capability for more precise diagnostic functionality.
A trend in the field of medical imaging is to automate the role of a radiological technician by applying image-recognition software solutions. Such solutions have been developed for various medical imaging mediums, including specific vessel detection, such as carotid artery or jugular vein, in B-mode images, however, these solutions are parameterized for only structural 2D images and cannot be applied to vessel classification in general. Error in proper identification of artery and veins can be fatal; in the case of puncturing the jugular vein, it could be confused with carotid artery and that might result into severe complications.
The present invention improves on state-of-the-art solutions for automation of medical image analysis, specifically, classification of blood vessels using a combination of B-mode and spectral Doppler ultrasound imaging for any blood vessel. In standard procedures, multiple steps require manual manipulation or analysis by a trained technician, and in the present invention, all of the manual steps are replaced by machine learning algorithms Comparison of the steps that are automated in the present invention will become apparent in the below description.
A system and a method hereby described are intended to detect blood vessels by applying trained deep learning (DL) algorithms to ultrasound structural B-mode images and subsequently to identity vessel character (vein or artery) using Doppler spectrogram image analysis by trained machine learning (ML) classification models. The DL detection in B-mode images is followed by automatic positioning of PW Doppler gates followed by PW Doppler imaging scans. The resulting Doppler images of identified blood vessels are used for further classification of the vessels as either vein or artery using either spectrogram feature extraction combined with machine learning predictions or image recognition using deep learning predictions. Such classification is important for successful catheter insertion (cannulation) under ultrasound guidance or other procedures which requires differentiation between arteries and veins or quantitative characterization of blood flow. Previous studies have shown that usage of PW Doppler during vein cannulation increases first pass success rate.
The proposed system allows fully automated detection of vessels and vessel type differentiation and does not require manual Doppler gate placement, and therefore there is no need for a highly qualified technician with deep knowledge of spectral Doppler to perform the procedure. The presented solution is based on machine learning principles. It works in real-time with B-mode imaging having a frame rate close to 50 frames/second. Deep leaning-based detection is highly accurate compared with known object detection techniques. The differentiation of arteries and veins is typically done by evaluating blood flow velocity, but this feature alone is not sufficient for accurate classification, because in the diastolic periods blood flow velocity is comparable for arteries and veins. Herein, it is taught that a PW Doppler spectrogram of a blood vessel contains more features that can be used to more accurately classify a blood vessel.
The image analysis system consists of an ultrasound scanner, an array probe capable of performing B-mode and PW mode scanning, and one or more computer processors configured with computer-implemented methods 1) for automatic vessel tracking in real-time based on DL, 2) for Doppler spectrogram quality assessment, and 3) for classification of a detected vessel as an artery or a vein based on machine learning models.
The invention can be best understood by referring to the drawings, which depict preferred embodiments of the present invention.
The presented figures are for illustration and the scale, the proportions, and the other aspects do not necessarily correspond to the actual technical solution.
The present invention is best described by its preferred embodiments, which are exemplified by the figures. According to the schematic block diagram of
In a preferred embodiment, the image analysis system 100 as described above, is configured to execute the following procedure: an operator slowly moves the ultrasound probe 102 coupled to the tissue surface 108 and obtains one or more B-mode images, which represent a vessel or few vessels that could be an artery 106 or a vein 104. The scanning plane could be a transverse view (as illustrated in
The vessel classification module 122 could be implemented in two ways: by calculating a set of statistical quantities and inputting said statistical quantities into a trained ML algorithm, or by passing the obtained spectrogram as an image into a trained convolutional neural network. Statistical quantities such as the periodicity parameter of the envelope of the spectrogram are used in the first embodiment. The module formulates classifications of artery or vein, the result of which is overlaid with the B-mode image and displayed on the monitor.
The trained classifier 216 could be implemented in two ways: by calculating statistical quantities and inputting the statistical quantities into a trained machine learning algorithm or by imputing the spectrogram image into a trained deep learning algorithm. The details of the preferred embodiments for classification are described in more detail below. If the output of the trained classifier 216 exceeds a predefined threshold 218, which was obtained through the training procedure, the vessel is classified as artery, otherwise it is classified as a vein. The steps of classification, 208-218 are repeated for each detected vessel. The procedure is concluded 220 when all the detected vessels are analysed and classified, and the procedure can be repeated for a new set of B-mode images.
The vessel detection DL module 114 is dedicated for use with structural B-mode images. The module utilizes deep learning principles, and the preferred embodiments use trained convolutional neural networks. In a preferred embodiment, deep learning networks that utilize fast architecture are used so that the computation time for vessel detection is commensurate with real-time brightness mode image processing frame rate, at least 50 frames/second, such as YOLO, Fast R-CNN, Faster R-CNN, or other comparatively fast architectures.
A sketch of the preferred convolutional network architecture with base components is shown in
Z=ω
T
·X+b,
where Z is the output of convolutional layer, X is the input image matrix of I×J×K size, I is the number of image columns, J is the number of image rows, K is the number of channels, ω is the matrix of weighting coefficients, which could also be called L×L convolutional filter kernel, L is the size of a filter, and b is the bias coefficients vector, which is also obtained in the training phase. Multi-dimensional convolution is followed by the ReLU function. The ReLU function changes negative values of convolutional layer output to zeros and could be expressed as follows:
ReLU(Z)=max(0,Z).
A max pooling layer 306 is dedicated for down-sampling of the detection features in feature maps. It is realized by a maximum detector and selector in the predefined size window in feature maps. The convolution and ReLU function layer 308 is then repeated followed by a max pooling layer and so on until a prediction is made 312.
Training of the vessel detection deep learning network can be completed either offline or online using training data in a server-based database. For the offline training scheme, a representative database of B-mode image sequences and annotations must be collected. The annotations mean the bounding boxes of the vessels detected and outlined manually in B-mode images by an expert who visually evaluates the images. Optionally the vessels could be verified by using spectral Doppler to verify if the detected structure is a vessel. The collection of images and annotations are passed to the training procedure. Optimal weighing coefficients are obtained by using stochastic gradient descent (SGD) method or other techniques such as the Adam optimization algorithm. In the case of SGD, neural network weighing coefficients are updated by the following formula:
where η is a learning rate, i is the training iteration number, L is the loss function, and n is the number of observations.
The online training option requires a server-based database, which is connected to the ultrasound machine controlling PC (hereafter referred to as a workplace) to obtain new images and annotations performed by an expert; a picture archiving and communication system will serve for the purpose. The weighting coefficients of the neural network in such cases are updated with each new example received from the workplace. Continual training of the neural network produces a more reliable outcome.
The Doppler beam control 116 is dedicated for automatic adjustment of sample volume (Doppler gates) position and size, which are calculated based on the predictions of the vessel detection DL module 114.
The spectrogram quality evaluation module 120 assesses whether the spectrogram is of sufficient quality for quantitative analysis. Venous flow is sometimes very weak and cannot be detected by spectral Doppler ultrasound especially if the imaging and Doppler scanning is performed in the transverse plane. The obtained spectrogram could be classified into two classes according to pixel intensities: blood flow information and background noise. For this purpose, the spectrogram 502 is binarized 504 to obtain a mask for blood flow related information extraction. The procedure of binarization is illustrated by 500 in
σw2(t)=w0(t)·σ02(t)+w1(t)·σ12(t),
where w0, w1 are probabilities of the two classes 1 and 2 separated by a threshold t, and σ02, σ12 are variances of the classes.
In the next stage, the parameters of blood flow related pixel intensities are extracted. Two parameters of the proportion of blood flow related pixels in comparison to background are used: the ratio between the detected foreground pixels and a total number of pixels in a spectrogram, and the ratio between the sum of the intensities in the foreground and the sum of all intensities of the spectrogram. Finally, the parameters are combined into a vector and used for spectrogram classification into two classes: 1) sufficient quality and 2) insufficient quality. The optimal weights for the parameters are obtained through a training procedure. The parameters could be combined by using linear technique:
Y=w·X,
where Y is the output of a linear classifier and w is a weighting coefficients vector of a parameter X. Or by using non-linear classifiers such as support vector machines or others. Output of the classifier are compared to threshold values obtained through a training procedure. If the spectrogram is classified as of sufficient quality, the spectrogram passes to the vessel classification module 122. Spectrogram of insufficient quality cannot be used because the blood flow of veins is relatively weak and could be misidentified in a spectrogram with insufficient quality.
The vessel classification module 122 can be implemented by the following two embodiments. In the first embodiment, first mean velocity is calculated from the spectrogram then the spectrogram is parametrized. Next, four statistical quantities are calculated for blood flow characterization and classification:
where γ is delay, 0≤y≤N, N is the number of samples of the mean velocity curve, x is the mean velocity value at a certain time instance. The obtained function is multiplied by a triangular window function in order to supress the peak at zero delay and to enhance peaks arising due to heart beat related pulsatility:
where 0≤n≤N, N is the number of samples in the autocorrelation function. Finally, the presence of periodicity is evaluated by finding the maximum peak of the windowed autocorrelation function. The value of the peak serves as a statistical quantity for spectrogram characterization. A higher peak value indicates that there is a periodic pattern in the mean velocity curve, which is characteristic for arteries.
where M is the mobility parameter, x is the mean velocity vs. time curve. Mobility is then calculated as follows:
Where var is statistical variance. For mean velocity in veins, the signal complexity is lower while in the case of arteries, the mean velocity curve shape closely resembles a sine wave.
The statistical quantities are combined into a feature function and passed to the machine learning-based classifier, which determines if the scanned vessel belongs to an artery class or to a vein class. The trained classification algorithm could be a machine learning technique such as linear regression, non-linear, support vector machine classifier, or others.
In the second embodiment, the vessel classification module 122 is implemented by using deep learning principles, which evaluate the spectrogram directly using the principles of image recognition in a trained deep learning algorithm rather than evaluating the statistical quantities separately and combining each statistical quantity into a simplified feature function. In such case the spectrogram is passed into a trained convolutional neural network as an image and the network classifies the detected vessel to be an artery or a vein. The convolutional neural network for vessel classification architecture must be fast and the number of layers should not exceed 30. The feature extraction layers, including convolutional+ReLU, max pooling, are of a similar structure as shown in
Finally, the results of the method are represented on the display monitor 124, typical on a PC or laptop monitor via a graphical user interface 600 (