Blood vessels are typically difficult to see on an x-ray angiogram without the use of a radiocontrast agent (which may be referred to more simply as “contrast agent” or just as “contrast”) within the vessels, such as an iodinated contrast. When delivering an intravascular device, such as a balloon, stent or probe, within a blood vessel, physicians typically refer to a vessel tree generated from an x-ray angiogram (XA) captured with a high level of radiocontrast agent present to better define the vessel tree in the XA. However, the intravascular device itself is typically not visible in an x-ray angiogram captured using such a high level of contrast agent, and repeated or continuous use of contrast agent during a procedure is undesirable at least for medical reasons. Typically, when delivering an intravascular device, x-ray angiograms may be captured with low levels of, or no, contrast agent present in the blood vessels of interest, so that the position of the device can be readily seen using the x-ray angiograms. Then from time to time, to better see where the device is located within the vessel tree, the physician may inject more contrast agent into the blood vessel(s) of interest. The heart function typically clears the contrast agent after less than a second. To provide for continued observation of the vessel tree during the delivery of the device, the physician can either repeatedly inject more contrast agent, which can lead to unwanted side-effects, or be left unable to see the vessel tree when delivering the device.
The disclosure is generally directed to dynamically visualizing the delivery of a device within a blood vessel, for example within the context of a blood vessel tree or other region of interest, by correlating at least one first extraluminal image with second extraluminal images. The extraluminal images may be correlated based on motion features, visible within or derivable from the extraluminal images without the use of other sensors, such as readings from an electrocardiogram, or time stamps. The at least one first extraluminal image may be a high dose contrast x-ray angiogram (“XA”) captured using a high level of contrast agent present so as to make the blood vessel or blood vessel tree clearly visible in this XA (referred to herein as a high dose contrast XA or a high dose XA). The second extraluminal images may then be low dose XAs taken using little or no contrast agent present in the blood vessel or blood vessel tree so that the intravascular device is more clearly visible (referred to herein as low dose contrast XAs or low dose XAs). Typically for example, the amount or concentration of contrast agent present in the blood vessels or vessels of interest when the at least one high dose XA is taken may be at least five times or at least ten times the amount or concentration present when the low dose XAs are taken. The second extraluminal images may be, in some examples, one or more subsequent images captured after the first extraluminal image is captured. In some examples, the second extraluminal images may take the form of a low dose contrast XA video or video stream which may be displayed in real time, as a live low dose XA. In some examples, the live low dose XA may be low dose XAs that are captured during an intravascular procedure, such as intravascular imaging, stent placement, etc. In some examples, live low dose XAs may be referred to as fluoroscopy images.
The at least one high dose contrast XA may be used to generate a vessel map. The low dose contrast XAs may be taken during the delivery of an intravascular device, such as a balloon, stent, probe, or the like. Correlating, or maintaining a mapping between, the at least one high dose XA and the low dose XAs can be implemented using motion features, which are typically features which are visible or detectable within, or derivable from, both the high and low dose XAs. For example, the motion features can include at least one of a guide catheter tip, a distal endpoint of a working vessel, an optical flow at the guide catheter tip, or an optical flow at the distal endpoint of the working vessel.
The correlation or mapping then allows for the vessel map to be overlaid on the low dose XAs, or vice versa, to provide a physician with visualization of where the intravascular device is within the vessel or vessel tree in real time. For example, this may allow for the intravascular device to be tracked on the live low dose XA with reference to the vessel map created using the at least one high dose XA. In some examples, a position of the intravascular device may be determined based on the live low dose XAs and an indication of the position may be provided on both the low dose XAs and the vessel map determined using the at least one high dose XA. According to some examples, correlating or mapping between the at least one high dose XA and the low dose XAs may allow for a treatment zone, such as a stent or balloon landing zone, identified on the high dose XA to be transferred, or superimposed, on the low dose XAs.
According to some examples, during a percutaneous coronary intervention (PCI) workflow, the at least one high dose contrast XAs may form at least part of a high dose XA video which may contain both pre-contrast, low dose XAs captured before contrast agent has been released into the blood vessel, and post-contrast, high dose XAs captured once contrast agent has been released. The post-contrast, high dose XAs may provide a clear view of the vessel tree, including side branches that are not typically visible in the pre-contrast, low dose XAs. The post-contrast, high dose XAs may be used to generate a vessel map. A vessel of interest, such as a working vessel, may be identified based on an intravascular device such as a guide catheter and guide wire in the pre-contrast, low dose, XAs. The identified working vessel may then be used to correlate, or map between, low dose and high dose XAs captured during the delivery of a device within the vessel.
Such a correlation or mapping provides for particular physical positions, such as a position within the working vessel, as captured in a low dose XA to be mapped to the same physical positions, such as the same position within the working vessel, in a high dose XA or vessel map. In this way features such as position of an intravascular device as seen or detected in the low dose XAs can be stably depicted within the context of one or more high dose XAs or a corresponding vessel map, or features such as features of a vessel map determined using the one or more high dose XAs can be stably depicted within the context of one or more low dose XAs.
The working vessel may be identified using one or more artificial intelligence (AI) models, such as machine learning (ML) models, that use the shape prior, e.g., the prediction of the location of the guide wire and/or guide catheter in the previous frame(s), to predict the location of the guide wire and/or guide catheter in the current frame. The AI model may be a spatial temporal model. The AI model may be trained based on annotated XAs. The annotations may be, in some examples, a set of line strips that follow the trajectory of the guide catheter and guide wire in pre-contrast XAs.
Motion features, which are typically features which are visible or detectable within, or derivable from, both the high and low dose XAs, may be used to correlate, or determine a mapping between, the high and low dose XAs to provide for dynamic visualization during the delivery of a device within a vessel. Because of ongoing movement of the blood vessels during capture of the low dose XAs, for example due to heart movement if the blood vessels are within cardiac tissue, this correlation or mapping may continuously or rapidly change, albeit on a repeated cycle such as that caused by a heartbeat, and the motion features of the low dose XAs may then be used to continuously update the correlation or mapping. The motion features may be identified in the high and low dose XAs. The motion features may be determined on a vessel level, vessel region of interest level, and a vessel pixel level to characterize the motion phase of the high and low dose XAs. The motion features may be used to perform an initial search to correlate high and low dose XAs having substantially the same temporal phase. A normalized cross correlation may be performed on pairs of motion features in the high and low dose XAs. For a given motion feature pair, samples of the low dose XA feature that are within a predetermined time window may be identified. The time window may be compared with all windows of substantially the same time length from the corresponding high dose XA feature. A maximum correlation coefficient may be determined based on the more correlated pair of high and low dose motion features. In some examples, an initial time shift may be determined. The initial time shift may be filtered using an online update. The online update may use a filter, such as a linear Kalman filter, to predict a time shift that would most likely match the most recently obtained low dose XA with the next available high dose XA.
The high and low dose XAs may be correlated with, or mapped between each other, based on the temporal correlation of motion features, without using any external sensor information. In some examples, in addition to or as an alternative of, the high and low dose XAs may be correlated using spatial landmarks.
According to some examples, the intravascular device may include one more markers, such as radiopaque markers, that are visible on low dose XAs. The markers may be detected and used to track the position of the intravascular device with respect to the vessel map generated based on the high dose XA. Tracking the device markers may allow a physician to determine whether the device is nearing and/or reached a treatment zone that was identified on the vessel map. In some examples, the treatment zone identified on the vessel may be reflected on the live low dose XA based on the correlation of the low and high dose XAs. By transferring annotations generated on the vessel map, e.g., the high dose XA, to the live low dose XAs, a physician may be able to dynamically visualize and track the position of the device with respect to the vessel morphology during the delivery of the device.
One aspect of the technology is generally directed to a computer implemented method (for example implemented using one or more suitable computer processors), comprising: receiving at least one first extraluminal image; receiving second extraluminal images captured during delivery of an intravascular device; detecting motion features in the at least one first extraluminal image and the second extraluminal images; correlating, based the detected motion features, the at least one first extraluminal image and the second extraluminal images; and providing for output, real-time visualization of a position of the intravascular device on the at least one first extraluminal image or one of the second extraluminal images including the intravascular device. The method may also comprise displaying, on a suitable display device such as a computer monitor, the output real-time visualization.
The at least one first extraluminal image may be a high dose contrast x-ray angiogram. The second extraluminal images may be low dose contrast x-ray angiograms. The intravascular device may be at least one of a stent delivery device, a balloon device, an intravascular imaging probe, a vessel prep device, or a pressure wire. The method may further comprise generating, by the one or more processors based on the at least one first extraluminal image, a vessel map.
The method may further comprise automatically detecting, by the one or more processors using a AI model, a working vessel. Training the AI model to automatically detect the working vessel may comprise: annotating pre-contrast extraluminal images as line strips following a trajectory of at least one of a guide wire or a guide catheter, labeling a path of the working vessel in post-contrast extraluminal images, providing the annotated pre-contrast extraluminal images and labeled post-contrast extraluminal images as training data to the AI model, and training the AI model to predict a working vessel trajectory. The method may further comprise receiving, by the one or more processors as input into the AI model at least one pre-contrast extraluminal image, training, by the one or more processors augmenting high-dose extraluminal images into low-dose extraluminal images as input, the AI model to segment at least one of a guide catheter, guide wire, stent marker, or balloon marker on the low-dose extraluminal images, detecting, by the one or more processors executing the AI model and based on the at least one pre-contrast extraluminal image, a guide wire of the intravascular device, propagating, by the one or more processors on a frame by frame basis executing the AI model and based on the detected guide wire, wire information, and automatically predicting, by the one or more processors executing the AI model based on the propagated wire information, a working vessel trajectory.
The motion features may include at least one of a guide catheter tip, a distal endpoint of a working vessel, an optical flow at the guide catheter tip, or an optical flow at the distal endpoint of the working vessel. Detecting the motion features may further comprise automatically detecting, by the one or more processors executing a AI model, the working vessel. Correlating the at least one first extraluminal image and the second extraluminal images may comprise: determining, by the one or more processors, vessel level motion, determining, by the one or more processors, wire tip level motion, and determining, by the one or more processors, vessel pixel level motion. Determining the vessel level motion may further comprise determining, by the one or more processors, a two-dimensional translation vector. The two-dimensional translation vector may correspond to two-dimensional translation at an n-th frame with respect to a first frame. Determining the wire tip level motion may comprise determining, by the one or more processors, absolute spatial information and relative spatial information. The absolute spatial information may correspond to coordinates and the relative spatial information corresponds to optical flow between adjacent image frames. Determining vessel pixel level motion may comprise determining, by the one or more processors, absolute spatial information and relative spatial information for a plurality of points on a working vessel. The absolute spatial information may correspond to coordinates and the relative spatial information corresponds to optical flow between adjacent image frames. The method may further comprise determining, based on the detected motion features, a heartbeat period of a patient.
The method may further comprise determining, by the one or more processors based on the detected motion features, a spatial-temporal phase match between the at least one first extraluminal image and at least one of the second extraluminal images. The method may further comprise resampling, by the one or more processors, the detected motion features at a common frame rate, determining, by the one or more processors, a maximum correlation coefficient for pairs of motion features in the at least one first extraluminal image and the second extraluminal images, and determining, by the one or more processors based on the maximum correlation coefficient, a time shift. The method may further comprise determining, by the one or more processors, a drift between detected motion features in the at least one first extraluminal image, and adjusting, by the one or more processors based on the determined drift, the time shift. The method may further comprise iteratively predicting, by the one or more processors, the time shift; and updating, by the one or more processors based on the iteratively predicted time shift, the time shift to a corrected time shift.
The method may further comprise tuning, by the one or more processors, the spatial-temporal phase match. Tuning the spatial-temporal phase match may comprise identifying, by the one or more processors, another first extraluminal image different than the at least one of the first extraluminal images and tuning, by the one or more processors based on the other first extraluminal image, the spatial-temporal phase match. The method may further comprise updating, by the one or more processors based on the tuning, the real-time visualization to include the other first extraluminal image.
The method may further comprise detecting, by the one or more processors, the intravascular device. Detecting the intravascular device in the second extraluminal images may further include executing a AI model. The method may further comprise training the AI model, wherein training the AI model comprises providing as input to the AI model a co-registration dataset comprising a plurality of intraluminal images and extraluminal images, wherein the plurality of intraluminal and extraluminal images are annotated images; and training the AI model to predict a position of the intravascular device. The annotated images include annotations identifying one or more intravascular device markers. The method may further comprise detecting, by the one or more processors, an optical flow of the intravascular device, determining, by the one or more processors based on the detected optical flow, a position of the intravascular device in a first frame of the second extraluminal images, and predicting, by the one or more processors based on the detected optical flow, the position of the intravascular device in a subsequent frame of the second extraluminal images.
The method may further comprise providing for output, by the one or more processors, a treatment zone on at least one of the second extraluminal images or the at least one first extraluminal image. The treatment zone is at least one of a treatment device landing zone, a stent landing zone, a balloon device zone, a vessel prep device zone, or a lesion related zone. The lesion related zone may be at least one of calcification frames, lipid frames, or dissected frames, and the lesion related zone may be identified from another imaging modality and co-registered, by the one or processors, to the first extraluminal image. The other imaging modality may be an intravascular imaging modality.
The method may further comprise receiving, by the one or more processors, annotations of the at least one first extraluminal image or the second extraluminal images. When receiving the annotations the method may further comprises receiving, by the one or more processors, one or more inputs from a user corresponding to the annotations, or automatically determining, by the one or more processors based on vessel data, the annotations. The annotations may include one or more of a plaque burden, fractional flow reserve (“FFR”) measurements at one or more locations along the vessel, calcium angles, EEL detections, calcium detections, proximal frames, distal frames, EEL-based metrics, stent decisions, scores, recommendations for debulking, recommendations for subsequent procedures, stent placement zone, treatment device landing zone, balloon device zone, vessel prep device zone, or lesion related zone. The method may further comprise updating, by the one or more processors based on the received annotations, a second one of the at least one first extraluminal image or the second extraluminal images.
The method may further comprise automatically capturing, by the one or more processors, a screen capture of the real-time visualization of the position of the intravascular device. The screen capture may be automatically captured when the intravascular device is within a threshold distance of a region of interest. The region of interest may be a treatment zone. The treatment zone may be at least one of a treatment device landing zone, a balloon device zone, a vessel prep device zone, or a lesion related zone. Determining the threshold distance may comprise at least one of determining, by the one or more processors, a number of pixels between an outer boundary of the region of interest and at least one detected marker on the intravascular device, or determining, by the one or more processors, a spatial distance between the at least one detected marker on the intravascular device and the region of interest. The spatial distance may be a Euclidean distance, a geodesic distance, etc.
The method may further comprise automatically zooming, by the one or more processors, a portion of the real-time visualization. The automatic zooming may focus on the portion of the real-time visualization when the intravascular device is within a threshold distance of a region of interest. The portion of the real-time visualization may correspond to the region of interest. The region of interest may be a treatment zone. The treatment zone may be at least one of a treatment device landing zone, a balloon device zone, a vessel prep device zone, or a lesion related zone. Determining the threshold distance may comprise at least one of determining, by the one or more processors, a number of pixels between an outer boundary of the region of interest and at least one detected marker on the intravascular device, or determining, by the one or more processors, a spatial distance between the at least one detected marker on the intravascular device and the region of interest. The spatial distance may be, for example, a Euclidean distance or a geodesic distance. In some examples, the automatic zooming may further comprise sorting, by the one or more processors, pixel values of the real-time visualization by their intensity, normalizing pixel intensity values lower than a predetermined threshold, applying a median filter to the normalized pixel intensity values of the real-time visualization.
Another aspect of the disclosure is directed to a system comprising one or more processors. The one or more processors may be configured to receive at least one first extraluminal image, receive second extraluminal images captured during delivery of an intravascular device, detect motion features in the at least one first extraluminal image and the second extraluminal images, correlate, based the detected motion features, the at least one first extraluminal image and the second extraluminal images, and provide for output real-time visualization of a position of the intravascular device on the at least one first extraluminal image or one of the second extraluminal images including the intravascular device.
Yet another aspect of the disclosure is directed to a non-transitory computer readable medium, which when executed by one or more processors cause the one or more processors to receive at least one first extraluminal image, receive second extraluminal images captured during delivery of an intravascular device, detect motion features in the at least one first extraluminal image and the second extraluminal images, correlate, based the detected motion features, the at least one first extraluminal image and the second extraluminal images, and provide for output real-time visualization of a position of the intravascular device on the at least one first extraluminal image or one of the second extraluminal images including the intravascular device.
One aspects of the technology relates to a method comprising receiving extraluminal images captured during the delivery of an intravascular device, wherein the intravascular device has a radio-opaque marker, detecting a plurality of device marker candidates corresponds to the radio-opaque marker of the intravascular device, automatically detecting, using an artificial intelligence (AI) model, a working vessel using a plurality of virtual boxes and at least one of the plurality of device marker candidates. The at least one of the plurality of virtual boxes contains the at least one of the plurality of device marker candidates. The method further comprises selecting, by the one or more processors, at least one of the plurality of virtual boxes that includes a region of interest of the working vessel that contains the at least one of the plurality of device markers.
In some examples, the method may further comprise predicting center points of each of the plurality of virtual boxes, pairing the detected device markers with the nearest of the plurality of virtual boxes, filtering the device marker candidates that are beyond the boundaries of the plurality of virtual boxes, updating center points of each of the plurality of virtual boxes using the filtered device marker candidate points, determining the displacement of the predicted center points and updated center points and repositioning, based on the determined displacement, the plurality of virtual boxes. The updating the center point may include approximating the center point of the box using the device marker candidate. The one or more device marker candidate may be within the at least one of the plurality of boxes.
In some examples, the AI model may be trained using annotations on high dose angiographs. In some examples, the extraluminal images may be live x-ray angiographs or fluoroscopy images.
In some examples, the method may further comprise tracking the region of interest during a percutaneous coronary intervention procedure. The method may further include enhancing the region of interest using local contrast stretching by selectively enhancing the local contrast between the detected device marker candidate and surrounding regions. In some examples, the method may further comprise automatically zooming, by the one or more processors, the region of interest. The automatic zooming may comprise sorting, by the one ore more processors, pixel values of the real-time visualization by their intensity, normalizing pixel intensity values lower than a predetermined threshold, and applying a median filter to the normalized pixel intensity values of the real time visualization.
In another aspect of the disclosure is directed to a system comprising one or more processors configured to receive extraluminal images captured during the delivery of an intravascular device, wherein the intravascular device has a radio-opaque marker, detect a plurality of device marker candidates, wherein at least one of the device marker candidates corresponds to the radio-opaque marker of the intravascular device, automatically detect, using an artificial intelligence model, a working vessel using a plurality of virtual boxes and at least one of the plurality of device marker candidates, wherein at least one of the plurality of virtual boxes contains the at least one of the plurality of device marker candidates, and select at least one of the plurality of virtual boxes that includes a region of interest of the working vessel that contains the at least one of the plurality of device markers.
In yet another aspect of the disclosure is directed to a non-transitory computer-readable medium storing instructions, which when executed by one or more processors causes the one or more processors receive extraluminal images captured during the delivery of an intravascular device, wherein the intravascular device has a radio-opaque marker, detect a plurality of device marker candidates, wherein at least one of the device marker candidates corresponds to the radio-opaque marker of the intravascular device, automatically detect, using an artificial intelligence model, a working vessel using a plurality of virtual boxes and at least one of the plurality of device marker candidates, wherein at least one of the plurality of virtual boxes contains the at least one of the plurality of device marker candidates, and select at least one of the plurality of virtual boxes that includes a region of interest of the working vessel that contains the at least one of the plurality of device markers.
The technology is generally directed to dynamically visualizing an intravascular device within a blood vessel by correlating or mapping between extraluminal images. The extraluminal images may be correlated based on motion features visible or detectable within the images themselves, without the use of other sensors, such as readings from an electrocardiogram, and without the use of time stamps. The extraluminal images may include pre-contrast, high dose post-contrast, and subsequent low dose x-ray angiograms (“XA”). For example, a high dose XA may be used to generate a blood vessel map, pre-contrast XAs may be used to identify a vessel of interest or working vessel, and low dose XAs may be taken during the delivery of the device and used to track the position of the intravascular device.
High dose contrast XA images use more contrast agent during imaging as compared to low dose contrast XA images. According to some examples, a high dose XA may use 8-10 cc of a contrast agent, where a low dose XA may use 2-3 cc of the contrast agent. In some examples, a high dose XA may use at least three to five times more contrast than a low dose XA. In yet another example, a high dose XA may use at least five or more time more contrast than a low dose XA. The high dose XA images may cause the patient to have a higher radiation dose than low dose contrast XA images. High dose XA images may have longer imaging times as compared to low dose contrast XA images. The high dose XA images may have better diagnostic quality due to the increased contrast and/or higher radiation dose as compared to the low dose XA images having a lower contrast and/or lower radiation dose. As used herein, high dose contrast XA may be interchangeable with high dose XA and low dose contrast XAs may be interchangeable with low dose XA.
According to some examples, a guide wire and/or guide catheter used for delivering the device may be used to identify the working vessel. The guide wire and guide catheter may be visible in pre-contrast XAs. For example, the system can integrate artificial intelligence (“AI”) to identify the working vessel. As an example, the AI may be a AI (“ML”) model trained to segment XAs to predict output classes, including guide catheter, guide wire, catheter marker, and background. The multi-label segmentation output by the ML model may be converted into a binary mask that illustrates the working vessel region as a shape prior to the subsequent XAs. According to some examples, the initial shape prior may be obtained from a pre-contrast XA. The initial shape prior determined based on the pre-contrast XA may be used to segment the working vessel in subsequent high dose XAs. According to some examples, the AI may include a second ML model that is trained to identify the working vessel based on the shape prior determined by the first ML model. The second ML model may be a spatial temporal model. The second ML model may receive, as input, the current XA frame plus prediction information of the previous XA frames. The prediction information may be, for example, the prediction of the location of the working vessel in the given XA frame. In some examples, the second ML model may receive, as an additional input, the shape prior of the previous frame(s). The second ML model may provide, as output, the identified working vessel. According to some examples, AI may be used to segment the guide catheter and wire tip such that the locations of ending points of guide catheter and guide wire tip may be automatically detected based on the segmentation results. The AI may include other ML models similar to the first and/or second ML models.
A motion feature bank may be generated based on motion features detected in the high dose and low dose XAs. The feature bank may provide motion information from various perspectives and may be used for phase matching the low dose and high dose XAs for a given patient. The motion features may be used to perform an initial phase matching of the low dose and high dose XAs. The motion features may include a guide catheter tip, a distal endpoint of a working vessel, an optical flow at the guide catheter tip, and/or an optical flow at the distal endpoint of the working vessel. The initial phase matching may be an initial temporal phase matching. The high and low dose XAs may be temporally matched without the use of EKG or other physiological signals that measure heartbeat or breath patterns of a patient or time stamps. Rather, the temporal phase matching may be based on image motion features.
According to some examples, an update may be performed to iteratively filter the initial time shift estimate. The initial time shift estimate may be updated during an online update. For example, a filter, such as a linear Kalman filter, may be used to predict a time shift based on the more current low dose XA and the next available high dose XA. The predicted time shift may be used to cross-correlate the low and high dose XAs to determine whether the filter should be updated. The time shift determined by the online update may be used to correlate the low and high dose XAs. In some examples, a spatial fitness between spatial landmarks, such as the guide catheter, wire tip, device, or the like, identified in the high and low dose XAs may be used to evaluate and/or refine the phase matching performed using the motion features.
According to some examples, the motion features may be used to derive the heartbeat period of the patient. For example, an auto-correlation function may be applied to the motion features. The heartbeat period may be detected based on voting and non-maximum suppression.
To provide a visualization of the device during the delivery and/or treatment stage, the markers on the device may be detected and tracked. The device may include one or more radiopaque markers that are able to be detected using the high and/or low dose XA. The device may be detected and tracked by iteratively estimating the position of the marker in the XAs. In some examples, the motion of the marker within the sampling period is estimated by an optical flow method. The motion estimates may be used to predict the position of the marker in the most recent XA.
Correlating live low dose XAs with high dose XAs may reduce the amount of contrast required during PCI. For example, by identifying motion features in both the low dose and high dose XAs, the low dose XAs may be temporally and spatially correlated such that the position of the device can be tracked real time in low dose XAs relative to a vessel map generated from a high dose XA.
In some examples, correlative live low dose XAs with high dose XAs using motion features may reduce the number of sensors and/or the amount of equipment required to co-register low dose and high dose XAs. For example, by using motion features that can be detected in the low and high dose XAs, there is no longer a need to synchronize the collection of images, time stamps, or the like. Further, the use of an EKG to co-register the images based on a heartbeat cycle is negated. According to some examples, by correlating the low and high dose XAs using motion features, the amount of contrast needed during a procedure may be reduced. For example, a physician will be able to dynamically visualize the location of the device within the vessel based on the correlated low and high dose XA without having to inject additional contrast to determine where the device is within the vessel. Rather, the dynamic visualization allows for a physician to see, in real time, the location of the device on the vessel map without additional contrast.
According to some examples, by using AI to correlate the low and high dose XAs to provide dynamic visualization of the delivery of a device within a vessel, the amount of computational resources required to produce the dynamic visualization is reduced. For example, the AI may include one or more AI models. The use of one or more AI models to correlate the high and low dose XAs may reduce the amount of memory and processing power required to dynamically visualize the device as it is being delivered within the vessel.
The non-invasive imaging system 120 may be, for example, a nuclear magnetic resonance, x-ray, computer aided tomography, or other suitable non-invasive imaging technology. In particular, the non-invasive imaging system may be an angiography system configured to generate pre-contrast X-ray angiograms (XAs) before a radio opaque contrast agent has been used, high dose contrast XAs taken while a radio-opaque contrast agent is present in one or more blood vessels of interest, and low dose contrast XAs taken when little or no contrast agent is present. The high dose contrast XAs may, in some examples, be cine images, and the low dose contrast XAs may be fluoroscopic images. For example, the angiography system 120 may include a fluoroscopy system. The angiography system 120 may be configured to noninvasively image the subject S such that frames of angiography data, typically in the form of frames of image data, are generated. According to some examples, the x-ray imaging may occur while a device 104 is being delivered via the vessel such that a blood vessel in region R of subject S is imaged using angiography. In some examples, high dose XAs may be generated before the intravascular device 104 is delivered via the vessel, while the device is being delivered, and/or after the device has been delivered. The imaging results of a non-invasive scan may be provided for output on display 118. The high dose and low dose XAs may be correlated such that the device and/or treatment zone may be dynamically visualized during a procedure. As discussed above and herein, the non-invasive imaging system may be an angiography system. However, the non-invasive imaging system 120 may use various other imaging technologies and, therefore, references to the non-invasive imaging system 120 being an angiography system are not intended to be limiting.
The non-invasive imaging system 120 may be in communication with a storage device 106 via network 102. In some examples, the imaging system 120 may be in direct communication with storage device 106, e.g., without having to be connected via network 102. The storage device 106 may be a workstation or server. The data collected and/or generated by imaging system 120 may be storage and managed by the storage system 106. In some examples, a subsystem, a server or workstation handle the functions of storage device 106. According to some examples, the imaging system 120 may generate electromagnetic radiation, such as x-rays. The imaging system 120 may, in some examples, receive such radiation after passing through the subject S. In turn, data storage 106 may use the signals from the imaging system 120 to image one or more regions of the subject S including region R. In some examples, storage device 106 may be integrated with imaging device 120.
The region of interest R may be a subset of the vascular or peripherally vascular system, such as a particular blood vessel. The region R may be a region for the delivery of a device 104. The device may be, for example, a stent, balloon, or the like. In some examples, the device 104 may be an intravascular device, such as an optical coherence tomography (“OCT”) probe, an intravascular ultrasound (“IVUS”) catheter, micro-OCT probe, near infrared spectroscopy (NIRS) sensor, optical frequency domain imaging (“OFDI”), or any other device that can be used to image a blood vessel. In some examples, the device 104 may be a pressure wire, a flow meter, etc. The device 104 may include a device tip, one or more radiopaque markers, an optical fiber, a torque wire, or the like. Additionally, the device tip may include one or more data collecting subsystems such as an optical beam director, an acoustic beam director, a pressure detector sensor, other transducers or detectors, and combinations of the foregoing.
In examples where the device 104 includes an optical beam director, the optical fiber may be in optical communication with the device 104 and/or with the beam director. The torque wire may define a bore in which an optical fiber is disposed. According to some examples, the device 104 may include the sheath such as a polymer sheath (not shown) which forms part of a catheter. The optical fiber, which in the context of an intravascular system is a portion of the sample arm of an interferometer, may be optically coupled to subsystem 108 and/or a patient interface unit (PIU).
A guide wire, not shown, may be used to introduce the device 104 into the blood vessel. In examples where the device 104 is an intravascular data collection device, the device 104 may be introduced and pulled back along a length of a blood vessel while collecting data. In examples where the device 104 is a treatment device, such as a device to deliver a balloon or stent, the device 104 may be introduced to the vessel and guided to a treatment zone.
The device 104 may be connected to a subsystem 108. According to some examples, the device 104 may be connected to subsystem 108 via an optical fiber. The subsystem 108 may include a light source, such as a laser, an interferometer having a sample arm and a reference arm, various optical paths, a clock generator, photodiodes, and other OCT, IVUS, micro-OCT, NIRS, and/or pressure wire components. The device 104 may be connected to an optical receiver 110. According to some examples, the optical receiver 110 may be a balanced photodiode based system. The optical receiver 110 may be configured to receive light collected by the probe 104. The device 104 may be coupled to the optical receiver 110 via a wired or wireless connection.
The system 100 may further include, or be configured to receive data from, an non-invasive imaging system 120. The non-invasive imaging system 120 may be, for example, an imaging system based on angiography, fluoroscopy, x-ray, nuclear magnetic resonance, computer aided tomography, etc. non-invasive imaging system 120 may be configured to noninvasively image the blood vessel 102. According to some examples, the non-invasive imaging system 120 may obtain one or more images before, during, and/or after a pullback of the data collection probe 104. Non-invasive imaging system 120 may be used to image a patient such that diagnostic decisions can be made and various possible treatment options such as stent placement can be carried out. These and other imaging systems can be used to image a patient externally or internally to obtain raw data, which can include various types of image data.
The non-invasive imaging system 120 may be in communication with subsystem 108. According to some examples, the non-invasive imaging system 120 may be wirelessly coupled to subsystem 108 via network 102. For example, the non-invasive imaging system 120 may be wirelessly coupled to subsystem 108 via a communications interface, such as Wi-Fi or Bluetooth. In some examples, the non-invasive imaging system 120 may be in communication with subsystem 108 via a wire, such as an optical fiber. In yet another example, external imaging device 120 may be indirectly communicatively coupled to subsystem 108 or computing device 112. For example, the non-invasive imaging device 120 may be coupled to a separate computing device (not shown) that is in communication with computing device 112. As another example, data from the imaging system 120 may be transferred to the computing device 112 using a computer-readable storage medium, from storage device 106 via network 102, or the like.
The subsystem 108 may include a computing device 112. One or more steps may be performed automatically or without user input to navigate images, input information, select and/or interact with an input, etc. In some examples, one or more steps may be performed based on receiving a user input by mouse clicks, a keyboard, touch screen, verbal commands, etc. The computing device 112 may include one or more processors 113, memory 114, instructions 115, data 116, and one or more modules 117.
The one or more processors 113 may be any conventional processors, such as commercially available microprocessors. Alternatively, the one or more processors may be a dedicated device such as an application specific integrated circuit (ASIC) or other hardware-based processor. Although
Memory 114 may store information that is accessible by the processors, including instructions 115 that may be executed by the processors 113, and data 116. The memory 114 may be a type of memory operative to store information accessible by the processors 113, including a non-transitory computer-readable medium, or other medium that stores data that may be read with the aid of an electronic device, such as a hard-drive, memory card, read-only memory (“ROM”), random access memory (“RAM”), optical disks, as well as other write-capable and read-only memories. The subject matter disclosed herein may include different combinations of the foregoing, whereby different portions of the instructions 115 and data 116 are stored on different types of media.
Memory 114 may be retrieved, stored or modified by processors 113 in accordance with the instructions 115. For instance, although the present disclosure is not limited by a particular data structure, the data 116 may be stored in computer registers, in a relational database as a table having a plurality of different fields and records, XML documents, or flat files. The data 116 may also be formatted in a computer-readable format such as, but not limited to, binary values, ASCII or Unicode. By further way of example only, the data 116 may be stored as bitmaps comprised of pixels that are stored in compressed or uncompressed, or various image formats (e.g., JPEG), vector-based formats (e.g., SVG) or computer instructions for drawing graphics. Moreover, the data 116 may comprise information sufficient to identify the relevant information, such as numbers, descriptive text, proprietary codes, pointers, references to data stored in other memories (including other network locations) or information that is used by a function to calculate the relevant data.
The instructions 115 can be any set of instructions to be executed directly, such as machine code, or indirectly, such as scripts, by the processor 113. In that regard, the terms “instructions,” “application,” “steps,” and “programs” can be used interchangeably herein. The instructions can be stored in object code format for direct processing by the processor, or in any other computing device language including scripts or collections of independent source code modules that are interpreted on demand or compiled in advance. Functions, methods and routines of the instructions are explained in more detail below.
According to some examples, the computing device 112 may include receive, either by a wired connection or via a wireless connection, data from the device 104. The data may include, for example, intravascular data including intravascular imaging data, pressure data, temperature data, flow data, or the like. The data received by the computing device 112 from imaging system 120, storage device 106, and/or device 104 may be used to dynamically visualize the device 104 on one or more extraluminal images. In some examples, where the device is an intravascular data collection device, the data received from the device may be used to determine plaque burden, fractional flow reserve (“FFR”) measurements at one or more locations along the vessel, calcium angles, EEL detections, calcium detections, proximal frames, distal frames, EEL-based metrics, stent/no stent decisions, scores, recommendations for debulking and other procedures, evidence based recommendations informed by automatic detection of regions/features of interest, stent planning, etc.
In examples, the device 104 is an intravascular data collection device, the data obtained by device 104 and/or imaging system 120 may processed by one or more modules to provide information about the blood vessel including lumen contours, vessel diameters, vessel cross-sectional areas, lumen area, EEL values, EEL diameters, EEL arcs, lesion locations, lesion size, plaque burdens, VFR, FFR, landing zones, treatment zones, virtual stents bounded by the landing zones or the like. According to some examples, the modules 117 may include a dynamic visualization module, an EEL detection module, a lumen detection module, a lesion detection module, and a co-registration module.
The dynamic visualization module may automatically correlate high dose and low dose XA such that a vessel map may be overlaid on the live low dose XA, a treatment zone identified on the high dose XA and/or vessel map may be overlaid on the live low dose XA, and/or the device 104 may be tracked on the live low dose XA with reference to the vessel map. The dynamic visualization module may correlate the high and low dose XA based on motion features detected in both the high and low dose XA. In some examples, the dynamic visualization module may usc AI, such as one or more AI (“ML”) models, to identify the working vessel, detect markers on the device, and track the motion of the device.
The EEL detection module may automatically detect and measure the EEL diameter of a given intravascular image frame taken during a pullback of a vessel. The lumen detection module may automatically detect and measure the lumen diameter of a given intravascular image frame taken during a pullback of the vessel. The lesion detection module may automatically detect lesions within the vessel based on vessel data obtained from the device 104 and/or imaging system 120. The side branch detection module may process the vessel data obtained from the device 104 and/or imaging system 120 to detect one or more side branches of the blood vessel.
The computing device 112 may be adapted to co-register vessel data obtained during a pullback of the device 104 with intravascular image and/or an extraluminal image. For example, the computing device 112 may be configured to receive and store extraluminal image data, such as image data generated by imaging system 120 and obtained by a frame grabber. The computing device 112 may be configured to receive and store intravascular image data, such as image data generated by device 104 and obtained by the frame grabber. In some examples, computing device 112 may access co-registration module to co-register the vessel data with the luminal image. The luminal image may be an extraluminal image, such as an angiograph, x-ray, or the like. The co-registration module may co-register intravascular data, such as an intravascular image, plaque burden, EEL measurement, lumen diameter measurements, pressure readings, virtual flow reserve (“VFR”), fractional flow reserve (“FFR”), resting full-cycle ration (“RFR”), flow rates, etc. with the extraluminal image. In some examples, the co-registration module may co-register intravascular data with an intraluminal image, such as an intraluminal image captured by an OCT probe, IVUS probe, micro-OCT probe, or the like.
In one example, the co-registration module may co-register intraluminal data captured during a pullback with one or more extraluminal images. For example, the extraluminal image frames may be pre-processed. Various matrices such as convolution matrices, Hessians, and others can be applied on a per pixel basis to change the intensity, remove, or otherwise modify a given angiography image frame. As discussed herein, the preprocessing stage may enhance, modify, and/or remove features of the extraluminal images to increase the accuracy, processing speed, success rate, and other properties of subsequent processing stages. A vessel centerline may be determined and/or calculated. In some examples, the vessel centerline may be superimposed or otherwise displayed relative to the pre-processed extraluminal image. According to some examples, the vessel centerline may represent a trajectory of the device 104, such as an intravascular device, through the blood vessel during a pullback. In some examples, the centerline may be referred to as a trace. Additionally or alternatively, marker bands or radiopaque markers may be detected in the extraluminal image frames. According to some examples, the extraluminal image frames and the data received by device 104 may be co-registered based on the determined location of the marker bands.
According to some examples, the modules may additionally or alternatively include a video processing software module, a preprocessing software module, an image file size reduction software module, a catheter removal software module, a shadow removal software module, a vessel enhancement software module, a blob enhancement software module, a Laplacian of Gaussian filter or transform software module, a guide wire detection software module, an anatomic feature detection software module, stationary marker detection software module, a background subtraction module, a Frangi vesselness software module, an image intensity sampling module, a moving marker software detection module, iterative centerline testing software module, a background subtraction software module, a morphological close operation software module, a feature tracking software module, a catheter detection software module, a bottom hat filter software module, a path detection software module, a Dijkstra software module, a Viterbi software module, fast marching method based software modules, a vessel centerline generation software module, a vessel centerline tracking module software module, a Hessian software module, an intensity sampling software module, a superposition of image intensity software module and other suitable software modules as described herein. According to some examples, the modules may include software such as preprocessing software, transforms, matrices, and other software-based components that are used to process image data or respond to patient triggers to facilitate co-registration of different types of image data by other software-based components or to otherwise perform annotation of image data to generate ground truths and other software, modules, and functions suitable for implementing various features of the disclosure. The modules can include lumen detection using a scan line based or image based approach, stent detection using a scan line based or image based approach, indicator generation, apposition bar generation for stent planning, guide wire shadow indicator to prevent confusion with dissention, side branches and missing data, and others.
In some examples, the modules may be configured to process the vessel data obtained by the device 104 and/or imaging system 120 using AI algorithms, artificial intelligence, or the like.
The subsystem 108 may include a display 118 for outputting content to a user. The display 118 may be integrated with the computing device 112, or it may be a standalone unit electronically coupled to the computing device 112. The display 118 may output intravascular data relating to one or more features detected in the blood vessel and/or obtained during a pullback. For example, the output may include, without limitation, cross-sectional scan data, longitudinal scans, three-dimensional representations generated based on intraluminal and/or extraluminal images, diameter graphs, image masks, lumen border, plaque sizes, plaque circumference, visual indicia of plaque location, visual indicia of risk posed to stent expansion, flow rate, suggested treatment zones, or the like. The display 118 may identify features with text, arrows, color coding, highlighting, contour lines, or other suitable human or machine-readable indicia.
According to some examples the display 118 may include a graphic user interface (“GUI”). The display 118 may be a touchscreen display in which a user can provide an input to navigate images, input information, select and/or interact with an input, etc. In some examples, the display 118 and/or computing device 112 may include an input device, such as a trackpad, mouse, keyboard, etc. that allows a user to navigate images, input information, select and/or interact with an input, etc.
In some examples, the information input by the user may be annotations. For example, the system may be configured to receive annotations to the one or more representations of the vessel. The annotations may be, in some examples, an indication of plaque burden, fractional flow reserve (“FFR”) measurements at one or more locations along the vessel, calcium angles, EEL detections, calcium detections, proximal frames, distal frames, EEL-based metrics, stent/no stent decisions, scores, recommendations for debulking and other procedures, evidence based recommendations informed by automatic detection of regions/features of interest, stent planning, etc. In some examples, the annotations maybe a treatment device landing zone, balloon device zone, vessel prep device zone, or lesion related zone. For example, the system may receive an input corresponding to a proximal and distal location along the vessel corresponding to a proximal and distal location of a treatment device landing zone, balloon device zone, vessel prep device zone, or lesion related zone.
According to some examples, the annotations may be automatically determined by the system. For example, the system may, based on vessel data, determine one or more of plaque burden, FFR measurements at one or more locations along the vessel, calcium angles, EEL detections, calcium detections, proximal frames, distal frames, EEL-based metrics, stent/no stent decisions, scores, recommendations for debulking and other procedures, evidence based recommendations informed by automatic detection of regions/features of interest, stent planning, a treatment device landing zone, balloon device landing zone, vessel prep device zone, lesion related zone, etc. The system may automatically provide the plaque burden, FFR measurements at one or more locations along the vessel, calcium angles, EEL detections, calcium detections, proximal frames, distal frames, EEL-based metrics, stent/no stent decisions, scores, recommendations for debulking and other procedures, evidence based recommendations informed by automatic detection of regions/features of interest, stent planning, a treatment device landing zone, balloon device landing zone, vessel prep device zone, lesion related zone, etc. for output as one or more annotations on at least one of the vessel representations.
In some examples, inputs received by the system corresponding to input information and/or annotations (collectively “annotations”) may be provided for output on display 118. Annotations received with respect to one vessel representation may be provided for display on the vessel representation in which the annotations were received. In some examples, the annotations may be provided for display on a plurality of vessel representations. For example, if the system receives annotations with respect to the high dose XA, the annotations may be provided for output on one or both the high dose XA and the low dose XA. In examples where there are additional representations, such as longitudinal representations, three-dimensional representations, or the like, the annotations may be provided for output on one, some, or all the representations. According to some examples, when the system receives updated annotations on at least one of the representations, the annotations on the other representations may be updated to correspond to the updated annotations.
The display 118 alone or in combination with computing device 112 may allow for toggling between one or more viewing modes in response to user inputs. For example, a user may be able to toggle between different intravascular data, images, etc. recorded during each of the pullbacks. In some examples, the user may be able to toggle between different representations, such as a longitudinal representation, a cross-sectional representation, a three-dimensional representation, intravascular images, color images, black and white images, live images, or the like.
In some examples, the display 118, alone or in combination with computing device 112, may present one or more menus as output to the physician, and the physician may provide input in response by selecting an item from the one or more menus. For example, the menu may allow a user to show or hide various features. As another example, there may be a menu for selecting blood vessel features to display.
The output may include one or more representations of the vessel. For example, the representations may include images data, longitudinal representations, three-dimensional representations, live representations, or the like. The image data may include, for example, pre-contrast, high dose contrast, and low dose contrast XA images obtained and/or generated by the imaging system 120. In some examples, the output may include cross-sectional images obtained by the device 104 during a pullback. In some examples, the output extraluminal images may include a vessel map generated based on high dose XA and live low dose XA images obtained during delivery of a device. The low dose XA may include an indication of the location of the device 104, an indication of the treatment zone originally identified on the high dose XA, or the like. In some examples, the output may include a longitudinal representation of the vessel, a graphical representation of the vessel data, a three-dimensional representation of the vessel and/or vessel data, or the like. The representations of the vessel may provide various viewing angles and section views. In some examples, EEL positions, diameters thereof, or other EEL based parameters may be output for display relative to angular measurements, detected calcium arcs, plaque burden, or the like.
In some examples, one or more visual representations of the images may include an indication of a lesion location, lesion severity, lesion length, or the like. Additionally or alternatively, the indication of the lesion may be color coded, where each color represents the severity, length, or other measurement related to the lesion.
In some examples, the output may include candidate treatment zones. The candidate treatment zone may be, for example, a candidate stent landing zone. For example, the output may include an indication corresponding to a candidate proximal landing zone for a stent and a candidate distal landing zone for a stent. The candidate treatment zone may be determined based on the determined plaque burden, lesion locations, lesion length, or the like. The indications may be provided on any of the vessel representations, e.g., the three-dimensional representation, the longitudinal representation, the graphical representation, image data such as the external images, etc.
According to some examples, the display 118 and/or computing device 112 may be configured to receive one or more inputs corresponding to a selection on one or more representations. For example, an input may be received corresponding to a selection of an image frame on the longitudinal representation. In response, the other representations provided for output may be updated to display a corresponding indication or image frame. For example, the extraluminal image may be updated to have an indication along the vessel corresponding to the location of the image frame selected in the longitudinal representation, a circumferential indication may be provided on a three-dimensional representation corresponding to the location of the image frame selected in the longitudinal representation, the cross-sectional image frame may be updated to correspond to the image frame selected in the longitudinal representation, etc. In some examples, the vessel data associated with the selected location may be updated and provided for display.
Dynamic visualization may occur in real-time, as the low dose XAs are being captured. The dynamic visualization may allow for a physician to track the position of the device being delivered in the vessel in real time on one or both the low dose XA and a high dose XA. For example, dynamic visualization may allow for a physician to track the position of the device with respect to the vessel tree by providing an indication of the position of the device on the high dose XA, e.g., a vessel map. In some examples, dynamic visualization may allow for a treatment zone identified on a high dose XA to be superimposed on the live low dose XAs such that a physician can track the position of the device relative to the treatment zone.
To dynamically visualize the delivery of a device, first extraluminal images 202 and second extraluminal images 204 may be obtained. The first extraluminal images 202 may be high dose XAs and the second extraluminal images 204 may be low dose XAs. According to some examples, third extraluminal images may be obtained. The third extraluminal images may include one or more XAs obtained prior to the injection of contrast. An XA obtained prior to the injection of contrast may be a pre-contrast image. In some examples, the third extraluminal images may be generated, or captured, at the time the first or second first extraluminal images 202, 204 are generated. For example, a sequence of XA images may be captured prior to contrast being injected. The sequence of XAs may continue to be captured after the contrast is injected. The XAs captured prior to the contrast being injected may be pre-contrast XAs and XAs captured after contrast is injected may be post-contrast XAs. Post-contrast XAs may include low and/or high dose XAs.
The first extraluminal images 202, e.g., high dose XAs, may be used to generate a vessel map. Pre-contrast XAs may be used to identify the working vessel. The second extraluminal images 204, e.g., low dose XAs, may be used to detect and track the device.
The first and second extraluminal images 202, 204 may be correlated based on motion features. The motion features may include, for example, the guide catheter tip, distal endpoint of the working vessel, i-th sample point of the working vessel, optical flow at i-th sample point of the working vessel, translation of the working vessel, optical flow at guide catheter tip, optical flow at distal endpoint of the working vessel, the length of the wire tip, the wire tip at a regional location, the optical flow at the wire tip in the regional location, the proximal endpoint of the wire tip, the distal endpoint of the wire tip, the optical flow at the proximal endpoint of the wire tip, the optical flow at the distal endpoint of the wire tip, or the like. According to some examples, some motion features may be able to be detected in one of or both the first and second extraluminal images 202, 204. For example, the guide catheter tip, distal endpoint of the working vessel, i-th sample point of the working vessel, optical flow at i-th sample point of the working vessel, translation of the working vessel, optical flow at guide catheter tip, and optical flow at distal endpoint of the working vessel may be detected in the first extraluminal images 202. In some examples, the guide catheter tip, distal endpoint of the working vessel, optical flow at guide catheter tip, optical flow at distal endpoint of the working vessel, the length of the wire tip, the wire tip at a regional location, the optical flow at the wire tip in the regional location, the proximal endpoint of the wire tip, the distal endpoint of the wire tip, the optical flow at the proximal endpoint of the wire tip, and the optical flow at the distal endpoint of the wire tip may be detected in the second extraluminal images 204.
The motion features may be used to match the motion phase between the first and second extraluminal images 202, 204. For example, the motion features may include global vessel motion, vessel region of interest motion, and vessel pixel motion. The global vessel motion may be a two-dimensional translation vector which represents the two-dimensional translation at the t-th frame with respect to the first frame. The vessel region of interest motion may be represented by the wire tip region motion. The vessel pixel motion may generally correspond to “n” number of sampled points on the working vessel. Absolute spatial information, e.g., x and y coordinates, and relative spatial information, e.g., optical flow may be determined for the global vessel motion, vessel region of interest motion, and/or vessel pixel motion.
According to some examples, the motion features detected from the first and second extraluminal images 202, 204 may be used to spatially and temporally correlate the motion phase of the first and second extraluminal images 202, 204. The correlation of the first and second extraluminal images 202, 204 may allow for the dynamic visualization of a device within the vessel using a minimum amount of contrast agent. Typically to visualize the location of the device within the vessel, contrast must be injected to be able to identify the contours of the vessel. Accordingly, to consistently view the device during the delivery of the device within the vessel, contrast may have to be continuously injected. Large amounts of contrast may result in negative side effects. Therefore, by correlating a vessel map generated based on the first extraluminal images 202 with the live low dose XA images, e.g., the second extraluminal images 204, the amount of contrast required to track the position of the device during delivery of said device may be minimized.
To identify the working vessel, an extraluminal image frame 302 may be received. In some examples, the image frame 302 may be a pre-contrast XA. The pre-contrast image frame 302 may be captured prior to the delivery of the device. For example, the image frame 302 may be captured after the guide wire and/or guide catheter is inserted but before the device is delivered.
The guide catheter and/or guide wire within the image frame 302 may be segmented using AI. The AI may include a deep learning model, such as VNet model 304. The VNet model 304 may be trained to predict a plurality of classes, such as guide catheter, guide wire, device marker, and background. The VNet model 304 may receive, as input, the first image frame 302 and may output multi-label segmentation.
The multi-label segmentation output by the VNet model may be converted into a multi-label mask 306. The multi-label mask 306 may illustrate the working vessel region as a shape prior to the subsequent process. For example, as shown in
As shown in
Referring back to
The MaskTrack+Vnet model 310 may be trained with a plurality of annotations, as illustrated in
When executing the MaskTrack+Vnet model 310, the input into the MaskTrack+Vnet model 310 may be an image frame and prediction information from one or more previous frames. For example, if the current frame is frame “t”, the input into the MaskTrack+Vnet model 310 may be frame “t” and the prediction information from previous frames “t−n”, “t−n+1”, . . . “t−1”, when “n” is greater than 1. According to some examples, the shape prior 308 of the previous frame may be an additional input into the MaskTrack+Vnet model 310. The shape prior propagation may correspond to the “MaskTrack” of the MaskTrack+Vnet model 310. The “VNet” of the MaskTrack+Vnet model 310 is one type of encoder-decoder segmentation network. However, other segmentation networks may be used as part of the shape prior propagation workflow.
Using the shape prior 308 may allow for the MaskTrack+Vnet model 310 may learn, or predict, the anatomical context of the working vessel with respect to the whole vessel tree. According to some examples, the MaskTrack+Vnet model 310, e.g., the spatial-temporal segmentation model, may not be limited by the previous frame. Rather, in some examples, the shape prior 308 may be formulated as previous consecutive frames based on the data, or images, to be processed. The MaskTrack+Vnet model 310 may be executed with a plurality of subsequent frames 324 being used as input in addition to first frame 302. The subsequent frames 324 may be frames captured after the first frame 302 was captured. In some examples, the subsequent frames 324 may include contrast. For example, the subsequent frames 324 may be any combination of high or low dose XA, only low dose XA, or only high dose XA. Based on all the frames, e.g., first frame 302 and the subsequent frames 324, the output of the MaskTrack+Vnet model 310 may be a vessel map 316 of the working vessel.
Device marker candidate points, 2241, 2242, and 2243, shown in frame 2203 may be extrapolated from the low-dose XA image 2201. The portion of the working vessel in an extraluminal image, such as pre-contrast high-dose XA frame 2202, may be used to guide the extrapolation of device marker candidate points 2241, 2242 and 2243. In such a situation, the working vessel may serve as a shape prior. To determine the initial shape prior, a pretrained neural network may be used. Various methods, such as rigid-body registration and elastic registration, could be applied to the initial shape prior to alter its shape and increase its diversity. To detect device marker candidates, another neural network may be used. In some examples, the neural network may be a Vnet convolutional neural network. In another example, a MaskTrack+Vnet model, similar to the MaskTrack+Vnet model 310 described in
The shape prior used by a MaskTrack+Vnet model may require additional annotation. According to some examples, the shape prior for a high-dose extraluminal image may be obtained by directly annotating the device from the same image. In some examples, the shape prior for a low-dose extraluminal image may be obtained by directly annotating the intraluminal device in a pre-contrast, high-dose XA during the same PCI procedure where the low-dose XA was obtained. In other examples where there are no pre-contrast XAs, the shape prior may be directly annotated onto the working vessel in a post-contrast high-dose XA. In some examples, annotation of the working vessel in post-contrast XAs may involve another pretrained neural network, such as an SVS-Net neural network, shown in mask frame 2302A. In the example of
The trained model may be applied to low-dose XAs from the treatment stage of a PCI to detect candidate markers. In some examples, the probability map output from the neural network model may be thresholded, and candidate marker positions may be extracted by finding center points of each connected component in the thresholded image.
Referring back to
According to some examples, at least two factors may be applied to the contrast and signal-noise-ration (“SNR”) to transform high dose XAs into augmented low dose XAs 322 that can be used to train the MaskTrack+Vnet model 310, or a similar model, to segment the guide catheter and wire tip. The high dose XAs captured before delivery of a device or during a PCI workflow, may be processed to extract the neighborhood of target objects, such as the guide catheter, wire tip, or the like. In some examples, the high dose XAs may be processed using dilation to extract the neighborhood of target objects. The intensity contrast of the target objects in the high dose XAs may be determined. In some examples, a signal-noise-ration (“SNR”) may be determined based on the high dose XA and low dose subsequent frames 324. According to some examples, the augmented low dose XAs 322 may be generated using generative adversarial network (“GAN”) type models, or other generative type models.
The MaskTrack+Vnet model 310 may be trained using the augmented low dose XAs 322 as well as captured low dose XAs, such as subsequent frames 324, as input. The MaskTrack+Vnet model 310 may be trained to automatically detect the guide catheter and wire tip. The MaskTrack+Vnet model 310 may be trained to predict a segmentation of the guide catheter and wire tip. For example, when executed, the MaskTrack+Vnet model 310 may receive subsequent frames 324 and provide frame 312 as output, where frame 312 may be the segmented guide catheter and/or the wire tip.
In some examples, frame 312 may be a skeleton frame. The skeleton frame 312 may be analyzed 326 for reference point propagation. For example, as shown in
According to some examples, the frames 324 may be used to detect and track 318 the position of the marker on the device being delivered within the vessel. The markers on the device may be detected and tracked 318 using a AI model, a convolution neural network, or the like. Details pertaining to how the marker is detected and tracked is discussed in more detail with respect to
The working vessel map 316 along with the guide catheter tip location may be obtained using MaskTrack+Vnet 310 for the high dose XA, as illustrated in frame 360 in
In some examples, the region of interest may be determined using detected device marker candidates along a working vessel. A candidate region of interest (cROI) may denote a box that contains one or more device marker candidate's center points as detected by the neural network model. According to some examples, for each device marker candidate detected within a frame of low-dose XA, a cROI box may be determined with its center point set to the detected device marker candidate's center point. To reduce the number of cROI boxes without losing track of the device markers, various methods, such as a non-maximum suppression step (NMS) may be applied. In another example, the cROI boxes may be determined from the working vessel. In some examples, the working vessel may be determined in a pre-contrast, high-dose XA and overlaid on live feed of the low-dose XA.
The device marker candidate points may be the center of their associated cROIs. Initial working vessel trace points are determined from the high dose XA image's working vessel detection 2504. Given the working vessel detection 2504 from the high dose XA image used as the shape prior, a breadth first search (BFS) algorithm may be used to sort trace points along its path. Associated cROIs with minimum overlap, nearest to each detected marker in a new frame of low dose XA image and along this working vessel trace may be selected. In some examples, the device marker candidate may not align with the center of the nearest box. In this case the AI model may estimate a closest center point to the device marker candidate along the path of the vessel as the center of the box. For example, device marker candidate 2542 is nearest to box 2523. Because device marker candidate does not fall at the center of box 2523, the model may create center point 2544 nearest to the device marker candidate 2542. The Kalman filter in
In another example, a cROI or box may be determined as a region of interest, if it passes a persistency criterion. In some examples, a persistency criterion may be a check for the presence of a region of interest for more than 5 frames of low-dose XAs among the past and present 9 consecutive frames. In some examples, a cROI may be determined as a region of interest if it is distal to the guide catheter tip. In the situation where contrast is injected during the treatment phase, the pixel value distributions within a given region of interest may be compared with those from the same region of interest in past frames. In some examples, frames with contrast injection may be detected across the low-dose XA series.
One major limitation of prior methods is the need of a device-specific template to automate device marker detection and tracking. For advanced treatment techniques, such as kissing balloon, that require delivery and coordination of multiple devices, each device may need to be registered separately before use, complicating the procedure's overall workflow. This fully automated method improves real-time visualization of device markers in low-dose XAs, by detecting regions of interest containing device marker candidates within each frame of the live feed of low-dose Xas. This method is fast and robust to variation in the treatment workflow, thereby expanding the usability of device marker dynamic visualization to help physicians in complex clinical situations. This method may also be extended to other applications requiring device marker visualization, such as electrophysiology (EP), pulmonary and peripheral vascular angiography applications, or involving other imaging modalities such as MRI or IVUS in image guidance.
In block 2600, a model may estimate box center points of candidate regions of interest, based on extraluminal images. In some examples, a Kalman filter may be used to estimate these box center points. According to some examples, to reduce false positives, after the prediction step of each Kalman filter iteration, the predicted marker position may be compared to the detected marker positions 2601. In the prediction step, according to some examples, a loss function between pixel values within each box and within an adjacent box of the same size, from two subsequent frames of low dose XA images may be computed. In some examples, the loss function may be a 2-D cross correlation. In another example, the loss may be some other similarity or dissimilarity criteria used for image registration. The displacement between these two boxes with the minimized loss function, either the dissimilarity criterion or the negative of the similarity criterion, may be used to predict the motion of the box. According to another example, the displacement may be computed using an optical flow to predict the motion of the box. In some examples, the Lucas-Kanade (LK) method may be used to compute the optical flow.
In block 2602, the model may determine pairwise distances between the predicted box center points and detected device marker center points in the current frame. The model detects device marker candidates' positions on the extraluminal images.
In block 2604, the model may identify and pair each newly detected device marker center point to its nearest box. Among all the pairs, the corresponding distance may then be compared with the box width and height to determine how close a device marker is relative to the paired nearest box.
In block 2606, the model may filter out pairs with device markers outside, or with a distance larger than a predetermined threshold to its nearest box, or pairs with device markers inside the guide catheter. In the step, the model reduces the number of false positives detected by the model by removing any markers that are far from the boxes. The false positives may be a result of foreign matter or the anatomy of subject S.
In block 2608, a determination is made regarding whether the pairs with markers close to the boxes require updating or adjustments on the box center points. In some examples the detected markers may not be the natural center point of the nearest box. In some examples, multiple markers detected within a box, wherein only one marker can be the natural center point of the nearest box. This is allowed as the method may only average all displacements into one constant displacement, see block 2610 of
In block 2614, no pairs remain to be adjusted, no position update will be made to the boxes in the current frame. In block 2610, if adjustments need to be made to the box center points, an update step of the Kalman filter may be performed. In the update step, each detected marker's position may be used to update the Kalman prediction 2600 of its nearest box center position, and displacements between each markers' nearest box center points before and after the Kalman update may be averaged across all remaining pairs.
In block 2612, the model may update the box center points based on the determined average displacement.
Referring to
Referring to
XAs taken during PCI typically present mixed motion of a patient's heartbeat and breath. The motion makes it challenging to correlate live XA images with a previously captured XA image. Further, there is typically an inherent projection property that the vessel shown in the high and low dose XAs is a two-dimensional projected geometry, rather than a three-dimensional geometry. Due to such, accurately identifying physical pixel correspondences of the high and low dose XAs may be challenging and impractical. However, different pixels may present different motion patterns and the motion of different pixels can show similar patterns. The multi-scale motions features may, therefore, be used to correlate the high and low dose XAs. By using multi-scale motion features, the process for correlating high and low dose XAs may be completed without using external sensors, such as an EKF, time stamps, or time and computationally intensive methods.
To correlate the motion phase of a high dose XA and low dose XAs, image features may be identified that correspond to the mixed motion. A multiscale motion feature bank 602 may be generated. The multiscale motion feature bank may provide motion information from various perspectives and may be used for the subsequent phase matching process. The multiscale motion features may include global vessel motion, vessel region of interest motion, vessel pixel motion, or the like.
The vessel region of interest motion 704 may correspond to the wire tip level motion. The wire tip level motion may include, for example, absolute spatial information and relative spatial information. The absolute spatial information may correspond to coordinates, such as x and y coordinate. Relative spatial information may correspond to the optical flow between two adjacent frames.
The vessel pixel motion 706 may correspond to “n” number of sampled points along the working vessel. According to some examples, the sampled points may be between the guide catheter tip to the distal endpoint of the vessel. The sample points may be identified, or detected, using the VNet model 304 and/or the MaskTrack+VNet model 310. According to some examples, absolute spatial information and relative spatial information may be determined for the vessel level pixel motion.
Referring back to
The motion features may include, for example, the guide catheter tip, distal endpoint of the working vessel, i-th sample point of the working vessel, optical flow at i-th sample point of the working vessel, translation of the working vessel, optical flow at guide catheter tip, optical flow at distal endpoint of the working vessel, the length of the wire tip, the wire tip at a regional location, the optical flow at the wire tip in the regional location, the proximal endpoint of the wire tip, the distal endpoint of the wire tip, the optical flow at the proximal endpoint of the wire tip, the optical flow at the distal endpoint of the wire tip, or the like. According to some examples, some motion features may be able to be detected in one of or both the high dose and low dose XAs. For example, the guide catheter tip, distal endpoint of the working vessel, i-th sample point of the working vessel, optical flow at i-th sample point of the working vessel, translation of the working vessel, optical flow at guide catheter tip, and optical flow at distal endpoint of the working vessel may be detected in the high dose XAs. In some examples, the guide catheter tip, distal endpoint of the working vessel, optical flow at guide catheter tip, optical flow at distal endpoint of the working vessel, the length of the wire tip, the wire tip at a regional location, the optical flow at the wire tip in the regional location, the proximal endpoint of the wire tip, the distal endpoint of the wire tip, the optical flow at the proximal endpoint of the wire tip, and the optical flow at the distal endpoint of the wire tip may be detected in low dose XAs.
The motion features may form a periodic pattern when graphed. For example, as shown in
As illustrated in
After voting 1106, a NMS 1108 may be applied. The NMS 1108 may refine the heartbeat period in both the peak and valley domains. In some examples, the voting results may still present multiple peaks around a heartbeat moment, depending on the binning size. The NMS may be performed to suppress noisy votes and preserve the maximal vote in a local temporal neighborhood. As a result, the heartbeat moment and cycle can be extracted more reliably.
A filter 1110, such as a median filter or other statistical filter, may be used to identify the heartbeat period. The heartbeat period may be used when correlating the motion phases of the low and high dose XAs.
Referring back to
In block 902, extraluminal images, such as low and high dose XAs, are obtained. Motion features may be derived from the high and low dose XAs, respectively.
In block 904, the motion features may be selected from the high and low dose XAs. According to some examples, multiple motion features may be selected from both the high and low dose XAs. In some examples, the motion features may be resampled at a common frame rate. The frame rate may, in some examples, be used to estimate a time interval between incoming low dose XAs during the subsequent online update 812.
A correlation coefficient may be determined for the motion features in the high and low dose XAs. The correlation coefficient may be determined using a normalized cross correlation (“NCC”). The NCC may be determined, for example, using the following equation:
In the above equation, X and Y are the high-dose and low-dose XA motion features, Xk and Yk are the k-th samples of these motion features, and
For a given motion feature pair, samples of the low dose XA feature falling into a fixed time window may be identified. The window may be compared with all time windows of the same length from the corresponding high dose XA feature. Between each time window from the high dose XA feature and the time window from the low dose XA feature, a time shift may be identified. The time shift may, in some examples, be identified as the order of each time window of the high dose XA feature relative to the full time window of the same high dose XA feature. In some examples, the order of a time window of the high dose XA feature may be determined by the order of its last feature sample relative its full time window. The comparison may be performed in terms of the correlation coefficient in the NCC equation. The process may be performed for one or more combinations of motion features from high and low dose XAs to determine one pair of high and low dose XA features having a maximum correlation coefficient. According to some examples, the process may be performed for one, some, most, or all combinations of motion features from high and low dose XAs to determine one pair of high and low dose XA features having a maximum correlation coefficient. In some examples, the optimal time shift may be determined as the time shift between time windows of high and low dose XA features having the maximum correlation coefficient.
The NCC may be used to identify motion features in pairs of low and high dose XAs that have the maximum cross-correlation. For example, in block 906, the correlation coefficient may be compared to a threshold. If the correlation coefficient is below the threshold, the initial search may be indicated as a failure 918. If the correlation coefficient is above a threshold, the initial search may continue. According to some examples, the threshold may be between about 0.6-0.8.
In block 908, the motion features used to determine the correlation coefficient may be compared. If the motion features in the low dose and high dose XA are the same type of motion feature, the initial search may be completed, and an initial time shift 916 may be determined. If the motion features in the high and low dose XAs used to determine the correlation coefficients are different types of motion features, the initial search may proceed to block 910.
In block 910, a second motion feature of the high dose XA motion features may be selected. The selected second motion feature of the high dose XA motion features may be of the same type as the selected low dose XA motion feature. The selected first and second features of the high dose XA may be used to determine a phase shift offset caused by the mismatch in the motion feature types. Low dose XAs typically include less information as compared to high dose XAs. To account for this, the initial search 802 expands the usable feature pairs such that the success rate of the initial phase correlation is increased. The estimated time shift may, in some examples, include additional drift in examples where the high dose and low dose XA features are different. The additional drift may bias the motion phase correlation result. To account for the bias, potential drifts may be estimated, using a second NCC, between the selected high dose XA features and then offset from all time shifts estimated from the initial search 802 and online update 812.
In block 912, cross correlation, or correlation coefficient, for the selected high dose XA motion features may be determined. The cross correlation may be determined using the second NCC. The second NCC may be, in some examples, the same as the NCC equation provided above or may be a different NCC. In addition to the NCC, other metrics, such as various distance-based metrics may be used to temporally correlate the high and low dose XA, or estimate the phase shift offset between different types of motion features as in blocks 904 and 912.
In block 914, the correlation coefficient determined using the second NCC, e.g., a second correlation coefficient, may be compared to a threshold. If the second correlation coefficient is below the threshold, the initial search may be indicated as a failure 918. In such examples, a second search may be performed including a subset of features from block 904. In some examples, the second search may only use the same type of high and low dose XA features, thereby skipping blocks 910-914. If the second correlation coefficient meets or exceeds the threshold, the initial time shift 916 may be determined. At each of block 904 and 912, a time shift that maximizes the corresponding NCC may be obtained. The initial time shift may then be determined as the difference between these two time shifts, e.g., the time shift at block 904 and the time shift at block 912. According to some examples, the time shift determined from the second NCC may be used as an offset and subtracted from the time shift determined from the first NCC. In some examples, the initial time shift accepted in block 908 may be a case where the time shift from block 912 is 0. The online update 812, illustrated in
According to some examples, the high-dose XA motion features may have a finite sample size. One heart cycle from the initial search may be determined and high-dose XAs from within this cycle may be used and/or included as the map for the dynamic visualization. The starting time of the heart cycle may be determined as the initial time shift. In some examples, the length of the cycle may be determined as the heartbeat period, as described with respect to
In block 1002, the initial time shift is received. The initial time shift may be represented as “n−1”. The online update may be performed iteratively. For example, the online update may take the time shift obtained at time n−1 and estimate the time shift at time n. The “n” may denote the n-th low-dose XA motion feature sample during PCI. The initial time shift may, in some examples, correspond to n=1.
According to some examples, the maximum NCC between the low dose and high dose XA features may decrease over time, which may lead to errors in matching low and high dose XAs. In some examples, the ongoing process of low dose XA acquisition may result in the number of high dose XA available for correlation to be depleted if previous high dose XAs are not reused. To account for this, when correlating the high and low dose XAs temporally, a single heart cycle of the high dose XAs frames may be used. The low dose XAs from the same phase of distinct heart cycles may then be correlated with the same high dose XAs taken from a single heart cycle.
In block 1004, a determination may be made regarding whether the new low-dose XA is out of map cycle. For example, for each new low-dose XA sampled during PCI, if the time shift, e.g., “n−1”, is larger than the initial time shift by more than one heartbeat period, the low-dose XA may be considered out of the map cycle. If the new low dose XA is out of map cycle, then the online update may proceed to block 1006. If the new low dose XA is not out of map cycle, then the initial time shift may be input to block 1008 for filter predict.
In block 1006, the time shift may be reset by offsetting the heartbeat period from time shift “n−1”. In some examples, the time shift may be reset using other criteria. This time shift may then be input to block 1008 for filter predict.
In block 1008, a filter may be used to predict a new time shift. The filter may be, in some examples, a linear Kalman filter. The new time shift may be proposed to correlate a new low dose XA frame with the next available high dose XA frame in time order. The new low dose XA frame may be, for example, the most recently obtained low dose XA frame.
In block 1010, the sample window may be updated. In some examples, the sample window may be updated by removing the earliest low dose XA feature sample from and adding the most recent low dose XA feature sample to the last sample window. The optimal time shift may be the one that maximizes NCC and this is detailed in 1014. Time shift correction is detailed in 1018.
In block 1014, cross correlation is used to refine the time shift based on the updated sample window. For example, the time shift may be refined by moving the time window of the low dose XA motion feature until a maximum NCC is identified.
The cross correlation may generate a correlation coefficient. In block 1016, the correlation coefficient may be compared to a threshold. If the correlation coefficient is below the threshold, the time shift may be determined 1020, without updating the filter. If the correlation coefficient meets or exceeds the threshold, the filter may be updated 1018.
In block 1018, the filter may be updated when the correlation coefficient meets or exceeds the threshold. The filter may be, for example, the linear Kalman filter. According to some examples, the time shift may be updated using the following Kalman filter equation: {circumflex over (x)}n={circumflex over (x)}′n+Kn(zn−{circumflex over (x)}′n), where {circumflex over (x)}′n is the predicted new time shift (block 1008), zn is the time shift optimized by the NCC (block 1014) and Ky is the Kalman gain balancing the prediction and NCC measurement. The filter update 1018 may be used to reduce the prediction uncertainty.
In block 1020, the determined time shift may be provided after the filter has been updated.
Referring back to
Intrinsic motion states of the heart vessel in high and low dose XAs may also be captured using AI-based approaches. In some examples, these intrinsic motion states may be captured directly from XAs by using a convolutional neural network to extract deep features. These deep features may be used in addition to, as an alternative of or as refinement of the multiscale motion features to correlate the high and low dose XA images using temporal, spatial based metrics, or their combination.
According to some examples, for each low dose XA, given a correlated high dose XA, a loss function may be determined. The loss function may be, in some examples, a distance transform value. If the correlation between the high and low dose XAs is substantially accurate, the spatial loss may be a low value, as compared to a spatial loss value when the correlation between the high and low dose XAs is not accurate. A low loss value may correspond to the correlation of spatial landmarks and the working vessel. If the spatial mismatch is noticeable, e.g., the loss value is high, the correlation between the high and low dose XAs may be determined to be inaccurate.
In some examples, the system may automatically provide a correction when the correlation between the high and low dose XAs is inaccurate. For example, the system may provide for output correlated frames from the most recent accurate correlation. In some examples, the system may not show the current frame due to the inaccurate correlation. In another example, the system may automatically perform spatial registration to identify a correlated phase frame.
According to some examples, one or more landmarks along the working vessel may be identified. The landmarks may be automatically identified and/or identified based on a received input. For example, the system may receive an input via the computing device, touch screen, etc. corresponding to the selection of landmarks. The landmarks may correspond to a working zone or a device landing zone. The landing zone may correspond to coordinates relative to the catheter tip. In some examples, the distance between the landmarks and the catheter tip may be determined in each high dose XA. Pixels along the working vessel trajectory within distance range, e.g., the distance between the landmarks, may correspond to the determined landing zone.
The spatial correlation and/or temporal correlation of the high and low dose XAs may be fused 618 to provide dynamic visualization 620 of the delivery of a device within the vessel. For example, the system may provide for output the spatially and/or temporally correlated low and high dose XAs. The output may include any of a vessel map, live tracking of the location of the device, an indication of a treatment zone, an indication of the position of the device with respect to the treatment zone, or the like. In some examples, the output may include a plurality of images, videos, video stills, or the like. For example, two or more images and/or videos may be output relative to each other. One image may be a vessel map while another image may be a live image tracking the location of the device. The output may, in some examples, allow for a physician to dynamically visualize the delivery of a device within a vessel using less contrast. As dynamic visualization correlates high and low dose XAs, dynamic visualization may use less contrast as compared to having to continuously inject contrast to be able to visualize the vessel.
Device marker detection and tracking may allow for the device to be dynamically visualized during the treatment stage of PCI. The device may be represented by its distinct radiopaque markers. The detection and tracking of the markers may be performed by iteratively estimating the position of the markers and the motion state in the low dose XAs. Iteratively estimating the marker position allows for detection and tracking to be performed accurately and quickly. Moreover, the process can recover from errors quickly.
According to some examples, the device may be a balloon device. However, the device may be any device, such as an imaging probe, pressure wire, stent delivery device, a vessel prep device, such as an atherectomy device or lithotripsy device, or the like. Therefore, the example provided herein with respect to a balloon device is just one example and is not intended to be limiting. The balloon device provided in this example may be replaced with any device having radiopaque markers or markers that are visible in XAs.
To predict the position of the marker in the respective low dose XA, the motion of the marker in the most recent sampling period may be estimated using an optical flow estimate 1208. Optical flow assumes the local intensity of moving objects in an image to remain approximately constant for at least a short time duration. According to some examples, the optical flow equation may be: I(x,t)≠I(x+δx, t+δt), where δx is the displacement of local image region at location x from time t to t+δt. The aforementioned equation may be applied to each pair of low dose XAs sampled at an interval of δt. By applying the optical flow equation for each detected marker's position, δx may be determined and used as the motion of the marker. According to some examples, δx may be determined from the optical flow equation using the iterative Lucas-Kanade (ILK) method. In some examples, δx may contain a horizontal motion component u and a vertical motion component v. The motion estimates may be used to predict the position of the marker in the most recent low dose XA. According to some examples, a filter, such as a linear Kalman filter 1210, may be used to predict the position of the marker based on the motion estimate.
To update the position of the marker in the respective low dose XA, a neural network, such as a V-Net convolution neural network, may be used to detect marker candidates from the same low dose XA. The V-Net marker detector 1204 may be trained using annotated XAs. The XAs may be annotated to identify marker positions of devices within the XA. According to some examples, the V-Net marker detector 1204 may be trained based on annotated marker positions of an intraluminal imaging probe, such as an OCT probe, in XAs. Additionally or alternatively, the marker detector 1204 may be trained based on annotated marker positions of another device, such as a balloon delivery device, in XAs. Training the marker detector 1204 based on marker positions of any type of device, such as an OCT probe, may allow for the marker detector 1204 to detect markers from a different type of device, such as a balloon delivery device. According to some examples, the high and low dose XAs in the training data may be separated to adopt a transfer learning strategy. The marker detector 1205 may be trained to predict a probability map measuring the membership of each pixel to the marker class. According to some examples, the V-Net model may predict each position on the low dose XA image's probability of being from the marker. The “membership” may correspond to whether a position on the XA image is from the marker or not. The higher the probability value (over a scale of 0-1), the more likely the position on the XA is to be from the marker.
According to some examples, when executing the marker detector 1204, the marker detector 1204 may receive low dose XAs 1202 as input. The marker detector 1204 may output a probability map measuring the membership of each pixel to the marker class. According to some examples, the output of the marker detector 1204 may be processed to integrate the model inference into the Kalman filter 1210.
The processing may include detecting center points 1206. The center points may represent the detected markers. For example, the probability map output by the marker detector 1204 may be thresholded to segment candidate marker regions. Center points of each candidate region may be extracted. The extracted center points of each region may correspond to a detected marker.
According to some examples, to reduce false positives, after the prediction step of each Kalman filter iteration, the predicted marker position 1302 may be compared to the detected marker positions 1304 determined by the marker detector 1204. The comparisons may be assigned a unique candidate marker position. According to some examples, to determine the assignment of the marker position, a pairwise distance 1306 may be evaluated based on the set of predicted and detected marker positions 1302, 1304. The pairwise distance may be, in some examples, an Euclidean pairwise distance.
The predicted position may then be sorted 1308 in ascending order based on their distances to the nearest detected positions. A filter update, such as a Kalman filter update, may be iteratively performed in the sorted order 1310. The filter update is performed on each predicted position with the nearest detected position. To avoid conflict, assigned detected marker positions and those in a surrounding region may not be considered for new assignments.
If the distance between a predicted marker position and its nearest detected marker position is greater than a threshold distance 1312, no further updates to the position may be performed 1318. If the distance between a predicted marker position and its nearest detected marker position is less than the threshold distance 1312, the neighboring, or nearby, positions may be cleared 1314. The filter may be updated 1316 based on the cleared neighbors. The marker position 1318 may then be determined.
According to some examples, a bounding box 1406 may be provided for output. The bounding box 1406 may be sized based on the type of device. The bounding box 1406 may define the region of the device. The region of the device may be determined based on the detected marker positions. The markers 1408 detected in the real-time low dose XA 1402 may be superimposed on the high dose XA 1404 to provide for dynamic visualization of the location of the device within the working vessel.
The workflow 1700 may be, for example, used as part of an extraluminal image driven PCI. The extraluminal images may be angiography images. The workflow 1700 may begin by obtaining pre-contrast high dose extraluminal image(s) 1702 and post-contrast high dose extraluminal image(s) 1704. Pre-contrast high dose extraluminal images may be, for example, extraluminal images obtained, or captured, prior to contrast being injected into the vessel. Post-contrast high dose extraluminal images may be, for example, extraluminal images obtained, or captured, after contrast is injected into the vessel. Post contrast high dose extraluminal images may be captured when the contrast is in the vessel being imaged.
According to some examples, the workflow 1700 may include obtaining intravascular data 1722. Intravascular data 1722 may be obtained using an intravascular device, such as an OCT probe, IVUS catheter, micro-OCT probe, NIRS sensor, OFDI, pressure wire, flow meter, or the like. The intravascular data 1722 may include, for example, intravascular images, pressure measurements, flow measurements, etc.
The intravascular data 1722 may be co-registered 1724 with the extraluminal images, such as pre- and/or post-contrast high dose extraluminal images 1702, 1704.
According to some examples, if intravascular data 1722 is not co-registered with extraluminal images, the landing zone may be determined 1726. According to some examples, the landing zone 1726 may be determined by a user, such as a physician. In some examples, an occlusion or other disease-related features may be identified in the post-contrast angiogram. The occlusion or disease-related feature may, in some examples, be automatically identified based on vessel data captured by an intraluminal and/or extraluminal device. For example, an occlusion or disease-related feature may be identified using a constrained contrast flow in the post-contrast extraluminal image. In some examples, the system may receive an input via the GUI corresponding to the identification of the occlusion or disease related feature. The identified occlusion or other disease-related feature may be when determining a landing zone.
The co-registered extraluminal images and intravascular data 1724 and/or the landing zone determination 1726 may be used to determine information 1728 related to lesions, treatments, etc. The information related to the lesions may include, for example, a calcification area, lipidic area, plaque length, plaque location, percent stenosis, etc. The information related to a treatment may include, for example, identifying one or more candidate landing zones for a stent, balloon, vessel prep device, or the like. The information 1728 may, in some examples, include plaque burden, fractional flow reserve (“FFR”) measurements at one or more locations along the vessel, calcium angles, EEL detections, calcium detections, proximal frames, distal frames, EEL-based metrics, stent decisions, scores, recommendations for debulking, recommendations for subsequent procedures, or the like.
A deep learning model 1706 may be applied to the post-contrast high dose extraluminal images to obtain the vessel tree morphology. The deep learning model may be, for example, an SVS-Net convolution neural network. The deep learning model may be used to segment the vessel tree from the high dose extraluminal images. The vessel tree may include, for example, the working vessel, side branches, and any other vessel of the like. The deep learning model 1706 may be trained using annotated extraluminal images. In some examples, the annotations may be done automatically by using another pre-trained neutral network. The pre-trained neural network may be, for example, the same SVS-Net neural network configured for a different software or hardware.
The deep learning model 1706 may be trained to predict a probability map measuring the membership of each pixel to the vessel tree class. Membership may, for example, correspond to whether a position on the extraluminal image is from the vessel tree or not. For example, the deep learning model may predict the probability of each position on the high dose extraluminal image as being from the vessel tree. According to some examples, a greater probability value indicates that the position on the high dose extraluminal image is more likely to be from the vessel tree. In contrast, a lower probability value may indicate that it is less likely for the position on the high dose extraluminal image to be from the vessel tree. The scale of probability values may be, for example, zero to one (0-1). In such an example, a greater probability value may be a value closer to one (1) whereas a lower probability value may be a value closer to zero (0).
According to some examples, the scale can be from 0-100, 12-22, etc. Additionally or alternatively, a lower probability value may indicate that that the position is more likely to be from the vessel tree and a higher probably value may indicate that the position is less likely to be from the vessel tree. Accordingly, a scale of 0-1 with a higher probability indicating that the position is more likely and a lower probability indicating that the position is less likely is just one example and is not intended to be limiting.
The vessel tree may be filtered. For example, the vessel tree may be filtered by prior knowledge of the vessel morphology. The vessel morphology may be probed from pre-contrast extraluminal images and propagated to subsequent post-contrast extraluminal images. In some examples, a deep learning model, such as the MaskTrack+Vnet model 310, may be used to propagate the vessel information.
The output of the deep learning model 1706 may be, for example, a segmented vessel tree. The segmented vessel tree may exclude the regions from the guide catheter. In some examples, the segmented vessel tree may include the regions from the intraluminal imaging catheter. The regions from the intraluminal imaging catheter may be detected from the pre-contrast extraluminal images. The regions may include, for example, the portion of the image frame that include the imaging catheter but not the guide catheter. According to some examples, the output of the deep learning model 1706, e.g., the segmented vessel tree, may be used for dynamic visualization.
The heartbeat cycle 1708 may be determined. The high dose extraluminal sequence obtained prior to the insertion of device and/or prior to PCI may be cached for fast access during the treatment phase. The treatment phase may include, for example, the delivery of a device in the vessel, such as a stent, balloon, or the like. The sequence of extraluminal images may cover at least one cardiac cycle. The length of the cardiac cycle may be determined using heartbeat estimation. An example of heartbeat estimation is described with respect to
According to some examples, heartbeat estimation may use a multi-scale feature bank, such as multiscale motion features 1102. The multi-scale feature bank may include motion features at the working vessel level 702, wire tip level 704, vessel pixel level 706, or any other levels manifesting the heartbeat. In some examples, the motion features may be absolute and relative motion of the working vessel, guide catheter tip, distal wire tip, intermediate points along the working vessel, or the like. The absolute and relative motion features may correspond to, for example, the coordinates and the optical flow of a point in the high dose extraluminal image and/or low dose extraluminal images.
An ACF, such as the ACF 1104 described in
As part of the workflow 1700, low dose extraluminal images 1710 may be obtained. Low dose extraluminal images 1710 may be captured during the delivery of an intravascular device. The low dose extraluminal images 1710 may be, for example, fluoroscopic images. According to some examples, the low dose extraluminal images 1710 may substantially correspond to frames 324 in
The low dose extraluminal images 1710 may be used to detect the device marker 1712. The device may be, for example, a balloon device, an imaging probe, a pressure wire, a stent delivery device, a vessel prep device, such as an atherectomy device or lithotripsy device, or the like. The device may include radiopaque markers or markers that are visible in extraluminal images.
The markers on the device may be detected 1712 using artificial intelligence, such as a AI model, a convolution neural network, or the like. For example, a V-Net convolutional neural network, the MaskTrack+Vnet model, or the like may be used to detect the marker candidates from the same low dose extraluminal images. According to some examples, the V-Net marker detector may be trained based on annotated marker positions of an intraluminal device, such and an intraluminal imaging probe or balloon delivery device, in extraluminal images. The high and low dose extraluminal images in the training data may be separated to adopt a transfer learning strategy.
The marker detector, e.g., the V-Net model, may be trained to predict a probability map. According to some examples, the marker detector may be trained to predict the probability of the position on the low dose extraluminal image is a marker. In some examples, the probability map may be thresholded to segment the device markers. Thresholding may include, for example, segmenting the pixels inside the device markers based on the image's probability map. For example, the marker detector may predict a map of each pixel's probability of being inside the device marker. The device marker center point coordinates may, in some examples, be determined during post-processing of the image based on the thresholded probability map. For example, the device marker center coordinates may be determined by identifying connected regions and detecting the center points of each region. The marker positions may be propagated from previous frames using the motion estimates of the markers. In some examples, the motion may be determined from optical flow. According to some examples, the propagated device marker positions from previous frames may be updated by the neural network-detected marker positions in the current frame. For example, a linear Kalman filter may be used to update the device marker positions. Details pertaining to how the marker is detected and tracked are discussed in more detail with respect to
According to some examples, the detected markers 1712 may be enhanced. For example, the detected markers may be enhanced to increase the visualization of the markers on the low and/or high dose extraluminal images. Enhancing the markers may increase the visualization of the markers by differentiating the markers in terms of contrast, color, etc. as compared to surrounding image features. This may make it easier for the user to identify the markers and focus on a given area of the high and/or low dose extraluminal images provided for output as part of the dynamic visualization.
The detected markers may be enhanced using local image processing. In some examples, the local image processing may be through local contrast stretching. Various intensity transformation-based methods may be applied to stretch local image contrast.
In examples where intravascular data 1722 is obtained and information 1728 is determined, the information 1728, heartbeat cycle 1708, and detected device markers 1712 may be phase matched to provide dynamic visualization 1714.
Phase matching may include, for example, matching the phase of the high dose extraluminal images 1702, 1704 and the low dose extraluminal images 1710. The phase matching may be done using motion features. For example, the high and low dose extraluminal images may be temporally phase matched without the use of EKG or other physiological signals that measure heartbeat or breath patterns of a patient or time stamps. Rather, the temporal phase matching may be based on image motion features. According to some examples, the high and low dose extraluminal images may be spatially phase matched. Details pertaining to how the high and low dose extraluminal images are phased matched are discussed in more detail with respect to
Dynamic visualization may include, for example, providing as output a road map image relative to the low dose extraluminal image. The road map image may be, for example, a vessel map. The detected markers may, in some examples, be provided for output on the vessel map and/or the low dose extraluminal image to provide an indication of the location of the device in real time, without having to inject additional contrast. According to some examples, the information 1728 may be provided for output on or relative to the vessel map and/or low dose extraluminal image. For example, a treatment zone, initially identified on the vessel map, may be provided for display on the low dose extraluminal image such that the location of the device on the low dose extraluminal image can be seen relative to the treatment zone. In some examples, the marker may be provided for output on the vessel map to provide an indication of the location of the device in real time with respect to the treatment zone.
According to some examples, the temporal phase matching may have to be tuned. Tuning the phase matching may include, for example, adjusting, selecting, or matching another high dose extraluminal image with the low dose extraluminal image. Tuning the temporal synchronization 1716 may account for errors that may occur during automatic temporal phase matching. For example, if the patient has an irregular heartbeat, goes into an irregular rhythm during the procedure, or the like, tuning for temporal synchronization 1716 may be used to account for these differences.
In frame 1804, the temporal synchronization of the vessel tree and the low dose extraluminal image was tuned. For example, the selected vessel tree may be two frames ahead of the vessel tree from the phase matching 1714. In frame 1806, the selected vessel tree may be five frames ahead of the vessel tree from phase matching 1714.
According to some examples, the tuning may be done by the user, such as a physician. The physician may select a vessel tree from a heartbeat cycle, a specific point in the heartbeat cycle, etc. In some examples, a vessel tree may be chosen as a static image such that the vessel tree provided for output does not change regardless of the portion of the heartbeat cycle. In such an example, the low dose extraluminal image, e.g., the live fluoroscopic image, may align with the static vessel tree during at least a portion of the heartbeat cycle. In some examples, the tuning 1716 may allow for the static vessel tree to be adjusted or changed. According to some examples, the tuning 1716 may be done using various user control interfaces, such as a knob control, keyboard, joystick, touch screen, or the like.
Referring to
To improve the temporal synchronization of the example of
In
Referring back to
According to some examples, the output may be screen-captured and/or zoomed in 1720. The output may be automatically screen captured when the device reaches a region of interest. For example, the detected markers 1712 may be used to determine the location of the device relative to a region of interest. The region of interest may be, for example, a treatment device landing zone, a balloon device zone, a vessel prep device zone, or a lesion related zone. In some examples, the region of interest may be a selected portion of the vessel, a vessel line, or the like.
When the detected marker is within a threshold distance of the region of interest, the system may automatically screen-capture the output. The threshold distance may be determined based on a number of pixels between an outer boundary of the region of interest and the detected marker of the device. In some examples, the threshold distance may be determined based on a spatial distance, such as Euclidean distance or geodesic distance, between the detected markers and the region of interest.
According to some examples, the user, such as the physician, may provide an input to screen-capture the output. The screen-capture may be provided for output. According to some examples, the user may confirm the location of the device based on the screen-capture without having to inject additional contrast agents.
The dynamic visualization and/or the screen-capture may be manually or automatically zoomed in to display a high-resolution view of a region of interest. The region of interest may, in some examples, correspond to the location of the device, a treatment zone, or the like. According to some examples, the output of the dynamic visualization and/or the output of the screen-capture may automatically zoom in on the location of the device. In one example, the output of the dynamic visualization may automatically zoom in to focus on the device as the device approaches the treatment zone. The spatial distance between the detected markers and the treatment zone may be used to determine whether the device is approaching the treatment zone and/or has reached the treatment zone. In some examples, a number of pixels between the outer boundary of the treatment zone and the detected marker may be used to determine whether the device is approaching the treatment zone and/or has reached the treatment zone.
In some examples, a selected portion of the output of the dynamic visualization and or the output of the screen-capture, such as a region of interest, may be enhanced using local contrast stretching. Local contrast between the device markers and background may be stretched within the detected and tracked regions of interest. By selectively enhancing the local contrast between markers and surrounding regions, the model reserved computational resources as left computational resources are expended compared to enhancing the entire working vessel.
According to some examples, the pixel values within a region of interest may be populated and sorted into an approximate representation, such as a data graph or a histogram. In some examples, a threshold t on the intensity may be set by a user or by the AI model. According to some examples, the threshold t may be the pixel value corresponding to the 1.5-th percentile in the data representation. The choice of 1.5-th percentile as the threshold t is not fixed. Rather, threshold t may be varied to ensure that it stays above pixel values of the device marker. In some examples, pixels with intensity values lower than threshold t may be normalized within the interval [0, t]. A median filter may be applied to the intensity-stretched region of interest to smooth out noise exhibiting similar pixel values to that of the device markers. According to some examples, the filter size may be set to 3×3.
To ensure consistency in the background pixel value distributions inside and outside a given region of interest, the local contrast stretch may be further localized to the vicinity of the device markers. According to some examples, for each region of interest, the pixel values before and after the local contrast stretch may be cached, and the difference may be applied in a spatial-dependent manner as follows:
where I0 and I′ are the pixel values from the same region of interest before and after the local contrast stretch, and W is a spatial weighting function. According to some examples, W may be a Gaussian kernel centered at the device marker's center point in reference to the region of interest. In case multiple markers are present within a region of interest, W may be a mixture of Gaussian kernels each centered at one marker's center point. According to some examples, W may be normalized to [0, 1] to selectively apply the pixel-level intensity changes within the region of interest.
The high dose extraluminal images in the selected sequence may be phase matched, using phase match 1714 and/or tuning 1716, to the low dose extraluminal images based on their temporal order. When the selected high dose extraluminal images are used once for phase matching, the system may loop back to the first high dose extraluminal image to reset phase matching 2106. Phase matching of incoming low dose extraluminal images, e.g., live fluoroscopic images, may be repeated on the same sequence of high dose extraluminal images.
According to some examples, the system may perform a treatment zone check 2108. Treatment zone check 2108 may determine whether the detected device markers 1712 are close to a predefined treatment zone, such as a stent landing zone or a balloon landing zone. In some examples, the treatment zone check 2108 may be based on a spatial distance, such as a Euclidean distance, geodesic distance, pixel distance, or the like. If the determined distance is within a threshold distance of the treatment zone, an output may be provided. In some examples, the output may be a notification or alert. The notification may be a visual, audible, or haptic notification. The notification may indicate the closeness of the device to the treatment zone.
According to some examples, the system may detect when the low dose extraluminal images are not updated. The low dose extraluminal images may not be updated due to an interruption in image acquisition. In examples where the low dose extraluminal images are not updated, the system may enter replay mode 2110. Replay mode 2110 may include a replay period. The replay period may be, for example, the number of frames between the current low dose extraluminal image frame and the nearest previous low dose extraluminal image frame it repeats. The nearest previous low dose extraluminal image may be, for example, the most recent previously captured low dose extraluminal image frame. In some examples, the nearest previous low dose extraluminal image may be the nearest low dose extraluminal image frame based on distance, such as based on the image capture position. According to some examples, the elements of the dynamic visualization, e.g., treatment zones, device locations, etc., may be provided for output during replay mode 2110.
The system may compare the current low dose extraluminal frame with previous low dose extraluminal image frames in a given sequence. When the current low dose extraluminal image frame corresponds to at least one of the low dose extraluminal image frames in the sequence, the system may remain in replay mode 2110. The system may exit replay mode 2110 when a low dose extraluminal image frame is different from the previous low dose extraluminal frames in the sequence. In such an example, once the system exits replay mode 2110, the system may return to online matching 2100.
In block 1610, at least one first extraluminal image may be received. The at least one first extraluminal image may be, for example, a high dose contrast XA. According to some examples, a vessel map may be generated based on the first extraluminal image.
According to some examples, the vessel map may include an indication of the working vessel. The working vessel may be automatically detected using AI trained to automatically detect the working vessel. Training the AI may comprise annotating pre-contrast extraluminal images as line strips following a trajectory of at least one of a guide wire or a guide catheter, labeling a path of the working vessel in post-contrast extraluminal images, providing the annotated pre-contrast extraluminal images and labeled post-contrast extra luminal images as training data to the AI model, and training the AI model to predict a vessel trajectory. According to some examples, the AI may include a AI model trained to automatically detect the working vessel.
The AI model may be executed to predict, or identify, the working vessel trajectory. For example, at least one pre-contrast extraluminal image may be received as input into the AI model. The AI model, based on the at least one pre-contrast extraluminal image, may detect a guide wire of the intravascular device. The AI model, based on the detected guide wire, may propagate wire information. The AI model may automatically predict, based on the propagated wire information, a working vessel trajectory.
In block 1620, second extraluminal images captured during delivery of an intravascular device may be received. The second extraluminal images may be, for example, low dose contrast XAs. The intravascular device may be, for example, a stent delivery device, a balloon device, an intravascular imaging probe, or a pressure wire.
The at least one first extraluminal image may be captured, or generated, at a different time than the second extraluminal images. For example, the at least one first extraluminal image may be generated during the diagnosis stage, the procedure planning stage, or the like whereas the second extraluminal images may be captured during the procedure, e.g., the delivery of the device within the vessel.
In block 1630, motion features in the at least one first extraluminal image and the second extraluminal images may be detected. The motion features may include at least one of a guide catheter tip, a distal endpoint of a working vessel, an optical flow at the guide catheter tip, or an optical flow at the distal endpoint of the working vessel. According to some examples, detecting the motion features may further comprise automatically detecting, by executing a AI model, the working vessel. In some examples, a heartbeat period of a patient may be determined based on the detected motion features.
In block 1640, the at least one first extraluminal image and the second extraluminal images may be correlated based on the detected motion features. Correlating the at least one first extraluminal image and the second extraluminal image may comprise determining vessel level motion, wire tip level motion, and vessel pixel level motion. Determining vessel level motion may comprise determining a two-dimensional translation vector. The two-dimensional translation vector may correspond to two-dimensional translation at an n-th frame with respect to a first frame. Determining the wire tip level motion may comprise determining absolute spatial information and relative spatial information. Absolute spatial information may correspond to coordinates and relative spatial information may correspond to optical flow between adjacent image frames. Determining vessel pixel level motion may comprise determining absolute spatial information and relative spatial information.
According to some examples, a spatial-temporal phase match between the at least one first extraluminal image and at least one of the second extraluminal images may be determined based on the detected motion features. For example, the detected motion features may be resampled at a common frame rate. A maximum correlation coefficient for pairs of motion features in the at least one first extraluminal image and the second extraluminal images may be determined. A time shift may be determined based on the maximum correlation coefficient. In some examples, a drift between the detected motion features in the at least one first extraluminal image may be determined. The time shift may be adjusted based on the determined drift. According to some examples, the time shift may be iteratively filtered. The time shift may be updated to a corrected time shift based on the iteratively filtered time shift.
According to some examples, the spatial-temporal phase match may be tuned. For example, another first extraluminal image different than the at least one of the first extraluminal images may be identified. The spatial-temporal phase match may be tuned based on the other first extraluminal image. The other first extraluminal image may be, for example, another first extraluminal images from the sequence of first extraluminal images. In such an example, tuning the spatial-temporal phase match may include shifting the sequence of the first extraluminal images based on the identified another first extraluminal images. The real-time visualization may be updated based on the other first extraluminal image.
In block 1650, a real-time visualization of a location of the intravascular device may be provided for output on the at least one first extraluminal image one of the second extraluminal images including the intravascular device.
According to some examples, the intravascular device may be detected in the second extraluminal images. Detecting the intravascular device may include, for example, detecting one or more markers of the intravascular device. The intravascular device and/or the markers of the intravascular device may be detected by executing a AI model. The AI model may be trained by providing, as input, a co-registration dataset comprising a plurality of intraluminal images and extraluminal images. The intraluminal and extraluminal images may be annotated images. The annotated images may include annotations identifying one or more intravascular device markers. The AI model may be trained to predict a position of the intravascular device.
According to some examples, detecting the intravascular device may include detecting an optical flow of the intravascular device. The position of the intravascular device in a first frame of the second intraluminal images may be determined based on the optical flow. The position of the intravascular device in a subsequent frame may be determined based on the optical flow. In some examples, the position of the intravascular device in the subsequent frame may be further based on the position of the intravascular device in the first frame.
In some examples, detecting the intravascular device may include detecting an optical flow of the intravascular device. The position of the intravascular device in a first frame of the second extraluminal images may be determined by executing a AI model based on at least a first frame of the second extraluminal images. The position of the intravascular device in a subsequent frame of the extraluminal images may be predicted based on the detected optical flow and the position of the intravascular device in a previous frame of the second extraluminal images. The position of the intravascular device in the subsequent frame of the second extraluminal images may be updated by executing the AI model based on at least a subsequent frame of the second extraluminal images and the predicted position of the intravascular device in a subsequent frame of the second extraluminal images.
In some examples, the output may include an indication of a treatment zone on at least one of the second extraluminal images or the at least one first extraluminal image. The treatment zone may be, for example, a stent landing zone or a balloon landing zone.
According to some examples, the system may provide, as output, a notification. The notification may be provided when the intravascular device is inserted prior to, approaching, within, or past a treatment zone.
According to some examples, a screen capture of the real-time visualization of the position of the intravascular device may be captured. The screen capture may, in some examples, be automatically captured when the intravascular device is within a threshold distance of a region of interest. The region of interest may be a treatment zone. The treatment zone may be at least one of a treatment device landing zone, a balloon device zone, a vessel prep device zone, or a lesion related zone. According to some examples, the region of interest may be automatically determined based on intravascular data captured by one or more extraluminal and/or intraluminal devices. For example, a treatment zone, such as a stent landing zone, may be automatically determined based on one or more EEL values, calcium values, plaque burden, vessel diameter values, or the like. In some examples, the region of interest may be received as an input. For example, the system may receive an input corresponding to a selection of a region of interest. The input may be, for example, a user, such as a physician, selecting locations on or along the vessel via the user interface.
Determining the threshold distance may include at least one of determining a number of pixels between an outer boundary of the region of interest and at least one detected marker on the intravascular device or determining a spatial distance between the at least one detected marker on the intravascular device and the region of interest. The spatial distance may be a Euclidean distance, a geodesic distance, etc.
According to some examples, a zoom feature may be implemented to enlarge a portion of the real-time visualization. The portion of the real-time visualization may be automatically zoomed when the intravascular device is within a threshold distance of a region of interest. The portion of the real-time visualization may correspond to the region of interest. The region of interest may be, for example, a treatment zone or a location of the intravascular device. The treatment zone may be at least one of a treatment device landing zone, a balloon device zone, a vessel prep device zone, or a lesion related zone. Determining the threshold distance may include at least one of determining a number of pixels between an outer boundary of the region of interest and at least one detected marker on the intravascular device or determining a spatial distance between the at least one detected marker on the intravascular device and the region of interest. The spatial distance may be a Euclidean distance or a geodesic distance.
The systems and methods described above and herein provide for motion stabilized treatment visualization and PCI navigation, e.g., dynamic visualization. The dynamic visualization may be implemented at different zoom-in scales, allowing for physicians to view as much or as little detail as necessary. In addition, the dynamic visualization may also be implemented at different instances during a procedure. In some examples, the system may receive an input corresponding to a request for dynamic visualization for a threshold period of time, e.g., a few seconds. The dynamic visualization may, in some examples, be used as a replacement to injecting additional contrast in order. For example, the dynamic visualization may correspond to a virtual puff that simulates and, therefore, replaces a puff that injects small amount of contrast to visualize vessels. In such virtual puff scenarios, the system would include at least some, if not all, processing steps as described above and herein to detect landmarks in low dose XAs, calculate motion features, and perform phase matching. The vessels, lesions, devices, landing zones and relevant information from the phase-matched high dose XAs are then displayed in low dose XA views. The system can activate the virtual puff in response to an input. For example, the system may include inputs that are activated by the physician, such as a button, a computer mouse, a joystick, or a virtual injector. In some examples, the virtual puff can also be triggered by following standard puff steps/devices, but without the injection or use real contrast. The period of visualization can be estimated based on look up table approaches. For example, if a predetermined amount of contrast (e.g., X ml of contrast) can last a threshold period of time (e.g., Y seconds), the visualization system can be configured to provide dynamic information for the threshold period of time (e.g., Y seconds) and then fade out. Further, the output provided by the dynamic visualization allows the intravascular device to be detected, tracked, and projected on a vessel map. This allows for physicians to be able to clearly visualize the vessel morphology while delivering an intravascular device while taking only low dose XAs. Moreover, dynamic visualization allows for the efficient and almost fully automatic fusion of information of low and high dose XAs.
The aspects, features, and examples of the disclosure are to be considered illustrative in all respects and are not intended to limit the disclosure, the scope of which is defined only by the claims. Other examples, modifications, and usages will be apparent to those skilled in the art without departing from the spirit and scope of the claimed invention.
Throughout the application, where compositions are described as having, including, or comprising specific components, or where processes are described as having, including or comprising specific process steps, it is contemplated that compositions of the present teachings also consist essentially of, or consist of, the recited components, and that the processes of the present teachings also consist essentially of, or consist of, the recited process steps.
In the application, where an element or component is said to be included in and/or selected from a list of recited elements or components, it should be understood that the element or component can be any one of the recited elements or components and can be selected from a group consisting of two or more of the recited elements or components. Further, it should be understood that elements and/or features of a composition, an apparatus, or a method described herein can be combined in a variety of ways without departing from the spirit and scope of the present teachings, whether explicit or implicit herein.
The use of the terms “include,” “includes,” “including,” “have,” “has,” or “having” should be generally understood as open-ended and non-limiting unless specifically stated otherwise.
The use of the singular herein includes the plural (and vice versa) unless specifically stated otherwise. Moreover, the singular forms “a,” “an,” and “the” include plural forms unless the context clearly dictates otherwise. In addition, where the use of the term “about” is before a quantitative value, the present teachings also include the specific quantitative value itself, unless specifically stated otherwise. As used herein, the term “about” refers to a ±10% variation from the nominal value. All numerical values and ranges disclosed herein are deemed to include “about” before each value.
It should be understood that the order of steps or order for performing certain actions is immaterial so long as the present teachings remain operable. Moreover, two or more steps or actions may be conducted simultaneously.
Where a range or list of values is provided, each intervening value between the upper and lower limits of that range or list of values is individually contemplated and is encompassed within the invention as if each value were specifically enumerated herein. In addition, smaller ranges between and including the upper and lower limits of a given range are contemplated and encompassed within the invention. The listing of exemplary values or ranges is not a disclaimer of other values or ranges between and including the upper and lower limits of a given range.
The present application is a continuation of International Application No. PCT/US2024/039350, filed Jul. 24, 2024, which claims the benefit of the filing date of U.S. Provisional Patent Application No. 63/531,345, filed Aug. 8, 2023, U.S. Provisional Patent Application No. 63/603,931, filed Nov. 29, 2023, and U.S. Provisional Patent Application No. 63/659,536, filed Jun. 13, 2024, the disclosures of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
63659536 | Jun 2024 | US | |
63603931 | Nov 2023 | US | |
63531345 | Aug 2023 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/US2024/039350 | Jul 2024 | WO |
Child | 18798322 | US |