Automatic stent detection

Information

  • Patent Grant
  • 9307926
  • Patent Number
    9,307,926
  • Date Filed
    Wednesday, October 2, 2013
    11 years ago
  • Date Issued
    Tuesday, April 12, 2016
    8 years ago
Abstract
This invention relates generally to the detection of objects, such as stents, within intraluminal images using principal component analysis and/or regional covariance descriptors. In certain aspects, a training set of pre-defined intraluminal images known to contain an object is generated. The principal components of the training set can be calculated in order to form an object space. An unknown input intraluminal image can be obtained and projected onto the object space. From the projection, the object can be detected within the input intraluminal image. In another embodiment, a covariance matrix is formed for each pre-defined intraluminal image known to contain an object. An unknown input intraluminal image is obtained and a covariance matrix is computed for the input intraluminal image. The covariances of the input image and each image of the training set are compared in order to detect the presence of the object within the input intraluminal image.
Description
TECHNICAL FIELD

This invention generally relates to the automatic detection of stents in intraluminal imaging.


BACKGROUND

Tomographic imaging is a signal acquisition and processing technology that allows for high-resolution cross-sectional imaging in biological systems. Tomographic imaging systems include, for example, optical coherence tomography systems, ultrasound imaging systems, and computed tomography. Tomographic imaging is particularly well-suited for imaging the subsurface of a vessel or lumen within the body, such as a blood vessel, using probes disposed within a catheter through a minimally invasive procedure.


Typical tomographic imaging catheters consist of an imaging core that rotates and moves longitudinally through a blood vessel, while recording an image video loop of the vessel. The motion results in a 3D dataset, where each frame provides a 360 degree slice of the vessel at different longitudinal section. These frames provide cardiologists with invaluable information such as the location and severity of the stenosis in a patient, the presence of vulnerable plagues, and changes in the disease over time. The information also assists in determining the appropriate treatment plan for the patient, such as drug therapy, stent placement, angioplasty, or bypass surgery.


One of the most common analyses performed is the placement and apposition of stents. A stent is a small, typically meshed or slotted, tube-like structure made of a metal or polymer that is inserted into the blood vessel to hold vessel open and keep it from occluding and provides a framework for arterial lesions that are likely to embolize after balloon angioplasty. During placement, the stent should be placed in parallel within the vessel and contact the vessel wall. Apposition of a coronary artery stent is the term for how well stent lies against the wall of the artery. When the stent as placed does not mesh completely against the blood vessel, the stent is in ‘incomplete apposition’. Incomplete apposition may raise the risk of a subsequent blockage or thrombus because of blood pooling or stagnating in the dead space between the stent and the coronary artery wall. Therefore, it is critical to verify that the stent is properly employed.


In order to identify and measure stent opposition in intravascular images, a cardiologist typically has to manually locate the stent struts, which are the framework of the stent visible in the tomographic image. Generally, identification of at least two stent struts is required to determine stent apposition. This process can be a very time consuming and is prone to user error.


SUMMARY

This invention generally improves the ability of user of a tomographic imaging system to quickly assess a deployed stent by providing a method for detecting the stent location. Through use of the image processing techniques, the stent locations for all frames or a subset of frames in a recorded dataset for an imaging run are detected and provided to the user. The resulting stent detections may be displayed on the tomographic image, the image longitudinal display (ILD) or displayed on 3-D reconstructions of the vessel. This advantageously eliminates the need for the user to manually locate the stent struts in order to quantify the apposition. Moreover, automatically detecting stents reduces error associated with manual detection and provides a more reliable means to detect and remedy mal-apposed stents.


Tomographic imaging systems suitable for use in aspects of the invention include any tomographic cross-sectional imaging system capable of obtaining intraluminal images, such as optical coherence tomography systems, ultrasound imaging systems and combined OCT-ultrasound imaging systems. Intraluminal images are typically intravascular images, but also include any image taken within a biological lumen, such as an intestinal lumen.


This invention relates to computing systems and methods for computer-assisted detection of a stent, a stent strut, or a portion of the stent, and can also be used to detect other objects within intraluminal images such as tissue or a guidewire. Objects are identified based on the locations in the polar coordinate system using data obtained from one-dimensional, two dimensional or three-dimensional images. Once stent struts are identified, measurements of the stent apposition or coverage relative to the lumen border can be easily computed.


In one aspect, a set of pre-defined intraluminal images that are known to display a object are generated to train a processor to identify or recognize the object in intraluminal images unknown to contain the object, for example, input intraluminal images of a patient undergoing an OCT examination. For example, in this step, a set of pre-defined intraluminal data images can include a plurality of intraluminal images known to display a stent strut so that a processor can be trained to identify the stent strut. After a training set of the pre-defined intraluminal images is generated, the principal components for the set can be computed to create an object space for the object. The principal components for the object can be stored and used to detect the object in input intraluminal images that are unknown to contain the object. By projecting the input intraluminal image onto the object space, the object can be detected within the input intraluminal image.


In certain embodiments, after the input intraluminal image is projected onto the object space, the error, for example, the Euclidean distance, between the input intraluminal image and the object space image is determined. A small error can constitute a positive detection of the object within input intraluminal image. The image can then be post-processed to identify or highlight the detected object within the input intraluminal image. Post-processing can also include removing false object detections from the input intraluminal image.


In some embodiments, at least two sets of pre-defined intraluminal images known to display different objects are generated, for example, a set of pre-defined intraluminal images known to display stents and a set of pre-defined intraluminal images known to display tissue. The principal components for each set can be computed to generate an object space for each object. An input intraluminal image unknown to display either object is projected onto each object space and the objects are detected within the input intraluminal images. The objects can be detected by calculating an error between the input intraluminal image and each object space. The object space that most accurately represents the input intraluminal image, for example, has the smallest error, is indicative of a positive detection of the corresponding object to the object space. The object space with the larger error can be indicative of a negative detection for its corresponding object. This advantageously increases the accuracy of the detection because instead of detecting based on error alone, detection is based on the combination of error and a comparison of the errors.


In another aspect, an object, such as stent, can be detected within an input intraluminal image by generating a training set of intraluminal images of an object, where each image is defined by one or more features. A covariance matrix can be computed for a feature within each pre-defined intraluminal image of the training set. The covariance for a feature within the input intraluminal image can be calculated and compared to the covariances of the training set. From the comparison, the object can be detected within the input intraluminal image. In certain aspects, the feature can be the Cartesian coordinates of a pixel, the intensity at each pixel, or the first and second order derivatives of the image in the x and y direction.


Other and further aspects and features of the invention will be evident from the following detailed description and accompanying drawings, which are intended to illustrate, not limit, the invention.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 is a perspective view of a vessel.



FIG. 2 is a cross sectional view of the vessel shown in FIG. 1.



FIG. 3 is a diagram of components of an optical coherence tomography (OCT) system.



FIG. 4 is a diagram of the imaging engine shown in FIG. 3.



FIG. 5 is a diagram of a light path in an OCT system of certain embodiments of the invention.



FIG. 6 is a patient interface module of an OCT system.



FIG. 7 is an illustration of the motion of parts of an imaging catheter according to certain embodiments of the invention.



FIG. 8 shows an array of A scan lines of a three-dimensional imaging system according to certain embodiments of the invention.



FIG. 9 shows the positioning of A scans with in a vessel.



FIG. 10 illustrates a set of A scans used to compose a B scan according to certain embodiments of the invention.



FIG. 11 shows the set of A scans shown in FIG. 10 within a cross section of a vessel.



FIG. 12 shows a sample OCT B-Scan image calculated from 660 A-scans.



FIG. 13 shows a scan-converted OCT image from the B-scan of FIG. 12.



FIG. 14 depicts a basic flow chart for principal component analysis for stent detection.



FIG. 15 depicts an example OCT B-Scan with stent struts detected following the principal component analysis outlined in FIG. 14.



FIG. 16 depicts the error of projected data using stent and tissue principal components.



FIG. 17 depicts the resulting stent detections in scan-converted image of FIG. 15.



FIG. 18 depicts the tissue error and stent error from FIG. 16 for all frames in a pullback.



FIG. 19 depicts the corresponding stent detections from for all frames in a pullback.



FIG. 20 depicts the resulting stent detections using regional covariance analysis.



FIG. 21 is a system diagram according to certain embodiments.





DESCRIPTION

This invention generally relates to automatically detecting stents in intraluminal medical imaging. Medical imaging is a general technology class in which sectional and multidimensional anatomic images are constructed from acquired data. The data can be collected from a variety of signal acquisition systems including, but not limited to, magnetic resonance imaging (MRI), radiography methods including fluoroscopy, x-ray tomography, computed axial tomography and computed tomography, nuclear medicine techniques such as scintigraphy, positron emission tomography and single photon emission computed tomography, photo acoustic imaging ultrasound devices and methods including, but not limited to, intravascular ultrasound spectroscopy (IVUS), ultrasound modulated optical tomography, ultrasound transmission tomography, other tomographic techniques such as electrical capacitance, magnetic induction, functional MRI, optical projection and thermo-acoustic imaging, combinations thereof and combinations with other medical techniques that produce one-, two- and three-dimensional images. Although the exemplifications described herein are drawn to the invention as applied to OCT, at least all of these techniques are contemplated for use with the systems and methods of the present invention.


Systems and methods of the invention have application in intravascular imaging methodologies such as intravascular ultrasound (IVUS) and optical coherence tomography (OCT) among others that produce a three-dimensional image of a lumen. A segment of a lumen 101 is shown in FIG. 1 having a feature 113 of interest. FIG. 2 shows a cross-section of lumen 101 through feature 113. In certain embodiments, intravascular imaging involves positioning an imaging device near feature 113 and collecting data representing a three-dimensional image.


OCT is a medical imaging methodology using a specially designed catheter with a miniaturized near infrared light-emitting probe attached to the distal end of the catheter. As an optical signal acquisition and processing method, it captures micrometer-resolution, three-dimensional images from within optical scattering media (e.g., biological tissue). Commercially available OCT systems are employed in diverse applications, including art conservation and diagnostic medicine, notably in ophthalmology where it can be used to obtain detailed images from within the retina. The detailed images of the retina allow one to identify several eye diseases and eye trauma. Recently it has also begun to be used in interventional cardiology to help diagnose coronary artery disease. OCT allows the application of interferometric technology to see from inside, for example, blood vessels, visualizing the endothelium (inner wall) of blood vessels in living individuals.


Other applications of OCT and other signal processing imaging systems for biomedical imaging include use in: dermatology in order to image subsurface structural and blood flow formation; dentistry in order to image the structure of teeth and gum line to identify and track de-mineralization and re-mineralization, tarter, caries, and periodontal disease; gastroenterology in order to image the gastrointestinal tract to detect polyps and inflammation, such as that caused by Crohn's disease and ulcerative colitis; cancer diagnostics in order to discriminate between malignant and normal tissue.


Generally, an OCT system comprises three components which are 1) an imaging catheter 2) OCT imaging hardware, 3) host application software. When utilized, the components are capable of obtaining OCT data, processing OCT data, and transmitting captured data to a host system. OCT systems and methods are generally described in Milner et al., U.S. Patent Application Publication No. 2011/0152771, Condit et al., U.S. Patent Application Publication No. 2010/0220334, Castella et al., U.S. Patent Application Publication No. 2009/0043191, Milner et al., U.S. Patent Application Publication No. 2008/0291463, and Kemp, N., U.S. Patent Application Publication No. 2008/0180683, the content of each of which is incorporated by reference in its entirety. In certain embodiments, systems and methods of the invention include processing hardware configured to interact with more than one different three dimensional imaging system so that the tissue imaging devices and methods described here in can be alternatively used with OCT, IVUS, or other hardware.


Various lumen of biological structures may be imaged with aforementioned imaging technologies in addition to blood vessels, including, but not limited, to vasculature of the lymphatic and nervous systems, various structures of the gastrointestinal tract including lumen of the small intestine, large intestine, stomach, esophagus, colon, pancreatic duct, bile duct, hepatic duct, lumen of the reproductive tract including the vas deferens, vagina, uterus and fallopian tubes, structures of the urinary tract including urinary collecting ducts, renal tubules, ureter, and bladder, and structures of the head and neck and pulmonary system including sinuses, parotid, trachea, bronchi, and lungs.


The arteries of the heart are particularly useful to examine with imaging devices such as OCT. OCT imaging of the coronary arteries can determine the amount of plaque built up at any particular point in the coronary artery. The accumulation of plaque within the artery wall over decades is the setup for vulnerable plaque which, in turn, leads to heart attack and stenosis (narrowing) of the artery. OCT is useful in determining both plaque volume within the wall of the artery and/or the degree of stenosis of the artery lumen. It can be especially useful in situations in which angiographic imaging is considered unreliable, such as for the lumen of ostial lesions or where angiographic images do not visualize lumen segments adequately. Example regions include those with multiple overlapping arterial segments. It is also used to assess the effects of treatments of stenosis such as with hydraulic angioplasty expansion of the artery, with or without stents, and the results of medical therapy over time. In an exemplary embodiment, the invention provides a system for capturing a three dimensional image by OCT.


In OCT, a light source delivers a beam of light to an imaging device to image target tissue. Light sources can be broad spectrum light sources, pulsating light sources, continuous wave light sources, and include superluminescent diodes, ultrashort pulsed lasers and supercontinuum. Within the light source is an optical amplifier and a tunable filter that allows a user to select a wavelength of light to be amplified. Wavelengths commonly used in medical applications include near-infrared light, for example between about 800 nm and about 1700 nm.


Methods of the invention apply to image data obtained from obtained from any OCT system, including OCT systems that operate in either the time domain or frequency (high definition) domain. Basic differences between time-domain OCT and frequency-domain OCT is that in time-domain OCT, the scanning mechanism is a movable mirror, which is scanned as a function of time during the image acquisition. However, in the frequency-domain OCT, there are no moving parts and the image is scanned as a function of frequency or wavelength.


In time-domain OCT systems an interference spectrum is obtained by moving the scanning mechanism, such as a reference mirror, longitudinally to change the reference path and match multiple optical paths due to reflections within the sample. The signal giving the reflectivity is sampled over time, and light traveling at a specific distance creates interference in the detector. Moving the scanning mechanism laterally (or rotationally) across the sample produces two-dimensional and three-dimensional images.


In frequency domain OCT, a light source capable of emitting a range of optical frequencies excites an interferometer, the interferometer combines the light returned from a sample with a reference beam of light from the same source, and the intensity of the combined light is recorded as a function of optical frequency to form an interference spectrum. A Fourier transform of the interference spectrum provides the reflectance distribution along the depth within the sample.


Several methods of frequency domain OCT are described in the literature. In spectral-domain OCT (SD-OCT), also sometimes called “Spectral Radar” (Optics letters, Vol. 21, No. 14 (1996) 1087-1089), a grating or prism or other means is used to disperse the output of the interferometer into its optical frequency components. The intensities of these separated components are measured using an array of optical detectors, each detector receiving an optical frequency or a fractional range of optical frequencies. The set of measurements from these optical detectors forms an interference spectrum (Smith, L. M. and C. C. Dobson, Applied Optics 28: 3339-3342), wherein the distance to a scatterer is determined by the wavelength dependent fringe spacing within the power spectrum. SD-OCT has enabled the determination of distance and scattering intensity of multiple scatters lying along the illumination axis by analyzing a single the exposure of an array of optical detectors so that no scanning in depth is necessary. Typically the light source emits a broad range of optical frequencies simultaneously. Alternatively, in swept-source OCT, the interference spectrum is recorded by using a source with adjustable optical frequency, with the optical frequency of the source swept through a range of optical frequencies, and recording the interfered light intensity as a function of time during the sweep. An example of swept-source OCT is described in U.S. Pat. No. 5,321,501.


Generally, time domain systems and frequency domain systems can further vary in type based upon the optical layout of the systems: common beam path systems and differential beam path systems. A common beam path system sends all produced light through a single optical fiber to generate a reference signal and a sample signal whereas a differential beam path system splits the produced light such that a portion of the light is directed to the sample and the other portion is directed to a reference surface. Common beam path systems are described in U.S. Pat. No. 7,999,938; U.S. Pat. No. 7,995,210; and U.S. Pat. No. 7,787,127 and differential beam path systems are described in U.S. Pat. No. 7,783,337; U.S. Pat. No. 6,134,003; and U.S. Pat. No. 6,421,164, the contents of each of which are incorporated by reference herein in its entirety.


In certain embodiments, the invention provides a differential beam path OCT system with intravascular imaging capability as illustrated in FIG. 3. For intravascular imaging, a light beam is delivered to the vessel lumen via a fiber-optic based imaging catheter 826. The imaging catheter is connected through hardware to software on a host workstation. The hardware includes an imagining engine 859 and a handheld patient interface module (PIM) 839 that includes user controls. The proximal end of the imaging catheter is connected to PIM 839, which is connected to an imaging engine as shown in FIG. 3.


As shown in FIG. 4, the imaging engine 859 (e.g., a bedside unit) houses a power supply 849, light source 827, interferometer 831, and variable delay line 835 as well as a data acquisition (DAQ) board 855 and optical controller board (OCB) 851. A PIM cable 841 connects the imagine engine 859 to the PIM 839 and an engine cable 845 connects the imaging engine 859 to the host workstation.



FIG. 5 shows light path in a differential beam path system according to an exemplary embodiment of the invention. Light for image capture originates within the light source 827. This light is split between an OCT interferometer 905 and an auxiliary, or “clock”, interferometer 911. Light directed to the OCT interferometer is further split by splitter 917 and recombined by splitter 919 with an asymmetric split ratio. The majority of the light is guided into the sample path 913 and the remainder into a reference path 915. The sample path includes optical fibers running through the PIM 839 and the imaging catheter 826 and terminating at the distal end of the imaging catheter where the image is captured.


Typical intravascular OCT involves introducing the imaging catheter into a patient's target vessel using standard interventional techniques and tools such as a guide wire, guide catheter, and angiography system. The imaging catheter may be integrated with IVUS by an OCT-IVUS system for concurrent imaging, as described in, for example, Castella et al. U.S. Patent Application Publication No. 2009/0043191 and Dick et al. U.S. Patent Application Publication No. 2009/0018393, both incorporated by reference in their entirety herein.


Rotation of the imaging catheter is driven by spin motor 861 while translation is driven by pullback motor 865, shown in FIG. 6. This results in a motion for image capture described by FIG. 7. Blood in the vessel is temporarily flushed with a clear solution for imaging. When operation is triggered from the PIM or control console, the imaging core of the catheter rotates while collecting image data. Using light provided by the imaging engine, the inner core sends light into the tissue in an array of A-scan lines as illustrated in FIG. 8 and detects reflected light.



FIG. 9 shows the positioning of A-scans within a vessel. Each place where one of A-scans A11, A12, . . . , AN intersects a surface of a feature within vessel 101 (e.g., a vessel wall) coherent light is reflected and detected. Catheter 826 translates along axis 117 being pushed or pulled by pullback motor 865.


The reflected, detected light is transmitted along sample path 913 to be recombined with the light from reference path 915 at splitter 919 (FIG. 5). A variable delay line (VDL) 925 on the reference path uses an adjustable fiber coil to match the length of reference path 915 to the length of sample path 913. The reference path length is adjusted by a stepper motor translating a mirror on a translation stage under the control of firmware or software. The free-space optical beam on the inside of the VDL 925 experiences more delay as the mirror moves away from the fixed input/output fiber.


The combined light from splitter 919 is split into orthogonal polarization states, resulting in RF-band polarization-diverse temporal interference fringe signals. The interference fringe signals are converted to photocurrents using PIN photodiodes 929a, 929b, . . . on the OCB 851 as shown in FIG. 5. The interfering, polarization splitting, and detection steps are done by a polarization diversity module (PDM) on the OCB. Signal from the OCB is sent to the DAQ 855, shown in FIG. 4. The DAQ includes a digital signal processing (DSP) microprocessor and a field programmable gate array (FPGA) to digitize signals and communicate with the host workstation and the PIM. The FPGA converts raw optical interference signals into meaningful OCT images. The DAQ also compresses data as necessary to reduce image transfer bandwidth to 1 Gbps (e.g., compressing frames with a glossy compression JPEG encoder).


Data is collected from A-scans A11, A12, . . . , AN and stored in a tangible, non-transitory memory. Typically, rotational systems consist of an imaging core which rotates and pulls back (or pushes forward) while recording an image video loop. This motion results in a three dimensional dataset of two dimensional image frames, where each frame provides a 360° slice of the vessel at different longitudinal locations.


A set of A-scans generally corresponding to one rotation of catheter 826 around axis 117 collectively define a B-scan. FIG. 10 illustrates a set of A-scans A11, A12, . . . , A18 used to compose a B-scan according to certain embodiments of the invention. These A-scan lines are shown as would be seen looking down axis 117 (i.e., longitudinal distance between then is not shown). While eight A-scan lines are illustrated in FIG. 10, typical OCT applications can include between 300 and 1,000 A-scan lines to create a B-scan (e.g., about 660). Reflections detected along each A-scan line are associated with features within the imaged tissue. Reflected light from each A-scan is combined with corresponding light that was split and sent through reference path 915 and VDL 925 and interference between these two light paths as they are recombined indicates features in the tissue.


The data of all the A-scan lines together represent a three-dimensional image of the tissue. The data of the A-scan lines generally referred to as a B-scan can be used to create an image of a cross section of the tissue, sometimes referred to as a tomographic view. For example, FIG. 11 shows the set of A-scans shown in FIG. 10 within a cross section of a vessel.


The set of A-scans obtained by rotational imaging modality can be combined to form a B-scan. FIG. 12 is an example of an OCT polar coordinate B-Scan with 660 A-scans. To create a final tomographic view of the vessel, the B-scan is scan converted to a Cartesian coordinate system. FIG. 13 displays the scan-converted image of the B-scan in FIG. 12.


Systems and methods of the invention include image-processing techniques that provide automatic detection of objects, such as stents, within intraluminal images. Typically, the OCT intraluminal image is an intravascular image taken within a lumen of a blood vessel, but the detection methods described herein can be used to detect objects within other biological lumens, such as the intestine. Although the following description is directed towards detecting objects in OCT images, one skilled in the art would readily recognize that methods and systems of intention can be utilized to detect objects in any intraluminal images obtained from any other imaging technique, such as intravascular ultrasound imaging (IVUS) and combined OCT-IVUS.


Embodiments of the invention provide for algorithms to detect a stents location within the polar coordinate system using features within one-dimensional images, such as A-scan, two-dimensional images, such as a B-scan, and/or three-dimensional images. Once the polar coordinates of the object are detected, the polar coordinates can be converted to the Cartesian coordinates and displayed as a tomographic image. Thus, a three-dimensional profile of the stent can be detected and displayed to a user. In addition, with the polar coordinates of the stent automatically detected, the strent stut apposition or coverage relative to the lumen border can easily be computed. Additionally, these algorithms can be applied to pre-scan converted data and to scan converted data.


Because the algorithms disclosed herein can be applied to every frame taken during an OCT imaging run, the location of the object can be detected in one or more frames can be computed and provided to the user on a graphic display.


The 1-D, 2-D or 3-D images include data, such as pixel data, which includes pixel locations, pixel intensity, color intensities, which includes the RGB color channel for the pixels, and/or volumetric data, which includes the x, y, z coordinates. The data obtained from the image are considered features within the image that can be used to classify or detect the object. Images can be associated with other data features such as amplitude, phase, frequency, polarity, velocity, weight, density, transparency, reflectance, hardness, and temperature.



FIG. 14 exemplifies the steps employed in an embodiment for detecting stents using an adaptation of principal component analysis, a known signal processing approach. Exemplary principal components analysis techniques can be found in M. Turk and A. Pentland “Eigenfaces for Recognition” and Pentland et al. “View-based and modular eigenspaces for face recognition” in Proc. IEEE Conf. Comput Vision. @ Pattern Recogn. These techniques have been adapted to of detection of object in intraluminal images. The first step 30 includes generating a training set of pre-defined intraluminal images for an object. The second step 32 involves computing the principal components for the object to create an object space. The third step 34 involves projecting an input intraluminal image onto the object space. The fourth step 36 involves detecting the object within the intraluminal image. Methods of the invention are not limited to stent detection, but can be used to detect any object within an intraluminal image, such as guidewire, lumen border, and tissue.


The first step 30 is to generate a training set of pre-defined intraluminal images known to contain an object so that the images can be used to train the processor to identify the object within images or regions of images not known to contain the object. Images known to contain the object include images in which the object was manually located and detected. The images known to contain the object can be obtained online or compiled off-line. In certain aspects, the training set of pre-defined intraluminal images can be pre-processed by, for example filtering the images prior to generating a training set.


In certain aspects, the images for the training set are all the same size or interpolated to the same size. Data, such as pixel intensity, can be arranged into a matrix for principal component analysis. In one aspect, data from each image in the training set can be taken and stacked into a matrix, wherein each column in the matrix represents a different image and each row within a column of the matrix represents the same pixel location in each training image. The rows can be through of as features and each column is a sample of that feature.


It should be noted that in all training sets generated for use in the embodiments described herein are not limited to a fixed amount of pre-defined images. Rather, the training sets can be made to have an adaptive training process. For example, by continually updating training sets with input intraluminal images that have been positively identified for a specific object. As the training set database grows, so does the accuracy of the detection.


Once a matrix for the training set of pre-defined matrix is compiled, the principal components for the training set matrix are computed to create an object space, as in the second step 32. The principal components can be computed by directly computing the eigenvectors of a covariance matrix computed from the training data set matrix or by utilizing Singular Value Decomposition (SVD) of the training set data matrix, which is described in detail in, for example, Abdi, H, & Williams, L. J. (2010). “Principal component analysis.” Wiley Interdisciplinary Reviews: Computational Statistics, 2: 433-459. By calculating the principal components, one can determine which vectors out of the set of pre-defined images best account for the distribution of object images within the entire object space. Therefore, only the top n eigenvectors are kept in order to create a basis which accounts for most of the variance within the image set. These vectors define the subspace of object images, and constitute the object space. The principal components are stored within a memory and utilized later on to detect an object within input intraluminal images.


In order to utilize the principal components to detect object in unknown input images, a threshold error can be computed. In one aspect, the threshold value is computed by determining the amount of error between one or more of pre-defined images known to contain the object and the object space. This threshold error can be obtained using the same pre-defined images that were used to create the object space or another image known to create the same object. In order to determine error, a pre-defined image can be projected onto the object space in order to determine the distance between points in the pre-defined image in comparison to the object space. In certain aspects, the error is the Euclidean distance between the training set image and the object space. This error computation can be repeated for each pre-defined image in the training set, a portion of pre-defined images in the training set, or for multiple pre-defined images outside of the training set.


Using the computed errors, one can calculate a threshold error value that can be used to determine if an unknown image contains the object. For example, unknown images that are projected against the object space that have an error greater than the threshold error will not be determined to contain the object and unknown images with an error smaller than the threshold error will return a positive detection for the object. The threshold error can be the maximum error, minimum error, or an average computation of the error, such as the quadratic mean, arithmetic mean, mode, medium, or any other statistical average known in the art.


The third step 34 involves projecting an input intraluminal image onto the object space. The input intraluminal image can be an image taken during an OCT procedure in real-time or a previously taken image, for example, an image that was stored or uploaded onto the computing system. The error between the input intraluminal image and the object space is computed in a similar manner the error was computed for each pre-defined image to determine a threshold error. In some embodiments, the error is the Euclidean distance between the input image and the object space. After the error is computed, the error of the input intraluminal image can be used to detect the object, as in the fourth step 36. For example, if the error is below a threshold value, the object is positively detected in the input intraluminal image. If the error is above the threshold image, then the object is negatively detected within input intraluminal image.


In certain aspects, an object space is created for two or more objects in order to compare the input intraluminal image to two or more object spaces. Step 30 and step 32 are repeated for at least one other object. In one embodiment, a training set is generated of pre-defined intraluminal images known to contain stents and a training set is generated of pre-defined intraluminal images known to contain tissue, for example, imaging of a blood vessel without stents. The principal components are generated for both training sets to compute a tissue space and a stent space. For step 36, an input intraluminal image can be projected onto both the tissue space and the stent space, and a tissue error and a stent error can be calculated by comparing the input intraluminal image to both spaces. The set of principal components, tissue or stent, which most accurately represents the original image, is selected as the class matching the input intraluminal image, and the corresponding object is positively detected. For example, if the error between the stent and the input intraluminal image is less than the error between the tissue and the input intraluminal image, the stent is positively detected within the input intraluminal image.


In addition, a threshold error value can also be computed for each of the plurality of object spaces. A comparison between the input intraluminal image and each object space's threshold value can determine whether or not the object is present in the input intraluminal image. If the error is significantly high for both classes, this can indicate that the input intraluminal image does not match any of the data that was used in the training set. Comparing threshold error reduces the risk of misclassification when comparing simply comparing the magnitudes of the error. If the input intraluminal image does not match any of the training sets, an indicator can appear on an OCT graphical display to indicate to the user that manual detection may be required with respect to the unclassified input intraluminal image.


In a specific embodiment, method of FIG. 14 is adapted to train the classifier to detect tissue, stents, and guidewires. Guidewires are often misclassified as a stent strut because its features appear stent like in the intraluminal images. This prevents the likelihood that a positive detection for a stent is actually a guidewire.


In some embodiments, the input intraluminal image may be defined using the lumen border. In order to improve performance, the detected lumen border can be used to identify search regions for the object, such as a stent strut, within the image. In this aspect, the training sets of pre-defined images generated for an object will also be defined by the lumen border. For example, if a lumen border is detected within a region around +30 pixels, −200 pixels within an A-line, a training set can be formed using only the lumen border region of pre-defined images and an object space for that region can be generated. The same region of the A-line intraluminal image can be projected onto the object space to detect the object in that region. Detection occurs using the same error methods as previously described. The lumen border can be automatically or semi-automatically detected in an image using any method known in the art, such as the techniques disclosed in U.S. Pat. No. 7,978,916, S. Tanimoto, G. Rodriguez-Granillo, P. Barlis, S. de Winter, N. Bruining, R. Hamers, M. Knappen, S. Verheye, P. W. Serruys, and E. Regar, “A novel approach for quantitative analysis of intracoronary optical coherencetomography: High inter-observer agreement with computer-assisted contour detection,” Cathet. Cardiovasc. Intervent. 72, 228-235 (2008); K. Sihan, C. Botka, F. Post, S. de Winter, E. Regar, R. Hamers, and N. Bruining, “A novel approach to quantitative analysis of intraluminal optical coherence tomography imaging,” Comput. Cardiol. 1089-1092 (2008); J. Canny, “A computational approach to edge detection,” IEEE Trans. Pattern Anal. Mach. Intell. 8, 679-698 (1986).


Additionally, methods of the invention provide for a post-processing step to, for example, detect the location of the stent within the image, for example, the stent depth. Any method known in the art can be used to locate the depth position of the stent, such as peak detection within some maximum distance of the detected lumen border can be used to identify the final location of the stent. Post-processing can be used to verify the detection of the stent within the image. For example, methods of the invention can also be combined with other algorithms or methods used to detect guidewires or other false stent detections. In one aspect, after detection of stents using the methods of the invention, a guidewire detection/tracking algorithm can be applied to the image to remove the false stent detections. Post-processing can also be used to visually illustrate the resulting stent detections within the intravascular image on a graphical user interface. For example, detected portions of the stent can be highlighted with a bolded line or circled within the image.


The following description and figures illustrate stent detection following the block diagram in FIG. 14. FIG. 15 depicts an example of an OCT B-scan highlighting both detected stent struts and the automatically detected lumen border. A training set for stents and tissue were generated defined by the lumen border and a tissue space and stent space were computed. The A-line input intraluminal image around the border was projected onto a stent space and a tissue space. The error between the input intraluminal image and both the tissue space and stent space were computed, and plotted in FIG. 16. As shown in FIG. 16, the stent error is smaller at the location of the stents and the tissue error is higher at the locations of the stents. The difference between the stent error and the tissue error is also plotted in FIG. 16. Locations along the lumen border where the stent error is lower than the tissue error are classified as stents, and thus a stent is positively detected within the image. FIG. 17 displays the corresponding scan-converted image of the B-scan shown in FIG. 15. Post-processing was utilized to highlight the stent detections within the scan-converted image. The difference between the stent and tissue error for all frames in a pullback is plotted in a 2D splayed map in FIG. 18. The corresponding stent detections for all frames in this pull-back are provided in FIG. 19.


In another embodiment, objects are detected within an intraluminal image using region covariance descriptors to detect objects in images or in regions of images. This approach can be adapted to both 1D, 2D, and 3D intraluminal images. Similar to the detection method outlined in FIG. 13, this algorithm requires generating a training step of pre-defined intraluminal images and determining a compute for the training set that is compared to an input intraluminal image for detection. Regional covariance imaging techniques are known in the art and are described in, for example, Tuzel et al. “Region Covariance: A Fast Descriptor for Detection and Classification,” European Conference on Computer Vision (ECCV), May 2006; Forstner and Moonen, “A metric for covariance matrices,” Technical Report, Dept. of Geodesy and Geoinformatics, Stuttgart University (1999).


To detect stents or objects within intraluminal images using regional covariance, the first step is to generate a training set of pre-defined intraluminal images known to contain an object, for example, stent, tissue, or guidewire. A feature matrix can then be generated for the training set using a number of features within each pre-defined intraluminal image of the training set, e.g., the x and y coordinates of the pixel location, intensity of each pixel, and the first and second order derivatives of the image in the x and y direction. These features are computes for each pre-defined image within the training set. A pre-defined image of the training set can be any size image of m×n, and in one aspect m and n correspond to a dimension that is slightly larger than the width of a stent and the depth of tissue in an intraluminal image, and all images of the training set should be the same size. Although it is possible to perform regional covariance analysis on the entire image, use of m×n regions allows for targeted search of stents and other objects located on the lumen border. For example, the m×n image region can be created around the lumen border detected within an input intraluminal image, using any method of detecting the lumen border known in the art and discussed above.


Each pixel of a pre-defined intraluminal image is converted in to a feature matrix, using the equation listed below.









Equation





1


:






Feature





Matrix





for





Region





Covariance






Tracking




[
1
]













F


(

x
,
y

)


=

[







x





y






I


(

x
,
y

)













I


(

x
,
y

)





x















I


(

x
,
y

)





y















2



I


(

x
,
y

)






x
2
















2



I


(

x
,
y

)






y
2







]





Equation





1








In the above equation, x and y are indices and I is the intensity. For purposes of stent detection, the feature equation can be adapted to have more or less features or contain additional feature data, for example, the addition of RGB color values. Each input intraluminal image within the training set will have the same (x, y) pixel locations, and although not distinguishing, these coordinates are useful to correlate other features that vary from image to image. Once the feature matrix is computed, a covariance matrix for the set of features for each image can be computed using the following equation, where z represents the features, u is the mean of the feature samples, and T is a transpose operator.










C
R

=


1

n
-
1







k
=
1

n








(


z
k

-
μ

)




(


z
k

-
μ

)

T








Equation





2







The above process is repeated for each pre-defined intraluminal image within the training set, and the covariance matrices are saved in the memory for later use during detection of objects of unknown input intraluminal images. The covariance matrices represent subspaces of the object.


In order to detect an object in the input intraluminal image, the input intraluminal image is broken down into the same m×n regions as the pre-defined images of the training set to identify, for example, stent locations. The covariance matrix of the input intraluminal image for the region is computed and compared to each covariance matrix of the training set. The comparison of the covariance matrix involves performing a distance calculation, or error calculation, between feature points within the covariance matrices. Any method known in the art for calculating the distance between covariance matrices can be used or adapted to calculate the distance between the unknown input intraluminal image covariance matrix and the covariance matrices of the pre-defined training. See for example, J. N. L. Brümmer and L. R. Strydom, “An euclidean distance measure between covariance matrices of speechcepstra for text-independent speaker recognition,” in Proc. 1997 South African Symp. Communications and Signal Processing, 1997, pp. 167-172; W. Förstner and B. Moonen, A Metric for Covariance Matrices Dept. Geodesy and Geoinformatics, Stuttgart Univ., Stuttgart, Germany, 1999; Ö Tüzel, F. Porikli, and P. Meer, “Region covariance: A fast descriptor and for detection and classification,” in Proc. Image and Vision Computing, Auckland, New Zealand, 2004.


A threshold error can be determined for the training set and used to determine whether the distance between the input intraluminal image and training set images are indicative of a positive detection of the object. Any method can be used to create a threshold distance. For example, the threshold distance is obtained by calculating the covariance distance between the training set images and selecting the maximum distance as a threshold distance, or calculating an average value as a threshold distance.


Like previous embodiments, the regional covariance approach can also be used to detect one or more objects within an input intravascular image by generating covariance matrices for more than one object. For example, a training set of pre-defined images can be generated for tissue and stents, features matrices can be computed for each pre-defined image within a training set, and a covariance matrix can be calculated from each feature matrix. A covariance matrix calculated for an input intraluminal image is then compared to the stent and tissue covariance matrices. The training set that minimizes the distance from input intraluminal image indicates a positive detection of the object corresponding to the training set within the input intraluminal image. In addition, a threshold error can be computed for each object, and used to determine if the either object is present in the intravascular image.



FIG. 20 displays detected stent struts within A-scan-converted image using the regional covariance approach. The bolded lines indicate the stent detections. Like the previously discussed embodiments, post-processing can be applied to identify the location of the stent in depth and remove false detections.


In addition, other algorithm image processing techniques known in the art that utilize subspaces for object recognition within images can adapted to detect stents and other objects in intraluminal images. For a concise overview of various object recognition techniques, see Bain and Tao, Chapter 3: “Face Subspace Learning”, Handbook of Face Recognition, 2011. For example, Fisher's linear discriminate analysis (FLDA) can be used to detect stents. Linear discriminant analysis is primarily used to reduce the number of features, such as pixel values, to a more manageable number before classification or detection as compared to using principal component analysis. Each of the new dimensions is a linear combination of pixel values, which form a template. The linear combinations obtained using Fisher's linear discriminant is called a linear classifier and can be used in comparison to input intraluminal images to detect stents.


In certain aspects, FDLA can be combined with other algorithmic techniques to improve the accuracy of object detection using the technique. For example, FDLA can be combined with general mean criterion and max-min distance analysis (MMDA), discriminatory locality alignment analysis (DLA), and manifold elastic net (MNE).


Another detection method that can be used or adapted to detect stents or objects in intraluminal images includes using statistical model-based image recognition algorithms. See, for example, Felzenszwalb and Huttenlocher, “Pictorial Structures for Object Recognition,” Volume 61, Number 1, 55-79, DOI: 10.1023/B:VISI.0000042934.15159.49; A. A. Amini, T. E. Weymouth, and R. C. Jain. “Using dynamic programming for solving variational problems in vision,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 12(9):8551{867, September 1990; M. A. Fischler and R. A. Elschlager, “The representation and matching of pictorial structures,” IEEE Transactions on Computer, 22(1):67-92, January 1973.


With respect to the methods of detecting objects within intraluminal images discussed herein, various computer or processor-based systems are suitable for compiling data from intraluminal images, interfacing with an OCT probe to obtain input intraluminal images, applying the disclosed algorithms to detect objects, and displaying the detected objects to a user of the OCT system. The systems and methods of use described herein may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. The systems and methods of use described herein can be performed using any type of computing device, such as a computer, that includes a processor or any combination of computing devices where each device performs at least part of the process or method.


In some embodiments, a device of the invention includes an OCT imaging system and obtains a three-dimensional data set through the operation of OCT imaging hardware. In some embodiments, a device of the invention is a computer device such as a laptop, desktop, or tablet computer, and obtains a three-dimensional data set by retrieving it from a tangible storage medium, such as a disk drive on a server using a network or as an email attachment.


Methods of the invention can be performed using software, hardware, firmware, hardwiring, or combinations of any of these. Features implementing functions can also be physically located at various positions, including being distributed such that portions of functions are implemented at different physical locations (e.g., imaging apparatus in one room and host workstation in another, or in separate buildings, for example, with wireless or wired connections).


In some embodiments, a user interacts with a visual interface to view images from the imaging system. Input from a user (e.g., parameters or a selection) are received by a processor in an electronic device. The selection can be rendered into a visible display. An exemplary system including an electronic device is illustrated in FIG. 21. As shown in FIG. 21, imaging engine 859 communicates with host workstation 433 as well as optionally server 413 over network 409. In some embodiments, an operator uses computer 449 or terminal 467 to control system 400 or to receive images. An image may be displayed using an I/O 454, 437, or 471, which may include a monitor. Any I/O may include a keyboard, mouse or touchscreen to communicate with any of processor 421, 459, 441, or 475, for example, to cause data to be stored in any tangible, nontransitory memory 463, 445, 479, or 429. Server 413 generally includes an interface module 425 to effectuate communication over network 409 or write data to data file 417.


Processors suitable for the execution of computer program include, by way of example, both general and special purpose microprocessors, and any one or more processor of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of computer are a processor for executing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks. Information carriers suitable for embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, (e.g., EPROM, EEPROM, solid state drive (SSD), and flash memory devices); magnetic disks, (e.g., internal hard disks or removable disks); magneto-optical disks; and optical disks (e.g., CD and DVD disks). The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.


To provide for interaction with a user, the subject matter described herein can be implemented on a computer having an I/O device, e.g., a CRT, LCD, LED, or projection device for displaying information to the user and an input or output device such as a keyboard and a pointing device, (e.g., a mouse or a trackball), by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well. For example, feedback provided to the user can be any form of sensory feedback, (e.g., visual feedback, auditory feedback, or tactile feedback), and input from the user can be received in any form, including acoustic, speech, or tactile input.


The subject matter described herein can be implemented in a computing system that includes a back-end component (e.g., a data server 413), a middleware component (e.g., an application server), or a front-end component (e.g., a client computer 449 having a graphical user interface 454 or a web browser through which a user can interact with an implementation of the subject matter described herein), or any combination of such back-end, middleware, and front-end components. The components of the system can be interconnected through network 409 by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include cell network (e.g., 3G or 4G), a local area network (LAN), and a wide area network (WAN), e.g., the Internet.


The subject matter described herein can be implemented as one or more computer program products, such as one or more computer programs tangibly embodied in an information carrier (e.g., in a non-transitory computer-readable medium) for execution by, or to control the operation of, data processing apparatus (e.g., a programmable processor, a computer, or multiple computers). A computer program (also known as a program, software, software application, app, macro, or code) can be written in any form of programming language, including compiled or interpreted languages (e.g., C, C++, Perl), and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. Systems and methods of the invention can include instructions written in any suitable programming language known in the art, including, without limitation, C, C++, Perl, Java, ActiveX, HTML5, Visual Basic, or JavaScript.


A computer program does not necessarily correspond to a file. A program can be stored in a portion of file 417 that holds other programs or data, in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub-programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network.


A file can be a digital file, for example, stored on a hard drive, SSD, CD, or other tangible, non-transitory medium. A file can be sent from one device to another over network 409 (e.g., as packets being sent from a server to a client, for example, through a Network Interface Card, modem, wireless card, or similar).


Writing a file according to the invention involves transforming a tangible, non-transitory computer-readable medium, for example, by adding, removing, or rearranging particles (e.g., with a net charge or dipole moment into patterns of magnetization by read/write heads), the patterns then representing new collocations of information about objective physical phenomena desired by, and useful to, the user. In some embodiments, writing involves a physical transformation of material in tangible, non-transitory computer readable media (e.g., with certain optical properties so that optical read/write devices can then read the new and useful collocation of information, e.g., burning a CD-ROM). In some embodiments, writing a file includes transforming a physical flash memory apparatus such as NAND flash memory device and storing information by transforming physical elements in an array of memory cells made from floating-gate transistors. Methods of writing a file are well-known in the art and, for example, can be invoked manually or automatically by a program or by a save command from software or a write command from a programming language.


Suitable computing devices typically include mass memory, at least one graphical user interface, at least one display device, and typically include communication between devices. The mass memory illustrates a type of computer-readable media, namely computer storage media. Computer storage media may include volatile, nonvolatile, removable, and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data. Examples of computer storage media include RAM, ROM, EEPROM, flash memory, or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, Radiofrequency Identification tags or chips, or any other medium which can be used to store the desired information and which can be accessed by a computing device.


It will be understood that each block of the FIG. 14, as well as any portion of the systems and methods disclosed herein, can be implemented by computer program instructions. These program instructions may be provided to a processor to produce a machine, such that the instructions, which execute on the processor, create means for implementing the actions specified in the FIG. 14 or described for the systems and methods disclosed herein. The computer program instructions may be executed by a processor to cause a series of operational steps to be performed by the processor to produce a computer implemented process. The computer program instructions may also cause at least some of the operational steps to be performed in parallel. Moreover, some of the steps may also be performed across more than one processor, such as might arise in a multi-processor computer system. In addition, one or more processes may also be performed concurrently with other processes or even in a different sequence than illustrated without departing from the scope or spirit of the invention.


INCORPORATION BY REFERENCE

References and citations to other documents, such as patents, patent applications, patent publications, journals, books, papers, web contents, have been made throughout this disclosure. All such documents are hereby incorporated herein by reference in their entirety for all purposes.


EQUIVALENTS

Various modifications of the invention and many further embodiments thereof, in addition to those shown and described herein, will become apparent to those skilled in the art from the full contents of this document, including references to the scientific and patent literature cited herein. The subject matter herein contains important information, exemplification and guidance that can be adapted to the practice of this invention in its various embodiments and equivalents thereof.

Claims
  • 1. A computer-readable, non-transitory medium storing software code representing instructions that when executed by a computing system cause the computing system to perform a method of detecting an object within an intraluminal image, the method comprising generating a set of pre-defined intraluminal images known to display an object;computing principal components for the set to create an object space for the object;projecting an input intraluminal image onto the object space; anddetecting the object in the input intraluminal image,wherein the object is selected from the group comprising a stent strut or a guidewire.
  • 2. The computer-readable, non-transitory medium of claim 1, wherein the step of detecting further comprises calculating an error between the input intraluminal image and the object space.
  • 3. The computer-readable, non-transitory medium of claim 2, wherein a small error constitutes a positive detection of the object in the input intraluminal image.
  • 4. The computer-readable, non-transitory medium of claim 1, wherein the pre-defined intraluminal images and input intraluminal image are one-dimensional, two-dimensional, or three-dimensional.
  • 5. The computer-readable, non-transitory medium of claim 1, further comprising post-processing the input intraluminal image.
  • 6. The computer-readable, non-transitory medium of claim 5, wherein the step of post-processing comprises removing false object detections and highlighting the object detections.
  • 7. The computer-readable, non-transitory medium of claim 1, further comprising performing the steps of generating, identifying, and projecting for at least one other object; anddetecting the at least one other object in the input intraluminal image.
  • 8. The computer-readable, non-transitory medium of claim 7, wherein the step of detecting the at least one other object further comprises calculating an error between the input intraluminal image and the object space for the object and between the input intraluminal image and the object space for the at least one other object.
  • 9. The computer-readable, non-transitory medium of claim 7, further wherein the smaller error constitutes a positive detection for the corresponding object.
  • 10. The computer-readable, non-transitory medium of claim 7, wherein the pre-defined intraluminal images and input intraluminal image are one-dimensional, two-dimensional, or three-dimensional.
  • 11. The computer-readable, non-transitory medium of claim 7, wherein the object is selected from the group consisting of a tissue, a stent strut, or a guidewire.
  • 12. The computer-readable, non-transitory medium of claim 7, further comprising post-processing the input intraluminal image.
  • 13. The computer-readable, non-transitory medium of claim 12, wherein post-processing comprises removing false object detections and highlighting the object detections.
  • 14. A system for automatically detecting an object within an intraluminal image, comprising: a central processing unit (CPU); anda storage device coupled to the CPU and having stored there information for configuring the CPU to:generate a set of pre-defined intraluminal images known to display a object;compute principal components for the set to create an object space for the object; andproject an input intraluminal data image onto the object space;detect the object in the input intraluminal image,wherein the object is selected from the group comprising a stent strut or a guidewire.
  • 15. The system of claim 14, wherein detecting the object further comprises calculating an error between the input intraluminal image and the object space.
  • 16. The system of claim 15, wherein a small error as compared to the object space constitutes a positive detection of the object in the input intraluminal image.
  • 17. The system of claim 14, wherein the pre-defined intraluminal images and input intraluminal image are one-dimensional, two-dimensional, or three-dimensional.
  • 18. The system of claim 14, further comprising post-processing the input intraluminal image.
  • 19. The system of claim 18, wherein post-processing comprises removing false object detections and highlighting the object detections.
  • 20. The system of claim 14, further comprising performing the steps of generating, identifying, and projecting for at least one other object; anddetecting the at least one other object in the input intraluminal image.
  • 21. The system of claim 20, wherein the detecting the at least one other object further comprises calculating an error between the input intraluminal image and the object space for the object and between the input intraluminal image and the object space for the at least one other object.
  • 22. The system of claim 21, further wherein the smaller error constitutes a positive detection for the corresponding object.
  • 23. The system of claim 20, wherein the pre-defined intraluminal images and input intraluminal image are one-dimensional, two-dimensional, or three-dimensional.
  • 24. The system of claim 20, wherein the object is selected from the group consisting of a lumen border, a stent strut, or a guidewire.
  • 25. The system of claim 20, further comprising post-processing the input intraluminal image.
  • 26. The system of claim 25, wherein post-processing comprises removing false object detections and highlighting the object detections.
  • 27. A computer-readable, non-transitory medium storing software code representing instructions that when executed by a computing system cause the computing system to perform a method of detecting an object within a blood vessel, the method comprising: generating a training set of pre-defined intraluminal images known to display an object, each intraluminal image comprising a feature;computing a covariance for a feature within each intraluminal image of the training set; anddetecting the object within an input intraluminal image, the detecting step comprising: computing a covariance for a feature within the intraluminal input image; andcomparing the covariance of the input intraluminal image to the covariances of the training set to detect the object in the input image,wherein the object is selected from the group comprising a stent strut or a guidewire.
  • 28. The computer-readable, non-transitory medium of claim 27, wherein a feature is selected from the group consisting of the Cartesian coordinates of a pixel, the intensity at each pixel, and the first and second order derivatives of the image in the x and y direction.
  • 29. The computer-readable, non-transitory medium of claim 27, wherein the pre-defined intraluminal images and input intraluminal image are one-dimensional, two-dimensional, or three-dimensional.
  • 30. A method for detecting an object in an intraluminal image, the method comprising the steps of: generating a set of pre-defined intraluminal images known to display an object;computing principal components for the set to create an object space for the object;projecting an input intraluminal image onto the object space; anddetecting the object in the input intraluminal image,wherein the object is selected from the group comprising a stent strut or a guidewire.
  • 31. The method of claim 30, further comprising processing the input intraluminal image.
  • 32. The method of claim 31, wherein said processing step processing comprises removing false object detections and highlighting the object detections.
  • 33. The method of claim 30, wherein the step of detecting the at least one other object further comprises calculating an error between the input intraluminal image and the object space for the object and between the input intraluminal image and the object space for the at least one other object.
RELATED APPLICATION

This application claims the benefit of and priority to U.S. Provisional No. 61/710,429, filed Oct. 5, 2012, which is incorporated by reference in its entirety.

US Referenced Citations (1002)
Number Name Date Kind
3301258 Werner Jan 1967 A
3617880 Cormack et al. Nov 1971 A
3789841 Antoshkiw Feb 1974 A
3841308 Tate Oct 1974 A
4140364 Yamashita et al. Feb 1979 A
4274423 Mizuno et al. Jun 1981 A
4344438 Schultz Aug 1982 A
4398791 Dorsey Aug 1983 A
4432370 Hughes et al. Feb 1984 A
4552554 Gould et al. Nov 1985 A
4577543 Wilson Mar 1986 A
4676980 Segal et al. Jun 1987 A
4682895 Costello Jul 1987 A
4733665 Palmaz Mar 1988 A
4744619 Cameron May 1988 A
4762129 Bonzel Aug 1988 A
4766386 Oliver et al. Aug 1988 A
4771774 Simpson et al. Sep 1988 A
4794931 Yock Jan 1989 A
4800886 Nestor Jan 1989 A
4803639 Steele et al. Feb 1989 A
4816567 Cabilly et al. Mar 1989 A
4819740 Warrington Apr 1989 A
4821731 Martinelli et al. Apr 1989 A
4824435 Giesy et al. Apr 1989 A
4830023 de Toledo et al. May 1989 A
4834093 Littleford et al. May 1989 A
4841977 Griffith et al. Jun 1989 A
4864578 Proffitt et al. Sep 1989 A
4873690 Adams Oct 1989 A
4877314 Kanamori Oct 1989 A
4887606 Yock et al. Dec 1989 A
4917085 Smith Apr 1990 A
4917097 Proudian et al. Apr 1990 A
4928693 Goodin et al. May 1990 A
4932413 Shockey et al. Jun 1990 A
4932419 de Toledo Jun 1990 A
4948229 Soref Aug 1990 A
4951677 Crowley et al. Aug 1990 A
4969742 Falk et al. Nov 1990 A
4987412 Vaitekunas et al. Jan 1991 A
4993412 Murphy-Chutorian Feb 1991 A
4998972 Chin et al. Mar 1991 A
5000185 Yock Mar 1991 A
5024234 Leary et al. Jun 1991 A
5025445 Anderson et al. Jun 1991 A
5032123 Katz et al. Jul 1991 A
5037169 Chun Aug 1991 A
5039193 Snow et al. Aug 1991 A
5040548 Yock Aug 1991 A
5041108 Fox et al. Aug 1991 A
5054492 Scribner et al. Oct 1991 A
5065010 Knute Nov 1991 A
5065769 de Toledo Nov 1991 A
5085221 Ingebrigtsen et al. Feb 1992 A
5095911 Pomeranz Mar 1992 A
5100424 Jang et al. Mar 1992 A
5120308 Hess Jun 1992 A
5125137 Corl et al. Jun 1992 A
5135486 Eberle et al. Aug 1992 A
5135516 Sahatjian et al. Aug 1992 A
5155439 Holmbo et al. Oct 1992 A
5158548 Lau et al. Oct 1992 A
5163445 Christian et al. Nov 1992 A
5167233 Eberle et al. Dec 1992 A
5174295 Christian et al. Dec 1992 A
5176141 Bom et al. Jan 1993 A
5176674 Hofmann Jan 1993 A
5178159 Christian Jan 1993 A
5183048 Eberle Feb 1993 A
5188632 Goldenberg Feb 1993 A
5201316 Pomeranz et al. Apr 1993 A
5202745 Sorin et al. Apr 1993 A
5203779 Muller et al. Apr 1993 A
5220922 Barany Jun 1993 A
5224953 Morgentaler Jul 1993 A
5226421 Frisbie et al. Jul 1993 A
5240003 Lancee et al. Aug 1993 A
5240437 Christian Aug 1993 A
5242460 Klein et al. Sep 1993 A
5243988 Sieben et al. Sep 1993 A
5257974 Cox Nov 1993 A
5266302 Peyman et al. Nov 1993 A
5267954 Nita Dec 1993 A
5301001 Murphy et al. Apr 1994 A
5312425 Evans et al. May 1994 A
5313949 Yock May 1994 A
5313957 Little May 1994 A
5319492 Dorn et al. Jun 1994 A
5321501 Swanson et al. Jun 1994 A
5325198 Hartley et al. Jun 1994 A
5336178 Kaplan et al. Aug 1994 A
5346689 Peyman et al. Sep 1994 A
5348017 Thornton et al. Sep 1994 A
5348481 Ortiz Sep 1994 A
5353798 Sieben Oct 1994 A
5358409 Obara Oct 1994 A
5358478 Thompson et al. Oct 1994 A
5368037 Eberle et al. Nov 1994 A
5373845 Gardineer et al. Dec 1994 A
5373849 Maroney et al. Dec 1994 A
5375602 Lancee et al. Dec 1994 A
5377682 Ueno et al. Jan 1995 A
5383853 Jung et al. Jan 1995 A
5387193 Miraki Feb 1995 A
5396328 Jestel et al. Mar 1995 A
5397355 Marin et al. Mar 1995 A
5405377 Cragg Apr 1995 A
5411016 Kume et al. May 1995 A
5419777 Hofling May 1995 A
5421338 Crowley et al. Jun 1995 A
5423806 Dale et al. Jun 1995 A
5427118 Nita et al. Jun 1995 A
5431673 Summers et al. Jul 1995 A
5436759 Dijaili et al. Jul 1995 A
5439139 Brovelli Aug 1995 A
5443457 Ginn et al. Aug 1995 A
5453575 O'Donnell et al. Sep 1995 A
5456693 Conston et al. Oct 1995 A
5459570 Swanson et al. Oct 1995 A
5480388 Zadini et al. Jan 1996 A
5485845 Verdonk et al. Jan 1996 A
5492125 Kim et al. Feb 1996 A
5496997 Pope Mar 1996 A
5507761 Duer Apr 1996 A
5512044 Duer Apr 1996 A
5514128 Hillsman et al. May 1996 A
5529674 Hedgcoth Jun 1996 A
5541730 Chaney Jul 1996 A
5546717 Penczak et al. Aug 1996 A
5546948 Hamm et al. Aug 1996 A
5565332 Hoogenboom et al. Oct 1996 A
5573520 Schwartz et al. Nov 1996 A
5581638 Givens et al. Dec 1996 A
5586054 Jensen et al. Dec 1996 A
5592939 Martinelli Jan 1997 A
5596079 Smith et al. Jan 1997 A
5598844 Diaz et al. Feb 1997 A
5609606 O'Boyle Mar 1997 A
5630806 Inagaki et al. May 1997 A
5651366 Liang et al. Jul 1997 A
5660180 Malinowski et al. Aug 1997 A
5667499 Welch et al. Sep 1997 A
5667521 Keown Sep 1997 A
5672877 Liebig et al. Sep 1997 A
5674232 Halliburton Oct 1997 A
5693015 Walker et al. Dec 1997 A
5713848 Dubrul et al. Feb 1998 A
5745634 Garrett et al. Apr 1998 A
5771895 Slager Jun 1998 A
5779731 Leavitt Jul 1998 A
5780958 Strugach et al. Jul 1998 A
5798521 Froggatt Aug 1998 A
5800450 Lary et al. Sep 1998 A
5803083 Buck et al. Sep 1998 A
5814061 Osborne et al. Sep 1998 A
5817025 Alekseev et al. Oct 1998 A
5820594 Fontirroche et al. Oct 1998 A
5824520 Mulligan-Kehoe Oct 1998 A
5827313 Ream Oct 1998 A
5830222 Makower Nov 1998 A
5848121 Gupta et al. Dec 1998 A
5851464 Davila et al. Dec 1998 A
5857974 Eberle et al. Jan 1999 A
5872829 Wischmann et al. Feb 1999 A
5873835 Hastings et al. Feb 1999 A
5882722 Kydd Mar 1999 A
5912764 Togino Jun 1999 A
5916194 Jacobsen et al. Jun 1999 A
5921931 O'Donnell et al. Jul 1999 A
5925055 Adrian et al. Jul 1999 A
5949929 Hamm Sep 1999 A
5951586 Berg et al. Sep 1999 A
5974521 Akerib Oct 1999 A
5976120 Chow et al. Nov 1999 A
5978391 Das et al. Nov 1999 A
5997523 Jang Dec 1999 A
6021240 Murphy et al. Feb 2000 A
6022319 Willard et al. Feb 2000 A
6031071 Mandeville et al. Feb 2000 A
6036889 Kydd Mar 2000 A
6043883 Leckel et al. Mar 2000 A
6050949 White et al. Apr 2000 A
6059738 Stoltze et al. May 2000 A
6068638 Makower May 2000 A
6074362 Jang et al. Jun 2000 A
6078831 Belef et al. Jun 2000 A
6080109 Baker et al. Jun 2000 A
6091496 Hill Jul 2000 A
6094591 Foltz et al. Jul 2000 A
6095976 Nachtomy et al. Aug 2000 A
6097755 Guenther, Jr. et al. Aug 2000 A
6099471 Torp et al. Aug 2000 A
6099549 Bosma et al. Aug 2000 A
6102938 Evans et al. Aug 2000 A
6106476 Corl et al. Aug 2000 A
6120445 Grunwald Sep 2000 A
6123673 Eberle et al. Sep 2000 A
6134003 Tearney et al. Oct 2000 A
6139510 Palermo Oct 2000 A
6141089 Thoma et al. Oct 2000 A
6146328 Chiao et al. Nov 2000 A
6148095 Prause et al. Nov 2000 A
6151433 Dower et al. Nov 2000 A
6152877 Masters Nov 2000 A
6152878 Nachtomy et al. Nov 2000 A
6159225 Makower Dec 2000 A
6165127 Crowley Dec 2000 A
6176842 Tachibana et al. Jan 2001 B1
6179809 Khairkhahan et al. Jan 2001 B1
6186949 Hatfield et al. Feb 2001 B1
6190353 Makower et al. Feb 2001 B1
6200266 Shokrollahi et al. Mar 2001 B1
6200268 Vince et al. Mar 2001 B1
6203537 Adrian Mar 2001 B1
6208415 De Boer et al. Mar 2001 B1
6210332 Chiao et al. Apr 2001 B1
6210339 Kiepen et al. Apr 2001 B1
6212308 Donald Apr 2001 B1
6231518 Grabek et al. May 2001 B1
6245066 Morgan et al. Jun 2001 B1
6249076 Madden et al. Jun 2001 B1
6254543 Grunwald et al. Jul 2001 B1
6256090 Chen et al. Jul 2001 B1
6258052 Milo Jul 2001 B1
6261246 Pantages et al. Jul 2001 B1
6275628 Jones et al. Aug 2001 B1
6283921 Nix et al. Sep 2001 B1
6283951 Flaherty et al. Sep 2001 B1
6295308 Zah Sep 2001 B1
6299622 Snow et al. Oct 2001 B1
6312384 Chiao Nov 2001 B1
6325797 Stewart et al. Dec 2001 B1
6328696 Fraser Dec 2001 B1
6343168 Murphy et al. Jan 2002 B1
6343178 Burns et al. Jan 2002 B1
6350240 Song et al. Feb 2002 B1
6364841 White et al. Apr 2002 B1
6366722 Murphy et al. Apr 2002 B1
6367984 Stephenson et al. Apr 2002 B1
6373970 Dong et al. Apr 2002 B1
6375615 Flaherty et al. Apr 2002 B1
6375618 Chiao et al. Apr 2002 B1
6375628 Zadno-Azizi et al. Apr 2002 B1
6376830 Froggatt et al. Apr 2002 B1
6379352 Reynolds et al. Apr 2002 B1
6381350 Klingensmith et al. Apr 2002 B1
6387124 Buscemi et al. May 2002 B1
6396976 Little et al. May 2002 B1
6398792 O'Connor Jun 2002 B1
6417948 Chowdhury et al. Jul 2002 B1
6419644 White et al. Jul 2002 B1
6421164 Tearney et al. Jul 2002 B2
6423012 Kato et al. Jul 2002 B1
6426796 Pulliam et al. Jul 2002 B1
6428041 Wohllebe et al. Aug 2002 B1
6428498 Uflacker Aug 2002 B2
6429421 Meller et al. Aug 2002 B1
6440077 Jung et al. Aug 2002 B1
6443903 White et al. Sep 2002 B1
6450964 Webler Sep 2002 B1
6457365 Stephens et al. Oct 2002 B1
6459844 Pan Oct 2002 B1
6468290 Weldon et al. Oct 2002 B1
6475149 Sumanaweera Nov 2002 B1
6480285 Hill Nov 2002 B1
6491631 Chiao et al. Dec 2002 B2
6491636 Chenal et al. Dec 2002 B2
6501551 Tearney et al. Dec 2002 B1
6504286 Porat et al. Jan 2003 B1
6508824 Flaherty et al. Jan 2003 B1
6514237 Maseda Feb 2003 B1
6520269 Geiger et al. Feb 2003 B2
6520677 Iizuka Feb 2003 B2
6535764 Imran et al. Mar 2003 B2
6538778 Leckel et al. Mar 2003 B1
6544217 Gulachenski Apr 2003 B1
6544230 Flaherty et al. Apr 2003 B1
6545760 Froggatt et al. Apr 2003 B1
6546272 MacKinnon et al. Apr 2003 B1
6551250 Khalil Apr 2003 B2
6566648 Froggatt May 2003 B1
6570894 Anderson May 2003 B2
6572555 White et al. Jun 2003 B2
6579311 Makower Jun 2003 B1
6584335 Haar et al. Jun 2003 B1
6592612 Samson et al. Jul 2003 B1
6594448 Herman et al. Jul 2003 B2
6602241 Makower et al. Aug 2003 B2
6611322 Nakayama et al. Aug 2003 B1
6611720 Hata et al. Aug 2003 B2
6612992 Hossack et al. Sep 2003 B1
6615062 Ryan et al. Sep 2003 B2
6615072 Izatt et al. Sep 2003 B1
6621562 Durston Sep 2003 B2
6631284 Nutt et al. Oct 2003 B2
6638227 Bae Oct 2003 B2
6645152 Jung et al. Nov 2003 B1
6646745 Verma et al. Nov 2003 B2
6655386 Makower et al. Dec 2003 B1
6659957 Vardi et al. Dec 2003 B1
6660024 Flaherty et al. Dec 2003 B1
6663565 Kawagishi et al. Dec 2003 B2
6665456 Dave et al. Dec 2003 B2
6669716 Gilson et al. Dec 2003 B1
6671055 Wavering et al. Dec 2003 B1
6673015 Glover et al. Jan 2004 B1
6673064 Rentrop Jan 2004 B1
6685648 Flaherty et al. Feb 2004 B2
6689056 Kilcoyne et al. Feb 2004 B1
6689144 Gerberding Feb 2004 B2
6696173 Naundorf et al. Feb 2004 B1
6701044 Arbore et al. Mar 2004 B2
6701176 Halperin et al. Mar 2004 B1
6709444 Makower Mar 2004 B1
6712836 Berg et al. Mar 2004 B1
6714703 Lee et al. Mar 2004 B2
6719717 Johnson et al. Apr 2004 B1
6725073 Motamedi et al. Apr 2004 B1
6726677 Flaherty et al. Apr 2004 B1
6730107 Kelley et al. May 2004 B2
6733474 Kusleika May 2004 B2
6738144 Dogariu May 2004 B1
6740113 Vrba May 2004 B2
6746464 Makower Jun 2004 B1
6780157 Stephens et al. Aug 2004 B2
6795188 Ruck et al. Sep 2004 B2
6795196 Funakawa Sep 2004 B2
6798522 Stolte et al. Sep 2004 B2
6822798 Wu et al. Nov 2004 B2
6830559 Schock Dec 2004 B2
6832024 Gerstenberger et al. Dec 2004 B2
6842639 Winston et al. Jan 2005 B1
6847449 Bashkansky et al. Jan 2005 B2
6855115 Fonseca et al. Feb 2005 B2
6856138 Bohley Feb 2005 B2
6856400 Froggatt Feb 2005 B1
6856472 Herman et al. Feb 2005 B2
6860867 Seward et al. Mar 2005 B2
6866670 Rabiner et al. Mar 2005 B2
6878113 Miwa et al. Apr 2005 B2
6886411 Kjellman et al. May 2005 B2
6891984 Petersen et al. May 2005 B2
6895106 Wang et al. May 2005 B2
6898337 Averett et al. May 2005 B2
6900897 Froggatt May 2005 B2
6912051 Jensen Jun 2005 B2
6916329 Zhao Jul 2005 B1
6922498 Shah Jul 2005 B2
6937346 Nebendahl et al. Aug 2005 B2
6937696 Mostafavi Aug 2005 B1
6943939 DiJaili et al. Sep 2005 B1
6947147 Motamedi et al. Sep 2005 B2
6947787 Webler Sep 2005 B2
6949094 Yaron Sep 2005 B2
6952603 Gerber et al. Oct 2005 B2
6954737 Kalantar et al. Oct 2005 B2
6958042 Honda Oct 2005 B2
6961123 Wang et al. Nov 2005 B1
6966891 Ookubo et al. Nov 2005 B2
6969293 Thai Nov 2005 B2
6969395 Eskuri Nov 2005 B2
6985234 Anderson Jan 2006 B2
7004963 Wang et al. Feb 2006 B2
7006231 Ostrovsky et al. Feb 2006 B2
7010458 Wilt Mar 2006 B2
7024025 Sathyanarayana Apr 2006 B2
7027211 Ruffa Apr 2006 B1
7027743 Tucker et al. Apr 2006 B1
7033347 Appling Apr 2006 B2
7035484 Silberberg et al. Apr 2006 B2
7037269 Nix et al. May 2006 B2
7042573 Froggatt May 2006 B2
7044915 White et al. May 2006 B2
7044964 Jang et al. May 2006 B2
7048711 Rosenman et al. May 2006 B2
7049306 Konradi et al. May 2006 B2
7058239 Singh et al. Jun 2006 B2
7060033 White et al. Jun 2006 B2
7060421 Naundorf et al. Jun 2006 B2
7063679 Maguire et al. Jun 2006 B2
7068852 Braica Jun 2006 B2
7074188 Nair et al. Jul 2006 B2
7095493 Harres Aug 2006 B2
7110119 Maestle Sep 2006 B2
7113875 Terashima et al. Sep 2006 B2
7123777 Rondinelli et al. Oct 2006 B2
7130054 Ostrovsky et al. Oct 2006 B2
7139440 Rondinelli et al. Nov 2006 B2
7153299 Tu et al. Dec 2006 B1
7171078 Sasaki et al. Jan 2007 B2
7175597 Vince et al. Feb 2007 B2
7177491 Dave et al. Feb 2007 B2
7190464 Alphonse Mar 2007 B2
7215802 Klingensmith et al. May 2007 B2
7218811 Shigenaga et al. May 2007 B2
7236812 Ballerstadt et al. Jun 2007 B1
7245125 Harer et al. Jul 2007 B2
7245789 Bates et al. Jul 2007 B2
7249357 Landman et al. Jul 2007 B2
7291146 Steinke et al. Nov 2007 B2
7292715 Furnish Nov 2007 B2
7292885 Scott et al. Nov 2007 B2
7294124 Eidenschink Nov 2007 B2
7300460 Levine et al. Nov 2007 B2
7335161 Von Arx et al. Feb 2008 B2
7337079 Park et al. Feb 2008 B2
7355716 de Boer et al. Apr 2008 B2
7356367 Liang et al. Apr 2008 B2
7358921 Snyder et al. Apr 2008 B2
7359062 Chen et al. Apr 2008 B2
7359554 Klingensmith et al. Apr 2008 B2
7363927 Ravikumar Apr 2008 B2
7366376 Shishkov et al. Apr 2008 B2
7382949 Bouma et al. Jun 2008 B2
7387636 Cohn et al. Jun 2008 B2
7391520 Zhou et al. Jun 2008 B2
7397935 Kimmel et al. Jul 2008 B2
7399095 Rondinelli Jul 2008 B2
7408648 Kleen et al. Aug 2008 B2
7414779 Huber et al. Aug 2008 B2
7440087 Froggatt et al. Oct 2008 B2
7447388 Bates et al. Nov 2008 B2
7449821 Dausch Nov 2008 B2
7450165 Ahiska Nov 2008 B2
RE40608 Glover et al. Dec 2008 E
7458967 Appling et al. Dec 2008 B2
7463362 Lasker et al. Dec 2008 B2
7463759 Klingensmith et al. Dec 2008 B2
7491226 Palmaz et al. Feb 2009 B2
7515276 Froggatt et al. Apr 2009 B2
7527594 Vardi et al. May 2009 B2
7534251 WasDyke May 2009 B2
7535797 Peng et al. May 2009 B2
7547304 Johnson Jun 2009 B2
7564949 Sattler et al. Jul 2009 B2
7577471 Camus et al. Aug 2009 B2
7583857 Xu et al. Sep 2009 B2
7603165 Townsend et al. Oct 2009 B2
7612773 Magnin et al. Nov 2009 B2
7633627 Choma et al. Dec 2009 B2
7645229 Armstrong Jan 2010 B2
7658715 Park et al. Feb 2010 B2
7660452 Zwirn et al. Feb 2010 B2
7660492 Bates et al. Feb 2010 B2
7666204 Thornton et al. Feb 2010 B2
7672790 McGraw et al. Mar 2010 B2
7680247 Atzinger et al. Mar 2010 B2
7684991 Stohr et al. Mar 2010 B2
7711413 Feldman et al. May 2010 B2
7720322 Prisco May 2010 B2
7728986 Lasker et al. Jun 2010 B2
7734009 Brunner et al. Jun 2010 B2
7736317 Stephens et al. Jun 2010 B2
7742795 Stone et al. Jun 2010 B2
7743189 Brown et al. Jun 2010 B2
7762954 Nix et al. Jul 2010 B2
7766896 Kornkven Volk et al. Aug 2010 B2
7773792 Kimmel et al. Aug 2010 B2
7775981 Guracar et al. Aug 2010 B1
7777399 Eidenschink et al. Aug 2010 B2
7781724 Childers et al. Aug 2010 B2
7783337 Feldman et al. Aug 2010 B2
7787127 Galle et al. Aug 2010 B2
7792342 Barbu et al. Sep 2010 B2
7801343 Unal et al. Sep 2010 B2
7801590 Feldman et al. Sep 2010 B2
7813609 Petersen et al. Oct 2010 B2
7831081 Li Nov 2010 B2
7846101 Eberle et al. Dec 2010 B2
7853104 Oota et al. Dec 2010 B2
7853316 Milner et al. Dec 2010 B2
7860555 Saadat Dec 2010 B2
7862508 Davies et al. Jan 2011 B2
7872759 Tearney et al. Jan 2011 B2
7880868 Aoki Feb 2011 B2
7881763 Brauker et al. Feb 2011 B2
7909844 Alkhatib et al. Mar 2011 B2
7921854 Hennings et al. Apr 2011 B2
7927784 Simpson Apr 2011 B2
7929148 Kemp Apr 2011 B2
7930014 Huennekens et al. Apr 2011 B2
7930104 Baker et al. Apr 2011 B2
7936462 Jiang et al. May 2011 B2
7942852 Mas et al. May 2011 B2
7947012 Spurchise et al. May 2011 B2
7951186 Eidenschink et al. May 2011 B2
7952719 Brennan, III May 2011 B2
7972353 Hendriksen et al. Jul 2011 B2
7976492 Brauker et al. Jul 2011 B2
7977950 Maslen Jul 2011 B2
7978916 Klingensmith et al. Jul 2011 B2
7981041 McGahan Jul 2011 B2
7981151 Rowe Jul 2011 B2
7983737 Feldman et al. Jul 2011 B2
7993333 Oral et al. Aug 2011 B2
7995210 Tearney et al. Aug 2011 B2
7996060 Trofimov et al. Aug 2011 B2
7999938 Wang Aug 2011 B2
8021377 Eskuri Sep 2011 B2
8021420 Dolan Sep 2011 B2
8025622 Rold et al. Sep 2011 B2
8036732 Milner Oct 2011 B2
8040586 Smith et al. Oct 2011 B2
8047996 Goodnow et al. Nov 2011 B2
8049900 Kemp et al. Nov 2011 B2
8050478 Li et al. Nov 2011 B2
8050523 Younge et al. Nov 2011 B2
8052605 Muller et al. Nov 2011 B2
8057394 Dala-Krishna Nov 2011 B2
8059923 Bates et al. Nov 2011 B2
8070800 Lock et al. Dec 2011 B2
8080800 Hoctor et al. Dec 2011 B2
8088102 Adams et al. Jan 2012 B2
8100838 Wright et al. Jan 2012 B2
8104479 Glynn et al. Jan 2012 B2
8108030 Castella et al. Jan 2012 B2
8114102 Galdonik et al. Feb 2012 B2
8116605 Petersen et al. Feb 2012 B2
8125648 Milner et al. Feb 2012 B2
8126239 Sun et al. Feb 2012 B2
8133199 Weber et al. Mar 2012 B2
8133269 Flechsenhar et al. Mar 2012 B2
8140708 Zaharia et al. Mar 2012 B2
8148877 Jiang et al. Apr 2012 B2
8167932 Bourang et al. May 2012 B2
8172757 Jaffe et al. May 2012 B2
8177809 Mavani et al. May 2012 B2
8187191 Hancock et al. May 2012 B2
8187267 Pappone et al. May 2012 B2
8187830 Hu et al. May 2012 B2
8199218 Lee et al. Jun 2012 B2
8206429 Gregorich et al. Jun 2012 B2
8208995 Tearney et al. Jun 2012 B2
8222906 Wyar et al. Jul 2012 B2
8233681 Aylward et al. Jul 2012 B2
8233718 Klingensmith et al. Jul 2012 B2
8238624 Doi et al. Aug 2012 B2
8239938 Simeral et al. Aug 2012 B2
8277386 Ahmed et al. Oct 2012 B2
8280470 Milner et al. Oct 2012 B2
8289284 Glynn et al. Oct 2012 B2
8289522 Tearney et al. Oct 2012 B2
8298147 Huennekens et al. Oct 2012 B2
8298149 Hastings et al. Oct 2012 B2
8301000 Sillard et al. Oct 2012 B2
8309428 Lemmerhirt et al. Nov 2012 B2
8317713 Davies et al. Nov 2012 B2
8323201 Towfiq et al. Dec 2012 B2
8329053 Martin et al. Dec 2012 B2
8336643 Harleman Dec 2012 B2
8349000 Schreck Jan 2013 B2
8353945 Andreas et al. Jan 2013 B2
8353954 Cai et al. Jan 2013 B2
8357981 Martin et al. Jan 2013 B2
8361097 Patel et al. Jan 2013 B2
8386560 Ma et al. Feb 2013 B2
8398591 Mas et al. Mar 2013 B2
8412312 Judell et al. Apr 2013 B2
8417491 Trovato et al. Apr 2013 B2
8449465 Nair et al. May 2013 B2
8454685 Hariton et al. Jun 2013 B2
8454686 Alkhatib Jun 2013 B2
8475522 Jimenez et al. Jul 2013 B2
8478384 Schmitt et al. Jul 2013 B2
8486062 Belhe et al. Jul 2013 B2
8486063 Werneth et al. Jul 2013 B2
8491567 Magnin et al. Jul 2013 B2
8500798 Rowe et al. Aug 2013 B2
8550911 Sylla Oct 2013 B2
8594757 Boppart et al. Nov 2013 B2
8597349 Alkhatib Dec 2013 B2
8600477 Beyar et al. Dec 2013 B2
8600917 Schimert et al. Dec 2013 B1
8601056 Lauwers et al. Dec 2013 B2
8620055 Barratt et al. Dec 2013 B2
8644910 Rousso et al. Feb 2014 B2
20010007940 Tu et al. Jul 2001 A1
20010029337 Pantages et al. Oct 2001 A1
20010037073 White et al. Nov 2001 A1
20010046345 Snyder et al. Nov 2001 A1
20010049548 Vardi et al. Dec 2001 A1
20020034276 Hu et al. Mar 2002 A1
20020041723 Ronnekleiv et al. Apr 2002 A1
20020049375 Strommer et al. Apr 2002 A1
20020069676 Kopp et al. Jun 2002 A1
20020089335 Williams Jul 2002 A1
20020099289 Crowley Jul 2002 A1
20020163646 Anderson Nov 2002 A1
20020186818 Arnaud et al. Dec 2002 A1
20020196446 Roth et al. Dec 2002 A1
20020197456 Pope Dec 2002 A1
20030004412 Izatt et al. Jan 2003 A1
20030016604 Hanes Jan 2003 A1
20030018273 Corl et al. Jan 2003 A1
20030023153 Izatt et al. Jan 2003 A1
20030032886 Dgany et al. Feb 2003 A1
20030050871 Broughton Mar 2003 A1
20030065371 Satake Apr 2003 A1
20030069723 Hegde Apr 2003 A1
20030077043 Hamm et al. Apr 2003 A1
20030085635 Davidsen May 2003 A1
20030090753 Takeyama et al. May 2003 A1
20030092995 Thompson May 2003 A1
20030093059 Griffin et al. May 2003 A1
20030103212 Westphal et al. Jun 2003 A1
20030152259 Belykh et al. Aug 2003 A1
20030181802 Ogawa Sep 2003 A1
20030187369 Lewis et al. Oct 2003 A1
20030194165 Silberberg et al. Oct 2003 A1
20030195419 Harada Oct 2003 A1
20030208116 Liang et al. Nov 2003 A1
20030212491 Mitchell et al. Nov 2003 A1
20030219202 Loeb et al. Nov 2003 A1
20030220749 Chen et al. Nov 2003 A1
20030228039 Green Dec 2003 A1
20040015065 Panescu et al. Jan 2004 A1
20040023317 Motamedi et al. Feb 2004 A1
20040028333 Lomas Feb 2004 A1
20040037742 Jen et al. Feb 2004 A1
20040042066 Kinoshita et al. Mar 2004 A1
20040054287 Stephens Mar 2004 A1
20040067000 Bates et al. Apr 2004 A1
20040068161 Couvillon Apr 2004 A1
20040082844 Vardi et al. Apr 2004 A1
20040092830 Scott et al. May 2004 A1
20040106853 Moriyama Jun 2004 A1
20040111552 Arimilli et al. Jun 2004 A1
20040126048 Dave et al. Jul 2004 A1
20040143160 Couvillon Jul 2004 A1
20040146546 Gravett et al. Jul 2004 A1
20040186369 Lam Sep 2004 A1
20040186558 Pavcnik et al. Sep 2004 A1
20040195512 Crosetto Oct 2004 A1
20040220606 Goshgarian Nov 2004 A1
20040225220 Rich Nov 2004 A1
20040239938 Izatt Dec 2004 A1
20040242990 Brister et al. Dec 2004 A1
20040248439 Gernhardt et al. Dec 2004 A1
20040260236 Manning et al. Dec 2004 A1
20050013778 Green et al. Jan 2005 A1
20050031176 Hertel et al. Feb 2005 A1
20050036150 Izatt et al. Feb 2005 A1
20050078317 Law et al. Apr 2005 A1
20050101859 Maschke May 2005 A1
20050140582 Lee et al. Jun 2005 A1
20050140682 Sumanaweera et al. Jun 2005 A1
20050140981 Waelti Jun 2005 A1
20050140984 Hitzenberger Jun 2005 A1
20050147303 Zhou et al. Jul 2005 A1
20050165439 Weber et al. Jul 2005 A1
20050171433 Boppart et al. Aug 2005 A1
20050171438 Chen et al. Aug 2005 A1
20050182297 Gravenstein et al. Aug 2005 A1
20050196028 Kleen et al. Sep 2005 A1
20050197585 Brockway et al. Sep 2005 A1
20050213103 Everett et al. Sep 2005 A1
20050215942 Abrahamson et al. Sep 2005 A1
20050234445 Conquergood et al. Oct 2005 A1
20050243322 Lasker et al. Nov 2005 A1
20050249391 Kimmel Nov 2005 A1
20050251567 Ballew et al. Nov 2005 A1
20050254059 Alphonse Nov 2005 A1
20050264823 Zhu et al. Dec 2005 A1
20060013523 Childlers et al. Jan 2006 A1
20060015126 Sher Jan 2006 A1
20060029634 Berg et al. Feb 2006 A1
20060036167 Shina Feb 2006 A1
20060038115 Maas Feb 2006 A1
20060039004 de Boer et al. Feb 2006 A1
20060041180 Viswanathan et al. Feb 2006 A1
20060045536 Arahira Mar 2006 A1
20060055936 Yun et al. Mar 2006 A1
20060058622 Tearney et al. Mar 2006 A1
20060064009 Webler et al. Mar 2006 A1
20060067620 Shishkov et al. Mar 2006 A1
20060072808 Grimm et al. Apr 2006 A1
20060074442 Noriega et al. Apr 2006 A1
20060098927 Schmidt et al. May 2006 A1
20060100694 Globerman May 2006 A1
20060106375 Werneth et al. May 2006 A1
20060132790 Gutin Jun 2006 A1
20060135870 Webler Jun 2006 A1
20060142703 Carter et al. Jun 2006 A1
20060142733 Forsberg Jun 2006 A1
20060173299 Romley et al. Aug 2006 A1
20060179255 Yamazaki Aug 2006 A1
20060184048 Saadat Aug 2006 A1
20060187462 Srinivasan et al. Aug 2006 A1
20060187537 Huber et al. Aug 2006 A1
20060195269 Yeatman et al. Aug 2006 A1
20060204119 Feng et al. Sep 2006 A1
20060229591 Lee Oct 2006 A1
20060239312 Kewitsch et al. Oct 2006 A1
20060241342 Macaulay et al. Oct 2006 A1
20060241465 Huennekens et al. Oct 2006 A1
20060241503 Schmitt et al. Oct 2006 A1
20060244973 Yun et al. Nov 2006 A1
20060258895 Maschke Nov 2006 A1
20060264743 Kleen et al. Nov 2006 A1
20060267756 Kates Nov 2006 A1
20060270976 Savage et al. Nov 2006 A1
20060276709 Khamene et al. Dec 2006 A1
20060279742 Tearney et al. Dec 2006 A1
20060279743 Boesser et al. Dec 2006 A1
20060285638 Boese et al. Dec 2006 A1
20060287595 Maschke Dec 2006 A1
20060293597 Johnson et al. Dec 2006 A1
20070015969 Feldman et al. Jan 2007 A1
20070016029 Donaldson et al. Jan 2007 A1
20070016034 Donaldson Jan 2007 A1
20070016062 Park et al. Jan 2007 A1
20070027390 Maschke et al. Feb 2007 A1
20070036417 Argiro et al. Feb 2007 A1
20070038061 Huennekens et al. Feb 2007 A1
20070038121 Feldman et al. Feb 2007 A1
20070038125 Kleen et al. Feb 2007 A1
20070043292 Camus et al. Feb 2007 A1
20070043597 Donaldson Feb 2007 A1
20070049847 Osborne Mar 2007 A1
20070060973 Ludvig et al. Mar 2007 A1
20070065077 Childers et al. Mar 2007 A1
20070066888 Maschke Mar 2007 A1
20070066890 Maschke Mar 2007 A1
20070066983 Maschke Mar 2007 A1
20070084995 Newton et al. Apr 2007 A1
20070100226 Yankelevitz et al. May 2007 A1
20070135887 Maschke Jun 2007 A1
20070142707 Wiklof et al. Jun 2007 A1
20070156019 Larkin et al. Jul 2007 A1
20070161893 Milner et al. Jul 2007 A1
20070161896 Adachi et al. Jul 2007 A1
20070161963 Smalling Jul 2007 A1
20070162860 Muralidharan et al. Jul 2007 A1
20070165141 Srinivas et al. Jul 2007 A1
20070167710 Unal Jul 2007 A1
20070167804 Park et al. Jul 2007 A1
20070191682 Rolland et al. Aug 2007 A1
20070201736 Klingensmith et al. Aug 2007 A1
20070206193 Pesach Sep 2007 A1
20070208276 Kornkven Volk et al. Sep 2007 A1
20070225220 Ming et al. Sep 2007 A1
20070225590 Ramos Sep 2007 A1
20070229801 Tearney et al. Oct 2007 A1
20070232872 Prough et al. Oct 2007 A1
20070232874 Ince Oct 2007 A1
20070232890 Hirota Oct 2007 A1
20070232891 Hirota Oct 2007 A1
20070232892 Hirota Oct 2007 A1
20070232893 Tanioka Oct 2007 A1
20070232933 Gille et al. Oct 2007 A1
20070238957 Yared Oct 2007 A1
20070247033 Eidenschink et al. Oct 2007 A1
20070250000 Magnin et al. Oct 2007 A1
20070250036 Volk et al. Oct 2007 A1
20070258094 Izatt et al. Nov 2007 A1
20070260138 Feldman et al. Nov 2007 A1
20070278389 Ajgaonkar et al. Dec 2007 A1
20070287914 Cohen Dec 2007 A1
20080002183 Yatagai et al. Jan 2008 A1
20080013093 Izatt et al. Jan 2008 A1
20080021275 Tearney et al. Jan 2008 A1
20080027481 Gilson et al. Jan 2008 A1
20080043024 Schiwietz et al. Feb 2008 A1
20080045842 Furnish Feb 2008 A1
20080051660 Kakadaris et al. Feb 2008 A1
20080063304 Russak et al. Mar 2008 A1
20080085041 Breeuwer Apr 2008 A1
20080095465 Mullick et al. Apr 2008 A1
20080095714 Castella et al. Apr 2008 A1
20080097194 Milner Apr 2008 A1
20080101667 Begelman et al. May 2008 A1
20080108867 Zhou May 2008 A1
20080114254 Matcovitch et al. May 2008 A1
20080119739 Vardi et al. May 2008 A1
20080124495 Horn et al. May 2008 A1
20080125772 Stone et al. May 2008 A1
20080139897 Ainsworth et al. Jun 2008 A1
20080143707 Mitchell Jun 2008 A1
20080146941 Dala-Krishna Jun 2008 A1
20080147111 Johnson et al. Jun 2008 A1
20080154128 Milner Jun 2008 A1
20080161696 Schmitt et al. Jul 2008 A1
20080171944 Brenneman et al. Jul 2008 A1
20080175465 Jiang et al. Jul 2008 A1
20080177183 Courtney et al. Jul 2008 A1
20080180683 Kemp Jul 2008 A1
20080181477 Izatt et al. Jul 2008 A1
20080187201 Liang et al. Aug 2008 A1
20080228086 Ilegbusi et al. Sep 2008 A1
20080247622 Aylward et al. Oct 2008 A1
20080247716 Thomas et al. Oct 2008 A1
20080262470 Lee et al. Oct 2008 A1
20080262489 Steinke Oct 2008 A1
20080269599 Csavoy et al. Oct 2008 A1
20080281205 Naghavi et al. Nov 2008 A1
20080281248 Angheloiu et al. Nov 2008 A1
20080285043 Fercher et al. Nov 2008 A1
20080287795 Klingensmith et al. Nov 2008 A1
20080291463 Milner et al. Nov 2008 A1
20080292173 Hsieh et al. Nov 2008 A1
20080294034 Krueger et al. Nov 2008 A1
20080298655 Edwards Dec 2008 A1
20080306766 Ozeki et al. Dec 2008 A1
20090009801 Tabuki Jan 2009 A1
20090018393 Dick et al. Jan 2009 A1
20090034813 Dikmen et al. Feb 2009 A1
20090043191 Castella et al. Feb 2009 A1
20090046295 Kemp et al. Feb 2009 A1
20090052614 Hempel et al. Feb 2009 A1
20090069843 Agnew Mar 2009 A1
20090079993 Yatagai et al. Mar 2009 A1
20090088650 Corl Apr 2009 A1
20090093980 Kemp et al. Apr 2009 A1
20090122320 Petersen et al. May 2009 A1
20090138544 Wegenkittl et al. May 2009 A1
20090149739 Maschke Jun 2009 A9
20090156941 Moore Jun 2009 A1
20090174886 Inoue Jul 2009 A1
20090174931 Huber et al. Jul 2009 A1
20090177090 Grunwald et al. Jul 2009 A1
20090177183 Pinkernell et al. Jul 2009 A1
20090195514 Glynn et al. Aug 2009 A1
20090196470 Carl et al. Aug 2009 A1
20090198125 Nakabayashi et al. Aug 2009 A1
20090203991 Papaioannou et al. Aug 2009 A1
20090264768 Courtney et al. Oct 2009 A1
20090269014 Winberg et al. Oct 2009 A1
20090270695 McEowen Oct 2009 A1
20090284322 Harrison et al. Nov 2009 A1
20090284332 Moore et al. Nov 2009 A1
20090284749 Johnson et al. Nov 2009 A1
20090290167 Flanders et al. Nov 2009 A1
20090292048 Li et al. Nov 2009 A1
20090299195 Muller et al. Dec 2009 A1
20090299284 Holman et al. Dec 2009 A1
20090318951 Kashkarov et al. Dec 2009 A1
20090326634 Vardi Dec 2009 A1
20100007669 Bethune et al. Jan 2010 A1
20100030042 Denninghoff et al. Feb 2010 A1
20100061611 Xu et al. Mar 2010 A1
20100063400 Hall et al. Mar 2010 A1
20100087732 Eberle et al. Apr 2010 A1
20100094125 Younge et al. Apr 2010 A1
20100094127 Xu Apr 2010 A1
20100094135 Fang-Yen et al. Apr 2010 A1
20100094143 Mahapatra et al. Apr 2010 A1
20100113919 Maschke May 2010 A1
20100125238 Lye et al. May 2010 A1
20100125268 Gustus et al. May 2010 A1
20100125648 Zaharia et al. May 2010 A1
20100128348 Taverner May 2010 A1
20100152717 Keeler Jun 2010 A1
20100160788 Davies et al. Jun 2010 A1
20100161023 Cohen et al. Jun 2010 A1
20100168714 Burke et al. Jul 2010 A1
20100179421 Tupin Jul 2010 A1
20100179426 Davies et al. Jul 2010 A1
20100220334 Condit et al. Sep 2010 A1
20100226607 Zhang et al. Sep 2010 A1
20100234736 Corl Sep 2010 A1
20100249601 Courtney Sep 2010 A1
20100256616 Katoh et al. Oct 2010 A1
20100272432 Johnson Oct 2010 A1
20100284590 Peng et al. Nov 2010 A1
20100290693 Cohen et al. Nov 2010 A1
20100331950 Strommer Dec 2010 A1
20110010925 Nix et al. Jan 2011 A1
20110021926 Spencer et al. Jan 2011 A1
20110025853 Richardson Feb 2011 A1
20110026797 Declerck et al. Feb 2011 A1
20110032533 Izatt et al. Feb 2011 A1
20110034801 Baumgart Feb 2011 A1
20110044546 Pan et al. Feb 2011 A1
20110066073 Kuiper et al. Mar 2011 A1
20110071401 Hastings et al. Mar 2011 A1
20110072405 Chen et al. Mar 2011 A1
20110077528 Kemp et al. Mar 2011 A1
20110080591 Johnson et al. Apr 2011 A1
20110087104 Moore et al. Apr 2011 A1
20110137140 Tearney et al. Jun 2011 A1
20110144502 Zhou et al. Jun 2011 A1
20110152771 Milner et al. Jun 2011 A1
20110157597 Lu et al. Jun 2011 A1
20110160586 Li et al. Jun 2011 A1
20110178413 Schmitt et al. Jul 2011 A1
20110190586 Kemp Aug 2011 A1
20110216378 Poon et al. Sep 2011 A1
20110220985 Son et al. Sep 2011 A1
20110238061 van der Weide et al. Sep 2011 A1
20110238083 Moll et al. Sep 2011 A1
20110245669 Zhang Oct 2011 A1
20110249094 Wang et al. Oct 2011 A1
20110257545 Suri Oct 2011 A1
20110264125 Wilson et al. Oct 2011 A1
20110274329 Mathew et al. Nov 2011 A1
20110282334 Groenhoff Nov 2011 A1
20110301684 Fischell et al. Dec 2011 A1
20110306995 Moberg Dec 2011 A1
20110319752 Steinberg Dec 2011 A1
20120004529 Tolkowsky et al. Jan 2012 A1
20120004668 Wallace et al. Jan 2012 A1
20120013914 Kemp et al. Jan 2012 A1
20120016344 Kusakabe Jan 2012 A1
20120016395 Olson Jan 2012 A1
20120022360 Kemp Jan 2012 A1
20120026503 Lewandowski et al. Feb 2012 A1
20120029007 Graham et al. Feb 2012 A1
20120059253 Wang et al. Mar 2012 A1
20120059368 Takaoka et al. Mar 2012 A1
20120062843 Ferguson et al. Mar 2012 A1
20120065481 Hunter et al. Mar 2012 A1
20120071823 Chen Mar 2012 A1
20120071838 Fojtik Mar 2012 A1
20120075638 Rollins et al. Mar 2012 A1
20120083696 Kitamura Apr 2012 A1
20120095340 Smith Apr 2012 A1
20120095372 Sverdlik et al. Apr 2012 A1
20120108943 Bates et al. May 2012 A1
20120113108 Dala-Krishna May 2012 A1
20120116353 Arnold et al. May 2012 A1
20120130243 Balocco et al. May 2012 A1
20120130247 Waters et al. May 2012 A1
20120136259 Milner et al. May 2012 A1
20120136427 Palmaz et al. May 2012 A1
20120137075 Vorbach May 2012 A1
20120155734 Barratt et al. Jun 2012 A1
20120158101 Stone et al. Jun 2012 A1
20120162660 Kemp Jun 2012 A1
20120165661 Kemp et al. Jun 2012 A1
20120170848 Kemp et al. Jul 2012 A1
20120172698 Teo et al. Jul 2012 A1
20120176607 Ott Jul 2012 A1
20120184853 Waters Jul 2012 A1
20120184859 Shah et al. Jul 2012 A1
20120184977 Wolf Jul 2012 A1
20120215094 Rahimian et al. Aug 2012 A1
20120220836 Alpert et al. Aug 2012 A1
20120220851 Razansky et al. Aug 2012 A1
20120220865 Brown et al. Aug 2012 A1
20120220874 Hancock et al. Aug 2012 A1
20120220883 Manstrom et al. Aug 2012 A1
20120224751 Kemp et al. Sep 2012 A1
20120226153 Brown et al. Sep 2012 A1
20120230565 Steinberg et al. Sep 2012 A1
20120232400 Dickinson et al. Sep 2012 A1
20120238869 Schmitt et al. Sep 2012 A1
20120238956 Yamada et al. Sep 2012 A1
20120244043 Leblanc et al. Sep 2012 A1
20120250028 Schmitt et al. Oct 2012 A1
20120253186 Simpson et al. Oct 2012 A1
20120253192 Cressman Oct 2012 A1
20120253276 Govari et al. Oct 2012 A1
20120257210 Whitney et al. Oct 2012 A1
20120262720 Brown et al. Oct 2012 A1
20120265077 Gille et al. Oct 2012 A1
20120265268 Blum et al. Oct 2012 A1
20120265296 McNamara et al. Oct 2012 A1
20120271170 Emelianov et al. Oct 2012 A1
20120271175 Moore et al. Oct 2012 A1
20120271339 O'Beirne et al. Oct 2012 A1
20120274338 Baks et al. Nov 2012 A1
20120276390 Ji et al. Nov 2012 A1
20120277722 Gerber et al. Nov 2012 A1
20120279764 Jiang et al. Nov 2012 A1
20120283758 Miller et al. Nov 2012 A1
20120289987 Wilson et al. Nov 2012 A1
20120299439 Huang Nov 2012 A1
20120310081 Adler et al. Dec 2012 A1
20120310332 Murray et al. Dec 2012 A1
20120319535 Dausch Dec 2012 A1
20120323075 Younge et al. Dec 2012 A1
20120323127 Boyden et al. Dec 2012 A1
20120330141 Brown et al. Dec 2012 A1
20130015975 Huennekens et al. Jan 2013 A1
20130023762 Huennekens et al. Jan 2013 A1
20130023763 Huennekens et al. Jan 2013 A1
20130026655 Lee et al. Jan 2013 A1
20130030295 Huennekens et al. Jan 2013 A1
20130030303 Ahmed et al. Jan 2013 A1
20130030410 Drasler et al. Jan 2013 A1
20130053949 Pintor et al. Feb 2013 A1
20130109958 Baumgart et al. May 2013 A1
20130109959 Baumgart et al. May 2013 A1
20130137980 Waters et al. May 2013 A1
20130150716 Stigall et al. Jun 2013 A1
20130158594 Carrison et al. Jun 2013 A1
20130218201 Obermiller et al. Aug 2013 A1
20130218267 Braido et al. Aug 2013 A1
20130223789 Lee et al. Aug 2013 A1
20130223798 Jenner et al. Aug 2013 A1
20130296704 Magnin et al. Nov 2013 A1
20130303907 Corl Nov 2013 A1
20130303920 Corl Nov 2013 A1
20130310698 Judell et al. Nov 2013 A1
20130331820 Itou et al. Dec 2013 A1
20130338766 Hastings et al. Dec 2013 A1
20130339958 Droste et al. Dec 2013 A1
20140039294 Jiang Feb 2014 A1
20140180067 Stigall et al. Jun 2014 A1
20140180128 Corl Jun 2014 A1
20140200438 Millett et al. Jul 2014 A1
Foreign Referenced Citations (79)
Number Date Country
1041373 Oct 2000 EP
01172637 Jan 2002 EP
2438877 Apr 2012 EP
2280261 Jan 1995 GB
2000-262461 Sep 2000 JP
2000-292260 Oct 2000 JP
2001-125009 May 2001 JP
2001-272331 Oct 2001 JP
2002-374034 Dec 2002 JP
2003-143783 May 2003 JP
2003-172690 Jun 2003 JP
2003-256876 Sep 2003 JP
2003-287534 Oct 2003 JP
2005-274380 Oct 2005 JP
2006-184284 Jul 2006 JP
2006-266797 Oct 2006 JP
2006-313158 Nov 2006 JP
2007-024677 Feb 2007 JP
2009-233001 Oct 2009 JP
2011-56786 Mar 2011 JP
9101156 Feb 1991 WO
9216865 Oct 1992 WO
9306213 Apr 1993 WO
9308829 May 1993 WO
9838907 Sep 1998 WO
9857583 Dec 1998 WO
0011511 Mar 2000 WO
0044296 Aug 2000 WO
0111409 Feb 2001 WO
03062802 Jul 2003 WO
03073950 Sep 2003 WO
2004010856 Feb 2004 WO
2004023992 Mar 2004 WO
2004096049 Nov 2004 WO
2005047813 May 2005 WO
2005106695 Nov 2005 WO
2006029634 Mar 2006 WO
2006037132 Apr 2006 WO
2006039091 Apr 2006 WO
2006061829 Jun 2006 WO
2006068875 Jun 2006 WO
2006111704 Oct 2006 WO
2006119416 Nov 2006 WO
2006121851 Nov 2006 WO
2006130802 Dec 2006 WO
2007002685 Jan 2007 WO
2007025230 Mar 2007 WO
2007045690 Apr 2007 WO
2007058895 May 2007 WO
2007067323 Jun 2007 WO
2007084995 Jul 2007 WO
2008058084 May 2008 WO
2008069991 Jun 2008 WO
2008107905 Sep 2008 WO
2009009799 Jan 2009 WO
2009009801 Jan 2009 WO
2009046431 Apr 2009 WO
2009121067 Oct 2009 WO
2009137704 Nov 2009 WO
2011006886 Jan 2011 WO
2011038048 Mar 2011 WO
2011081688 Jul 2011 WO
2012003369 Jan 2012 WO
2013033414 Mar 2012 WO
2012061935 May 2012 WO
2012071388 May 2012 WO
2012087818 Jun 2012 WO
2012098194 Jul 2012 WO
2012109676 Aug 2012 WO
2012130289 Oct 2012 WO
2012154767 Nov 2012 WO
2012155040 Nov 2012 WO
2013033415 Mar 2013 WO
2013033418 Mar 2013 WO
2013033489 Mar 2013 WO
2013033490 Mar 2013 WO
2013033592 Mar 2013 WO
2013126390 Aug 2013 WO
2014109879 Jul 2014 WO
Non-Patent Literature Citations (190)
Entry
Sihan et al., 2008, A novel approach to quantitative analysis of intraluminal optical coherence tomography imaging, Comput. Cardiol:1089-1092.
Siwy et al., 2003, Electro-responsive asymmetric nanopores in polyimide with stable ion-current signal, Applied Physics A: Materials Science & Processing 76:781-785.
Smith et al., 1989, Absolute displacement measurements using modulation of the spectrum of white light in a Michelson interferometer, Applied Optics, 28(16):3339-3342.
Smith, 1997, The Scientist and Engineer's Guide to Digital Signal Processing, California Technical Publishing, San Diego, CA:432-436.
Soller, 2003, Polarization diverse optical frequency domain interferometry:All coupler implementation, Bragg Grating, Photosensitivity, and Poling in Glass Waveguides Conference MB4:30-32.
Song et al., 2012, Active tremor cancellation by a “Smart” handheld vitreoretinal microsurgical tool using swept source optical coherence tomography, Optics Express, 20(21):23414-23421.
Stenqvist et al., 1983, Stiffness of central venous catheters, Acta Anaesthesiol Scand., 2:153-157.
Strickland, 1970, Time-Domain Reflectometer Measurements, Tektronix, Beaverton, OR, (107 pages).
Strobl et al., 2009, An Introduction to Recursive Partitioning:Rationale, Application and Characteristics of Classification and Regression Trees, Bagging and Random Forests, Psychol Methods., 14(4):323-348.
Sutcliffe et al., 1986, Dynamics of UV laser ablation of organic polymer surfaces, Journal of Applied Physics, 60(9):3315-3322.
Suzuki, 2013, A novel guidewire approach for handling acute-angle bifurcations, J Inv Cardiol 25(1):48-54.
Tanimoto et al., 2008, A novel approach for quantitative analysis of intracoronary optical coherence tomography: high inter-observer agreement with computer-assisted contour detection, Cathet Cardiovascular Intervent., 72(2):228-235.
Tearney et al., 1997, In vivo Endoscopic Optical Biopsy with Optical Coherence Tomography, Science, 276:2037-2039.
Tonino et al., 2009, Fractional flow reserve versus angiography for guiding percutaneous coronary intervention, The New England Journal of Medicine, 360:213-224.
Toregeani et al., 2008, Evaluation of hemodialysis arteriovenous fistula maturation by color-flow Doppler ultrasound, J Vasc. Bras. 7(3):203-213.
Translation of Notice of Reason(s) for Refusal dated Apr. 30, 2014, for Japanese Patent Application No. 2011-508677, (5 pages).
Translation of Notice of Reason(s) for Refusal dated May 25, 2012, for Japanese Patent Application No. 2009-536425, (3 pages).
Translation of Notice of Reason(s) for Refusal dated Nov. 22, 2012, for Japanese Patent Application No. 2010-516304, (6 pages).
Traunecker et al., 1991, Bispecific single chain molecules (Janusins) target cytotoxic lymphocytes on HIV infected cells, EMBO J., 10:3655-3659.
Trolier-McKinstry et. al., 2004, Thin Film Piezoelectric for MEMS, Journal of Electroceramics 12:7-17.
Tuniz et al., 2010, Weaving the invisible thread: design of an optically invisible metamaterial fibre, Optics Express 18 (17):18095-18105.
Turk et al., 1991, Eigenfaces for Recognition, Journal of Cognitive Neuroscience 3(1):71-86.
Tuzel et al., 2006, Region Covariance: A Fast Descriptor for Detection and Classification, European Conference on Computer Vision (ECCV).
Urban et al., 2010, Design of a Pressure Sensor Based on Optical Bragg Grating Lateral Deformation, Sensors (Basel), 10(12):11212-11225.
Vakhtin et al., 2003, Common-path interferometer for frequency-domain optical coherence tomography, Applied Optics, 42(34):6953-6958.
Vakoc et al., 2005, Phase-Resolved Optical Frequency Domain Imaging, Optics Express 13(14):5483-5493.
Verhoeyen et al., 1988, Reshaping human antibodies: grafting an antilysozyme activity, Science, 239:1534-1536.
Villard et al., 2002, Use of a blood substitute to determine instantaneous murine right ventricular thickening with optical coherence tomography, Circulation, 105:1843-1849.
Wang et al., 2002, Optimizing the Beam Patten of a Forward-Viewing Ring-Annular Ultrasound Array for Intravascular Imaging, Transactions on Ultrasonics, Ferroelectrics, and Frequency Control, 49(12).
Wang et al., 2006, Multiple biomarkers for the prediction of first major cardiovascular events and death, The New England Journal of Medicine, 355(25):2631-2639.
Wang et al., 2009, Robust Guidewire Tracking in Fluoroscopy, IEEE Conference on Computer Vision and Pattern Recognition—CCPR 2009:691-698.
Wang et al., 2011, In vivo intracardiac optical coherence tomography imaging through percutaneous access: toward image-guided radio-frequency ablation, J. Biomed. Opt. 0001 16(11):110505-1 (3 pages).
Waterhouse et. al., 1993, Combinatorial infection and in vivo recombination: a strategy for making large phage antibody repertoires, Nucleic Acids Res., 21:2265-2266.
Wegener, 2011, 3D Photonic Metamaterials and Invisibility Cloaks: The Method of Making, MEMS 2011, Cancun, Mexico, Jan. 23-27, 2011.
West et al., 1991, Arterial insufficiency in hemodialysis access procedures: correction by banding technique, Transpl Proc 23(2):1838-40.
Wyawahare et al., 2009, Image registration techniques: an overview, International Journal of Signal Processing, Image Processing and Pattern Recognition, 2(3):11-28.
Yaqoob et al., 2006, Methods and application areas of endoscopic optical coherence tomography, J. Biomed. Opt., 11, 063001-1-063001-19.
Yasuno et al., 2004, Polarization-sensitive complex Fourier domain optical coherence tomography for Jones matrix imaging of biological samples, Applied Physics Letters 85(15):3023-3025.
Zhang et al., 2004, Full range polarization-sensitive Fourier domain optical coherence tomography, Optics Express, 12(24):6033-6039.
Zitova et al., 2003, Image registration methods: A survey. Image and Vision Computing, 21(11):977-1000.
Abdi et al., 2010, Principal component analysis, Wiley Interdisciplinary Reviews: Computational Statistics 2:433-459.
Adler et al., 2007, Phase-Sensitive Optical Coherence Tomography at up to 370,000 Lines Per Second Using Buffered Fourier Domain Mode-Locked Lasers, Optics Letters, 32(6):626-628.
Agresti, 1996, Models for Matched Pairs, Chapter 8, An Introduction to Categorical Data Analysis, Wiley-Interscience A John Wiley & Sons, Inc., Publication, Hoboken, New Jersey.
Akasheh et al., 2004, Development of piezoelectric micromachined ultrasonic transducers, Sensors and Actuators A Physical, 111:275-287.
Amini et al., 1990, Using dynamic programming for solving variational problems in vision, IEEE Transactions on Pattern Analysis and Machine Intelligence, 12(9):855-867.
Bail et al., 1996, Optical coherence tomography with the “Spectral Radar”—Fast optical analysis in volume scatterers by short coherence interferometry, Optics Letters 21(14):1087-1089.
Bain, 2011, Privacy protection and face recognition, Chapter 3, Handbook of Face Recognition, Stan et al., Springer-Verlag.
Barnea et al., 1972, A class of algorithms for fast digital image registration, IEEE Trans. Computers, 21(2):179-186.
Blanchet et al., 1993, Laser Ablation and the Production of Polymer Films, Science, 262(5134):719-721.
Bonnema, 2008, Imaging Tissue Engineered Blood Vessel Mimics with Optical Tomography, College of Optical Sciences dissertation, University of Arizona (252 pages).
Bouma et al., 1999, Power-efficient nonreciprocal interferometer and linear-scanning fiber-optic catheter for optical coherence tomography, Optics Letters, 24(8):531-533.
Breiman, 2001, Random forests, Machine Learning 45:5-32.
Brown, 1992, A survey of image registration techniques, ACM Computing Surveys 24(4):325-376.
Bruining et al., 2009, Intravascular Ultrasound Registration/Integration with Coronary Angiography, Cardiology Clinics, 27(3):531-540.
Brummer, 1997, An euclidean distance measure between covariance matrices of speechcepstra for text-independent speaker recognition, in Proc. South African Symp. Communications and Signal Processing:167-172.
Burr et al., 2005, Searching for the Center of an Ellipse in Proceedings of the 17th Canadian Conference on Computational Geometry:260-263.
Canny, 1986, A computational approach to edge detection, IEEE Trans. Pattern Anal. Mach. Intel. 8:679-698.
Cavalli et al., 2010, Nanosponge formulations as oxygen delivery systems, International Journal of Pharmaceutics 402:254-257.
Choma et al., 2003, Sensitivity Advantage of Swept Source and Fourier Domain Optical Coherence Tomography, Optics Express 11(18):2183-2189.
Clarke et al., 1995, Hypoxia and myocardial ischaemia during peripheral angioplasty, Clinical Radiology, 50(5):301-303.
Collins, 1993, Coronary flow reserve, British Heart Journal 69:279-281.
Communication Mechanisms for Distributed Real-Time Applications, NI Developer Zone, http://zone.ni.eom/devzone/cda/tut/p/id/3105, accessed Jul. 23, 2007.
Cook, 2007, Use and misuse of receiver operating characteristic curve in risk prediction, Circulation 115(7):928-35.
D'Agostino et al., 2001, Validation of the Framingham coronary heart disease prediction score: results of a multiple ethnic group investigation, JAMA 286:180-187.
David et al., 1974, Protein iodination with solid-state lactoperoxidase, Biochemistry 13:1014-1021.
Davies et al., 1985, Plaque fissuring—the cause of acute myocardial infarction, sudden ischaemic death, and crescendo angina, British Heart Journal 53:363-373.
Davies et al., 1993, Risk of thrombosis in human atherosclerotic plaques: role of extracellular lipid, macrophage, and smooth muscle cell content, British Heart Journal 69:377-381.
Deterministic Data Streaming in Distributed Data Acquisition Systems, NI Developer Zone, “What is Developer Zone?”, http://zone.ni.eom/devzone/cda/tut/p/id/3105, accessed Jul. 23, 2007.
Eigenwillig, 2008, K-Space Linear Fourier Domain Mode Locked Laser and Applications for Optical Coherence Tomography, Optics Express 16(12):8916-8937.
Elghanian et al., 1997, Selective colorimetric detection of polynucleotides based on the distance-dependent optical properties of gold nanoparticles, Science, 277(5329):1078-1080.
Ergun et al., 2003, Capacitive Micromachined Ultrasonic Transducers:Theory and Technology, Journal of Aerospace Engineering, 16(2):76-84.
Evans et al., 2006, Optical coherence tomography to identify intramucosa carcinoma and high-grade dysplasia in Barrett's esophagus, Clin Gast Hepat 4(1):38-43.
Fatemi et al., 1999, Vibro-acoustography: an imaging modality based on ultrasound-stimulated acoustic emission, PNAS U.S.A., 96(12):6603-6608.
Felzenszwalb et al., 2005, Pictorial Structures for Object Recognition, International Journal of Computer Vision, 61 (1):55-79.
Ferring et al., 2008, Vasculature ultrasound for the pre-operative evaluation prior to arteriovenous fistula formation for haemodialysis: review of the evidence, Nephrol. Dial. Transplant. 23(6):1809-1815.
Fischler et al., 1973, The representation and matching of pictorial structures, IEEE Transactions on Computer 22:67-92.
Fleming et al., 2010, Real-time monitoring of cardiac radio-frequency ablation lesion formation using an optical coherence tomography forward-imaging catheter, Journal of Biomedical Optics 15 (3):030516-1 (3 pages).
Fookes et al., 2002, Rigid and non-rigid image registration and its association with mutual information:A review, Technical Report ISBN:1 86435 569 7, RCCVA, QUT.
Forstner & Moonen, 1999, A metric for covariance matrices, In Technical Report of the Dpt of Geodesy and Geoinformatics, Stuttgart University, 113-128.
Goel et al., 2006, Minimally Invasive Limited Ligation Endoluminal-assisted Revision (MILLER) for treatment of dialysis access-associated steal syndrome, Kidney Int 70(4):765-70.
Gotzinger et al., 2005, High speed spectral domain polarization sensitive optical coherence tomography of the human retina, Optics Express 13(25):10217-10229.
Gould et al., 1974, Physiologic basis for assessing critical coronary stenosis, American Journal of Cardiology, 33:87-94.
Griffiths et al., 1993, Human anti-self antibodies with high specificity from phage display libraries, The EMBO Journal, 12:725-734.
Griffiths et al., 1994, Isolation of high affinity human antibodies directly from large synthetic repertoires, The EMBO Journal, 13(14):3245-3260.
Grund et al., 2010, Analysis of biomarker data:logs, odds, ratios and ROC curves, Curr Opin HIV AIDS 5(6):473-479.
Harrison et al., 2011, Guidewire Stiffness: What's in a name?, J Endovasc Ther, 18(6):797-801.
Huber et al., 2005, Amplified, Frequency Swept Lasers for Frequency Domain Reflectometry and OCT Imaging: Design and Scaling Principles, Optics Express 13(9):3513-3528.
Huber et al., 2006, Fourier Domain Mode Locking (FDML): A New Laser Operating Regime and Applications for Optical Coherence Tomography, Optics Express 14(8):3225-3237.
International Search Report and Written Opinion mailed Mar. 11, 2014, for International Patent Application No. PCT/US13/75675, filed Dec. 17, 2013 (7 pages).
International Search Report and Written Opinion mailed Mar. 19, 2014, for International Patent Application No. PCT/US13/075353, filed Dec. 16, 2013 (8 pages).
International Search Report and Written Opinion mailed Nov. 2, 2012, for International Patent Application No. PCT/US12/53168, filed Aug. 30, 2013 (8 pages).
International Search Report and Written Opinion mailed on Apr. 14, 2014, for International Patent Application No. PCT/US2013/076148, filed Dec. 18, 2013 (8 pages).
International Search Report and Written Opinion mailed on Apr. 21, 2014, for International Patent Application No. PCT/US2013/076015, filed Dec. 18, 2013 (7 pages).
International Search Report and Written Opinion mailed on Apr. 23, 2014, for International Patent Application No. PCT/US2013/075328, filed Dec. 16, 2013 (8 pages).
International Search Report and Written Opinion mailed on Apr. 29, 2014, for International Patent Application No. PCT/US13/76093, filed Dec. 18, 2013 (6 pages).
International Search Report and Written Opinion mailed on Apr. 9, 2014, for International Patent Application No. PCT/US13/75089, filed Dec. 13, 2013 (7 pages).
International Search Report and Written Opinion mailed on Feb. 21, 2014, for International Patent Application No. PCT/US13/76053, filed Dec. 18, 2013 (9 pages).
International Search Report and Written Opinion mailed on Feb. 21, 2014, for International Patent Application No. PCT/US2013/076965, filed Dec. 20, 2013 (6 pages).
International Search Report and Written Opinion mailed on Feb. 27, 2014, for International Patent Application No. PCT/US13/75416, filed Dec. 16, 2013 (7 pages).
International Search Report and Written Opinion mailed on Feb. 28, 2014, for International Patent Application No. PCT/US13/75653, filed Dec. 17, 2013 (7 pages).
International Search Report and Written Opinion mailed on Feb. 28, 2014, for International Patent Application No. PCT/US13/75990, filed Dec. 18, 2013 (7 pages).
International Search Report and Written Opinion mailed on Jan. 16, 2009, for International Patent Application No. PCT/US08/78963 filed on Oct. 6, 2008 (7 pages).
International Search Report and Written Opinion mailed on Jul. 30, 2014, for International Patent Application No. PCT/US14/21659, filed Mar. 7, 2014 (15 pages).
International Search Report and Written Opinion mailed on Mar. 10, 2014, for International Patent Application No. PCT/US2013/076212, filed Dec. 18, 2013 (8 pages).
International Search Report and Written Opinion mailed on Mar. 11, 2014, for International Patent Application No. PCT/US13/76173, filed Dec. 16, 2013 (9 pages).
International Search Report and Written Opinion mailed on Mar. 11, 2014, for International Patent Application No. PCT/US13/76449, filed Dec. 19, 2013 (9 pages).
International Search Report and Written Opinion mailed on Mar. 18, 2014, for International Patent Application No. PCT/US2013/076502, filed Dec. 19, 2013 (7 pages).
International Search Report and Written Opinion mailed on Mar. 18, 2014, for International Patent Application No. PCT/US2013/076788, filed Dec. 20, 2013 (7 pages).
International Search Report and Written Opinion mailed on Mar. 19, 2014, for International Patent Application No. PCT/US13/75349, filed Dec. 16, 2013 (10 pages).
International Search Report and Written Opinion mailed on Mar. 19, 2014, for International Patent Application No. PCT/US2013/076587, filed Dec. 19, 2013 (10 pages).
International Search Report and Written Opinion mailed on Mar. 19, 2014, for International Patent Application No. PCT/US2013/076909, filed Dec. 20, 2013 (7 pages).
International Search Report and Written Opinion mailed on Mar. 7, 2014, for International Patent Application No. PCT/US2013/076304, filed Dec. 18, 2013 (9 pages).
International Search Report and Written Opinion mailed on Mar. 7, 2014, for International Patent Application No. PCT/US2013/076480, filed Dec. 19, 2013 (8 pages).
International Search Report and Written Opinion mailed on Mar. 7, 2014, for International Patent Application No. PCT/US2013/076512, filed Dec. 19, 2013 (8 pages).
International Search Report and Written Opinion mailed on Mar. 7, 2014, for International Patent Application No. PCT/US2013/076531, filed Dec. 19, 2013 (10 pages).
Jakobovits et al., 1993, Analysis of homozygous mutant chimeric mice:deletion of the immunoglobulin heavy-chain joining region blocks B-cell development and antibody production, PNAS USA 90:2551-255.
Jakobovits et al., 1993, Germ-line transmission and expression of a human-derived yeast artificial chromosome, Nature 362:255-258.
Jang et al., 2002, Visualization of Coronary Atherosclerotic Plaques in Patients Using Optical Coherence Tomography: Comparison With Intravascular Ultrasound, Journal of the American College of Cardiology 39:604-609.
Jiang et al., 1992, Image registration of multimodality 3-D medical images by chamfer matching, Proc. SPIE 1660, Biomedical Image Processing and Three-Dimensional Microscopy, 356-366.
Johnson et al., 1993, Human antibody engineering: Current Opinion in Structural Biology, 3:564-571.
Jones et al., 1986, Replacing the complementarity-determining regions in a human antibody with those from a mouse, Nature, 321:522-525.
Juviler et al., 2008, Anorectal sepsis and fistula-in-ano, Surgical Technology International, 17:139-149.
Karapatis et al., 1998, Direct rapid tooling:a review of current research, Rapid Prototyping Journal, 4(2):77-89.
Karp et al., 2009, The benefit of time-of-flight in PET imaging, J Nucl Med 49:462-470.
Kelly et al., 2005, Detection of Vascular Adhesion Molecule-1 Expression Using a Novel Multimodal Nanoparticle, Circulation Research 96:327-336.
Kemp et al., 2005, Depth Resolved Optic Axis Orientation in Multiple Layered Anisotropic Tissues Measured with Enhanced Polarization Sensitive Optical Coherence Tomography, Optics Express 13(12):4507-4518.
Kersey et al., 1991, Polarization insensitive fiber optic Michelson interferometer, Electron. Lett. 27:518-520.
Kheir et al., 2012, Oxygen Gas-Filled Microparticles Provide Intravenous Oxygen Delivery, Science Translational Medicine 4(140):140ra88 (10 pages).
Khuri-Yakub et al., 2011, Capacitive micromachined ultrasonic transducers for medical imaging and therapy, J Micromech Microeng. 21(5):054004-054014.
Kirkman, 1991, Technique for flow reduction in dialysis access fistulas, Surg Gyn Obstet, 172(3):231-3.
Kohler et al., 1975, Continuous cultures of fused cells secreting antibody of predefined specificity, Nature, 256:495-7.
Koo et al., 2011, Diagnosis of IschemiaCausing Coronary Stenoses by Noninvasive Fractional Flow Reserve Computed From Coronary Computed Tomographic Angiograms, J Am Coll Cardiol 58(19):1989-1997.
Kozbor et al., 1984, A human hybrid myeloma for production of human monoclonal antibodies, J. Immunol., 133:3001-3005.
Kruth et al., 2003, Lasers and materials in selective laser sintering, Assembly Automation, 23(4):357-371.
Kumagai et al., 1994, Ablation of polymer films by a femtosecond high-peak-power Ti:sapphire laser at 798 nm, Applied Physics Letters, 65(14):1850-1852.
Larin et al., 2002, Noninvasive Blood Glucose Monitoring with Optical Coherence Tomography: a pilot study in human subjects, Diabetes Care, 25(12):2263-7.
Larin et al., 2004, Measurement of Refractive Index Variation of Physiological Analytes using Differential Phase OCT, Proc of SPIE 5325:31-34.
Laufer, 1996, Introduction to Optics and Lasers in Engineering, Cambridge University Press, Cambridge UK:156-162.
Lefevre et al., 2001, Stenting of bifurcation lesions:a rational approach, J. Interv. Cardiol., 14(6):573-585.
Li et al., 2000, Optical Coherence Tomography: Advanced Technology for the Endoscopic Imaging of Barrett's Esophagus, Endoscopy, 32(12):921-930.
Little et al., 1991, The underlying coronary lesion in myocardial infarction:implications for coronary angiography, Clinical Cardiology, 14(11):868-874.
Loo, 2004, Nanoshell Enabled Photonics-Based Imaging and Therapy of Cancer, Technology in Cancer Research & Treatment 3(1):33-40.
Machine translation of JP 2000-097846.
Machine translation of JP 2000-321034.
Machine translation of JP 2000-329534.
Machine translation of JP 2004-004080.
Maintz et al., 1998, An Overview of Medical Image Registration Methods, Technical Report UU-CS, (22 pages).
Mamas et al., 2010, Resting Pd/Pa measured with intracoronary pressure wire strongly predicts fractional flow reserve, Journal of Invasive Cardiology 22(6):260-265.
Marks et al., 1991, By-passing Immunization Human Antibodies from V-gene Libraries Displayed on Phage, J. Mol. Biol. 222:581-597.
Marks et al., 1992, By-Passing Immunization:Building High Affinity Human Antibodies by Chain Shuffling, BioTechnol., 10:779-783.
Maruno et al., 1991, Fluorine containing optical adhesives for optical communications systems, J. Appl. Polymer. Sci. 42:2141-2148.
McCafferty et al., 1990, Phage antibodies: filamentous phage displaying antibody variable domains, Nature 348:552-554.
Mendieta et al., 1996, Complementary sequence correlations with applications to reflectometry studies, Instrumentation and Development 3(6):37-46.
Mickley, 2008, Steal Syndrome-strategies to preserve vascular access and extremity, Nephrol Dial Transplant 23:19-24.
Miller et al., 2010, The MILLER banding procedure is an effective method for treating dialysis-associated steal syndrome, Kidney International 77:359-366.
Milstein et al., 1983, Hybrid hybridomas and their use in immunohistochemistry, Nature 305:537-540.
Mindlin et al., 1936, A force at a point of a semi-infinite solid, Physics, 7:195-202.
Morrison et al., 1984, Chimeric human antibody molecules: mouse antigen-binding domains with human constant region domains, PNAS 81:6851-6855.
Munson et al., 1980, Ligand: a versatile computerized approach for characterization of ligand-binding systems, Analytical Biochemistry, 107:220-239.
Nezam, 2008, High Speed Polygon-Scanner-Based Wavelength-Swept Laser Source in the Telescope-Less Configurations with Application in Optical Coherence Tomography, Optics Letters 33(15):1741-1743.
Nissen, 2001, Coronary Angiography and Intravascular Ultrasound, American Journal of Cardiology, 87 (suppl):15A-20A.
Nitenberg et al., 1995, Coronary vascular reserve in humans: a critical review of methods of evaluation and of interpretation of the results, Eur Heart J. 16(Suppl 1):7-21.
Notice of Reason(s) for Refusal dated Apr. 30, 2013, for Japanese Patent Application No. 2011-508677 for Optical Imaging Catheter for Aberation Balancing to Volcano Corporation, which application is a Japanese national stage entry of PCT/US2009/043181 with international filing date May 7, 2009, of the same title, published on Nov. 12, 2009, as WO 2009/137704, and accompanying English translation of the Notice of Reason(s) for Refusal and machine translations of JP11-56786 and JP2004-290548 (56 pages).
Nygren, 1982, Conjugation of horseradish peroxidase to Fab fragments with different homobifunctional and heterobifunctional cross-linking reagents. A comparative study, J. Histochem. and Cytochem. 30:407-412.
Oesterle et al., 1986, Angioplasty at coronary bifurcations: single-guide, two-wire technique, Cathet Cardiovasc Diagn., 12:57-63.
Okuno et al., 2003, Recent Advances in Optical Switches Using Silica-based PLC Technology, NTT Technical Review 1(7):20-30.
Oldenburg et al., 1998, Nanoengineering of Optical Resonances, Chemical Physics Letters 288:243-247.
Oldenburg et al., 2003, Fast-Fourier-Domain Delay Line for In Vivo Optical Coherence Tomography with a Polygonal Scanner, Applied Optics, 42(22):4606-4611.
Othonos, 1997, Fiber Bragg gratings, Review of Scientific Instruments 68(12):4309-4341.
Owens et al., 2007, A Survey of General-Purpose Computation on Graphics Hardware, Computer Graphics Forum 26 (1):80-113.
Pain et al., 1981, Preparation of protein A-peroxidase mono conjugate using a heterobifunctional reagent, and its use in enzyme immunoassays, J Immunol Methods, 40:219-30.
Park et al., 2005, Real-time fiber-based multi-functional spectral-domain optical coherence tomography at 1.3 um., Optics Express 13(11):3931-3944.
Pasquesi et al., 2006, In vivo detection of exercise induced ultrastructural changes in genetically-altered murine skeletal muscle using polarization-sensitive optical coherence tomography, Optics Express 14(4)1 547-1556.
Pepe et al., 2004, Limitations of the odds ratio in gauging the performance of a diagnostic, prognostic, or screening marker, American Journal of Epidemiology 159(9):882-890.
Persson et al., 1985, Acoustic impedance matching of medical ultrasound transducers, Ultrasonics, 23(2):83-89.
Placht et al., 2012, Fast time-of-flight camera based surface registration for radiotherapy patient positioning, Medical Physics 39(1):4-17.
Rabbani et al., 1999, Review: Strategies to achieve coronary arterial plaque stabilization, Cardiovascular Research 41:402-417.
Radvany et al., 2008, Plaque Excision in Management of Lower Extremity Peripheral Arterial Disease with the SilverHawk Atherectomy Catheter, Seminars in Interventional Radiology, 25(1):11-19.
Reddy et al., 1996, An FFT-Based Technique for Translation, Rotation, and Scale-Invariant Image Registration, IEEE Transaction on Image Processing 5(8):1266-1271.
Riechmann et al., 1988, Reshaping human antibodies for therapy, Nature, 332:323-327.
Rivers et al., 1992, Correction of steal syndrome secondary to hemodialysis access fistulas: a simplified quantitative technique, Surgery, 112(3):593-7.
Robbin et al., 2002, Hemodialysis Arteriovenous Fistula Maturity: US Evaluation, Radiology 225:59-64.
Rollins et al., 1998, In vivo video rate optical coherence tomography, Optics Express 3:219-229.
Sarunic et al., 2005, Instantaneous Complex Conjugate Resolved Spectral Domain and Swept-Source OCT Using 3×3 Fiber Couplers, Optics Express 13(3):957-967.
Satiani et al., 2009, Predicted Shortage of Vascular Surgeons in the United States, J. Vascular Surgery 50:946-952.
Schneider et al., 2006, T-banding: A technique for flow reduction of a hyper-functioning arteriovenous fistula, J Vase Surg. 43(2):402-405.
Sen et al., 2012, Development and validation of a new adenosine-independent index of stenosis severity from coronary wave-intensity analysis, Journal of the American College of Cardiology 59(15):1392-1402.
Setta et al., 2005, Soft versus firm embryo transfer catheters for assisted reproduction: a systematic review and meta-analysis, Human Reproduction, 20(11):3114-3121.
Seward et al., 1996, Ultrasound Cardioscopy: Embarking on New Journey, Mayo Clinic Proceedings 71(7):629-635.
Shen et al., 2006, Eigengene-based linear discriminant model for tumor classification using gene expression microarray data, Bioinformatics 22(21):2635-2642.
Related Publications (1)
Number Date Country
20140100449 A1 Apr 2014 US
Provisional Applications (1)
Number Date Country
61710429 Oct 2012 US