Processes, arrangements and systems for providing a fiber layer thickness map based on optical coherence tomography images

Information

  • Patent Grant
  • 7782464
  • Patent Number
    7,782,464
  • Date Filed
    Friday, May 4, 2007
    17 years ago
  • Date Issued
    Tuesday, August 24, 2010
    14 years ago
Abstract
A system, arrangement, computer-accessible medium and process are provided for determining information associated with at least one portion of an anatomical structure. For example, an interference between at least one first radiation associated with a radiation directed to the anatomical structure and at least one second radiation associated with a radiation directed to a reference can be detected. Three-dimensional volumetric data can be generated for the at least one portion as a function of the interference. Further, the information can be determined which is at least one geometrical characteristic and/or at least one intensity characteristic of the portion based on the volumetric data.
Description
FIELD OF THE INVENTION

The present invention relates to processes, arrangements, computer-accessible medium and systems which can provide a fiber layer thickness map based on optical coherence tomography (“OCT”) images, and more particularly to such processes, systems, computer-accessible medium and arrangements that the boundaries and therefore the thickness of a layer in the sample can be automatically determined starting from OCT images.


BACKGROUND INFORMATION

Spectral-domain optical coherence tomography (“SD-OCT”) was recently established as a real-time technique for investigating the depth structure of biomedical tissue with the purpose of non-invasive optical diagnostics. A detailed description of SD-OCT techniques is described in Fercher et al. “Measurement of Intraocular Distances by Backscattering Spectral Interferometry”, Optics Communications, 117(1-2), 43 (1995) and Wojtkowski et al. “In vivo human retinal imaging by Fourier domain optical coherence tomography”, J. Biomed. Opt. 7(3), 457 (2002). Compared to the commercially available time-domain OCT systems, SD-OCT techniques provide for video-rate OCT scans, are relative fast, as shown in Nassif et al. “In vivo human retinal imaging by ultrahigh-speed spectral domain optical coherence tomography”, Opt. Lett. 29(5), 480 (2004), and provide a good sensitivity, as described in Leitgeb et al. “Performance of Fourier domain vs. time domain optical coherence tomography”, Opt. Express, 11(8), 889 (2003) and de Boer et al. “Improved signal-to-noise ratio in spectral-domain compared with time-domain optical coherence tomography”, Opt. Lett., 28(21), 2067 (2003). An exemplary arrangement which can be used for video-rate OCT scans has been described in International application number PCT/US03/02349 filed Jan. 24, 2003 and in Nassif et al. “In vivo human retinal imaging by ultrahigh-speed spectral domain optical coherence tomography”, Opt. Lett., 29(5), 480 (2004).


Similar technology, e.g., Optical Frequency Domain Imaging (“OFDI”), can use a rapidly tuned laser to measure the wavelength resolved interference as described in Chinn et al. “Optical coherence tomography using a frequency tunable optical source”, Opt. Lett. 22(5), 340 (1997), and Yun et al. “High-speed optical frequency-domain imaging”, Opt. Express 11(22), 2953 (2003) and International Application PCT/US04/029148 filed Sep. 8, 2004.


The depth profile in SD-OCT/OFDI can be obtained as the Fourier transform (“FFT”) of the spectral interference in a Michelson interferometer as described in Fercher et al. “Measurement of Intraocular Distances by Backscattering Spectral Interferometry”, Optics Communications, 117(1-2), 43 (1995) and Wojtkowski et al. “In vivo human retinal imaging by Fourier domain optical coherence tomography”, J. Biomed. Opt. 7(3), 457 (2002). The data processing steps to generate a good quality structural SD-OCT image have been described in Cense et al. “Ultrahigh-resolution high-speed retinal imaging using spectral-domain optical coherence tomography”, Opt. Express, 12(11), 2435 (2004), Yun et al. “High-speed optical frequency-domain imaging”, Opt. Express, 11(22), 2953 (2003), and Nassif et al. “In vivo high-resolution video-rate spectral-domain optical coherence tomography of the human retina and optic nerve”, Opt. Express, 12(3), 367 (2004). Various dispersion compensation techniques for OCT have been described in Marks et al. “Autofocus algorithm for dispersion correction in optical coherence tomography”, Appl. Opt., 42(16), 3038 (2003), Marks et al. “Digital algorithm for dispersion correction in optical coherence tomography for homogeneous and stratified media”, Appl. Opt., 42(2), 204 (2003), Wojtkowski et al. “Ultrahigh-resolution, high-speed, Fourier domain optical coherence tomography and methods for dispersion compensation”, Opt. Express, 12(11), 2404 (2004), and Fercher et al. “Numerical dispersion compensation for Partial Coherence Interferometry and Optical Coherence Tomography”, Opt. Express, 9(12), 610 (2001).


In ophthalmic applications, it has been suggested that OCT may be helpful for diagnosing glaucoma by measuring the thickness of the retinal nerve fiber layer (RNFL). In publications, the RNFL thickness has been evaluated with time-domain OCT commercial instruments for only a small number of circular scans, in general three, and not as a full map of the retina. A method to generate a large area thickness map of the RNFL is desirable. See e.g., Bourne et al. “Comparability of retinal nerve fiber layer thickness measurements of optical coherence tomography instruments” Invest. Opthalmol. Visual Sci., 46(4), 1280 (2005), Carpineto et al. “Reliability of nerve fiber layer thickness measurements using optical coherence tomography in normal and glaucomatous eyes” Opthalmology, 110(1), 190 (2003), Aydin et al. “Optical coherence tomography assessment of retinal nerve fiber layer thickness changes after glaucoma surgery”, Opthalmology, 110(8), 1506 (2003), and Guedes et al. “Optical coherence tomography measurement of macular and nerve fiber layer thickness in normal and glaucomatous human eyes”, Opthalmology, 110(1), 177 (2003).


Additional extensions of OCT techniques such as polarization-sensitive OCT (“PS-OCT”) can assist in identifying the properties of the RNFL including the layer's birefringence and boundaries as described in Cense et al. “In vivo birefringence and thickness measurements of the human retinal nerve fiber layer using polarization-sensitive optical coherence tomography”, J. Biomed. Opt., 9(1), 121 (2004), International Application PCT/US05/39374 filed Oct. 31, 2005, International Application PCT/US07/66017 filed Apr. 5, 2007 and International Application PCT/US06/15484 filed Apr. 24, 2006. It is believed that birefringence changes of the RNFL may preclude thickness changes and therefore, birefringence measurement can assist in early diagnosis of glaucoma.


Boundary detection has been studied since the early days of computer vision and image processing, and different approaches have been proposed. Segmentation procedures have also been applied to retinal imaging either for estimating the thickness of various retinal layers, as presented in Ishikawa et al. “Macular segmentation with optical coherence tomography”, Invest. Opthalmol. Visual Sci., 46(6), 2012 (2005) and Fernandez et al. “Automated detection of retinal layer structures on optical coherence tomography images”, Opt. Express, 13(25), 10200 (2005), or for evaluating the thickness of the retina, as presented in Koozekanani et al. “Retinal thickness measurements from optical coherence tomography using a Markov boundary model”, IEEE Trans. Medical Imag., 20(9), 900 (2001). Another segmentation technique based on a deformable spline (snake) algorithm has been described in details in Xu and Prince, “Snakes, shapes, and gradient vector flow” IEEE Trans. Image Process., 7(3), 359 (1998) and Kass et al. “Snakes—Active Contour Models”, Int. J. Comput. Vis., 1(4), 321 (1987). As the snake seeks to minimize its overall energy, its shape will converge on the image gradient contour. However, in general, the snake may not be allowed to travel extensively, and proper initialization may be needed. The snake parameters (elasticity, rigidity, viscosity, and external force weight) can be set to allow the snake to follow the boundary for a large number of retinal topographies. Deformable spline procedures are widely used in medical imaging.


A RNFL thickness map is a quantitative assessment and provides evaluation of large retinal areas as compared to a limited number of circular or radial scans measured with the current commercial instruments. The RNFL thickness maps can potentially be used for a thorough evaluation of the RNFL thickness in longitudinal studies of glaucoma progression. These procedures use large area RNFL thickness maps, which may allow for more accurate correlations of RNFL thinning with visual field defects as opposed to individual circular or radial scans that should be measured at precisely the same retinal location, which is very difficult and that give less information. Therefore, a methodology that allows a determination of RNFL thickness maps based on noise suppression and edge detection may be desirable. Also an intuitive representation of OCT data may be desirable for diagnostic purposes by correlating the quantitative RNFL thickness map with an ultra-high resolution OCT movie, therefore providing a comprehensive picture to clinicians.


Accordingly, there is a need to overcome the deficiencies as described herein above.


OBJECTS AND SUMMARY OF EXEMPLARY EMBODIMENTS

To address and/or overcome the above-described problems and/or deficiencies, exemplary embodiments of processes, systems, computer-accessible medium and arrangements that the boundaries and therefore the thickness of a layer in the sample can be automatically determined starting from OCT images. For example, exemplary processes, systems, computer-accessible medium and arrangements may be provided for determining the thickness of retinal layers including but not limited to the RNFL.


According to one exemplary embodiment of the present invention, it is possible to implement the procedures, systems and arrangements described in U.S. Pat. No. 6,980,299 and International Application No. PCT/US04/023585 filed Jul. 23, 2004


The exemplary embodiments of the procedures, systems, computer-accessible medium and arrangements according to the present invention can be used to identify the boundaries of the retinal layers, such as but not limited to the anterior and posterior RNFL boundaries. In further exemplary embodiments, the retinal layers can by defined and be differentiated based on various characteristics, including but not limited to the magnitude and/or standard deviation of intensity reflectance, polarization properties, texture, and/or Doppler properties.


In addition, the exemplary embodiments of the procedures, systems, computer-accessible medium and arrangements according to the present invention can be used to determine the retinal surface topography and the topography of the optic nerve head (“ONH”). High-resolution characterization of the ONH topography can be interesting for a quantitative assessment of glaucoma. Further, another exemplary embodiment of the procedures, systems and arrangements according to the present invention can be used to determine the boundaries of the retinal pigment epithelium (“RPE”). The shape of the RPE may be interesting in the analysis of retinal drusen. Yet further exemplary embodiments of the procedures, systems, computer-accessible medium and arrangements according to the present invention are capable of quantifying the retinal nerve fiber tissue by measuring the RNFL thickness and the thickness distribution across large areas of the retina.


Thus, according to the exemplary embodiments of the present invention, it is possible to:

    • a. determine layer boundaries, including but not limited to the anterior and posterior boundaries of the RNFL;
    • b. determine the properties of the RPE and the optic disc as the edge of the RPE;
    • c. determine the RNFL thickness and thickness distribution over large areas of the retina;
    • d. determine the geometry, the shape and volume of anatomical structures including but not limited to the ONH cup;
    • e. use model-based extraction of retinal properties starting from know structural characteristics.


Still another exemplary embodiment of the procedures, systems, computer-accessible medium and arrangements according to the present invention are capable of facilitating clinical interpretation of the OCT data. A display modality, as an exemplary embodiment, can combine the thickness map and a reflectivity map (e.g., a fundus-type image), together with the cross-sectional images of the retina (e.g., OCT movie).


In yet another exemplary embodiment of the procedures, systems, computer-accessible medium and arrangements according to the present invention, different images can be combined with their difference and/or ratio to illustrate features not evident from either of the two images. For example, the two images can be 3D volumes, cross-sectional OCT frames and or thickness maps. Such exemplary images may be obtained from measurements taken at different patient visits and the difference between them could indicate changes as a result of disease progression.


According to still further exemplary embodiment of the present invention, the measurement may be performed on the same eye using light from different spectral bands. The scattering/reflectivity/absorption properties of the ocular tissue can depend on the wavelength of light, and therefore, measurements performed within different wavelength bands could potentially reveal different structural and morphological information.


In yet another exemplary embodiment of the present invention, these exemplary images may be different due to, e.g., external stimuli or factors including but not limited to light, medication, or blood pressure, and therefore, the difference or ratio image could reveal functional properties of the ocular tissue.


Thus, according to one exemplary embodiment of the present invention, a system, arrangement, computer-accessible medium and process may be provided for determining information associated with at least one portion of an anatomical structure. For example, an interference between at least one first radiation associated with a radiation directed to the anatomical structure and at least one second radiation associated with a radiation directed to a reference can be detected. Three-dimensional volumetric data can be generated for the at least one portion as a function of the interference. Further, the information can be determined which is at least one geometrical characteristic and/or at least one intensity characteristic of the portion based on the volumetric data.


For example, the first radiation can be generated by a low coherence source, and the interference may be detected simultaneously for separate wavelengths that are different from one another. The first radiation can also be generated by an automatically wavelength-tuned light source. The anatomical structure can be an ocular structure. The geometrical characteristic can include at least one continuous boundary. The generation of the continuous boundary can be based on:

    • first data associated with a scattered intensity and/or a polarization state of the interference,
    • second data associated with a distribution which is a spatial distribution and/or a temporal distribution of intensity variations of the interference, and/or
    • third data associated with a motion of scattering objects within the anatomical structure.


Further, the boundary can define a topological structure, and the information may include a geometry, a curvature, a volume and/or a thickness of the topological structure. It is also possible to generate at least one visualization associated with the information. The visualization can be at least one image. In addition, it is possible to determine at least one change of the information as a function of at least one condition.


According to yet another exemplary embodiment of the present invention, it is possible to filter the three-dimensional volumetric data based on a priori knowledge associated with the anatomical structure, and the information may be determined as a function of the filtered volumetric data. The priori knowledge can be based on at least one characteristic of a system performing such functions. The priori knowledge can also be based on at least one characteristic of a known progression of at least one abnormality associated with the anatomical structure.


In still another exemplary embodiment of the present invention, the first radiation can comprise a first radiation signal provided at a first wavelength range and a second radiation signal provided at a second wavelength range which is different from the first range. The three-dimensional volumetric data may be generated as a function of the first and second radiation signals. The first radiation can further comprise a third radiation signal provided at a third wavelength range which is different from the first and second ranges. Thus, the three-dimensional volumetric data may be generated as a function of the first, second and third radiation signals, and the information can be a color volume of the portion of the anatomical structure.


These and other objects, features and advantages of the present invention will become apparent upon reading the following detailed description of embodiments of the invention, when taken in conjunction with the appended claims.





BRIEF DESCRIPTION OF THE DRAWINGS

Further objects, features and advantages of the invention will become apparent from the following detailed description taken in conjunction with the accompanying figures showing illustrative embodiments of the invention, in which:


FIGS. 1A(a) through 1A(l) are exemplary illustrations of images takes during an implementation of an exemplary embodiment of a process for locating, e.g., RNFL boundaries and thickness according to the present invention;



FIG. 1B is a top level flow diagram of the exemplary embodiment of the process according to the present invention;



FIG. 1C is a detailed flow diagram of a first step of the exemplary process shown in FIG. 1B;



FIG. 1D is a detailed flow diagram of a second step of the exemplary process shown in FIG. 1B;



FIGS. 2A and 2B are selected cross-sectional frames of moving images of the OCT scans indicating the anterior and posterior boundary of the RNFL obtained using the exemplary procedures, systems and arrangements according to the present invention;



FIGS. 3B and 3A are exemplary integrated reflectance images obtained using the exemplary procedures, systems and arrangements according to the present invention as compared to the resultant images obtained using other technologies on the same eye;



FIG. 4 is an exemplary RNFL thickness map obtained using the exemplary procedures, systems and arrangements according to the present invention;


FIGS. 5A(top) and 5B(top) are combined exemplary representations of an integrated reflectance map obtained using the exemplary procedures, systems and arrangements according to the present invention;


FIGS. 5A(bottom) and 5B(bottom) are combined exemplary representations of a RNFL thickness map using the exemplary procedures, systems and arrangements according to the present invention;



FIGS. 5C and 5D are combined exemplary representations of retinal cross-sectional images obtained using exemplary procedures, systems and arrangements according to the present invention;



FIGS. 6A and 6B are images of two exemplary RNFL thickness maps;



FIG. 6C is an exemplary image map with is indicative of the difference between the exemplary images of FIGS. 6A and 6B; and



FIG. 7 is an exemplary embodiment of the system/arrangement according to the exemplary embodiment of the present invention which is configured to perform exemplary processes and/or procedures according to the present invention.





Throughout the figures, the same reference numerals and characters, unless otherwise stated, are used to denote like features, elements, components or portions of the illustrated embodiments. Moreover, while the subject invention will now be described in detail with reference to the figures, it is done so in connection with the illustrative embodiments. It is intended that changes and modifications can be made to the described embodiments without departing from the true scope and spirit of the subject invention as defined by the appended claims.


DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS

The exemplary embodiments of procedures, systems and arrangements according to the present invention which is configured to process data for determining an RNFL thickness can involve, e.g., certain steps of edge detection. In one such exemplary embodiment, the exemplary process can be performed frame by frame in 2D by identifying the edges corresponding to the anterior and posterior boundaries of the RNFL 1000 as shown in FIG. 1A(b). According to another exemplary embodiment, the analysis can be performed in 3D thereby extending the exemplary embodiment described herein above from being a 2D analysis to a 3D analysis.


Particularly, FIGS. 1A(a) through 1A(l) are exemplary illustrations of images takes during an implementation of an exemplary embodiment of a process for locating, e.g., RNFL boundaries and thickness according to the present invention. FIG. 1B is a top level flow diagram of the exemplary embodiment of the process according to the present invention. FIG. 1C is a detailed flow diagram of a first step of the exemplary process shown in FIG. 1B, and FIG. 1D is a detailed flow diagram of a second step of the exemplary process shown in FIG. 1B.


The first step (100) of the exemplary process of FIG. 1B is provided to identify the anterior boundary (“AB”) 1010 of the RNFL 1000 shown in FIG. 1A(b). Certain preliminary steps can be performed to determine an initial guess result for AB 1010. The image can be blurred using a Gaussian kernel (step 100-10 of FIG. 1C) for smoothing out the edges present in the image while suppressing spurious artifacts due to imaging noise. The standard deviation of the kernel can be set statistically with respect to the depth size of the image. The edge image is calculated as the magnitude of the image gradient, and rescaled between 0 and 1 by subtracting its minimum value and normalizing to its maximum value. The edge image is then converted to a binary image by keeping all edges above a threshold value. (Step 100-20 of FIG. 1C). This threshold can be determined such that AB 1010 remains a continuous line in the edge image.


Certain areas in the OCT scan, e.g., around the ONH and the fovea, may contain weak edges that should still to be identified, and such procedure can indicate how low the threshold should be set. Certain false edges are, however, preserved this way due to the noise in the image. They could be eliminated or reduced by removing any object in the binary image that has an area smaller than a certain percentage of the total image size determined based on analyzing a large number of images. The exemplary purpose is to preserve preferably the continuous lines across the image, and therefore this value could be set based on the size of the image. The exemplary result is shown in the image of FIG. 1A(a).


The initial guess for AB 1010 can then be determined as the first unity pixel from the top in the binary edge image along each column (step 100-30 of FIG. 1C). To avoid holes in the identified AB 1010, the columns that contain no edge may be removed and then AB 1010 can be linearly interpolated over the removed A-lines. A median filter with a moving window may be applied to smooth out the AB 1010. (Step 100-40 of FIG. 1C). The size of the moving window can be set statistically with respect to the lateral size of the image.


This exemplary initial guess of AB 1010 may be used as initialization for a multiresolution deformable spline algorithm/procedure. The external force field for the snake algorithm/procedure may be obtained as the gradient of the edge image that was generated as described above for a Gaussian kernel (step 100-60 of FIG. 1C) with a standard deviation radius set statistically with respect to the size of the image. The purpose of the Gaussian blur, as described earlier, is to make sure the initial guess for AB 1010 is within the capture range of the true boundary at AB 1010. This exemplary procedure also determines the resolution in finding AB 1010. Further the multiresolution deformable spline algorithm/procedure can be applied to the result (step 100-70 of FIG. 1C). The exemplary procedure can then be repeated (step 100-80 of FIG. 1C) with a smaller value of the standard deviation of the kernel (see FIG. 1A(b)), therefore a better resolution, using as initialization the value of AB 1010 obtained from the previous run of the algorithm with a coarse resolution.


The anterior boundary of the RNFL 1000 can be used to create and display the 3D topography of the retinal surface and of the ONH. (Step 100-90 of FIG. 1C). The knowledge of the topography of the retinal surface can be important for determining the geometry, the shape and volume of retinal structures including but not limited to the ONH cup. The geometrical properties of the ONH cup may be used for quantitative characterization of glaucoma.


The second step in determining the RNFL thickness is to identify the posterior boundary (“PB”) 1140 of the nerve fiber layer (step 110 of FIG. 2B). As the transition between the RNFL 1000 and the ganglion cell layer may not be as sharp and uniform as between the vitreous humor and the retina, another exemplary procedure according to one exemplary embodiment of the present invention can be used for identifying the PB 1140 (FIG. 1A(l)). There may be a shadow cast by the blood vessels 1020 from the RNFL 1000 that may generate “holes” 1040 in the PB 1140. Further, there may be little or no RNFL 1000 in the ONH area (e.g., the nerve fibers may be provided along the optic nerve, perpendicular to the retina, and therefore parallel to the OCT beam). If the nerve fibers are perpendicular to the incident laser beam, they can be detected as a layer. However, if these nerve fibers are positioned parallel to the OCT beam (as may be in the optic nerve), it may be difficult to detect them, as there is no “horizontal” boundary (e.g., perpendicular to the laser beam) to reflect the light.


Certain exemplary preliminary steps can be taken before the actual estimation of the PB 1140. As an initial matter, everything above the AB 1010 can be removed (step 110-5 of FIG. 1D). The image may be realigned to AB 1010 based on the assumption that the anterior and posterior boundaries are relatively parallel and the procedure in determining PB 1140 may be more stable for primarily horizontal edges. A certain depth can be maintained along each A-line. This exemplary depth may be selected to include certain relevant part of the depth profiles. Most or all of the images shown in FIGS. 1A(c) to 1A(k) can have AB 1010 as the top edge, and may preferably represent the steps in determining the PB 1140.


Further, the image can be processed to generate the smoothed field f (FIG. 1A(c)) and the edge field s (FIG. 1A(d)) by using an algorithm for joint anisotropic smoothing and edge-preservation. (Step 110-10 of FIG. 1D). The edge field s is the rescaled (between 0 and 1) magnitude of the gradient of f, as described above. The exemplary procedure can be based on a conjugate gradient based iterative minimization of a cost functional described in Barrett et al. “Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods”, 2nd ed. 1994, Philadelphia, Pa.: SIAM, which can balance a data fidelity term against prior models for image smoothness and edge density as described in Chan et al. “Anisotropic edge-preserving smoothing in carotid B-mode ultrasound for improved segmentation and intima-media thickness (IMT) measurement” in IEEE Computers in Cardiology 2000, Cambridge, Mass., USA, and Chan et al. “A variational energy approach for estimating vascular structure and deformation from B-mode ultrasound imagery” in IEEE International Conference on Image Processing. 2000. Vancouver, BC, Canada. The cost functional is provided by:










E


(

f
,
s

)


=




(


β





f
-
g



1


+



α


(

1
-
s

)


2







f



1


+


ρ
2







s



2
2


+


1
ρ



s
2



)




A







(
1
)








where the notation ∥.∥k represents an lk norm (k=1, 2), and the integration is done over the entire image.


The first term can represent the data fidelity and controls the degree to which the smoothed field f resembles the original image g. The second term, representing the smoothness constraint, may penalize large gradients in f except where edges exist (s 1), generating the greatest anisotropic smoothing far away from the edges. The last two terms, representing the edge penalty, can control the edge width and prevent the minimization process from placing edges everywhere in the edge field s. The real positive scalars α, β, and ρ can adjust the relative weighting between these competing terms. The solution of Eq. (1) may be obtained by iterative minimization of the cost functional until convergence to a predefined tolerance level is achieved or an exemplary maximum number of iterations can be exceeded. The subsequent steps in the image processing may be based on f and s, and preferably not the original image.


In a further exemplary embodiment of the present invention, the original image g in Eq. (1) can be replaced by a model image based on a priori knowledge of the retinal structure. In a normal retina, the structure, the properties, and the number of the layers are known, and a model based extraction of the layers' boundaries can be used. Deviations from the model can indicate pathologies and could help in diagnosing retinal diseases.


An identification of blood vessels 1020 position in the RNFL 1000 can also be beneficial prior to the estimation of the PB 1140. This can be done based on the analysis of the RPE 1030. Scattering and absorption of the OCT beam due to the blood can significantly reduce the ability of the OCT beam to probe behind the blood vessels 1020, and the structural image (FIG. 1A(b)) appears to have a “shadow” underneath the blood vessels 1020. This can generate “holes” 1040 in the PB 1140 as well as in the RPE 1030. The RPE 1030 can be identified in the image, and the index of the A-lines corresponding to these “holes” 1040 can be estimated. These indices, that may be declared as invalid, can later be used in the analysis of the PB 1140.


The identification of the RPE 1030 may be based on both the magnitude of the smoothed field f and its gradient, e.g., the edge field s. An intensity mask (FIG. 1A(e)) may be generated from the smoothed field f based on the statistical properties of f, mean and standard deviation, using mean(f)−std(f) as threshold. (Step 110-15 of FIG. 1D). A binary image may also be created by thresholding the edge field s. (Step 110-17 of FIG. 1D). The threshold value can be selected as described above to preserve the weaker edges. Small patches may be removed from the binary edge image to clear out spurious points. The lowest (deepest) pixel with value one along each A-line can be selected from the two binary images as the posterior boundary of RPE 1030. (Step 110-20 of FIG. 1D). This boundary is then smoothed out with a moving window median filter. A band 1070 (FIG. 1A(f)) with a predefined thickness above the posterior RPE boundary 1030 is then selected from the smoothed field f and is averaged along the A-lines to generate a mean value of the RPE reflectance 1080 along each A-line (see dots in FIG. 1A(f)). (Step 110-25 of FIG. 1D). The mean RPE reflectance 1080 is then filtered out 1090 using a Savitzky-Golay FIR smoothing filter (black line in FIG. 1A(f)) that fits the data with a third order polynomial in a moving window. The index of an A-line is determined as invalid if the mean RPE value 1080 drops on that A-line by more than a certain predefined value as compared to the filtered mean RPE 1090. (Step 110-30 of FIG. 1D). As mentioned above, these invalid indices can correspond to the “holes” 1040 in the RPE 1030 generated by the shadow of the blood vessels 1020. Such indices may also correspond to the ONH area where there is no RPE 1030. The invalid indices are extended by a number of 10 A-lines on both sides of the shadows, as identified above, just to make sure that the vertical edges corresponding to the boundary of the shadows are removed from the image. Generally, the blood vessels 1020 have a larger diameter than the holes 1040 identified in the RPE 1030.


In yet another exemplary embodiment of the present invention, a map of the blood vessels network and the boundaries of the optic disc can be determined from the entire 3D OCT scan rather than frame by frame. After identifying the AB 1010, it is possible to remove from the 3D OCT scan a band below the AB 1010 that may include the RNFL 1000 and then integrate the depth profiles generating a RPE map. Integrating preferably only the layers below the RNFL 1000 can increase the contrast of the RPE map since the RPE 1030 is likely the strongest reflecting layer below the RNFL 1000. Appropriate thresholding and segmentation of the RPE map may provide the invalid A-lines (indices) described above, as well as the boundaries of the optic disc that can be used subsequently for identifying the PB 1140.


The RPE area 1060 can be removed from the smoothed field f, the intensity mask, and the binary edge image, and the rest of the processing is focused on the posterior boundary 1140 of the RNFL 1000. (Step 110-35 of FIG. 1D). The intensity mask 1050 may be applied to the edge field s in order to remove the edges that are outside the interest area. (Step 110-40 of FIG. 1D). To avoid broken lines, the binary edge image may be dilated and then eroded (FIG. 1A(g)). A morphological thinning operation is then performed on the binary edge image. (Step 110-45 of FIG. 1D). Vertical and negative edges are removed since predominantly horizontal edges with the correct slope from the RNFL 1000 to the ganglion cell layer (FIG. 1A(h)) are being looked for. (Step 110-50 of FIG. 1D).


An initial guess of the PB 1140 can be estimated as the first pixel of value one from the top along each A-line (see white dots 1100 in FIG. 1A(i)). (Step 110-55 of FIG. 1D). For the A-lines that have no pixel of value 1 in the binary edge image, the PB 1140 can be set to coincide with AB 1010. This exemplary situation may correspond to the areas around the ONH and fovea. The columns with invalid indices may be removed and the PB 1140 is linearly interpolated over the invalid regions. (Step 110-60 of FIG. 1D). To confirm that the PB 1140 is relatively smooth, the smoothing procedure described above for the RPE analysis employing a Savitzky-Golay FIR smoothing filter can be used, as well with a moving window (see black line 1110 in FIG. 1A(i)). This exemplary step can be used to remove points that are too far from the smoothed version of PB 1140.


At this point, the deformable spline algorithm/procedure can be applied. The intensity mask 1050 is applied to the original edge field s, and the edge field is then blurred with a Gaussian kernel. The external forces are calculated as gradient of the rescaled edge field, and they are set to zero for the A-lines with invalid indices. (Step 110-65 of FIG. 1D). The A-lines with invalid indices are removed from the result of the snake algorithm (1120 in FIG. 1A(j)) and PB 1140 is linearly interpolated over the invalid regions. (Steps 110-70 and 110-75 of FIG. 1D). A final median filter with a moving window can be applied to the calculated PB 1130 (FIG. 1A(k)). (Step 110-75 of FIG. 1D). Since the image has been realigned to the anterior boundary, the result obtained can preferably be the RNFL thickness. With respect to the original image, it is possible to add this result to the AB 1010 values to obtain the true or accurate PB 1140 (FIG. 1A(l)). (Step 110-80 of FIG. 1D).


The exemplary results of the exemplary procedure according to the present invention described above are shown in FIG. 1A(l) by displaying the identified AB 1010 and PB 1140 on the corresponding structural image. The exemplary parameters in the process, system and arrangement according to the exemplary embodiments of the present invention, such as threshold values, snake parameters for anterior and posterior boundaries, and RPE thickness, can be established based on a large number of OCT scans in different areas of the retina and for different subjects to account for a statistically significant variability in the boundaries' characteristics. The snake parameters can be set differently for AB 1010 and PB 1140 given the different properties of the two boundaries. The exemplary data sets described herein can be processed with fixed settings for AB 1010 and PB 1140.


According to another exemplary embodiment of the present invention, as shown in FIGS. 2A and 2B, can provide examples of RNFL boundaries AB 2000 and 2020 and PB 2010 and 2030 as determined by the exemplary embodiment of an automated procedure according to the present invention. For example, a video-rate OCT movie can be difficult to follow and interpret. It is not always clear what is occurring with the blood vessels moving laterally in the image and with the continuous and sometimes sudden change in their apparent diameter. If the orientation of the blood vessel changes from frame to frame, it may appear in the OCT-movie that the diameter of the blood vessel changes. There bay be certain situation that in the OCT-movie, the vessel can be provided parallel to the frame, and it may “appear” very thick, and then in the next frame it can be visualized as being very thin, since it changed its orientation with respect to the frame. Additional information may be useful for determining the correct interpretation of OCT-movies.


Previously, a visualization has been demonstrated where a fundus-type image was shown as a still picture simultaneously with the OCT movie—see Jiao et al. “Simultaneous acquisition of sectional and fundus ophthalmic images with spectral-domain optical coherence tomography”, Opt. Express, 13(2) 444 (2005). A line across this picture indicated the position of the OCT scan for orientation along the retinal surface. This picture was obtained by integrating the depth profiles (the reflectance) and was displayed on a logarithmic scale. This operation creates enough contrast to differentiate between the A-lines that correspond to blood vessels' location and those that do not intersect blood vessels. The scattering and absorption on blood reduce the total reflectance in the area corresponding to blood vessels and the vessels appear darker than the surroundings.


According to the exemplary embodiment of the present invention, a smoother image can be obtained by, e.g., integrating the logarithmic depth profile and displaying it in linear scale as shown in the exemplary images on the right sides of FIGS. 3A and 3B. The integrated reflectance image can be compared in FIGS. 3A and 3B to images of the same area in the same eye obtained with different procedures, e.g., fundus imaging and fluoresce in angiography, respectively.


As compared to a fundus image of the same eye (as shown in the left side of FIG. 3A), the integrated reflectance image illustrates the same structure of the blood vessels 3000 with a very good quality, approaching that of a fundus photo. FIG. 3B shows an angiogram (side left) and the integrated reflectance image (right side) for the same area around the fovea 3030. The images of FIGS. 3A and 3B demonstrate that the integrated reflectance map can be used as a reliable representation of the retinal vasculature 3000 and 3040, ONH 3010, and fovea 3020 and 3030. The integrated reflectance image can be obtained directly from the OCT data, and may not need to utilize additional imaging procedures. The ability to provide an accurate representation of the retina it is also beneficial to clinicians.


The experimental measurements described herein above were performed on the right eye of a healthy volunteer. According to another exemplary embodiment of the present invention as shown in FIG. 4, a map of the RNFL thickness can be obtained using the exemplary procedure described herein. For example, the RNFL thickness map can be smoothed out with a median filter. The darkest areas in FIG. 4 represent the absence of the RNFL, corresponding to the fovea 4000 and ONH 4010. The lighter area corresponds to thinner areas of the RNFL, while other darker areas represent thicker RNFL areas. Such RNFL map may be consistent with known normal retinal anatomy, as the RNFL thickness may have greater superior and inferior to the ONH. This exemplary image map can provide a quantitative assessment of the RNFL thickness and a comprehensive evaluation of large retinal areas as opposed to a limited number of circular or radial scans measured with the current commercial instruments.


In another exemplary embodiment of the present invention FIGS. 5A-5D illustrate the combined retinal information analyzed herein. For example, FIGS. 5A(top) and 5B(top) show the exemplary integrated reflectance as a map of the retina. FIGS. 5A(bottom) and 5B(bottom) show the exemplary RNFL thickness map. FIGS. 5C and 5D shown frames of a movie of a video-rate OCT scan indicating the boundaries of the RNFL, e.g., AB 5040 and 5090 (FIGS. 5C and 5D, respectively), and PB 5050 and 5100 (FIGS. 5C and 5D, respectively). Using this exemplary display modality, it is possible to better follow in the movie the structure of the blood vessels pattern 5020 and 5080 (FIGS. 5A and 5B, respectively). The position of each depth scan, indicated by the horizontal red line 5010 and 5060 (FIGS. 5A and 5B, respectively) across the maps, can be related to retinal morphology, i.e. the ONH 5030 (FIG. 5A) and the fovea 5000 and 5070 (FIGS. 5A and 5B, respectively).


The position of the blood vessels across the cross-sectional images, indicated by their “shadow”, can be correlated with the intersection of the horizontal line with the vasculature evident in the integrated reflectance map. The integrated reflectance map may also illustrate the orientation of the blood vessels with respect to the cross-sectional scans, thus allowing for a clear interpretation of the continuous and sometimes sudden change in the apparent diameter of the blood vessels. The association of the integrated reflectance map and of the RNFL thickness map with the OCT movie can provide the clinicians a more intuitive way of interpreting the OCT data for diagnosing retinal diseases such as glaucoma. FIGS. 5A and 5C show illustrations of a large area OCT scan including the ONH 5030 and the fovea 5000. FIGS. 5B and 5D show an exemplary OCT scan centered on the fovea 5070, which also illustrate the vasculature around the fovea and the retinal depth structure with a greater detail as compared to FIGS. 5A and 5C.


The dark band on the center left side of the RNFL thickness map shown in FIG. 5B(bottom) can correspond to the temporal raphe, a structure located temporal to the macula. Since it may be difficult to distinguish individual fibers, it is also difficult to see the structure and the direction of the retinal nerve fibers in the thickness map. However, the RNFL thickness is small in the raphe area since there are a limited number of fibers. Moving away from the raphe, additional fibers comprise the RNFL and the thickness likely increases. The exemplary RNFL thickness map obtained according to the exemplary embodiments of the present invention may be consistent with the RNFL distribution pattern described in Vrabec “Temporal Raphe of Human Retina”, Am. J. Opthalmol., 62(5), 926 (1966) that was confirmed later on by opthalmoscopy as described in Sakai et al. “Temporal Raphe of the Retinal Nerve-Fiber Layer Revealed by Medullated Fibers”, Jpn. J. Opthalmol., 31(4), 655 (1987).


According to yet a further exemplary embodiment of the present invention, two exemplary RNFL thickness maps 6000, 6010 corresponding to the same scan area on the same eye can be obtained as shown in FIGS. 6A and 6B, respectively, and the difference between them is shown in an image 6020 of FIG. 6C. These exemplary RNFL thickness maps 6000, 6010 can be obtained based on the exemplary measurements taken at different patient visits, and the difference image 6020 between them may indicate changes as a result of disease progression.


In still another exemplary embodiment of the present invention, the measurement can be performed on the same eye using light from different spectral bands. The scattering/reflectivity/absorption properties of the ocular tissue can depend on the wavelength of light, and therefore, measurements performed with different wavelength bands may potentially reveal different structural and morphological information. The exemplary images 6000, 6010 shown in FIGS. 6A and 6B can be the result of such measurements, and the exemplary image 6020 of FIG. 6C may be the difference or the ratio of the images 6000, 6010 of FIGS. 6A and 6B illustrating exemplary features not evident in either FIG. 6A or 6B. In yet another exemplary embodiment, the exemplary RNFL thickness maps 6000, 6010 shown in FIGS. 6A and 6B can be different due to external stimuli or factors such as light, medication, or blood pressure, and therefore, the exemplary difference or ratio image 6020 of FIG. 6C can reveal functional properties of the ocular tissue.


In still another exemplary embodiment of the present invention, the exemplary 3D OCT scans can be acquired for multiple (e.g., three) different wavelength bands and the structural information obtained in each wavelength band can be mapped to a color system including but not limited to the RGB system. A color volume representation of the OCT data may provide structural and morphological information not otherwise evident.


The above-described exemplary embodiments of the processes and procedures according to the present invention can be performed by processing arrangements described in a number of patent applications referenced herein. For example, one exemplary embodiment of the system/arrangement according to the exemplary embodiment of the present invention which is configured to perform such exemplary processes and/or procedures is shown in FIG. 7. In particular, an exemplary arrangement 7020 can be provided, which may include an interferometric arrangement 7030, a processing arrangement (e.g., a microprocessor, a computer, etc.) 7040, and a storage arrangement/computer accessible medium (e.g., hard disk, CD-ROM, RAM, ROM, etc.) 7050.


In the exemplary operation of the system/arrangement according to one embodiment of the present invention shown in FIG. 7, the interferometric arrangement 7020 can be configured to receive radiation from an anatomical structure 7000 and a radiation from a reference 7010, and detect an interference there between. The processing arrangement 7040 can access the storage arrangement/computer accessible medium 7050 to obtain instructions and/or software provided or stored thereon. As an alternative, the processing arrangement 7040 can have such instructions/software previously provided therein. These instructions/software can be executed by the processing arrangement 7040 so that it can be configured to receive the interference information associated with the determined interference, and generate three-dimensional volumetric data for at least one portion of the anatomical structure as a function of the interference. Then, such configured processing arrangement 7040 can determine and provide the information 7060 which is at least one geometrical characteristic and/or at least one intensity characteristic of such portion of the anatomical structure based on the volumetric data.


The foregoing merely illustrates the principles of the invention. Various modifications and alterations to the described embodiments will be apparent to those skilled in the art in view of the teachings herein. Indeed, the arrangements, systems and methods according to the exemplary embodiments of the present invention can be used with any OCT system, OFDI system, SD-OCT system or other imaging systems, and for example with those described in International Patent Application PCT/US2004/029148, filed Sep. 8, 2004, U.S. patent application Ser. No. 11/266,779, filed Nov. 2, 2005, and U.S. patent application Ser. No. 10/501,276, filed Jul. 9, 2004, the disclosures of which are incorporated by reference herein in their entireties. It will thus be appreciated that those skilled in the art will be able to devise numerous systems, arrangements and methods which, although not explicitly shown or described herein, embody the principles of the invention and are thus within the spirit and scope of the present invention. In addition, to the extent that the prior art knowledge has not been explicitly incorporated by reference herein above, it is explicitly being incorporated herein in its entirety. All publications referenced herein above are incorporated herein by reference in their entireties.

Claims
  • 1. A system for determining information associated with at least one portion of an anatomical structure, comprising: at least one arrangement configured to:a. detect an interference between at least one first radiation associated with a radiation directed to the anatomical structure and at least one second radiation associated with a radiation directed to a reference,b. generate three-dimensional volumetric data for the at least one portion as a function of the interference, andc. determine the information based on the volumetric data and at least one of (i) a spatial distribution or (ii) a temporal distribution of an intensity of the interference.
  • 2. The system according to claim 1, wherein the at least one first radiation is generated by a low coherence source.
  • 3. The system according to claim 2, wherein the interference is detected simultaneously for separate wavelengths that are different from one another.
  • 4. The system according to claim 1, wherein the at least one first radiation is generated by an automatically wavelength-tuned light source.
  • 5. The system according to claim 1, wherein the anatomical structure is an ocular structure.
  • 6. The system according to claim 1, wherein the information comprises at least one geometrical characteristic which includes at least one continuous boundary.
  • 7. The system according to claim 6, wherein the at least one arrangement is further configured to generate the at least one continuous boundary based on at least one of: first data associated with a distribution which is at least one of the spatial distribution or the temporal distribution of intensity variations of the interference, orsecond data associated with a motion of scattering objects within the anatomical structure.
  • 8. The system according to claim 6, wherein the at least one boundary is overlaid over an image of the at least one portion to generate a topological structure.
  • 9. The system according to claim 8, wherein the at least one arrangement is further configured to generate at least one visualization of the topological structure associated with the information.
  • 10. The system according to claim 9, wherein the at least one visualization is at least one image.
  • 11. The system according to claim 8, wherein the at least one arrangement is further configured to determine at least one change of the information as a function of at least one condition.
  • 12. The system according to claim 1, wherein the at least one arrangement is further configured to filter the three-dimensional volumetric data based on a priori knowledge associated with the anatomical structure, and wherein the information is determined as a function of the filtered volumetric data.
  • 13. The system according to claim 12, wherein the priori knowledge is based on at least one characteristic of the system anatomical structure.
  • 14. The system according to claim 12, wherein the priori knowledge is based on at least one characteristic of a known progression of at least one abnormality associated with the anatomical structure.
  • 15. The system according to claim 1, wherein the at least first radiation comprises a first radiation signal provided at a first wavelength range and a second radiation signal provided at a second wavelength range which is different from the first range, and wherein the three-dimensional volumetric data is generated as a function of the first and second radiation signals.
  • 16. The system according to claim 15, wherein the at least one first radiation further comprises a third radiation signal provided at a third wavelength range which is different from the first and second ranges, and wherein the three-dimensional volumetric data is generated as a function of the first, second and third radiation signals, and wherein the information is a color volume of the at least one portion of the anatomical structure.
  • 17. A process for determining information associated with at least one portion of an anatomical structure, comprising: detecting an interference between at least one first radiation associated with a radiation directed to the anatomical structure and at least one second radiation associated with a radiation directed to a reference;generating three-dimensional volumetric data for the at least one portion as a function of the interference; anddetermining the information based on the volumetric data and at least one of (i) a spatial distribution or (ii) a temporal distribution of an intensity of the interference.
  • 18. A computer accessible medium for determining information associated with at least one portion of an anatomical structure, and providing thereon a software program, which, when executed by a processing arrangement, is operable to perform the procedures comprising: detecting an interference between at least one first radiation associated with a radiation directed to the anatomical structure and at least one second radiation associated with a radiation directed to a reference;generating three-dimensional volumetric data for the at least one portion as a function of the interference; anddetermining the information based on the volumetric data and at least one of (i) a spatial distribution or (ii) a temporal distribution of an intensity of the interference.
  • 19. The computer accessible medium according to claim 18, wherein the processing arrangement, when the software is executed thereby, is configured to perform further procedures comprising: separating different wavelengths of the interference via a dispersive arrangement; andsimultaneously detecting the interference for the different wavelengths using a plurality of detectors.
  • 20. The system according to claim 1, wherein the at least one arrangement is further configured to: d. separate different wavelengths of the interference via a dispersive arrangement; ande. simultaneously detect the interference for the different wavelengths using a plurality of detectors.
CROSS-REFERENCE TO RELATED APPLICATION(S)

This application is based upon and claims the benefit of priority from U.S. Patent Application Ser. No. 60/800,088, filed May 12, 2006, the entire disclosure of which is incorporated herein by reference.

US Referenced Citations (294)
Number Name Date Kind
2339754 Brace Jan 1944 A
3090753 Matuszak et al. May 1963 A
3601480 Randall Aug 1971 A
3856000 Chikama Dec 1974 A
3872407 Hughes Mar 1975 A
3941121 Olinger Mar 1976 A
3973219 Tang et al. Aug 1976 A
3983507 Tang et al. Sep 1976 A
4030827 Delhaye et al. Jun 1977 A
4140364 Yamashita et al. Feb 1979 A
4141362 Wurster Feb 1979 A
4224929 Furihata Sep 1980 A
4295738 Meltz et al. Oct 1981 A
4300816 Snitzer et al. Nov 1981 A
4303300 Pressiat et al. Dec 1981 A
4428643 Kay Jan 1984 A
4479499 Alfano Oct 1984 A
4533247 Epworth Aug 1985 A
4585349 Gross et al. Apr 1986 A
4601036 Faxvog et al. Jul 1986 A
4607622 Fritch et al. Aug 1986 A
4631498 Cutler Dec 1986 A
4650327 Ogi Mar 1987 A
4744656 Moran et al. May 1988 A
4751706 Rohde et al. Jun 1988 A
4770492 Levin et al. Sep 1988 A
4827907 Tashiro et al. May 1989 A
4834111 Khanna et al. May 1989 A
4868834 Fox et al. Sep 1989 A
4890901 Cross, Jr. Jan 1990 A
4892406 Waters Jan 1990 A
4909631 Tan et al. Mar 1990 A
4925302 Cutler May 1990 A
4928005 Lefevre et al. May 1990 A
4965441 Picard Oct 1990 A
4965599 Roddy et al. Oct 1990 A
4993834 Carlhoff et al. Feb 1991 A
4998972 Chin et al. Mar 1991 A
5039193 Snow et al. Aug 1991 A
5040889 Keane Aug 1991 A
5045936 Lobb et al. Sep 1991 A
5046501 Crilly Sep 1991 A
5065331 Vachon et al. Nov 1991 A
5085496 Yoshida et al. Feb 1992 A
5120953 Harris Jun 1992 A
5121983 Lee Jun 1992 A
5127730 Brelje et al. Jul 1992 A
5197470 Helfer et al. Mar 1993 A
5202745 Sorin et al. Apr 1993 A
5212667 Tomlinson et al. May 1993 A
5214538 Lobb May 1993 A
5228001 Birge et al. Jul 1993 A
5241364 Kimura et al. Aug 1993 A
5248876 Kerstens et al. Sep 1993 A
5262644 Maguire Nov 1993 A
5291885 Taniji et al. Mar 1994 A
5293872 Alfano et al. Mar 1994 A
5293873 Fang Mar 1994 A
5304173 Kittrell et al. Apr 1994 A
5304810 Amos Apr 1994 A
5305759 Kaneko et al. Apr 1994 A
5317389 Hochberg et al. May 1994 A
5318024 Kittrell et al. Jun 1994 A
5321501 Swanson et al. Jun 1994 A
5353790 Jacques et al. Oct 1994 A
5383467 Auer et al. Jan 1995 A
5411016 Kume et al. May 1995 A
5419323 Kittrell et al. May 1995 A
5439000 Gunderson et al. Aug 1995 A
5441053 Lodder et al. Aug 1995 A
5450203 Penkethman Sep 1995 A
5454807 Lennox et al. Oct 1995 A
5459325 Hueton et al. Oct 1995 A
5459570 Swanson et al. Oct 1995 A
5465147 Swanson Nov 1995 A
5486701 Norton et al. Jan 1996 A
5491524 Hellmuth et al. Feb 1996 A
5491552 Knuttel Feb 1996 A
5526338 Hasman et al. Jun 1996 A
5555087 Miyagawa et al. Sep 1996 A
5562100 Kittrell et al. Oct 1996 A
5565986 Knüttel Oct 1996 A
5583342 Ichie Dec 1996 A
5590660 MacAulay et al. Jan 1997 A
5600486 Gal et al. Feb 1997 A
5601087 Gunderson et al. Feb 1997 A
5621830 Lucey et al. Apr 1997 A
5623336 Raab et al. Apr 1997 A
5635830 Itoh Jun 1997 A
5649924 Everett et al. Jul 1997 A
5697373 Richards-Kortum et al. Dec 1997 A
5698397 Zarling et al. Dec 1997 A
5710630 Essenpreis et al. Jan 1998 A
5716324 Toida Feb 1998 A
5719399 Alfano et al. Feb 1998 A
5730731 Mollenauer et al. Mar 1998 A
5735276 Lemelson Apr 1998 A
5740808 Panescu et al. Apr 1998 A
5748318 Maris et al. May 1998 A
5748598 Swanson et al. May 1998 A
5784352 Swanson et al. Jul 1998 A
5785651 Kuhn et al. Jul 1998 A
5795295 Hellmuth et al. Aug 1998 A
5801826 Williams Sep 1998 A
5803082 Stapleton et al. Sep 1998 A
5807261 Benaron et al. Sep 1998 A
5817144 Gregory Oct 1998 A
5840023 Oraevsky et al. Nov 1998 A
5840075 Mueller et al. Nov 1998 A
5842995 Mahadevan-Jansen et al. Dec 1998 A
5843000 Nishioka et al. Dec 1998 A
5843052 Benja-Athon Dec 1998 A
5847827 Fercher Dec 1998 A
5862273 Pelletier Jan 1999 A
5865754 Sevick-Muraca et al. Feb 1999 A
5867268 Gelikonov et al. Feb 1999 A
5871449 Brown Feb 1999 A
5872879 Hamm Feb 1999 A
5877856 Fercher Mar 1999 A
5887009 Mandella et al. Mar 1999 A
5892583 Li Apr 1999 A
5910839 Erskine et al. Jun 1999 A
5912764 Togino Jun 1999 A
5920373 Bille Jul 1999 A
5920390 Farahi et al. Jul 1999 A
5921926 Rolland et al. Jul 1999 A
5926592 Harris et al. Jul 1999 A
5949929 Hamm Sep 1999 A
5951482 Winston et al. Sep 1999 A
5955737 Hallidy et al. Sep 1999 A
5956355 Swanson et al. Sep 1999 A
5968064 Selmon et al. Oct 1999 A
5975697 Podoleanu et al. Nov 1999 A
5983125 Alfano et al. Nov 1999 A
5987346 Benaron et al. Nov 1999 A
5991697 Nelson et al. Nov 1999 A
5994690 Kulkarni et al. Nov 1999 A
6002480 Izatt et al. Dec 1999 A
6004314 Wei et al. Dec 1999 A
6006128 Izatt et al. Dec 1999 A
6010449 Selmon et al. Jan 2000 A
6014214 Li Jan 2000 A
6020963 DiMarzio et al. Feb 2000 A
6033721 Nassuphis Mar 2000 A
6044288 Wake et al. Mar 2000 A
6045511 Ott et al. Apr 2000 A
6048742 Weyburne et al. Apr 2000 A
6053613 Wei et al. Apr 2000 A
6069698 Ozawa et al. May 2000 A
6091496 Hill Jul 2000 A
6091984 Perelman et al. Jul 2000 A
6111645 Tearney et al. Aug 2000 A
6117128 Gregory Sep 2000 A
6120516 Selmon et al. Sep 2000 A
6134003 Tearney et al. Oct 2000 A
6134010 Zavislan Oct 2000 A
6134033 Bergano et al. Oct 2000 A
6141577 Rolland et al. Oct 2000 A
6151522 Alfano et al. Nov 2000 A
6159445 Klaveness et al. Dec 2000 A
6160826 Swanson et al. Dec 2000 A
6161031 Hochman et al. Dec 2000 A
6166373 Mao Dec 2000 A
6174291 McMahon et al. Jan 2001 B1
6175669 Colston et al. Jan 2001 B1
6185271 Kinsinger Feb 2001 B1
6191862 Swanson et al. Feb 2001 B1
6193676 Winston et al. Feb 2001 B1
6198956 Dunne Mar 2001 B1
6201989 Whitehead et al. Mar 2001 B1
6208415 De Boer et al. Mar 2001 B1
6208887 Clarke Mar 2001 B1
6245026 Campbell et al. Jun 2001 B1
6249349 Lauer Jun 2001 B1
6263234 Engelhardt et al. Jul 2001 B1
6264610 Zhu Jul 2001 B1
6272376 Marcu et al. Aug 2001 B1
6274871 Dukor et al. Aug 2001 B1
6282011 Tearney et al. Aug 2001 B1
6297018 French et al. Oct 2001 B1
6308092 Hoyns Oct 2001 B1
6324419 Guzelsu et al. Nov 2001 B1
6341036 Tearney et al. Jan 2002 B1
6353693 Kano et al. Mar 2002 B1
6359692 Groot Mar 2002 B1
6374128 Toida et al. Apr 2002 B1
6377349 Fercher Apr 2002 B1
6384915 Everett et al. May 2002 B1
6393312 Hoyns May 2002 B1
6394964 Sievert, Jr. et al. May 2002 B1
6396941 Bacus et al. May 2002 B1
6421164 Tearney et al. Jul 2002 B2
6445485 Frigo et al. Sep 2002 B1
6445944 Ostrovsky Sep 2002 B1
6459487 Chen et al. Oct 2002 B1
6463313 Winston et al. Oct 2002 B1
6469846 Ebizuka et al. Oct 2002 B2
6475159 Casscells et al. Nov 2002 B1
6475210 Phelps et al. Nov 2002 B1
6477403 Eguchi et al. Nov 2002 B1
6485413 Boppart et al. Nov 2002 B1
6485482 Belef Nov 2002 B1
6501551 Tearney et al. Dec 2002 B1
6501878 Hughes et al. Dec 2002 B2
6538817 Farmer et al. Mar 2003 B1
6549801 Chen et al. Apr 2003 B1
6552796 Magnin et al. Apr 2003 B2
6556305 Aziz et al. Apr 2003 B1
6556853 Cabib et al. Apr 2003 B1
6558324 Von Behren et al. May 2003 B1
6564087 Pitris et al. May 2003 B1
6564089 Izatt et al. May 2003 B2
6567585 Harris May 2003 B2
6615071 Casscells, III et al. Sep 2003 B1
6622732 Constantz Sep 2003 B2
6680780 Fee Jan 2004 B1
6685885 Varma et al. Feb 2004 B2
6687007 Meigs Feb 2004 B1
6687010 Horii et al. Feb 2004 B1
6687036 Riza Feb 2004 B2
6701181 Tang et al. Mar 2004 B2
6738144 Dogariu et al. May 2004 B1
6741355 Drabarek May 2004 B2
6790175 Furusawa et al. Sep 2004 B1
6806963 Wälti et al. Oct 2004 B1
6816743 Moreno et al. Nov 2004 B2
6839496 Mills et al. Jan 2005 B1
6903820 Wang Jun 2005 B2
6949072 Furnish et al. Sep 2005 B2
6980299 de Boer Dec 2005 B1
7006231 Ostrovsky et al. Feb 2006 B2
7019838 Izatt et al. Mar 2006 B2
7061622 Rollins et al. Jun 2006 B2
7190464 Alphonse Mar 2007 B2
7231243 Tearney et al. Jul 2007 B2
7242480 Alphonse Jul 2007 B2
7267494 Deng et al. Sep 2007 B2
7336366 Choma et al. Feb 2008 B2
7355716 De Boer et al. Apr 2008 B2
7359062 Chen et al. Apr 2008 B2
7366376 Shishkov et al. Apr 2008 B2
7391520 Zhou et al. Jun 2008 B2
20010047137 Moreno et al. Nov 2001 A1
20020016533 Marchitto et al. Feb 2002 A1
20020052547 Toida May 2002 A1
20020064341 Fauver et al. May 2002 A1
20020076152 Hughes et al. Jun 2002 A1
20020085209 Mittleman et al. Jul 2002 A1
20020093662 Chen et al. Jul 2002 A1
20020122246 Tearney et al. Sep 2002 A1
20020140942 Fee et al. Oct 2002 A1
20020158211 Gillispie Oct 2002 A1
20020161357 Anderson et al. Oct 2002 A1
20020163622 Magnin et al. Nov 2002 A1
20020172485 Keaton et al. Nov 2002 A1
20020183623 Tang et al. Dec 2002 A1
20020188204 McNamara et al. Dec 2002 A1
20020196446 Roth et al. Dec 2002 A1
20020198457 Tearney et al. Dec 2002 A1
20030023153 Izatt et al. Jan 2003 A1
20030026735 Nolte et al. Feb 2003 A1
20030082105 Fischman et al. May 2003 A1
20030108911 Klimant et al. Jun 2003 A1
20030135101 Webler Jul 2003 A1
20030164952 Deichmann et al. Sep 2003 A1
20030171691 Casscells, III et al. Sep 2003 A1
20030174339 Feldchtein et al. Sep 2003 A1
20030199769 Podoleanu et al. Oct 2003 A1
20030216719 Debenedictis et al. Nov 2003 A1
20030220749 Chen et al. Nov 2003 A1
20030236443 Cespedes et al. Dec 2003 A1
20040002650 Mandrusov et al. Jan 2004 A1
20040086245 Farroni et al. May 2004 A1
20040100631 Bashkansky et al. May 2004 A1
20040100681 Bjarklev et al. May 2004 A1
20040126048 Dave et al. Jul 2004 A1
20040133191 Momiuchi et al. Jul 2004 A1
20040150829 Koch et al. Aug 2004 A1
20040152989 Puttappa et al. Aug 2004 A1
20040166593 Nolte et al. Aug 2004 A1
20040212808 Okawa et al. Oct 2004 A1
20040239938 Izatt Dec 2004 A1
20050018133 Huang et al. Jan 2005 A1
20050018201 De Boer Jan 2005 A1
20050046837 Izumi et al. Mar 2005 A1
20050075547 Wang Apr 2005 A1
20050083534 Riza et al. Apr 2005 A1
20050165303 Kleen et al. Jul 2005 A1
20050171438 Chen et al. Aug 2005 A1
20060103850 Alphonse et al. May 2006 A1
20060155193 Leonardi et al. Jul 2006 A1
20060244973 Yun et al. Nov 2006 A1
20070019208 Toida et al. Jan 2007 A1
20070291277 Everett et al. Dec 2007 A1
Foreign Referenced Citations (71)
Number Date Country
4105221 Sep 1991 DE
4309056 Sep 1994 DE
19542955 May 1997 DE
10351319 Jun 2005 DE
0110201 Jun 1984 EP
0251062 Jan 1988 EP
0617286 Feb 1994 EP
0590268 Apr 1994 EP
0728440 Aug 1996 EP
0933096 Aug 1999 EP
1324051 Jul 2003 EP
1426799 Jun 2004 EP
2738343 Aug 1995 FR
1257778 Dec 1971 GB
2030313 Apr 1980 GB
2209221 May 1989 GB
2298054 Aug 1996 GB
6073405 Apr 1985 JP
4135550 May 1992 JP
4135551 May 1992 JP
5509417 Nov 1993 JP
7900841 Oct 1979 WO
9201966 Feb 1992 WO
9216865 Oct 1992 WO
9219930 Nov 1992 WO
9303672 Mar 1993 WO
9533971 Dec 1995 WO
9628212 Sep 1996 WO
9732182 Sep 1997 WO
9800057 Jan 1998 WO
9801074 Jan 1998 WO
9814132 Apr 1998 WO
9835203 Aug 1998 WO
9838907 Sep 1998 WO
9846123 Oct 1998 WO
9848838 Nov 1998 WO
9848846 Nov 1998 WO
9905487 Feb 1999 WO
9944089 Sep 1999 WO
9957507 Nov 1999 WO
0058766 Oct 2000 WO
0108579 Feb 2001 WO
0127679 Apr 2001 WO
0138820 May 2001 WO
0142735 Jun 2001 WO
0236015 May 2002 WO
0238040 May 2002 WO
02054027 Jul 2002 WO
03020119 Mar 2003 WO
03046636 Jun 2003 WO
03052478 Jun 2003 WO
03062802 Jul 2003 WO
03105678 Dec 2003 WO
2004034869 Apr 2004 WO
2004057266 Jul 2004 WO
2004066824 Aug 2004 WO
2004088361 Oct 2004 WO
2004015598 Dec 2004 WO
2005000115 Jan 2005 WO
2005047813 May 2005 WO
2005054780 Jun 2005 WO
2005082225 Sep 2005 WO
2006004743 Jan 2006 WO
2006014392 Feb 2006 WO
2006039091 Apr 2006 WO
2006059109 Jun 2006 WO
2006124860 Nov 2006 WO
2006130797 Dec 2006 WO
2007028531 Mar 2007 WO
2007038787 Apr 2007 WO
2007083138 Jul 2007 WO
Related Publications (1)
Number Date Country
20070263227 A1 Nov 2007 US
Provisional Applications (1)
Number Date Country
60800088 May 2006 US