Process, system and software arrangement for a chromatic dispersion compensation using reflective layers in optical coherence tomography (OCT) imaging

Information

  • Patent Grant
  • 8018598
  • Patent Number
    8,018,598
  • Date Filed
    Friday, July 23, 2004
    20 years ago
  • Date Issued
    Tuesday, September 13, 2011
    13 years ago
Abstract
A system, process and software arrangement are provided to compensate for a dispersion in at least one portion of an image. In particular, information associated with the portion of the image is obtained. The portion of the image can be associated with an interference signal that includes a first electromagnetic radiation received from a sample and a second electromagnetic radiation received from a reference. The dispersion in the at least one portion of the image can be compensated by controlling a phase of at least one spectral component of the interference signal.
Description
FIELD OF THE INVENTION

The present invention relates generally to chromatic dispersion compensation in optical coherence tomography (“OCT”) imaging, and more particularly to processes, systems and software arrangements which can compensate for dispersions in OCT images.


BACKGROUND OF THE INVENTION

The spectral-domain variant of optical coherence tomography (“OCT”), called spectral-domain optical coherence tomography (SD-OCT), is a technique is a technology that is suitable for ultrahigh-resolution ophthalmic imaging. This technique has been described in Cense, B. et al., “Ultrahigh-resolution high-speed retinal imaging using spectral-domain optical coherence tomography”, Optics Express, 2004 and in International Patent Publication No. WO 03/062802. In addition, U.S. patent application Ser. No. 10/272,171 filed on Oct. 16, 2002 also relates to this subject matter. The axial resolution of an OCT system may be defined in terms of the coherence length (Lcoh), which can be determined by the center wavelength and bandwidth of the source and the index of refraction of the medium, as described in greater detail in Swanson, E. A. et al., “High-Speed Optical Coherence Domain Reflectometry”, Optics Letters, 1992, 17(2), pp. 151-153. The axial resolution of the OCT system can be improved by using an ultra broadband source, as provided in further detail in Drexler, W. et al., “Enhanced Visualization of Macular Pathology with the Use of Ultrahigh-Resolution Optical Coherence Tomography”, Archives of Ophthalmology, 2003, 121(5), pp. 695-706.


One potential difficulty that may arises from using ultra-broadband sources in a fiber-based OCT setup in, e.g., ophthalmic imaging is the presence of a chromatic dispersion in optically-dense materials like glass, tissue and water. Chromatic dispersion can lead to smearing of the coherence function and/or point spread function in the axial direction, which can significantly affect the image quality. Considerable amounts of dispersion can be tolerated if the dispersion in the two arms of the interferometer is balanced, thus creating a coherence function that would likely to be free from dispersion artifacts. However, when sample and reference arms contain different lengths of optical fiber or other dispersive media, a dispersion mismatch can occur. For example, in the sample arm of an OCT system, the analysis of an eye as a sample with unknown axial length may introduce an unknown amount of chromatic dispersion. Thus, the coherence function may be broadened by an unbalanced dispersion, and the peak intensity of the coherence function can decrease as well. A second order or a group-velocity dispersion can be compensated for using hardware by, e.g., changing the lens to grating distance in a rapid scanning optical delay line. The above has been described in detail in Tearney, G. J. et al., “High-Speed Phase- and Group-Delay Scanning with a Grating-Based Phase Control Delay Line”, Optics Letters, 1997, 22(23), pp. 1811-1813. However, this technique generally does not compensate for higher orders of dispersion. Alternatively, it is possible to balance a dispersion in the OCT system by inserting variable-thickness optical materials with different dispersion properties (such as BK7 and fused silica prisms) in the path of the reference arm or the sample arm. The number of materials with different optical properties that are inserted in the path of the reference arm or the sample arm may determine the number of orders of dispersion one can compensate. The axial length of an eye can vary from one person to another, thus changing the amount of dispersion between patients. Therefore a flexible technique for a dispersion compensation is desirable.


It is possible that, instead of using hardware for such compensation, to use software, and thereby a more flexible compensation easy to adapt to different eyes. Another publication describes a technique to provide an induced dispersion in the delay line of a time domain OCT system that equipped with an optical amplifier based source (e.g., AFC technologies, λ0=1310 nm, Δλ=75 nm) and compensated for dispersion artifacts in structural intensity images obtained of an onion. See de Boer, J. F. et al., “Stable Carrier Generation and Phase-Resolved Digital Data Processing in Optical Coherence Tomography”, Applied Optics, 2001, 40(31), pp. 5787-5790. Another publication describes a dispersion compensation which is induced by a glass sample. See Fercher, A. F. et al., “Dispersion Compensation For Optical Coherence Tomography Depth-Scan Signals By A Numerical Technique”, Optics Communications, 2002, 204(1-6), pp. 67-74. Their broadband spectrum is generated using a high-pressure mercury lamp. Other dispersion compensation techniques are described in Marks, D. L. et al., “Autofocus Algorithm for Dispersion Correction in Optical Coherence Tomography”, Applied Optics, 2003. 42(16), pp. 3038-3046, Marks, D. L. et al., “Digital Algorithm for Dispersion Correction in Optical Coherence Tomography for Homogeneous and Stratified Media”, Applied Optics, 2003, 42(2), pp. 204-217, and U.S. Pat. No. 5,994,690 which describe an algorithm that used an autocorrelation function to correct image data. However, the above-described problems have not been addressed adequately. Accordingly, there is a need to overcome such deficiencies.


SUMMARY OF THE INVENTION

In contrast to the conventional techniques, the exemplary embodiment of a system, process and software arrangement according to the present invention is capable of using a dispersion broadened reflection of a layer or structure in the biological sample (e.g., retina, skin, coronary artery) to derive parameters to compensate for the chromatic dispersion. One of the advantages of the exemplary system, process and software arrangement according to the present invention is the ease of its implementation, the flexibility thereof, and its adaptation to individual patients or sample locations without the need to make hardware changes so as to compensate for the chromatic dispersion.


According to exemplary embodiments of the present invention, a process, system and software arrangement is provided which can compensate for a dispersion using a numerical technique (e.g., without the need to modify hardware), and can be configured to remove artifacts from OCT images.


In general, a dispersion mismatch between the sample arm and the reference arm of an interferometer may introduce a phase shift etθ(k) in the cross-spectral density I(k) as a function of wave vector k (k=2π/λ). In a spectral-domain OCT or optical frequency domain interferometry (“OFDI”) setup (as described in Wojtkowski et al., “In Vivo Human Retinal Imaging by Fourier Domain Optical Coherence Tomography”, Journal of Biomedical Optics, 2002, 7(3), pp. 457-463, Nassif, N. et al., “In Vivo Human Retinal Imaging by Ultrahigh-Speed Spectral Domain Optical Coherence Tomography”, Optics Letters, 2004, 29(5), pp. 480-482, Yun, S. H. et al., “High-Speed Optical Frequency-Domain Imaging”, Optics Express, 2003, 11(22), pp. 2953-2963, International Publication No. WO 03/062802 and U.S. Patent Application Ser. No. 60/514,769 filed on Oct. 27, 2004, the spectrometer data can be acquired as a function of a wavelength. Such data may be transformed to k-space. The relation between the phase θ(k) and the multiple orders of dispersion can be described by a Taylor series expansion:



















θ


(
k
)


=




θ


(

k
0

)


+




θ


(
k
)





k







k
0




(


k
0

-
k

)


+


1
2

·




2



θ


(
k
)






k
2








k
0














(


k
0

-
k

)

2

+

+


1

n
!


·




n



θ


(
k
)






k
n








k
0





(


k
0

-
k

)

n








(
1
)








with λ0 being the center wavelength, and k0 being equal to 2π/λ0. The first two terms generally describe a constant offset and group velocity, respectively, and are likely not related to dispersive broadening. The third term represents a second order or a group-velocity dispersion. A dispersion mismatch in the sample arm and the reference arm can to a large extend be attributed to this term. However, higher order dispersion terms may contribute to the dispersion mismatch as well, for example when an ultra-broadband source is used.


The dispersion can be removed by multiplying the dispersed cross-spectral density function I(k) with a phase term e−tθ(k). In order to determine the phase term e−tθ(k) to remove the chromatic dispersion and the resulting broadening of the coherence function, data may be obtained with the interferometer using an object in the sample arm with a reflection. This object can be a mirror or a biological sample with a distinct reflection. The spectrum, I(k), acquired with the spectral domain OCT system is Fourier transformed to z-space, resulting in a depth profile of the reflectivity of the sample. A single reflective peak is determined in the depth profile, and the remaining points in the depth profile are set to zero. An inverse transform can be performed to obtain cross spectral density for this single reflective peak. The phase term θ(k) can be approximately equal to the arctangent of the imaginary component divided by the real component.


In order to reduce noise on the phase function and avoid distortion in the image by introducing a group velocity and/or offset in the phase, this function can be fit to a polynomial expression yielding a set of N coefficients α1N. Individual spectra may be multiplied with a phase e−θ(k) as determined from the polynomial coefficients, where the first two coefficients of the polynomial fit that correspond to a phase offset and a group velocity are omitted. The chromatic dispersion corrected spectra may then be Fourier transformed to z-space into A-lines, thus resulting in A-lines or depth profiles, where the dispersion has been removed substantially.


In one exemplary embodiment of the present invention, a system, method and software arrangement can be provided to compensate for a dispersion in at least one portion of an image. For example, information associated with the portion of the image is obtained. The portion of the image can be associated with an interference signal that includes a first electromagnetic radiation received from a sample and a second electromagnetic radiation received from a reference. The dispersion in the at least one portion of the image can be compensated by controlling a phase of at least one spectral component of the interference signal. The dispersion may be an indication of a difference between the first and second electromagnetic radiations. In addition, the dispersion may be compensated by reducing and/or removing the dispersion in the portion of the image. Further, data associated with reflective layers in a tissue of the sample may be determined from the interference signal, and information associated with the dispersion that is provided in the data can be obtained. Such information may be used to reduce and/or remove the dispersion from the data. The phase of the spectral component of the portion of the image can be controlled using software.


According to another exemplary embodiment of the present invention, prior to controlling the phase of the at least one spectral component of the interference signal, the dispersion may be quantified, and the dispersion may be corrected for in the image based on the quantification. The dispersion can be a chromatic dispersion. Data associated with the dispersion of the image may also be determined, the dispersion quantified using the data, and the dispersion in the image corrected for based on the quantification. The sample may be a retina of an eye, and the information may include data associated with spectral reflections obtained from the retina. Further, an operator may be enable to select at least one dispersed spectral reflection of the spectral reflections. The dispersed spectral reflection may be selected using a graphical user interface, e.g., during an acquisition of the image and/or after the acquisition of the image. The dispersion can be quantified using the dispersed spectral reflection, and corrected for in the image based on the quantification. A brightest one of the spectral reflections may be interactively searched for, the dispersion quantified using the brightest one of the spectral reflections, and corrected for in the image based on the quantification.


According to still another exemplary embodiment of the present invention, the dispersion can be a depth dependent chromatic dispersion. The information may include dispersed image data, and the dispersion may be quantified using the dispersed image data, and corrected for in the image based on the quantification. The sample may be a retina of an eye, and the dispersed image data may includes spectral reflections. The dispersion may be quantified using the spectral reflections.


In a further exemplary embodiment of the present invention, the dispersion can be compensated for by correcting the dispersion in the image using predetermined constant chromatic dispersion parameters, e.g., based on an estimate of an axial eye length and/or an estimate of an axial eye length.


Other features and advantages of the present invention will become apparent upon reading the following detailed description of embodiments of the invention, when taken in conjunction with the appended claims.





BRIEF DESCRIPTION OF THE DRAWINGS

Further objects, features and advantages of the invention will become apparent from the following detailed description taken in conjunction with the accompanying figures showing illustrative embodiments of the invention, in which:



FIG. 1 is a block diagram of an exemplary embodiment of a spectral domain optical coherence tomography (“SD-OCT”) arrangement according to the present invention which is capable of implementing the exemplary embodiments of the system, process and software arrangement according to the present invention;



FIG. 2 is a block diagram of an exemplary embodiment of an optical frequency domain intereferometry (“OFDI”) arrangement according to the present invention which is capable of implementing the exemplary embodiments of the system, process and software arrangement according to the present invention;



FIG. 3 is an exemplary graph illustrating an absolute value/depth which can be used for the exemplary embodiments of the system, process and software arrangement according to the present invention;



FIG. 4 is an exemplary graph illustrating curves without dispersion compensation, and with the dispersion compensation applied according to the exemplary embodiment of the present invention;



FIG. 5 is an exemplary graph of a phase θ(k) obtained according to an exemplary embodiment of the present invention from a model eye and from a spectral reflection in a fovea;



FIG. 6 is a retinal image of a human subject which include spectral reflections that may be utilized according to an exemplary embodiment of the present invention;



FIG. 7 is an exemplary image that may be obtained from a human subject, which illustrates the fovea after the dispersion compensation according to an exemplary embodiment of the present invention has been applied;



FIG. 8 is an exemplary graph of a coherence function obtained from a reflective spot in the fovea obtained using an exemplary embodiment of the present invention;



FIG. 9 is a high level flow diagram of a process according to an exemplary embodiment of the present invention;



FIG. 10 is another exemplary image that may be obtained from a human subject, in which a portion of dispersion can be selected via software by an operator; and



FIG. 11 is a detailed flow diagram of a process according to yet another exemplary embodiment of the present invention.





DETAILED DESCRIPTION


FIG. 1 shows an exemplary embodiment of a sample configuration of a spectral domain optical coherence tomography (“SD-OCT”) arrangement which can be used for implementing the exemplary embodiments of the system, process and software arrangement according to the present invention. A detailed description of operation of this arrangement is described in detail in International Patent Publication No. WO 03/062802. In particular, as shown in FIG. 1, a high-powered superluminescent diode source (“HP-SLD”) 10 generates an electromagnetic radiation or light signal which is transmitted through a first polarization controller (“PC”) 20′ and an optical isolator 30 so as to facititate a one way propegation of an electromagnetic energy to reach a signal splitter 40. The signal splitter forwards one portion of the split signal to a reference arm (which includes a second PC 20″, a reference, certain optics and a neutral density filter (“NFD”) 50) and another portion of the split signal to a sample arm (which includes a third PC 20′″, certain optics and a sample 60 such as the eye). Thereafter, an electromagnetic signal is reflected from the sample 60 and is combined with the light from the reference arm to form an interference signal. This interference signal is forwarded to a fourth PC 20″″, and forwarded to a collimator (“Col”) 70, a transmission grating (“TG”) 80, an air-spaced focusing lens (“ASL”) 90, and a linescan camera (“LSC”) 100 to be detected by a detecting arrangement (e.g., provided in the linescan camera), and then analyzed by a processing arrangement, e.g., a computer (not shown). Such processing arrangement is capable of implementing the various exemplary embodiments of the system, process and software arrangement according to the present invention.



FIG. 2 shows an exemplary embodiment of an optical imaging frequency domain intereferometry (“OFDI”) arrangement according to the present invention which is capable of implementing the exemplary embodiments of the system, process and software arrangement according to the present invention. A detailed description of various embodiments of the OFDI arrangement is provided in U.S. Patent Application Ser. No. 60/514,769. For example, the light source may be a wavelength-swept source 110. In order to generate a synchronization signal, a portion of the laser output (for example −20%) is obtained, and detected using a fast InGaAs photo-detector through a narrowband fixed-wavelength filter. The detector generates a pulse when the output spectrum of the laser sweeps through the narrow passband of the filter. The detector pulse is fed to a digital circuit 120, e.g., a synchronous TTL pulse generator, for converting the resultant signal to a TTL pulse train. The TTL pulses are used to generate gating pulses for signal sampling. 90% of the remaining light is directed to the sample arm and 10% to the reference mirror 130. This exemplary arrangement can utilize an optical probe based on a galvanometer mirror (e.g., scanner) 140 and an imaging lens. The galvanometer-mounted mirror 140 is controlled by a glava-driver 145 so as to scan the probe light transversely on the sample 60. The total optical power illuminated on the sample 60 may be approximately 3.5 mW. The light reflected from the reference mirror 130 and the sample 60 is received through magneto-optic circulators 150′, 150″, and combined by a 50/50 coupler 160. A fiber-optic polarization controller may be used in the reference arm to align polarization states of the reference and sample arms.


In general, a relative intensity noise (“RIN”) of the received light signal may be proportional to a reciprocal of the linewidth, and the relatively high RIN can be reduced by dual balanced detection (e.g., using a dual balanced receiver 170). The differential current of two InGaAs detectors D1 and D2 in the receiver 170 may be amplified using trans-impedance amplifiers (“TIA”) having a total gain of 56 dB, and passed through a low pass filter (“LPF”) with a 3-dB cutoff frequency at approximately half the sampling rate. The common-noise rejection efficiency of the receiver 170 may be typically greater than 20 dB. In addition to the RIN reduction, the balanced detection may provide other significant benefits—a suppression of a self-interference noise originating from multiple reflections within the sample and optical components; an improvement in the dynamic range; and a reduction of a fixed-pattern noise by greatly reducing the strong background signal from the reference light. Thereafter, a detecting arrangement 180 receives such signals, and forward them to a processing arrangement 190 (e.g., a computer) which implements the exemplary embodiments of the system, process and software arrangement according to the present invention to reduce dispersion, and assist in displaying a resultant image that is based on the original image and the reduction of the dispersion.


Both of these exemplary arrangements, e.g., the SD-OCT arrangement described above with reference to FIG. 1 and the OFDI arrangement described above with reference to FIG. 2, are capable of generating a spectrum I(k) as a function of wave vector k. To determine the phase term, the spectrum I(k) can be Fourier transformed to z-space. FIG. 3 shows an exemplary graph 200 providing an illustration of the curve of an absolute value Abs(I(z)) for z>0 of the Fourier transformed spectrum I(k) vs. depth, with I(z)=FFT(I(k)). As shown in FIG. 3, a dispersion broadened peak may be observed at a depth of approximately 0.6 mm. The function I(z) may be shifted such that the coherence function is centered on the origin. A window can be selected around the coherence function so as to possibly eliminate coherence functions from other reflective structures in the depth profile, and all values outside the window may be set equal to zero. A complex spectrum in k-space may be obtained after an inverse Fourier transformation. The phase term θ(k) can be equal to the arctangent of the imaginary component divided by the real component. Such term can indicate by how much are the subsequent wave numbers k out of phase with each other. According to one example, this function can be provided to a polynomial expression of 9th order, yielding a set of coefficients α1-9.


According to one exemplary embodiment of the present invention, individual spectra may be multiplied with a phase e−iθ(k) as determined from the previous seven polynomial coefficients, where the first two polynomial coefficients may be set to zero, and then inversely Fourier transformed into A-lines, thus removing dispersion. The original and resulting exemplary coherence functions are illustrated in FIG. 4. In particular, the curve of FIG. 4 shows the results without the dispersion compensation is shown as a dashed line, and referred to by numeral 210, and the curve illustrating the results after the dispersion compensation has been successfully applied which is shown as a solid line, and referred to by numeral 220.



FIG. 5 shows a an illustration which aids in the determination of the phase function θ(k) based on certain measurements according to an exemplary embodiment of the present invention, as well as the phase function that subtracts the polynomial fit of 9th order to the phase function. The phase θ(k) may be obtained from a mirror in a model eye and from a spectral reflection in the fovea (e.g., the left axis).


In another example according to the present invention, in vivo human data may be used to determine the phase function for an optimal dispersion compensation. FIG. 6 shows an exemplary retinal image of a human subject, in which three spectral reflections 300, 310, 320 are marked with arrows. These exemplary reflections 300, 310, 320 originate from an internal limiting membrane on top of the retinal nerve fiber layer and the foveolar umbo and from the external limiting membrane. Unmarked, still visible is an exemplary spectral reflection on the surface between the inner and outer segments of the photoreceptor layer, just below the external limiting membrane. FIG. 6 shows typical examples of strong reflections in an image that can be used to determine the phase function for the optimal dispersion compensation.


In order to determine this phase term for the dispersion compensation of data obtained in the sample (e.g., the human eye) in vivo, it is preferable to use a coherence function obtained from a well-reflecting reference point in the eye. In this example, the reflection of the foveal umbo can be used. Other regions in the eye may also create spectral reflections. Spectral reflections may be present from the interface between the inner and outer segments of the photoreceptor layer (“IPRL”) and retinal pigmented epitheleum (“RPE”). In addition, spectral reflections may also be found on the inner limiting membrane, e.g., on top of the retinal nerve fiber layer. For example, five depth profiles may be selected that can illustrate a strong reflection from the foveal umbo. A window can be selected such that it is centered at these strong reflections, and the remaining points may be set to zero. The phase function θ(k) may then be determined as described herein above. In particular, the phase function θ(k) as shown in FIG. 5, can be determined from this measurement, as well as based on the phase function minus the polynomial fit of 9th order to the phase function.


Individual spectra of the image can be first multiplied with a phase e−iθ(k) as determined from the last seven polynomial coefficients, and then inversely Fourier transformed into A-lines, thus removing dispersion. The fit to the dispersion data as determined from the well reflecting reference point in the eye can be a polynomial of any order. Use of a 9th order polynomial was demonstrated. Instead of a polynomial, the data can be fitted to a Fourier series or any other known function set so as to determine a set of coefficients. One of the advantages of using e.g., a polynomial of limited order to filter the dispersion curve is a better immunity to noise of the determined phase correction function.



FIG. 7 shows an exemplary image that may be obtained from a human subject, which illustrates the fovea after the dispersion compensation. The spectral reflection marked with an R can be first used to determine the amount of a chromatic dispersion (as described above), and to remove the chromatic dispersion. The dimensions of the image illustrated in FIG. 7 are 3.1 mm×0.61 mm. Layers in this image are labeled as follows: RNFL—retinal nerve fiber layer; GCL—ganglion cell layer; IPL—inner plexiform layer; INL—inner nuclear layer; OPL—outer plexiform layer; ONL—outer nuclear layer; ELM—external limiting membrane; IPRL—interface between the inner and outer segments of the photoreceptor layer; RPE—retinal pigmented epithelium; C—choriocapillaris and choroid. A highly reflective spot in the center of the fovea is marked with an R. A blood vessel is marked with a large circle (BV) and structures in the outer plexiform layer are marked with smaller circles. FIG. 8 shows a graph of a coherence function obtained from a reflective spot in the fovea. For example, the coherence length is equal to 4.8 μm in air.


To summarize, in the graph shown in FIG. 5, the phase term θ(k) obtained from a mirror in a water-filled model eye (averaged over 100 A-lines) and from a spectral reflective spot in the human fovea (averaged over 5 A-lines, see FIG. 7) are illustrated. The differences between the measured phase terms and polynomial fits (9th order) to the data are also shown, with the corresponding axis provided on a right side thereof. Both phases show approximately the same pattern, which indicates that the model eye and the real eye generally experience similar amounts of dispersion. The phase term obtained from the spectral reflection of the fovea can be used (e.g., curve 270 of FIG. 5) to remove chromatic dispersion artifacts in data obtained from a human subject in vivo, as shown in the graph of FIG. 7 and quantified in the graph of FIG. 8.


The coherence length can be determined in vivo from the spectral reflection in the center of the fovea labeled as R in FIG. 7, averaged over 5 A-lines. This coherence function is shown as a graph in FIG. 8, and the coherence length after dispersion compensation as shown in FIG. 8 as being equal to 4.8 μm in air and 3.5 μm in tissue (n=1.38). It is clear that without dispersion compensation, the coherence length is significantly longer (e.g., by a factor of 2-3), thus reducing the axial resolution considerably. In particular, the curve of FIG. 8 (similarly to the graph in FIG. 4) shows the results without the dispersion compensation is illustrated as a dashed line, and referred to by numeral 410, and the curve illustrating the results after the dispersion compensation has been successfully applied which is shown as a solid line, and referred to by numeral 420.


Practically, an exemplary embodiment of the system, process and software arrangement according to the present invention can also be described with reference to FIG. 9 which illustrates a flow diagram according to one exemplary embodiment of the present invention. As previously described, an area in the image containing a spectral reflection is selected (step 510). After such selection, the existing algorithm determines the amount of chromatic dispersion (step 520) and removes such dispersion from the image (step 530). As previously described, the dispersion can be removed by multiplying spectra in k-space with a phase e−iθ(k). The earlier described polynomial fit can be used. Since the polynomial fit and the original phase are approximately similar (as shown in FIG. 5), it is also possible to use a measured phase curve. The selection procedure for selecting the location of the spectral reflection can be either a manual procedure or an automated procedure. Thereafter, a new image may be generated based on the originally-selected image, but with the dispersions that was removed according to the exemplary technique of the present invention (step 540).


The previously-described exemplary results may be obtained using a simple manual procedure, where the operator generally selects the specific portion of the image by hand, e.g., by determining the coordinates of the reflecting spots. Such procedure can be simplified with, e.g., MatLab software, in which the operator may be requested to draw a rectangular shape around a reflecting spot, (see numeral 600 in FIG. 10), thus selecting the location of the spectral spot. Using such exemplary selection of the portion of the image, the dispersion can be compensated using the compensation described above.


According to another exemplary embodiment of the present invention, spectral reflections can also be located automatically by using a particular technique. This exemplary technique can be based on an algorithm that finds a maximum signal For example, the signal returning from the spectral reflection, e.g., in the center of the fovea generally has a higher value than any of the other reflections. Using such exemplary technique, it is possible to select this reflecting spot automatically, and therefor manual input from an operator would not be necessary. with this technique, a feedback signal can be forwarded to the scanning apparatus, so that this apparatus monitors for the brightest spectral reflection in the sample 60 (e.g., the eye). For example, a series of smaller and smaller three-dimensional raster scans can be acquired, until the center of the fovea is located. If the subject moves during this operation, the raster scanning can expanded and confined the target image again. In another publication, an exemplary technique used to track the surface of the retina and compensate for motion artifacts has been described. See Cense, B. et al., “In Vivo Birefringence and Thickness Measurements of the Human Retinal Nerve Fiber Layer Using Polarization-Sensitive Optical Coherence Tomography”, Journal of Biomedical Optics, 2004, 9(1), pp. 121-125.


Another exemplary embodiment of the present invention uses compensated dispersion in dependence of depth. The technique according to the exemplary embodiment of the present invention described above is capable of compensating for a constant dispersion mismatch between the sample and the reference arm. However, with an increasing bandwidth available in the OCT imaging, dispersion broadening between superficial and deeper layers within an image may becomes important. The dispersion broadening may be due to the accumulated dispersion between the superficial and deeper layer.


Described herein below is a technique according to another exemplary embodiment of the present invention which is capable of depth dependent dispersion compensation, i.e., a correction for the dispersion that varies with depth. As is well known, the signal in SD-OCT and OFDI is defined by,










I


(
k
)


=



I
r



(
k
)


+

2





I
s



(
k
)





I
r



(
k
)








n




α
n



cos


(

kz
n

)





+


I
s



(
k
)







(
2
)








where Ir(k) and Is(k) are the wavelength-dependent intensities reflected from the reference and sample arms, respectively, and k is the wave number. The second term on the right hand side of Eq. (2) represents the interference between the light signal returning from the reference and sample arms. αn is the square root of the sample reflectivity at depth zn. As described in Hausler, G. et al., “Coherence Radar and Spectral Radar—New Tools for Dermatological Diagnosis”, J. Biomed. Opt., 1998, 3(1), pp. 21-31, depth information can be obtained by performing an inverse Fourier transform of Eq. (2), yielding the following convolution















FT

-
1




[

I


(
k
)


]




2

=



Γ
2



(
z
)




{





δ


(
0
)


+



n




α
n
2



δ


(

z
-

z
n


)




+









n




α
n
2



δ


(

z
+

z
n


)




+

O


[


I
s
2

/

I
r
2


]






}



,




(
3
)








with Γ(z) representing the envelope of the coherence function. The first term in the brackets on the right hand side refers to an autocorrelation signal from the reference arm, and has magnitude unity. The second and third terms are reflect the interference between light returning from the reference and sample arms and from two images, where each has magnitude on the order of Is/Ir. These two terms provide mirror images. The final term, with magnitude on the order of Is2/Ir2, describes autocorrelation noise due to interference within the sample arm. Is and Ir represent the total intensity reflected from sample and reference arms, respectively.


Retaining only the interference term







2





I
s



(
k
)





I
r



(
k
)








n




α
n



cos


(

kz
n

)





,





a constant dispersion mismatch can be described by introducing a phase term θ(k) in the cosine term,






2





I
s



(
k
)





I
r



(
k
)








n




α
n




cos


(


k






z
n


+

θ


(
k
)



)


.








The constant dispersion mismatch can be compensated for with the method described before. A depth dependent dispersion term is described by introducing a depth dependent phase term, f(k)zn in the cosine term,






2





I
s



(
k
)





I
r



(
k
)








n




α
n




cos


(


kz
n

+


f


(
k
)




z
n



)


.








The depth dependent dispersion term can be compensated for by a remapping operation of the data in k-space. The cosine term can be rewritten as






2





I
s



(
k
)





I
r



(
k
)








n




α
n



cos


(


k




z
n


)









with k′=k+f(k). After the remapping operation, the data can be linearized in k-space before the Fourier transform resulting in Eq. (3).


The function f(k) can be determined by measuring the dispersion F(k)n and F(k)m at two different locations, zn and zm using the method described for a constant dispersion term, where the function f(k) is given by







f


(
k
)


=





F


(
k
)


m

-


F


(
k
)


n




z
m

-

z
n



.






The locations for determining F(k)n and F(k)m are preferably locations in the material (tissue, vitrious, retina, coronary artery, etc) with strong reflections. Filtering the function f(k) to reject noise can be performed by retaining only a limited or predetermined number of coefficients from a polynomial or Fourier series fit to the data. This exemplary technique can be used to predetermine the dispersion in various materials or biological tissues, and utilize the determined values to implement depth dependent dispersion compensation during imaging or post processing of SD-OCT and OFDI data. For use in retinal data, several locations can provide strong reflections that can be used to determine the dispersion, such as the center of the fovea (fovealar umbo), external limiting membrane, interface between the inner and outer segments of the photoreceptor layer (“IPRL”) and retinal pigmented epitheleum (“RPE”). Spectral reflections can also be located on the inner limiting membrane, on top of the retinal nerve fiber layer. In order to see these reflections, the sample (e.g., the eye) should be tilted such that the surface thereof is exactly perpendicular to the beam.


A further technique according to yet another exemplary embodiment of the present invention can be used to determine a constant and depth dependent dispersion. For example, in the presence of constant and depth-dependent dispersion, the interference signal associated with the n-th reflection point in the sample can be defined by











I


(

k
,

z

s
,
n


,

z
r


)


=

2





I
s



(
k
)





I
r



(
k
)






α
n



cos


[


k


(


z

s
,
n


-

z
r


)


+


f


(
k
)




z

s
,
n



+

θ


(
k
)



]




,




(
4
)








where zs,n refers to the distance of the reflection point from the surface of the sample, and zr refers to the position of the reference mirror with respect to the sample surface. Shifting the position of the reference mirror to zr′=2zs,n−zr provides the following













I


(

k
,

z

s
,
n


,

z
r



)


=



2





I
s



(
k
)





I
r



(
k
)






α
n



cos


[





-

k


(


z

s
,
n


-

z
r


)



+


f


(
k
)




z

s
,
n



+







θ


(
k
)


+
δ




]










=



2





I
s



(
k
)





I
r



(
k
)






α
n



cos


[





k


(


z

s
,
n


-

z
r


)


-


f


(
k
)




z

s
,
n



-







θ


(
k
)


-
δ




]




,







(
5
)








where δ refers to any possible phase error introduced in the measurement. It is possible to determine the phase functions, φ(k,zs,n,zr) and φ(k,zs,n,zr′) of the interference signals in Eq. (4) and (5), respectively. It follows that

φ(k,zs,n,zr)−φ(k,zs,n,zr′)=2f(k)zs,n+2θ(k)+δ  (6)


The third-term on right hand side, a constant phase error, can be differentiated from the 1st and 2nd terms which are both dependent on k. By measuring the differential phase for multiple reflection points in the sample or for multiple zs,n where n=1 to N, it is possible to determine f(k) and θ(k).


If the constant dispersion is negligible or has been canceled, it is possible to locate the best or preferable mapping function that leads to transform-limited point spread function for each position of the reference mirror. The preferable mapping function may be defined by k′=k+j(k) for the signal represented in Eq. (4) and k′=k−f(k) for Eq. (5). Therefore, subtracting the two mapping functions can yield the depth-dependent dispersion f(k). Instead of shifting the reference mirror, the mirror can be placed so that the zero delay corresponds to (either approximately or exactly) the middle of the two reflection points in the sample. The interference signal associated with the two reflections can be simultaneously measured and analyzed to determine the dispersion.



FIG. 11 shows another exemplary embodiment of the process according to the present invention which can be used to control the dispersion of the data associated with the image obtained from the reference and sample arms. For example, a detector (e.g., the detectors of the arrangements shown in FIG. 1 and/or 2) received and detect an interference signal which contains data associated with the electromagnetic radiation received from the sample arm and the reference arm (step 605), and then generates a spectrum signal I(k) based on the detected interference signal (step 610). This spectrum signal I(k) is forwarded to the processing arrangement, e.g., as data, which performs a Fast Fourrier Transform (“FFT”) on the spectrum signal I(k) (step 615). Thereafter, an initial signal I(z) associated with the spectrum signal I(k) is set to 0 for z>0 and z<0 (step 620), and in step 625, an absolute values for the initial signal I(z) is set. In step 630, a signal I(k) is generated based on the detected signal, a window of interest of the image may be determined in step 635. Such are of interest can be a region of the peak of the absolute value signal (ABS(I(z))), a center peak at around z−0 by shifting the signal, etc. The window can be obtained automatically by the processing arrangement and/or manually by an operator.


In step 640, an inverse FFT is performed on the signal I(z), and a phase term θ(k) of the transformed I(z) signal is determined (step 645). In step 650, the exemplary process according to the present invention the phase function that can apply a polynomial of the order of N to θ(k), e.g., by subtracting the polynomial fit of 9th order, yielding a set of coefficients α1-9. As described herein, the phase θ(k) may be obtained from a mirror in a model eye and from a spectral reflection in the fovea. The filtered phase term can then be determined from the polynomial fit parameters/coefficients, e.g., by setting the first two polynomial coefficients to zero. In step 260, the filtered phase of the signal e−iθ(k) can be stored for use in multiple images. Then, in step 665, a correction curve of the filtered phase term θ(k) can be applied, e.g., by multiplying all spectra of the image may be multiplied by e−iθ(k). Finally, in step 670, dispersion corrected spectrum S′(k)=S(k) e−iθ(k) may be used to calculate image intensity, birefringence and/or flow information.


The foregoing merely illustrates the principles of the invention. Various modifications and alterations to the described embodiments will be apparent to those skilled in the art in view of the teachings herein. For example, the invention described herein is usable with the exemplary methods, systems and apparatus described in U.S. Patent Application No. 60/514,769. It will thus be appreciated that those skilled in the art will be able to devise numerous systems, arrangements and methods which, although not explicitly shown or described herein, embody the principles of the invention and are thus within the spirit and scope of the present invention. In addition, all publications, patents and patent applications referenced above are incorporated herein by reference in their entireties.

Claims
  • 1. A system to compensate for a dispersion in at least one portion of an image, comprising: a processing arrangement configured to obtain information associated with the at least one portion of the image, the at least one portion of the image being associated with the interference signal that includes a first electro-magnetic radiation received from a sample and a second electro-magnetic radiation received from a reference, wherein the processing arrangement is configured to determine (i) complex spectral data that is based the interference signal, and (ii) a phase of at least one spectral component of the complex spectral data, and wherein the processing arrangement is further configured to compensate for the dispersion in the at least one portion by controlling the phase of the at least one spectral component of the complex spectral data.
  • 2. The system according to claim 1, wherein the dispersion is an indication of a difference between the first and second electro-magnetic radiations.
  • 3. The system according to claim 1, wherein the processing arrangement is configured to control the dispersion by at least one of reducing and removing the dispersion in the at least one portion of the image.
  • 4. The system according to claim 1, wherein the processing arrangement is further configured to determine data associated with reflective layers in a tissue of the sample from the interference signal, and determining information associated with the dispersion that is provided in the data.
  • 5. The system according to claim 4, wherein the processing arrangement is further configured to utilize the information to at least one of reduce and remove the dispersion from the data.
  • 6. The system according to claim 1, wherein, when the processing arrangement executes software instructions, the processing arrangement is configured to control the phase of the at least one spectral component of the interference signal.
  • 7. The system according to claim 1, wherein, prior to controlling the phase of the at least one spectral component of the interference signal, the processing arrangement is configured to quantify the dispersion, and correct for the dispersion in the image based on the quantification.
  • 8. The system according to claim 1, wherein the dispersion is a chromatic dispersion.
  • 9. The system according to claim 1, wherein the processing arrangement is further configured to determine data associated with the dispersion of the image, quantify the dispersion using the data, and correct for the dispersion in the image based on the quantification.
  • 10. The system according to claim 9, wherein the sample is a retina of an eye.
  • 11. The system according to claim 10, wherein the information includes data associated with spectral reflections obtained from the retina.
  • 12. The system according to claim 11, wherein the processing arrangement is configured to enable an operator to select at least one dispersed spectral reflection of the spectral reflections.
  • 13. The system according to claim 12, wherein the at least one dispersed spectral reflection is selected using a graphical user interface.
  • 14. The system according to claim 12, wherein the at least one dispersed spectral reflection is selected at least one of during an acquisition of the image and after the acquisition of the image.
  • 15. The system according to claim 12, wherein the processing arrangement is further configured to quantify the dispersion using the at least one dispersed spectral reflection, and correct for the dispersion in the image based on the quantification.
  • 16. The system according to claim 12, wherein the processing arrangement is further configured to interactively search for a brightest one of the spectral reflections, quantify the dispersion using the brightest one of the spectral reflections, and correct for the dispersion in the image based on the quantification.
  • 17. The system according to claim 1, wherein the dispersion is a depth dependent chromatic dispersion.
  • 18. The system according to claim 17, wherein the information includes dispersed image data, and wherein the processing arrangement is further configured to quantify the dispersion using the dispersed image data, and correct for the dispersion in the image based on the quantification.
  • 19. The system according to claim 18, wherein the sample is a retina of an eye.
  • 20. The system according to claim 17, wherein the dispersed image data includes spectral reflections, and wherein the processing arrangement is configured to quantify the dispersion using the spectral reflections.
  • 21. The system according to claim 1, wherein the processing arrangement is configured to control the dispersion by correcting the dispersion in the image using predetermined constant chromatic dispersion parameters.
  • 22. The system according to claim 21, wherein the dispersion is compensated based on an estimate of an axial eye length.
  • 23. The system according to claim 21, wherein the dispersion is compensated based on an estimate of an axial eye length.
  • 24. A method to compensate for a dispersion in at least one portion of an image, comprising: obtaining information associated with the at least one portion of the image, the at least one portion of the image being associated with an interference signal that includes a first electro-magnetic radiation received from a sample and a second electro-magnetic radiation received from a reference;determining (i) complex spectral data that is based the interference signal, and (ii) a phase of at least one spectral component of the complex spectral data; andcompensating for the dispersion in the at least one portion of the image by controlling the phase of the at least one spectral component of the complex spectral data.
  • 25. The method according to claim 24, wherein the dispersion is an indication of a difference between the first and second electro-magnetic radiations.
  • 26. The method according to claim 24, wherein the controlling step includes the substep of at least one of reducing and removing the dispersion in the at least one portion of the image.
  • 27. The method according to claim 24, further comprising the steps of: determining data associated with reflective layers in a tissue of the sample from the interference signal; anddetermining information associated with the dispersion that is provided in the data.
  • 28. The method according to claim 27, further comprising the step of utilizing the information to at least one of reduce and remove the dispersion from the data.
  • 29. The method according to claim 24, wherein the controlling step is performed using software instructions.
  • 30. The method according to claim 24, further comprising the steps of, prior to the controlling step, quantifying the dispersion; andcorrecting for the dispersion in the image based on the quantification.
  • 31. The method according to claim 24, wherein the dispersion is a chromatic dispersion.
  • 32. The method according to claim 24, further comprising the steps of: determining data associated with the dispersion of the image;quantifying the dispersion using the data; andcorrecting for the dispersion in the image based on the quantification.
  • 33. The method according to claim 32, wherein the sample is a retina of an eye.
  • 34. The method according to claim 33, wherein the information includes data associated with spectral reflections obtained from the retina.
  • 35. The method according to claim 33, further comprising the step of enabling an operator to select at least one dispersed spectral reflection of the spectral reflections.
  • 36. The method according to claim 35, wherein the at least one dispersed spectral reflection is selected using a graphical user interface.
  • 37. The method according to claim 35, wherein the at least one dispersed spectral reflection is selected at least one of during an acquisition of the image and after the acquisition of the image.
  • 38. The method according to claim 35, further comprising the steps of: quantifying the dispersion using the at least one dispersed spectral reflection; andcorrecting for the dispersion in the image based on the quantification.
  • 39. The method according to claim 35, further comprising the steps of: interactively searching for a brightest one of the spectral reflections;quantifying the dispersion using the brightest one of the spectral reflections; andcorrecting for the dispersion in the image based on the quantification.
  • 40. The method according to claim 24, wherein the dispersion is a depth dependent chromatic dispersion.
  • 41. The method according to claim 40, wherein the information includes dispersed image data, and further comprising the steps of: quantifying the dispersion using the dispersed image data; andcorrecting for the dispersion in the image based on the quantification.
  • 42. The method according to claim 41, wherein the sample is a retina of an eye.
  • 43. The method according to claim 40, wherein the dispersed image data includes spectral reflections, and wherein the processing arrangement is adapted to quantify the dispersion using the spectral reflections.
  • 44. The method according to claim 43, further comprising the step of controlling the dispersion by correcting the dispersion in the image using predetermined constant chromatic dispersion parameters.
  • 45. The method according to claim 44, wherein the dispersion is corrected based on an estimate of an axial eye length.
  • 46. The method according to claim 44, wherein the dispersion is compensated based on an estimate of an axial eye length.
  • 47. A computer program adapted to compensate for a dispersion in at least one portion of an image for execution by a processing arrangement, said computer program comprising instructions for performing a method comprising: obtaining information associated with the at least one portion of the image, the at least one portion of the image being associated with an interference signal that includes a first electro-magnetic radiation received from a sample and a second electro-magnetic radiation received from a reference;determining (i) complex spectral data that is based the interference signal, and (ii) a phase of at least one spectral component of the complex spectral data; andcompensating for the dispersion in the at least one portion of the image by controlling the phase of the at least one spectral component of the complex spectral data.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority from U.S. Patent Application Ser. No. 60/575,773 filed on May 29, 2004, the entire disclosure of which is incorporated herein by reference.

PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/US2004/023585 7/23/2004 WO 00 4/11/2008
Publishing Document Publishing Date Country Kind
WO2005/117534 12/15/2005 WO A
US Referenced Citations (472)
Number Name Date Kind
2339754 Brace Jan 1944 A
3090753 Matuszak et al. May 1963 A
3601480 Randall Aug 1971 A
3856000 Chikama Dec 1974 A
3872407 Hughes Mar 1975 A
3941121 Olinger Mar 1976 A
3973219 Tang et al. Aug 1976 A
3983507 Tang et al. Sep 1976 A
4030827 Delhaye et al. Jun 1977 A
4030831 Gowrinathan Jun 1977 A
4140364 Yamashita et al. Feb 1979 A
4141362 Wurster Feb 1979 A
4224929 Furihata Sep 1980 A
4295738 Meltz et al. Oct 1981 A
4300816 Snitzer et al. Nov 1981 A
4303300 Pressiat et al. Dec 1981 A
4428643 Kay Jan 1984 A
4479499 Alfano Oct 1984 A
4533247 Epworth Aug 1985 A
4585349 Gross et al. Apr 1986 A
4601036 Faxvog et al. Jul 1986 A
4607622 Fritch et al. Aug 1986 A
4631498 Cutler Dec 1986 A
4639999 Daniele Feb 1987 A
4650327 Ogi Mar 1987 A
4734578 Horikawa Mar 1988 A
4744656 Moran et al. May 1988 A
4751706 Rohde et al. Jun 1988 A
4763977 Kawasaki et al. Aug 1988 A
4770492 Levin et al. Sep 1988 A
4827907 Tashiro et al. May 1989 A
4834111 Khanna et al. May 1989 A
4868834 Fox et al. Sep 1989 A
4890901 Cross, Jr. Jan 1990 A
4892406 Waters Jan 1990 A
4905169 Buican et al. Feb 1990 A
4909631 Tan et al. Mar 1990 A
4925302 Cutler May 1990 A
4928005 Lefevre et al. May 1990 A
4940328 Hartman Jul 1990 A
4965441 Picard Oct 1990 A
4965599 Roddy et al. Oct 1990 A
4966589 Kaufman Oct 1990 A
4984888 Tobias et al. Jan 1991 A
4993834 Carlhoff et al. Feb 1991 A
4998972 Chin et al. Mar 1991 A
5039193 Snow et al. Aug 1991 A
5040889 Keane Aug 1991 A
5045936 Lobb et al. Sep 1991 A
5046501 Crilly Sep 1991 A
5065331 Vachon et al. Nov 1991 A
5085496 Yoshida et al. Feb 1992 A
5120953 Harris Jun 1992 A
5121983 Lee Jun 1992 A
5127730 Brelje et al. Jul 1992 A
5197470 Helfer et al. Mar 1993 A
5202745 Sorin et al. Apr 1993 A
5202931 Bacus et al. Apr 1993 A
5208651 Buican May 1993 A
5212667 Tomlinson et al. May 1993 A
5214538 Lobb May 1993 A
5217456 Narciso, Jr. Jun 1993 A
5228001 Birge et al. Jul 1993 A
5241364 Kimura et al. Aug 1993 A
5248876 Kerstens et al. Sep 1993 A
5250186 Dollinger et al. Oct 1993 A
5251009 Bruno Oct 1993 A
5262644 Maguire Nov 1993 A
5275594 Baker Jan 1994 A
5281811 Lewis Jan 1994 A
5283795 Fink Feb 1994 A
5291885 Taniji et al. Mar 1994 A
5293872 Alfano et al. Mar 1994 A
5293873 Fang Mar 1994 A
5302025 Kleinerman Apr 1994 A
5304173 Kittrell et al. Apr 1994 A
5304810 Amos Apr 1994 A
5305759 Kaneko et al. Apr 1994 A
5317389 Hochberg et al. May 1994 A
5318024 Kittrell et al. Jun 1994 A
5321501 Swanson et al. Jun 1994 A
5348003 Caro Sep 1994 A
5353790 Jacques et al. Oct 1994 A
5383467 Auer et al. Jan 1995 A
5394235 Takeuchi et al. Feb 1995 A
5404415 Mori et al. Apr 1995 A
5411016 Kume et al. May 1995 A
5419323 Kittrell et al. May 1995 A
5424827 Horwitz et al. Jun 1995 A
5439000 Gunderson et al. Aug 1995 A
5441053 Lodder et al. Aug 1995 A
5450203 Penkethman Sep 1995 A
5454807 Lennox et al. Oct 1995 A
5459325 Hueton et al. Oct 1995 A
5459570 Swanson et al. Oct 1995 A
5465147 Swanson Nov 1995 A
5486701 Norton et al. Jan 1996 A
5491524 Hellmuth et al. Feb 1996 A
5491552 Knuttel Feb 1996 A
5522004 Djupsjobacka et al. May 1996 A
5526338 Hasman et al. Jun 1996 A
5555087 Miyagawa et al. Sep 1996 A
5562100 Kittrell et al. Oct 1996 A
5565983 Barnard et al. Oct 1996 A
5565986 Knüttel Oct 1996 A
5566267 Neuberger Oct 1996 A
5583342 Ichie Dec 1996 A
5590660 MacAulay et al. Jan 1997 A
5600486 Gal et al. Feb 1997 A
5601087 Gunderson et al. Feb 1997 A
5621830 Lucey et al. Apr 1997 A
5623336 Raab et al. Apr 1997 A
5635830 Itoh Jun 1997 A
5649924 Everett et al. Jul 1997 A
5697373 Richards-Kortum et al. Dec 1997 A
5698397 Zarling et al. Dec 1997 A
5710630 Essenpreis et al. Jan 1998 A
5716324 Toida Feb 1998 A
5719399 Alfano et al. Feb 1998 A
5730731 Mollenauer et al. Mar 1998 A
5735276 Lemelson Apr 1998 A
5740808 Panescu et al. Apr 1998 A
5748318 Maris et al. May 1998 A
5748598 Swanson et al. May 1998 A
5752518 McGee et al. May 1998 A
5784352 Swanson et al. Jul 1998 A
5785651 Kuhn et al. Jul 1998 A
5795295 Hellmuth et al. Aug 1998 A
5801826 Williams Sep 1998 A
5801831 Sargoytchev et al. Sep 1998 A
5803082 Stapleton et al. Sep 1998 A
5807261 Benaron et al. Sep 1998 A
5810719 Toida Sep 1998 A
5817144 Gregory Oct 1998 A
5836877 Zavislan Nov 1998 A
5840023 Oraevsky et al. Nov 1998 A
5840075 Mueller et al. Nov 1998 A
5842995 Mahadevan-Jansen et al. Dec 1998 A
5843000 Nishioka et al. Dec 1998 A
5843052 Benja-Athon Dec 1998 A
5847827 Fercher Dec 1998 A
5862273 Pelletier Jan 1999 A
5865754 Sevick-Muraca et al. Feb 1999 A
5867268 Gelikonov et al. Feb 1999 A
5871449 Brown Feb 1999 A
5872879 Hamm Feb 1999 A
5877856 Fercher Mar 1999 A
5887009 Mandella et al. Mar 1999 A
5892583 Li Apr 1999 A
5910839 Erskine et al. Jun 1999 A
5912764 Togino Jun 1999 A
5920373 Bille Jul 1999 A
5920390 Farahi et al. Jul 1999 A
5921926 Rolland et al. Jul 1999 A
5926592 Harris et al. Jul 1999 A
5949929 Hamm Sep 1999 A
5951482 Winston et al. Sep 1999 A
5955737 Hallidy et al. Sep 1999 A
5956355 Swanson et al. Sep 1999 A
5968064 Selmon et al. Oct 1999 A
5975697 Podoleanu et al. Nov 1999 A
5983125 Alfano et al. Nov 1999 A
5987346 Benaron et al. Nov 1999 A
5991697 Nelson et al. Nov 1999 A
5994690 Kulkarni et al. Nov 1999 A
5995223 Power Nov 1999 A
6002480 Izatt et al. Dec 1999 A
6004314 Wei et al. Dec 1999 A
6006128 Izatt et al. Dec 1999 A
6007996 McNamara et al. Dec 1999 A
6010449 Selmon et al. Jan 2000 A
6014214 Li Jan 2000 A
6016197 Krivoshlykov Jan 2000 A
6020963 DiMarzio et al. Feb 2000 A
6025956 Nagano et al. Feb 2000 A
6033721 Nassuphis Mar 2000 A
6037579 Chan et al. Mar 2000 A
6044288 Wake et al. Mar 2000 A
6045511 Ott et al. Apr 2000 A
6048742 Weyburne et al. Apr 2000 A
6053613 Wei et al. Apr 2000 A
6069698 Ozawa et al. May 2000 A
6078047 Mittleman et al. Jun 2000 A
6091496 Hill Jul 2000 A
6091984 Perelman et al. Jul 2000 A
6094274 Yokoi Jul 2000 A
6107048 Goldenring et al. Aug 2000 A
6111645 Tearney et al. Aug 2000 A
6117128 Gregory Sep 2000 A
6120516 Selmon et al. Sep 2000 A
6134003 Tearney et al. Oct 2000 A
6134010 Zavislan Oct 2000 A
6134033 Bergano et al. Oct 2000 A
6141577 Rolland et al. Oct 2000 A
6151522 Alfano et al. Nov 2000 A
6159445 Klaveness et al. Dec 2000 A
6160826 Swanson et al. Dec 2000 A
6161031 Hochmann et al. Dec 2000 A
6166373 Mao Dec 2000 A
6174291 McMahon et al. Jan 2001 B1
6175669 Colston et al. Jan 2001 B1
6185271 Kinsinger Feb 2001 B1
6191862 Swanson et al. Feb 2001 B1
6193676 Winston et al. Feb 2001 B1
6198956 Dunne Mar 2001 B1
6201989 Whitehead et al. Mar 2001 B1
6208415 De Boer et al. Mar 2001 B1
6208887 Clarke Mar 2001 B1
6245026 Campbell et al. Jun 2001 B1
6249349 Lauer Jun 2001 B1
6249381 Suganuma Jun 2001 B1
6249630 Stock et al. Jun 2001 B1
6263234 Engelhardt et al. Jul 2001 B1
6264610 Zhu Jul 2001 B1
6272376 Marcu et al. Aug 2001 B1
6274871 Dukor et al. Aug 2001 B1
6282011 Tearney et al. Aug 2001 B1
6297018 French et al. Oct 2001 B1
6301048 Cao et al. Oct 2001 B1
6308092 Hoyns Oct 2001 B1
6324419 Guzelsu et al. Nov 2001 B1
6341036 Tearney et al. Jan 2002 B1
6353693 Kano et al. Mar 2002 B1
6359692 Groot Mar 2002 B1
6374128 Toida et al. Apr 2002 B1
6377349 Fercher Apr 2002 B1
6384915 Everett et al. May 2002 B1
6393312 Hoyns May 2002 B1
6394964 Sievert, Jr. et al. May 2002 B1
6396941 Bacus et al. May 2002 B1
6421164 Tearney et al. Jul 2002 B2
6437867 Zeylikovich et al. Aug 2002 B2
6441892 Xiao et al. Aug 2002 B2
6441959 Yang et al. Aug 2002 B1
6445485 Frigo et al. Sep 2002 B1
6445939 Swanson et al. Sep 2002 B1
6445944 Ostrovsky Sep 2002 B1
6459487 Chen et al. Oct 2002 B1
6463313 Winston et al. Oct 2002 B1
6469846 Ebizuka et al. Oct 2002 B2
6475159 Casscells et al. Nov 2002 B1
6475210 Phelps et al. Nov 2002 B1
6477403 Eguchi et al. Nov 2002 B1
6485413 Boppart et al. Nov 2002 B1
6485482 Belef Nov 2002 B1
6501551 Tearney et al. Dec 2002 B1
6501878 Hughes et al. Dec 2002 B2
6516014 Sellin et al. Feb 2003 B1
6517532 Altshuler et al. Feb 2003 B1
6538817 Farmer et al. Mar 2003 B1
6540391 Lanzetta et al. Apr 2003 B2
6549801 Chen et al. Apr 2003 B1
6552796 Magnin et al. Apr 2003 B2
6556305 Aziz et al. Apr 2003 B1
6556853 Cabib et al. Apr 2003 B1
6558324 Von Behren et al. May 2003 B1
6564087 Pitris et al. May 2003 B1
6564089 Izatt et al. May 2003 B2
6567585 Harris May 2003 B2
6593101 Richards-Kortum et al. Jul 2003 B2
6611833 Johnson et al. Aug 2003 B1
6615071 Casscells, III et al. Sep 2003 B1
6622732 Constantz Sep 2003 B2
6654127 Everett et al. Nov 2003 B2
6657730 Pfau et al. Dec 2003 B2
6658278 Gruhl Dec 2003 B2
6680780 Fee Jan 2004 B1
6685885 Nolte et al. Feb 2004 B2
6687007 Meigs Feb 2004 B1
6687010 Horii et al. Feb 2004 B1
6687036 Riza Feb 2004 B2
6692430 Adler Feb 2004 B2
6701181 Tang et al. Mar 2004 B2
6721094 Sinclair et al. Apr 2004 B1
6738144 Dogariu et al. May 2004 B1
6741355 Drabarek May 2004 B2
6757467 Rogers Jun 2004 B1
6790175 Furusawa et al. Sep 2004 B1
6806963 Wälti et al. Oct 2004 B1
6816743 Moreno et al. Nov 2004 B2
6831781 Tearney et al. Dec 2004 B2
6839496 Mills et al. Jan 2005 B1
6882432 Deck Apr 2005 B2
6900899 Nevis May 2005 B2
6903820 Wang Jun 2005 B2
6909105 Heintzmann et al. Jun 2005 B1
6949072 Furnish et al. Sep 2005 B2
6961123 Wang et al. Nov 2005 B1
6980299 de Boer Dec 2005 B1
6996549 Zhang et al. Feb 2006 B2
7006231 Ostrovsky et al. Feb 2006 B2
7006232 Rollins et al. Feb 2006 B2
7019838 Izatt et al. Mar 2006 B2
7027633 Foran et al. Apr 2006 B2
7061622 Rollins et al. Jun 2006 B2
7072047 Westphal et al. Jul 2006 B2
7075658 Izatt et al. Jul 2006 B2
7099358 Chong Aug 2006 B1
7113288 Fercher Sep 2006 B2
7113625 Watson et al. Sep 2006 B2
7130320 Tobiason et al. Oct 2006 B2
7139598 Hull et al. Nov 2006 B2
7142835 Paulus Nov 2006 B2
7148970 de Boer Dec 2006 B2
7177027 Hirasawa et al. Feb 2007 B2
7190464 Alphonse Mar 2007 B2
7230708 Lapotko et al. Jun 2007 B2
7231243 Tearney et al. Jun 2007 B2
7236637 Sirohey et al. Jun 2007 B2
7242480 Alphonse Jul 2007 B2
7267494 Deng et al. Sep 2007 B2
7272252 De La Torre-Bueno et al. Sep 2007 B2
7304798 Izumi et al. Dec 2007 B2
7330270 O'Hara et al. Feb 2008 B2
7336366 Choma et al. Feb 2008 B2
7342659 Horn et al. Mar 2008 B2
7355716 de Boer et al. Apr 2008 B2
7355721 Quadling et al. Apr 2008 B2
7359062 Chen et al. Apr 2008 B2
7366376 Shishkov et al. Apr 2008 B2
7382809 Chong et al. Jun 2008 B2
7391520 Zhou et al. Jun 2008 B2
7458683 Chernyak et al. Dec 2008 B2
7530948 Seibel et al. May 2009 B2
7539530 Caplan et al. May 2009 B2
7609391 Betzig Oct 2009 B2
7630083 de Boer et al. Dec 2009 B2
7643152 de Boer et al. Jan 2010 B2
7643153 de Boer et al. Jan 2010 B2
7646905 Guittet et al. Jan 2010 B2
7649160 Colomb et al. Jan 2010 B2
7664300 Lange et al. Feb 2010 B2
7733497 Yun et al. Jun 2010 B2
7782464 Mujat et al. Aug 2010 B2
7805034 Kato et al. Sep 2010 B2
20010036002 Tearney et al. Nov 2001 A1
20010047137 Moreno et al. Nov 2001 A1
20020016533 Marchitto et al. Feb 2002 A1
20020024015 Hoffmann et al. Feb 2002 A1
20020048025 Takaoka Apr 2002 A1
20020048026 Isshiki et al. Apr 2002 A1
20020052547 Toida May 2002 A1
20020057431 Fateley et al. May 2002 A1
20020064341 Fauver et al. May 2002 A1
20020076152 Hughes et al. Jun 2002 A1
20020085209 Mittleman et al. Jul 2002 A1
20020086347 Johnson et al. Jul 2002 A1
20020091322 Chaiken et al. Jul 2002 A1
20020093662 Chen et al. Jul 2002 A1
20020109851 Deck Aug 2002 A1
20020122182 Everett et al. Sep 2002 A1
20020122246 Tearney et al. Sep 2002 A1
20020140942 Fee et al. Oct 2002 A1
20020158211 Gillispie Oct 2002 A1
20020161357 Anderson et al. Oct 2002 A1
20020163622 Magnin et al. Nov 2002 A1
20020168158 Furusawa et al. Nov 2002 A1
20020172485 Keaton et al. Nov 2002 A1
20020183623 Tang et al. Dec 2002 A1
20020188204 McNamara et al. Dec 2002 A1
20020196446 Roth et al. Dec 2002 A1
20020198457 Tearney et al. Dec 2002 A1
20030001071 Mandella et al. Jan 2003 A1
20030013973 Georgakoudi et al. Jan 2003 A1
20030023153 Izatt et al. Jan 2003 A1
20030026735 Nolte et al. Feb 2003 A1
20030028114 Casscells, III et al. Feb 2003 A1
20030030816 Eom et al. Feb 2003 A1
20030043381 Fercher Mar 2003 A1
20030053673 Dewaele et al. Mar 2003 A1
20030067607 Wolleschensky et al. Apr 2003 A1
20030082105 Fischman et al. May 2003 A1
20030097048 Ryan et al. May 2003 A1
20030108911 Klimant et al. Jun 2003 A1
20030120137 Pawluczyk et al. Jun 2003 A1
20030135101 Webler Jul 2003 A1
20030137669 Rollins et al. Jul 2003 A1
20030164952 Deichmann et al. Sep 2003 A1
20030165263 Hamer et al. Sep 2003 A1
20030171691 Casscells, III et al. Sep 2003 A1
20030174339 Feldchtein et al. Sep 2003 A1
20030199769 Podoleanu et al. Oct 2003 A1
20030216719 Debenedictis et al. Nov 2003 A1
20030220749 Chen et al. Nov 2003 A1
20030236443 Cespedes et al. Dec 2003 A1
20040002650 Mandrusov et al. Jan 2004 A1
20040039298 Abreu Feb 2004 A1
20040054268 Esenaliev et al. Mar 2004 A1
20040072200 Rigler et al. Apr 2004 A1
20040075841 Van Neste et al. Apr 2004 A1
20040076940 Alexander et al. Apr 2004 A1
20040077949 Blofgett et al. Apr 2004 A1
20040085540 Lapotko et al. May 2004 A1
20040086245 Farroni et al. May 2004 A1
20040100631 Bashkansky et al. May 2004 A1
20040100681 Bjarklev et al. May 2004 A1
20040110206 Wong et al. Jun 2004 A1
20040126048 Dave et al. Jul 2004 A1
20040126120 Cohen et al. Jul 2004 A1
20040133191 Momiuchi et al. Jul 2004 A1
20040150829 Koch et al. Aug 2004 A1
20040150830 Chan Aug 2004 A1
20040152989 Puttappa et al. Aug 2004 A1
20040165184 Mizuno Aug 2004 A1
20040166593 Nolte et al. Aug 2004 A1
20040189999 De Groot et al. Sep 2004 A1
20040212808 Okawa et al. Oct 2004 A1
20040239938 Izatt Dec 2004 A1
20040246490 Wang Dec 2004 A1
20040246583 Mueller et al. Dec 2004 A1
20040254474 Seibel et al. Dec 2004 A1
20040263843 Knopp et al. Dec 2004 A1
20050018133 Huang et al. Jan 2005 A1
20050018201 de Boer et al. Jan 2005 A1
20050035295 Bouma et al. Feb 2005 A1
20050036150 Izatt et al. Feb 2005 A1
20050046837 Izumi et al. Mar 2005 A1
20050057680 Agan Mar 2005 A1
20050057756 Fang-Yen et al. Mar 2005 A1
20050059894 Zeng et al. Mar 2005 A1
20050065421 Burckhardt et al. Mar 2005 A1
20050075547 Wang Apr 2005 A1
20050083534 Riza et al. Apr 2005 A1
20050119567 Choi et al. Jun 2005 A1
20050128488 Yelin et al. Jun 2005 A1
20050165303 Kleen et al. Jul 2005 A1
20050171438 Chen et al. Aug 2005 A1
20050190372 Dogariu et al. Sep 2005 A1
20050254061 Alphonse et al. Nov 2005 A1
20060033923 Hirasawa et al. Feb 2006 A1
20060093276 Bouma et al. May 2006 A1
20060103850 Alphonse et al. May 2006 A1
20060146339 Fujita et al. Jul 2006 A1
20060155193 Leonardi et al. Jul 2006 A1
20060164639 Horn et al. Jul 2006 A1
20060171503 O'Hara et al. Aug 2006 A1
20060184048 Saadat et al. Aug 2006 A1
20060193352 Chong et al. Aug 2006 A1
20060244973 Yun et al. Nov 2006 A1
20070019208 Toida et al. Jan 2007 A1
20070038040 Cense et al. Feb 2007 A1
20070070496 Gweon et al. Mar 2007 A1
20070076217 Baker et al. Apr 2007 A1
20070086013 De Lega et al. Apr 2007 A1
20070086017 Buckland et al. Apr 2007 A1
20070091317 Freischlad et al. Apr 2007 A1
20070133002 Wax et al. Jun 2007 A1
20070188855 Shishkov et al. Aug 2007 A1
20070223006 Tearney et al. Sep 2007 A1
20070236700 Yun et al. Oct 2007 A1
20070258094 Izatt et al. Nov 2007 A1
20070291277 Everett et al. Dec 2007 A1
20080002197 Sun et al. Jan 2008 A1
20080007734 Park et al. Jan 2008 A1
20080049220 Izzia et al. Feb 2008 A1
20080094613 de Boer et al. Apr 2008 A1
20080094637 de Boer et al. Apr 2008 A1
20080097225 Tearney et al. Apr 2008 A1
20080097709 de Boer et al. Apr 2008 A1
20080100837 de Boer et al. May 2008 A1
20080152353 de Boer et al. Jun 2008 A1
20080154090 Hashimshony Jun 2008 A1
20080204762 Izatt et al. Aug 2008 A1
20080265130 Colomb et al. Oct 2008 A1
20080308730 Vizi et al. Dec 2008 A1
20090011948 Uniu et al. Jan 2009 A1
20090196477 Cense et al. Aug 2009 A1
20090273777 Yun et al. Nov 2009 A1
20090290156 Popescu et al. Nov 2009 A1
20100086251 Xu et al. Apr 2010 A1
20100094576 de Boer et al. Apr 2010 A1
20100150467 Zhao et al. Jun 2010 A1
Foreign Referenced Citations (85)
Number Date Country
1550203 Dec 2004 CN
4105221 Sep 1991 DE
4309056 Sep 1994 DE
19542955 May 1997 DE
10351319 Jun 2005 DE
0110201 Jun 1984 EP
0251062 Jan 1988 EP
0617286 Feb 1994 EP
0590268 Apr 1994 EP
0728440 Aug 1996 EP
0933096 Aug 1999 EP
1324051 Jul 2003 EP
1426799 Jun 2004 EP
2738343 Aug 1995 FR
1257778 Dec 1971 GB
2030313 Apr 1980 GB
2209221 May 1989 GB
2298054 Aug 1996 GB
6073405 Apr 1985 JP
20040056907 Feb 1992 JP
4135550 May 1992 JP
4135551 May 1992 JP
5509417 Nov 1993 JP
02214127 Jul 2002 JP
20030035659 Feb 2003 JP
2007271761 Oct 2007 JP
7900841 Oct 1979 WO
9201966 Feb 1992 WO
9216865 Oct 1992 WO
9219930 Nov 1992 WO
9303672 Mar 1993 WO
9216865 Oct 1993 WO
9533971 Dec 1995 WO
9628212 Sep 1996 WO
9732182 Sep 1997 WO
9800057 Jan 1998 WO
9801074 Jan 1998 WO
9814132 Apr 1998 WO
9835203 Aug 1998 WO
9838907 Sep 1998 WO
9846123 Oct 1998 WO
9848838 Nov 1998 WO
9848846 Nov 1998 WO
9905487 Feb 1999 WO
9944089 Feb 1999 WO
9944089 Sep 1999 WO
9957507 Nov 1999 WO
0058766 Oct 2000 WO
0101111 Jan 2001 WO
0108579 Feb 2001 WO
0127679 Apr 2001 WO
0138820 May 2001 WO
0142735 Jun 2001 WO
0236015 May 2002 WO
0237075 May 2002 WO
0238040 May 2002 WO
0254027 Jul 2002 WO
02053050 Jul 2002 WO
2002084263 Oct 2002 WO
03020119 Mar 2003 WO
03046495 Jun 2003 WO
03046636 Jun 2003 WO
03052478 Jun 2003 WO
03062802 Jul 2003 WO
03105678 Dec 2003 WO
2004034869 Apr 2004 WO
2004057266 Jul 2004 WO
2004066824 Aug 2004 WO
2004088361 Oct 2004 WO
2004105598 Dec 2004 WO
2005000115 Jan 2005 WO
2005047813 May 2005 WO
2005054780 Jun 2005 WO
2005082225 Sep 2005 WO
2006004743 Jan 2006 WO
2006014392 Feb 2006 WO
2006038876 Apr 2006 WO
2006039091 Apr 2006 WO
2006059109 Jun 2006 WO
2006124860 Nov 2006 WO
2006130797 Dec 2006 WO
2007028531 Mar 2007 WO
2007038787 Apr 2007 WO
2007083138 Jul 2007 WO
2007084995 Jul 2007 WO
Related Publications (1)
Number Date Country
20090196477 A1 Aug 2009 US
Provisional Applications (1)
Number Date Country
60575773 May 2004 US