The present invention generally relates to external medical apparatus. Specifically, the present invention relates to retinal imaging.
Diseases involving retinal vasculature are one of the leading causes of blindness worldwide. Many of these diseases are both progressive and treatable. Thus, their early detection is highly desirable. Diagnoses are often made on the basis of the many structural changes which may occur in the retina as a consequence of problems with retinal blood flow. These problems include neovascularization (the growth of new blood vessels in an attempt to compensate for a reduction in flow through pre-existing vessels), “cotton-wool” patches (regions in which nerve fiber axoplasmic transport has failed), the degeneration of cells of the eye, (e.g., retinal nerve fibers), and/or other metabolic or biochemical abnormalities. Once observed, these and other phenomena may be used to diagnose retinal vascular disease, and treatment may be initiated to inhibit further degeneration. It is desirable to detect such problems early, if possible, before irreversible damage has occurred.
US Patent Application Publication 2006/0147897 to Grinvald et al., which is incorporated herein by reference, describes a method and system for detecting abnormalities in the properties of the walls of a subject's blood vessels by observing the characteristics of blood flow in vessels which are optically accessible, such as the retinal vasculature. A time sequenced series of images is taken, and the images are processed to eliminate the background and render erythrocyte motion visible. Information about the state of the inner wall of the blood vessel which has been imaged is obtained from the characteristics of this blood flow. This information is described as being capable of being extrapolated to provide information about the state of the blood vessels elsewhere in the subject. In addition, a system and method are described for detecting arteriosclerotic plaque on the walls of blood vessels by labeling the plaque with a molecular label having desired optical or radioactive properties, and directly imaging the plaque either in an optically accessible blood vessel, or by imaging radioactive label in the plaque in a blood vessel anywhere in the body.
US Patent Application Publication 2006/0131284 to Grinvald et al., which is incorporated herein by reference, describes a method and system for determining and mapping the quantity of chromophores having a distinct spectrum attached to moving objects in a spectrally rich environment that may include multiple chromophores attached to stationary objects. An area of interest is imaged at different times and different wavelengths, and the spectral properties of the one or more chromophores attached to the moving objects are separated from the stationary spectral properties of the background, followed by spectral analysis of the moving objects to determine their quantity. Application to the retinal vasculature is illustrated, showing the imaging, analyzing and quantifying the oxygen saturation of retinal blood, resolved for the different vascular compartments: capillaries, arterioles, venules, arteries, and veins. Changes in the structure of the vascular environment are also determined (whether growth of new vessels or the elimination of existing ones), by the generation of path maps based on analysis of differential images taken at a single wavelength of the moving components in the blood flow.
U.S. Pat. No. 6,588,901 to Grinvald et al., which is incorporated herein by reference, describes a system for directly imaging and analyzing the movement of individual erythrocytes in blood vessels, the system comprising imaging means for acquiring, within a predetermined time interval from each other, at least one pair of images of at least one same erythrocyte for producing at least two frames, each image representing an analog or digital image of the location of the erythrocyte in each of the frames at a predetermined time; image acquisition means for collecting and storing analog or digital images in machine-readable form, and a computer for controlling the operation of the imaging means and the image acquisition means, for processing the at least two frames, and for analyzing the movement of the erythrocyte in the blood vessels. A method for directly imaging and analyzing the movement of individual erythrocytes in blood vessels is also provided.
U.S. Pat. No. 6,478,424 to Grinvald et al., which is incorporated herein by reference, describes a system for imaging reflectance changes, intrinsic or extrinsic fluorescence changes of a retina due to retinal function, including an imaging illuminator for illumination of the retina; a retina-stimulating illuminator for inducing a functional response; an imaging device receiving light from the retina via retinal imaging optics; image acquisition means for digitizing and storing images received from the imaging device, and a computer for controlling the operation of the system and for processing the stored images to reveal a differential functional signal corresponding to the retina's function.
The following patents and patent applications may be of interest:
PCT Publication WO 06/121,984 to Nelson et al.
U.S. Pat. No. 5,463,426 to Grinvald
PCT Publication Wo 04/049899 to Grinvald et al.
PCT Publication WO 03/083434 to Grinvald et al.
PCT Publication WO 99/063882 to Grinvald et al.
PCT Publication Wo 00/006015 to Grinvald et al.
U.S. Pat. No. 6,351,663 to Flower et al.
U.S. Pat. No. 5,279,298 to Flower
US Patent Application Publication 2002/0016533 to Marchitto et al.
U.S. Pat. No. 5,784,162 to Cabib et al.
U.S. Pat. No. 5,983,120 to Groner et al.
U.S. Pat. No. 6,902,935 to Kaufman
US Patent Application Publication 2002/0111545 to Lindberg
U.S. Pat. No. 6,104,939 to Groner
The following articles may be of interest:
“Retinal microvascular abnormalities and incident stroke: the Atherosclerosis Risk in Communities Study,” by Wong, Lancet 2001: 358:1134-40
“Automatic Extraction and Measurement of Leukocyte Motion in Microvessels Using Spatiotemporal Image Analysis,” by Sato, IEEE Transactions on Biomedical Engineering, Vol. 44, No. 4, April 1997
“Visual stimulus induced changes in human near-infrared fundus reflectance,” by Abramoff, Invest Opthalmol Vis Sci. 2006 February; 47(2): 715-721
In some embodiments of the present invention, a clinical state of a portion of a subject's body, for example, the subject's retina, is diagnosed in accordance with the following method:
(1) A first pulse of electromagnetic energy (e.g., a first flash of light) is directed toward the portion.
(2) While the portion is in a functional state due to the first pulse of electromagnetic energy, a second pulse of electromagnetic energy (e.g., a second flash of light) is directed toward the portion. At a “second-pulse” time (e.g., a “second-flash” time), light of the second pulse of electromagnetic energy that is reflected from the portion is detected.
(3) While the portion is in a functional state due to the first and second pulses of electromagnetic energy, a third pulse of electromagnetic energy (e.g., a third flash of light) is directed toward the portion. At a “third-pulse” time (e.g., a “third-flash” time), light of the third pulse of electromagnetic energy that is reflected from the portion is detected.
(4) In response to detecting the pulses of electromagnetic energy (e.g., the flashes of light) at the second-pulse time and the third-pulse time, a characteristic of the portion resulting from (a) the functional state due to the first pulse of electromagnetic energy and (b) the functional state due to the first and second pulses of electromagnetic energy is quantified. The clinical state of the subject is identified in response to quantifying the characteristic of the portion.
In some embodiments, second-flash and third-flash images are produced respectively from light detected at second-flash and third-flash times. For some applications, the relationship between a characteristic of the second image and a characteristic of the third image is quantified. Alternatively or additionally, a differential image is formed from the second-flash and the third-flash images, and a characteristic of the differential image is quantified. Typically, a region of interest (ROI) is identified in at least one of the images, and a characteristic of the ROI is quantified.
For some applications, at a “first-flash” time, light of the first flash of light that is reflected from the retina while the retina is in a state that is not a functional state due to a flash of light is detected. A characteristic of the retina is quantified, the characteristic resulting from (a) the functional state due to the first flash of light, (b) the functional state due to the first and second flashes of light, and (c) the state that is not an functional state due to a flash of light. In some embodiments, a first-flash image of the retina is obtained using the light of the first flash of light that is reflected from the retina. For some applications, a differential image is formed using the first-flash image and at least one of the second-flash and third-flash images.
In some embodiments of the invention, a portion of a subject's body is stimulated by one or more stimuli. A response of the portion of the body is detected by detecting, at first and second times, respective parameters of light that are reflected from the portion of the body. Respective images of the portion of the body are formed by processing respective parameters of light that are detected. A region of interest (ROI) of the portion of the body is identified based on data from at least one of the images. A measurable parameter of the ROI in response to at least one of the one or more stimuli is quantified. A clinical state of the subject is identified in response to the quantification.
In some embodiments, the one or more stimuli are applied to the subject's retina. Typically, the retina is stimulated with at least first and second flashes of light, the flashes of light being separated in time by between 300 ms and 10 sec, between 10 ms and 300 ms, or between 1 ms and 20 ms, for example, between 1 ms and 10 ms. In some embodiments, each of the flashes of light has a duration of less than 1 ms, or a duration of 0.1 ms to 10 ms. Alternatively, the first flash of light has a duration of less than 1 ms and the second flash of light has a duration of less than 4 ms. In some embodiments, the retina is stimulated with at least three flashes of light, the lengths of the flashes and the intervals between the flashes being as described hereinabove with respect to the first and second flashes. Light of respective flashes of light that is reflected from the retina is detected using a detector, for example, (a) a fundus camera, (b) an opthalmoscope, or (c) a CCD that is not part of a fundus camera. (Typically, a fundus camera includes a CCD, or another type of detector.) Images of the retina are formed by processing the detected light.
In some embodiments, the retina is stimulated by modulating the temperature of the retina, and/or by modulating the oxygen saturation of the retina. For example, the oxygen saturation of the retina may be modulated by providing, for inhalation by the subject, oxygen, or a mixture of oxygen and another gas, e.g., nitrogen. In such embodiments, in addition to the retina being stimulated, the retina is illuminated. Light of the illuminating light that is reflected from the retina is detected using a fundus camera, or a CCD that is not part of a fundus camera, to form images of the retina. Alternatively or additionally, ambient light that is reflected from the retina is detected to form images of the retina.
In some embodiments, the ROI is the subject's optic nerve head, which is commonly referred to (and which is referred to in this application) as the optic disk. In an embodiment, darkening (or brightening) of the images of the optic disc in response to the one or more stimuli is quantified. In some embodiments, darkening of the optic disk relative to its surroundings is quantified. The quantification of the darkening is used to identify clinical states of the subject's eye, brain, cardiovascular system, nervous system, or entire body. For some applications, darkening of images of the optic disk, or another ROI, is quantified to identify that the subject is suffering from glaucoma, age related macular degeneration, and/or diabetic retinopathy.
There is therefore provided, in accordance with an embodiment of the present invention, a method for diagnosing a clinical state of a portion of a body of a subject, including:
(1) directing toward the portion a first pulse of electromagnetic energy;
(2) while the portion is in a functional state due to the first pulse of electromagnetic energy, directing toward the portion a second pulse of electromagnetic energy;
(3) detecting, at a second-pulse time, energy of the second pulse of electromagnetic energy that is reflected from the portion;
(4) while the portion is in a functional state due to the first and second pulses of electromagnetic energy, directing toward the portion a third pulse of electromagnetic energy;
(5) detecting, at a third-pulse time, energy of the third pulse of electromagnetic energy that is reflected from the portion;
(6) in response to detecting the pulse of electromagnetic energy at the second-pulse time and the third-pulse time, quantifying a characteristic of the portion resulting from (a) the functional state due to the first pulse of electromagnetic energy and (b) the functional state due to the first and second pulses of electromagnetic energy; and
(7) identifying a clinical state of the subject in response to quantifying the characteristic of the portion.
In an embodiment:
the first pulse of electromagnetic energy includes a first flash of light, and step (1) includes directing the first flash of light toward the portion;
the second pulse of electromagnetic energy includes a second flash of light, and step (2) includes, while the portion is in a functional state due to the first flash of light, directing the second flash of light toward the portion;
the second-pulse time includes a second-flash time, and step (3) includes detecting, at the second-flash time, light of the second flash of light that is reflected from the portion;
the third pulse of electromagnetic energy includes a third flash of light, and step (4) includes, while the portion is in a functional state due to the first and second flashes of light, directing toward the portion the third flash of light;
the third-pulse time includes a third-flash time, and step (5) includes detecting, at a third-flash time, light of the third flash of light that is reflected from the portion; and
step (6) includes, in response to detecting the flashes of light at the second-flash time and the third-flash time, quantifying a characteristic of the portion resulting from (a) the functional state due to the first flash of light and (b) the functional state due to the first and second flashes of light.
In an embodiment, the portion of the body includes a retina of the subject and directing the first, second, and third flashes of light toward the portion of the body includes directing the first, second, and third flashes of light toward the subject's retina.
In an embodiment, identifying the clinical state of the subject includes identifying a clinical state of a part of a body of the subject selected from the group consisting of: a brain, a cardiovascular system, and a nervous system.
In an embodiment, detecting the light of the second and third flashes of light reflected from the retina includes detecting the light of the second and third flashes of light reflected from the retina using a detecting device selected from the group consisting of: a fundus camera, a CCD, and an opthalmoscope.
In an embodiment, the method further includes detecting, at a first-flash time, light of the first flash of light that is reflected from the retina while the retina is in a state that is not a functional state due to a flash of light, and quantifying the characteristic of the retina includes quantifying a characteristic of the retina resulting from (a) the functional state due to the first flash of light, (b) the functional state due to the first and second flashes of light, and (c) the state that is not a functional state due to a flash of light.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, each of the flashes of light having a duration of less than 1 ms.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, each of the flashes of light having a duration of 0.1 ms to 10 ms.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, a time interval between directing the first and the second flashes being equal to a time interval between directing the second and third flashes.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, a time interval between directing the first and the second flashes being between 10 ms and 300 ms.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, a time interval between directing the first and the second flashes being between 300 ms and 10 s.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, a time interval between directing the first and the second flashes being between 1 ms and 20 ms.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, a time interval between directing the first and the second flashes being between 1 ms and 10 ms.
In an embodiment, directing the first, second, and third flashes of light toward the retina includes directing the first, second, and third flashes of light toward the retina, a time interval between directing the first and the second flashes being different from a time interval between directing the second and third flashes.
In an embodiment, the method further includes directing one or more additional flashes of light toward the retina between directing the second and third flashes of light toward the retina.
In an embodiment, identifying the clinical state of the subject includes identifying a clinical state of an eye of the subject.
In an embodiment, identifying the clinical state of the subject includes identifying glaucoma of a retina of the subject.
In an embodiment, identifying the clinical state of the subject includes identifying a clinical state of a retina of the subject selected from the group consisting of: age related macular degeneration, and diabetic retinopathy.
In an embodiment, step (3) includes obtaining a second-flash image of the retina based on the light of the second flash of light that is reflected from the retina, and step (5) includes obtaining a third-flash image of the retina based on the light of the third flash of light that is reflected from the retina.
In an embodiment,
step (6) includes:
In an embodiment, processing the second-flash and third-flash images of the retina to obtain a first differential image of the retina includes dividing the third-flash image by the second-flash image.
In an embodiment, processing the second-flash and third-flash images of the retina to obtain a first differential image of the retina includes subtracting the second-flash image from the third-flash image.
In an embodiment, the method further includes:
at a first-flash time, detecting light of the first flash of light that is reflected from the retina;
obtaining a first-flash image of the retina based on the light of the first flash of light that is reflected from the retina; and
obtaining a second differential image of the retina by processing the first-flash image of the retina and an image selected from the group consisting of: the second-flash and third-flash images of the retina.
In an embodiment, the method further includes:
identifying a region of interest (ROI) of the second differential image; and
computing a value that is representative of a darkness of the ROI of the second differential image.
In an embodiment,
In an embodiment, step (e) includes computing a ratio relating i-in to i-out in the second-flash image, and step (j) includes computing a ratio relating i-in to i-out in the third-flash image.
In an embodiment, step (e) includes computing an arithmetic difference between i-in and i-out in the second-flash image, and step (j) includes computing an arithmetic difference between i-in and i-out in the third-flash image.
In an embodiment, step (c) includes computing a mean intensity of the ROI in the second-flash image, step (d) includes computing a mean intensity of the OROI in the second-flash image, step (h) includes computing a mean intensity of the ROI in the third-flash image, and step (i) includes computing a mean intensity of the OROI in the third-flash image.
In an embodiment, step (c) includes computing a median intensity of the ROI in the second-flash image, step (d) includes computing a median intensity of the OROI in the second-flash image, step (h) includes computing a median intensity of the ROI in the third-flash image, and step (i) includes computing a median intensity of the OROI in the third-flash image.
In an embodiment, identifying the ROI in the second-flash image includes identifying a portion of the image that corresponds to an optic disk of the subject, and identifying the ROI in the third-flash image includes identifying a portion of the image that corresponds to an optic disk of the subject.
In an embodiment, identifying the OROI in the second-flash image includes identifying a strip that surrounds the ROI of the second-flash image, the strip having an area that is within 50% of an area of the ROI of the second-flash image, and identifying the OROI in the third-flash image includes identifying a strip that surrounds the ROI of the third-flash image, the strip having an area that is within 50% of an area of the ROI of the third-flash image.
In an embodiment, identifying the OROI in the second-flash image includes identifying an outer ring that is separated from the ROI of the second-flash image by an inner ring, and identifying the OROI in the third-flash image includes identifying an outer ring that is separated from the ROI of the third-flash image by an inner ring.
In an embodiment,
step (6) includes:
In an embodiment, quantifying the relationship between r-in and r-out includes computing a ratio relating r-in to r-out.
In an embodiment, quantifying the relationship between r-in and r-out includes computing an arithmetic difference between r-in and r-out.
In an embodiment, computing r-in includes computing a mean darkness of the ROI, and computing r-out includes computing a mean darkness of the OROI.
In an embodiment, computing r-in includes computing a median darkness of the ROI, and computing r-out includes computing a median darkness of the OROI.
In an embodiment, step (b) includes identifying a portion of the selected image that corresponds to an optic disk of the subject.
In an embodiment, identifying the OROI includes identifying a strip that surrounds the ROI, the strip having an area that is within 50% of an area of the ROI.
In an embodiment, identifying the OROI includes identifying an outer ring that is separated from the ROI by an inner ring.
In an embodiment, quantifying a relationship between r-in and r-out includes:
computing a ratio relating r-in to r-out; and
subtracting the ratio from 1 to produce a result.
In an embodiment, identifying the clinical state of the subject includes identifying that the subject's retina is glaucomatous in response to the relationship between r-in and r-out passing a threshold.
In an embodiment, the method further includes identifying a severity of the glaucoma of the retina in response to the relationship between r-in and r-out.
The present invention will be more fully understood from the following detailed description of embodiments thereof, taken together with the drawings, in which:
Reference is now made to
It is currently hypothesized by the inventors that the difference between the second image and the first image may be at least in part caused by the first flash of light, which was used to generate the first retinal image, also acting as a stimulus. The stimulus may cause a change to the functional state of the optic disk, which, in turn, causes a change in the appearance of the portion corresponding to the optic disk in the second retinal image. In the first retinal image, the eye was not stimulated, since the flash was too brief to evoke an activation before the first image was captured. Thus, the differential image may show the difference between the retina in activated and in inactivated states. Alternatively, differences between the first and second retinal images may have been for a different reason.
Reference is now made to
A first image 26 of series 24 is a retinal image obtained from light of the first flash of light that was reflected by the retina. A second image 28 of series 24 is a differential retinal image that was formed by dividing a second retinal image by first image 26, the second retinal image having been obtained from light of the second flash of light that was reflected by the retina. A third image 29 of series 24 is a differential retinal image that was formed by dividing a third retinal image by first image 26, and so on. The inventors observed that there is a darkening of a portion of the differential retinal images. The darkened portion corresponds to the subject's optic disk.
It should be noted that although in series 24 there is a progressive darkening of the optic disk portion of the images, differential images may exhibit other effects. For example, there may be a lightening of the optic disc portion, the darkening may not progress in progressive differential images of a series, and/or there may be a change to a different portion of the images. In addition, other differential images could be formed and may exhibit similar effects. For example, a differential image may be formed by dividing a sixth retinal image in a series, by a third retinal image in the series, or by dividing a second retinal image in a series by a fourth retinal image in the series, and so on. Dividing a later image in the series by an earlier image in the series may show the difference between the retina (a) in an activated state due to having been stimulated m number of times and (b) in an activated state due to having been stimulated n number of times (where n≠m). For example, dividing the third image in the series by the second image may show the difference between the retina being in a functional state (a) due to the first and the second flashes of light and (b) due to the first flash of light.
Furthermore, differential images could be formed using retinal images that are separated from each other by intervals of time that are greater or less than 17.5 ms. In some embodiments, a second retinal image is separated in time from a first retinal image such that a functional change in the retina, caused by the flash used to generate the first image, is present, or maximized, when the second retinal image is generated.
In general, the specific parameters that are described with reference to the retinal images of
Reference is now made to
Reference is now made to
In some embodiments of the invention, a subject's retina is stimulated by one or more stimuli, for example, by more than three stimuli. A response of the retina is detected by detecting light that is reflected from the retina. Respective images of the retina are obtained by detecting respective flashes of light that are reflected from the retina. For example, a first flash of light is directed toward the retina. While the retina is in a functional state due to the first flash, a second flash is directed toward the retina, and at a second-pulse time, light of the second flash that is reflected from the retina is detected. A second-flash image is obtained from the light detected at the second-flash time. While the portion is in a functional state due to the first and second flashes, a third flash is directed toward the retina and, at a third-pulse time, light of the third flash that is reflected from the portion is detected. A third-flash image is obtained from the light detected at the third-flash time.
In some embodiments, at a first-flash time, light of the first flash that is reflected from the retina is detected. For some applications, a first-flash image is obtained from the light detected at the first-flash time. In some embodiments, a differential image is formed by dividing one of the retinal images by a different retinal image, and/or by subtracting one of the retinal images from a different retinal image. For example, a differential image is formed by subtracting the second-flash image from the third-flash image, or by dividing the third-flash image by the second flash image. Alternatively or additionally, a differential image is formed by subtracting the first-flash image from the second-flash image, or by dividing the second-flash image by the first flash image
In some embodiments, the time interval between the second and third flashes being directed toward the retina is equal to the time interval between the first and second flashes being directed toward the retina. Alternatively, the time intervals between the flashes are not equal. In some embodiments, the “second” flash and the “third” flash are not consecutive flashes. For example, one or more other flashes directed toward the retina may intervene between the second and third flashes.
ROI 42 of the retina is identified based on data from at least one of the images (i.e., one of the raw retinal images, and/or one of the differential retinal images). A measurable parameter of the ROI in response to at least one of the one or more stimuli is quantified. A clinical state of the subject is identified in response to the quantification.
Typically, ROI 42 is the portion of a retinal image corresponding to the subject's optic disk. In some embodiments, OROI 44 is a strip that surrounds, and is adjacent to, ROI 42. Alternatively, OROI 44 is a different shape, is not adjacent to the ROI, and/or is in a different portion of the retinal image. For example, the OROI may be a ring that is separated from the ROI by an inner ring. In some embodiments, the area of the OROI corresponds to the area of the ROI; for example, the OROI may have an area that is approximately equal to the area of the ROI. In an embodiment, the ratio of the area of the OROI to that of the ROI is between 4:4 and 6:4, e.g., 5:4. Typically, the area of the OROI is within 50% of the area of the ROI (i.e., A(ROI)*0.5<A(OROI)<A(ROI)*1.5).
In some embodiments, ROI 42 is defined by analyzing a differential retinal image. Alternatively, ROI 42 is defined by defining, in a raw image of a subject's retina, an ROI that corresponds to the optic disk. The ROI of differential images of the subject's retina is defined by defining the region of the differential image that corresponds to the ROI of the raw image.
In some embodiments, a value that is representative of the darkness of ROI 42 of differential image 40 is computed, and a clinical state of the subject is identified in response to the computed darkness.
For some applications, an ROI and an OROI are defined in each of a first and a second raw retinal image, the first and the second images being separated in time. For example, the first and second raw retinal images may be the second and third images (and/or the first and second images) in a series of retinal images, acquired in accordance with techniques described herein. For each of the first and the second images, a value (i-in) that is representative of the intensity of the ROI of the image is computed. For example, i-in may be the mean intensity or the median intensity of the ROI. A second value (i-out) that is representative of the intensity of the OROI of the image is computed. For example, i-out may be the mean intensity or the median intensity of the OROI. The extent to which i-in and i-out in are distinct in each of the images is computed. For example, a ratio relating i-in to i-out, or an arithmetic difference between i-in and i-out, may be computed for each of the first and the second images. A relationship between the ratio relating i-in to i-out in the second image and the ratio relating i-in to i-out in the first image is quantified. Alternatively or additionally, a relationship between the arithmetic difference between i-in and i-out in the second image and the arithmetic difference between i-in and i-out in the first image is quantified. A clinical state of the subject is identified in response to the quantification.
In some embodiments, a value (r-in) that is representative of the darkness of ROI 42 of at least one differential image 40 is computed, the differential image having been formed in accordance with the techniques described herein. Similarly, a value (r-out) that is representative of the darkness of OROI 44 is computed. Typically, r-in and r-out are computed by finding the mean (or the median) intensity of ROI 42 and of OROI 44 respectively.
For some applications, a ratio relating r-in to r-out is calculated in order to provide an indication of the relative darkness of ROI 42 compared to OROI 44. Alternatively or additionally, an arithmetic difference between r-in and r-out is computed. Further alternatively or additionally, a disk darkness index (DDI) of the subject is calculated, in accordance with the following equation:
DDI=(1−c),
where c=r-in/r-out.
Reference is now made to
In some embodiments, quantification of the darkening of an ROI of differential retinal images is used to identify clinical states of the subject's eye, brain, cardiovascular system, nervous system, or entire body. For example, for any given clinical state of interest (e.g., extent of atherosclerosis, epilepsy, multiple sclerosis, Parkinson disease), a healthy population DDI (or related index) is calculated based on a modest number of healthy subjects (e.g., 100 subjects), and a corresponding DDI is determined for patients diagnosed with the given clinical state. Individual patients suspected of or suffering from the given clinical state are compared to the DDI's for the two populations, to determine to which group they belong and/or to determine the extent of the pathology. For some applications, darkening of images of the optic disk is quantified to identify that the subject is suffering from glaucoma, age related macular degeneration, and/or diabetic retinopathy.
It will be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope of the present invention includes both combinations and subcombinations of the various features described hereinabove, as well as variations and modifications thereof that are not in the prior art, which would occur to persons skilled in the art upon reading the foregoing description.
The present application claims the benefit of U.S. Provisional Patent Application 61/009,055 to Grinvald et al., filed Dec. 20, 2007, which is incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4336459 | Fay | Jun 1982 | A |
4621643 | New, Jr. et al. | Nov 1986 | A |
4718417 | Kittrell et al. | Jan 1988 | A |
4998533 | Winkelman | Mar 1991 | A |
5217456 | Narciso, Jr. | Jun 1993 | A |
5240006 | Fujii et al. | Aug 1993 | A |
5279298 | Flower | Jan 1994 | A |
5297554 | Glynn et al. | Mar 1994 | A |
5348003 | Caro | Sep 1994 | A |
5400091 | Okazaki | Mar 1995 | A |
5463426 | Grinvald et al. | Oct 1995 | A |
5515847 | Braig et al. | May 1996 | A |
5572996 | Doiron et al. | Nov 1996 | A |
5598842 | Ishihara et al. | Feb 1997 | A |
5666956 | Buchert | Sep 1997 | A |
5706821 | Matcher et al. | Jan 1998 | A |
5720291 | Schwartz | Feb 1998 | A |
5722398 | Ishihara et al. | Mar 1998 | A |
5741213 | Kouchi et al. | Apr 1998 | A |
5769076 | Maekawa et al. | Jun 1998 | A |
5784162 | Cabib et al. | Jul 1998 | A |
5787185 | Clayden | Jul 1998 | A |
5791345 | Ishihara et al. | Aug 1998 | A |
5811814 | Leone et al. | Sep 1998 | A |
5931779 | Arakaki et al. | Aug 1999 | A |
5934278 | Ishihara et al. | Aug 1999 | A |
5974338 | Asano et al. | Oct 1999 | A |
5983120 | Groner et al. | Nov 1999 | A |
6061583 | Ishihara et al. | May 2000 | A |
6081612 | Gotkowicz-Krusin et al. | Jun 2000 | A |
6088087 | Graves et al. | Jul 2000 | A |
6104939 | Groner et al. | Aug 2000 | A |
6116736 | Stark et al. | Sep 2000 | A |
6208749 | Gutkowicz-Krusin et al. | Mar 2001 | B1 |
6244712 | Smith et al. | Jun 2001 | B1 |
6350431 | Snow et al. | Feb 2002 | B1 |
6351663 | Flower et al. | Feb 2002 | B1 |
6362175 | Vinogradov et al. | Mar 2002 | B1 |
6478424 | Grinvald et al. | Nov 2002 | B1 |
6512936 | Monfre et al. | Jan 2003 | B1 |
6512937 | Blank et al. | Jan 2003 | B2 |
6567678 | Oosta et al. | May 2003 | B1 |
6571118 | Utzinger et al. | May 2003 | B1 |
6587701 | Stranc et al. | Jul 2003 | B1 |
6588901 | Grinvald et al. | Jul 2003 | B1 |
6721582 | Trepagnier et al. | Apr 2004 | B2 |
6782289 | Strauss | Aug 2004 | B1 |
6820979 | Stark et al. | Nov 2004 | B1 |
6826424 | Zeng et al. | Nov 2004 | B1 |
6840622 | Kutschbach et al. | Jan 2005 | B2 |
6844195 | Craine | Jan 2005 | B2 |
6859658 | Krug | Feb 2005 | B1 |
6869430 | Balbierz et al. | Mar 2005 | B2 |
6889075 | Marchitto et al. | May 2005 | B2 |
6898451 | Wuori | May 2005 | B2 |
6898458 | Zeng et al. | May 2005 | B2 |
6902935 | Kaufman et al. | Jun 2005 | B2 |
6917038 | Zheng et al. | Jul 2005 | B2 |
6957094 | Chance et al. | Oct 2005 | B2 |
7025765 | Balbierz et al. | Apr 2006 | B2 |
7054674 | Cane et al. | May 2006 | B2 |
7115841 | Zeng et al. | Oct 2006 | B2 |
7130672 | Pewzner et al. | Oct 2006 | B2 |
7190452 | Zeng et al. | Mar 2007 | B2 |
7217266 | Anderson et al. | May 2007 | B2 |
7225005 | Kaufman et al. | May 2007 | B2 |
7253894 | Zeng et al. | Aug 2007 | B2 |
7280678 | Haven et al. | Oct 2007 | B2 |
7334895 | Kandel et al. | Feb 2008 | B2 |
7418115 | Northcott et al. | Aug 2008 | B2 |
7488073 | Kandel et al. | Feb 2009 | B2 |
7490939 | Hirohara et al. | Feb 2009 | B2 |
7836894 | Brinkmann et al. | Nov 2010 | B2 |
20010031920 | Kaufman et al. | Oct 2001 | A1 |
20010056237 | Cane et al. | Dec 2001 | A1 |
20020007122 | Kaufman et al. | Jan 2002 | A1 |
20020007123 | Balas | Jan 2002 | A1 |
20020016533 | Marchitto et al. | Feb 2002 | A1 |
20020016534 | Trepagnier et al. | Feb 2002 | A1 |
20020024633 | Kim et al. | Feb 2002 | A1 |
20020026127 | Balbierz et al. | Feb 2002 | A1 |
20020026188 | Balbierz et al. | Feb 2002 | A1 |
20020035358 | Wang | Mar 2002 | A1 |
20020076820 | Craine | Jun 2002 | A1 |
20020103439 | Zeng et al. | Aug 2002 | A1 |
20020111545 | Lindberg et al. | Aug 2002 | A1 |
20020127735 | Kaufman et al. | Sep 2002 | A1 |
20020197728 | Kaufman et al. | Dec 2002 | A1 |
20030032064 | Soller et al. | Feb 2003 | A1 |
20030036693 | Avinash et al. | Feb 2003 | A1 |
20030036751 | Anderson et al. | Feb 2003 | A1 |
20030050541 | Wuori | Mar 2003 | A1 |
20030074028 | Breithardt et al. | Apr 2003 | A1 |
20030114762 | Balas et al. | Jun 2003 | A1 |
20030146385 | Zheng et al. | Aug 2003 | A1 |
20030163049 | Balas | Aug 2003 | A1 |
20030207250 | Kaufman et al. | Nov 2003 | A1 |
20040015211 | Nurmikko et al. | Jan 2004 | A1 |
20040044287 | Lin et al. | Mar 2004 | A1 |
20040054270 | Pewzner et al. | Mar 2004 | A1 |
20040105075 | Kandel et al. | Jun 2004 | A1 |
20040116814 | Stranc et al. | Jun 2004 | A1 |
20040170304 | Haven et al. | Sep 2004 | A1 |
20040189940 | Kutschbach et al. | Sep 2004 | A1 |
20050046793 | Hirohara et al. | Mar 2005 | A1 |
20050054936 | Balas | Mar 2005 | A1 |
20050064602 | Kaufman et al. | Mar 2005 | A1 |
20050090751 | Balas | Apr 2005 | A1 |
20050131284 | Grinvald et al. | Jun 2005 | A1 |
20050143662 | Marchitto et al. | Jun 2005 | A1 |
20050167621 | Zeng et al. | Aug 2005 | A1 |
20050177140 | Jay | Aug 2005 | A1 |
20050192493 | Wuori | Sep 2005 | A1 |
20050203421 | Zeng et al. | Sep 2005 | A1 |
20050203423 | Zeng et al. | Sep 2005 | A1 |
20050234315 | Mayevsky et al. | Oct 2005 | A1 |
20050251049 | Cane et al. | Nov 2005 | A1 |
20060082725 | Yamaguchi et al. | Apr 2006 | A1 |
20060122524 | Kawada et al. | Jun 2006 | A1 |
20060131284 | Sun et al. | Jun 2006 | A1 |
20060141633 | Balas | Jun 2006 | A1 |
20060147897 | Grinvald et al. | Jul 2006 | A1 |
20060184037 | Ince et al. | Aug 2006 | A1 |
20060195022 | Trepagnier et al. | Aug 2006 | A1 |
20060241577 | Balbierz et al. | Oct 2006 | A1 |
20070030447 | Yamaguchi et al. | Feb 2007 | A1 |
20070043341 | Anderson et al. | Feb 2007 | A1 |
20070056596 | Fanney et al. | Mar 2007 | A1 |
20070179366 | Pewzner et al. | Aug 2007 | A1 |
20070243521 | Zuckerman | Oct 2007 | A1 |
20080021515 | Horsager et al. | Jan 2008 | A1 |
20080068561 | Kandel et al. | Mar 2008 | A1 |
20080254531 | Zuckerman | Oct 2008 | A1 |
20080269629 | Reiner | Oct 2008 | A1 |
20090221912 | Nelson et al. | Sep 2009 | A1 |
20100103497 | Rehn | Apr 2010 | A1 |
Number | Date | Country |
---|---|---|
2151483 | Dec 1993 | CN |
1302158 | Apr 2003 | EP |
04-017076 | Jan 1992 | JP |
9715229 | May 1997 | WO |
9963882 | Dec 1999 | WO |
0006015 | Feb 2000 | WO |
0122741 | Mar 2001 | WO |
03083434 | Oct 2003 | WO |
2004004556 | Jan 2004 | WO |
2004049899 | Jun 2004 | WO |
2006121984 | Nov 2006 | WO |
Entry |
---|
Sato Y et al., “Automatic extraction and measurement of leukocyte motion in microvessels using spatiotemporal image analysis”, IEEE Trans. on Biomedical Engineering, IEEE Inc. New York, US, vol. 44, No. 4, Apr. 1, 1997. |
Holeman B et al., “Dynamic scene analysis for vessel structure determination”, Southeastcon '89 proceedings, Energy and information technologies in southeast, Columbia, SC USA, Apr. 9, 1989, pp. 1072-1073. |
Domingo J et al., “Irregular motion recovery in flourescein angiograms”, Pattern Recognition Letters, North-Holland Publ. Amsterdam, vol. 18, No. 8, Aug. 1, 1997, pp. 805-821. |
Sklenar J et al., “Parametric imaging for myocardial contrast echocardiography: pixel-by-pixel incorporation of information from both spatial and temporal domains”, Computer in Cardiology 1998 Cleveland, OH, USA, Sep. 13-16, 1998, pp. 461-464. |
R.A. Linsenmeier, et al., “Metabolic dependence of photoreceptors on the choroids in the normal and detached retina”, Published in Investigative Ophthalmology and Visual Science, vol. 41(10), pp. 3117-3123, Sep. 2000. |
R. A. Linsenmeier et al., “Retinal hypoxia in long-term diabetic cats”, Published in Investigative Ophthalmology and Visual Science, vol. 39(9), pp. 1647-1657, Aug. 1998. |
Kurt R. Denninghoff et al., “Retinal imaging techniques in diabetes”, Published in Diabetes Technology & Therapeutics, vol. 2, No. 1 (2000), pp. 111-113. |
M. Bruce Shields, “Textbook of Glaucoma”, Published by Lippincott Williams and Wilkins (Philadelphia), 1997. |
Wong T. Y. et al./, “White matter lesions, retinopathy, and incident clinical stroke”, JAMA 288(1), pp. 67-74 (2002). |
Wong T. Y et al., “Retinal arteriolar narrowing and risk coronary heart disease in men and women. The atherosclerosis risk in communities study”, JAMA; 287(9), pp. 1153-1159, (2002). |
H. H. Quick et al., “MR Imaging of the Vessel wall”, Published in Eur. Radio, vol. 12(4), pp. 889-900, Apr. 2002. |
Dormandy et al., “Lower-extremity arteiosclerosis as a reflection of a systemic process: implications for concomitant coronary and carotid disease”, Published in Semin. Vasc. Surg. pp. 118-112, vol. 12(2), Jun. 1999. |
Kutzner et al., “Fatal lipid embolism following intra-arterial angiography as an early stage of anteriosclerosis”, Published in British Journal of Radiology, vol. 73,(874), pp. 1108-1111, Oct. 2000. |
Wong T. Y. et al., “Retinal microvascular abnormalities and incident stroke: the atherosclerosis risk in communities study”, Lancet, 358(9288), pp. 1134-1140 (2001). |
“Finite element modeling of three-diimensional pulsatile flow in the abdominal aorta: Rellevance to atherosclerosis”, Taylor et al., Annals of Biomedical Engineering, vol. 26, pp. 975-987 (1998). |
“In vivo quantification of blood flow and wall shear stress in the human abdominal aorta during lower limb exercise”, Taylor et al., Annals of Biomedical Engineering, vol. 30, pp. 402-408 (2002). |
Michelson G. et al., “Flickering light increases retinal blood flow”, Retina, 22(3): 336-343, Jun. 2002. |
Grinvald A. et al., “In-vivo optical imaging of cortical architecture and dynamics”, Published in Modern Techniques in Neuroscience Research, U. Windhorst and H. Johansson (eds.), Springer Verlag. |
U. Seifert, W. Visler, “Retinal Vessel Analyzer (RVA)—Design and Function”, Biomed Tech vol. 47, suppl. 1, (2002). |
“Visual stimulus induced changes in human near-infrared fundus reflectance”, by Abramoff, Invest Ophthalmol Vis Sci. Feb. 2006 47(2): 715-721. |
Li J K-J, et al., “Noninvasive assessment of brain oxygenation and blood flow”. Proceedings of the Annual Northeast Bioengineering Conference, 1990, vol. 16, 107-108. |
Number | Date | Country | |
---|---|---|---|
20090163827 A1 | Jun 2009 | US |
Number | Date | Country | |
---|---|---|---|
61009055 | Dec 2007 | US |