This application claims priority from Japanese Patent Application No. 2021-160684 filed on Sep. 30, 2021, the entire subject-matter of which is incorporated herein by reference.
The present disclosure relates to a fundus image processing apparatus and a non-transitory computer-readable storage medium storing a fundus image processing program used for processing a fundus image of a subject eye.
In recent years, a technique for identifying a specific site of a fundus by analyzing a fundus image of a subject eye has been proposed. For example, an ophthalmologic imaging apparatus disclosed in JP2018-083106A performs image processing (edge detection, Hough transform, or the like) on a front image of a fundus of a subject eye, to detect a position of an optic disk (hereinafter, also referred to as a “optic nerve head”) of the fundus captured in the front image.
It is possible to detect an approximate position of an optic nerve head from a front image of a fundus, but it is difficult to detect a position of an end of the optic nerve head with high accuracy. Here, it is also conceivable to detect the position of the end of the optic nerve head from a plurality of two-dimensional tomographic images configuring a three-dimensional tomographic image of the fundus. In this case, the plurality of two-dimensional tomographic images, configuring the three-dimensional tomographic image of the fundus, include an image in which the optic nerve head is captured and an image in which the optic nerve head is not captured. Therefore, the end of the optic nerve head may be erroneously detected from the two-dimensional tomographic image in which the optic nerve head is not captured. In a case where a plurality of two-dimensional tomographic images configuring the three-dimensional tomographic image of the fundus are processed, a processing amount also increases. As described above, it was difficult to appropriately detect the end of the optic nerve head captured in the fundus image with high accuracy.
A typical object of the present disclosure is to provide a fundus image processing apparatus and a non-transitory computer-readable storage medium storing a fundus image processing program capable of appropriately detecting an end of an optic nerve head captured in a fundus image with high accuracy.
According to a first aspect of the present disclosure, there is provided a fundus image processing apparatus that processes a tomographic image of a fundus of a subject eye captured by an OCT apparatus, the fundus image processing apparatus including:
a controller configured to perform:
According to a second aspect of the present disclosure, there is provided a non-transitory computer-readable storage medium storing a fundus image processing program executed by a fundus image processing apparatus that processes a tomographic image of a fundus of a subject eye captured by an OCT apparatus, the fundus image processing program being executed by a controller of the fundus image processing apparatus to cause the fundus image processing apparatus to perform:
According to the fundus image processing apparatus and the non-transitory computer-readable storage medium storing the fundus image processing program related to the above aspect of the present disclosure, an end of an optic nerve head captured in a fundus image is appropriately detected with high accuracy.
<Outline>
(First Aspect)
A controller of a fundus image processing apparatus exemplified in the present disclosure performs image acquisition processing, deviation degree acquisition processing, and site identification processing. In the image acquisition processing, the controller acquires a fundus image captured by a fundus image capturing apparatus. In the deviation degree acquisition processing, the controller inputs a fundus image into a mathematical model trained by a machine learning algorithm to acquire a probability distribution for identifying a first site of the fundus captured in the fundus image and acquire the degree of deviation of the acquired probability distribution with respect to a probability distribution in a case where the first site is accurately identified. In the site identification processing, the controller identifies a second site of the fundus which is different from the first site, based on the degree of deviation.
In the fundus, a state of the first site may change between a position where the second site is present and a position where the second site is not present. For example, a state of at least any one of a layer and a boundary of the fundus (first site) differs between a position where an optic nerve head (second site) is present and a position where the optic nerve head is not present (for example, around the optic nerve head). In general, a plurality of layers and boundaries are normally present around the optic nerve head, but specific layers and boundaries are missing at the position of the optic nerve head.
Here, a case is assumed in which a fundus image in which both the first site and the second site are captured is input to a mathematical model for identifying the first site. In this case, at a position where the first site is present, the first site is easily identified accurately, and thus the degree of deviation tends to decrease. On the other hand, in a case where the first site is missing at a position where the second site is present, the degree of deviation tends to increase. This tendency is likely to appear regardless of the presence or absence of an eye disease or the like.
Based on the above findings, the controller of the fundus image processing apparatus of the present disclosure identifies the second site based on the degree of deviation of a probability distribution in a case where the fundus image is input to the mathematical model for identifying the first site. As a result, the identification accuracy of the second site is improved regardless of the presence or absence of an eye disease or the like.
The degree of deviation will be described in more detail. In a case where the first site is identified with high accuracy by the mathematical model, an acquired probability distribution is likely to be biased. On the other hand, in a case where the identification accuracy of the first site by the mathematical model is low, an acquired probability distribution is less likely to be biased. Therefore, the degree of deviation between a probability distribution in a case where the first site is accurately identified and a probability distribution actually acquired changes according to a state of the first site. Therefore, according to the fundus image processing apparatus of the present disclosure, the second site can be identified with high accuracy regardless of the presence or absence of a disease, by using the degree of deviation in a case where a state of the first site changes between the position where the second site is present and the position where the second site is not present.
The degree of deviation may be output by the mathematical model. The controller may calculate the degree of deviation based on a probability distribution output by the mathematical model.
The degree of deviation may also be expressed as the uncertainty of identification of the first site performed by the mathematical model on the fundus image. The same result can also be obtained in a case where, for example, a reciprocal of a high certainty of identification (certainty) by the mathematical model is used as the degree of deviation.
The degree of deviation may include entropy (average amount of information) of the acquired probability distribution. The entropy represents the degree of uncertainty, messiness, and disorder. In the present disclosure, the entropy of a probability distribution output in a case where the first site is accurately identified is 0. The more difficult it is to identify the first site, the greater the entropy.
However, a value other than entropy may be employed as the degree of deviation. For example, at least any one of a standard deviation, a coefficient of variation, and a variance indicating the degree of scatter of the acquired probability distribution may be used as the degree of deviation. KL divergence or the like, which is a measure for measuring a difference between probability distributions, may be used as the degree of deviation. The maximum value of the acquired probability distribution may be used as the degree of deviation.
In the deviation degree acquisition processing, the degree of deviation may be acquired with at least any one of a plurality of layers and boundaries in the fundus captured in the fundus image as the first site. That is, the first site may be at least any one of a plurality of layers and boundaries between the layers in the fundus (hereinafter, also referred to as “layer/boundary”). As described above, a state of at least any one (first site) of the layers and boundaries of the fundus may differ between the position where the second site is present and the position where the second site is not present. Therefore, by acquiring the degree of deviation with the layer/boundary as the first site, it becomes easier to appropriately identify the second site based on the degree of deviation.
However, a site other than the layer/boundary in the fundus may be used as the first site. For example, a state of a fundus blood vessel may differ between the position where the second site is present and the position where the second site is not present. In this case, the degree of deviation may be acquired with the fundus blood vessel as the first site.
In a case where the first site is a layer/boundary, the controller may identify the optic nerve head (optic disk) in the fundus as the second site based on the degree of deviation, in the site identification processing. As described above, a state of at least any one of the layers and boundaries of the fundus differs between the position where the optic nerve head is present and the position where the optic nerve head is not present. Therefore, by setting the layer/boundary as the first site and the optic nerve head as the second site, the optic nerve head is appropriately detected based on the degree of deviation.
In a case where the layer/boundary is the first site and the optic nerve head is the second site, in the deviation degree acquisition processing, the degree of deviation may be acquired with at least any one of layers and boundaries at positions deeper than a nerve fiber layer (NFL) among the plurality of layers and boundaries of the fundus captured in the fundus image as the first site. At the position where the optic nerve head is present, the NFL is present, and layers and boundaries at positions deeper than the NFL are missing. That is, at the position where the optic nerve head is present, the degree of deviation related to identification of layers and boundaries at positions deeper than the NFL is larger than that at the position where the optic nerve head is not present. Therefore, by setting at least any one of the layers and the boundaries at the positions deeper than the NFL as the first site, the identification accuracy of the optic nerve head is further improved.
In the deviation degree acquisition processing, the degree of deviation may be acquired with at least any one of the NFL and the layers and the boundaries at the positions deeper than the NFL as the first site. In the site detection processing, a site in which the degree of deviation related to identification of a layer/boundary at a position deeper than the NFL is more than a first threshold value and the degree of deviation related to identification of the NFL is less than a second threshold value, may be detected as the optic nerve head. In this case, a position where a plurality of layers/boundaries including the NFL are missing due to the influence of a disease or the like, and the position where the optic nerve head is present are appropriately distinguished. Therefore, the identification accuracy of the optic nerve head is further improved.
The controller may acquire a three-dimensional tomographic image of the fundus as a fundus image, in the fundus image acquisition processing. The controller may further perform reference position setting processing, radial pattern setting processing, image extraction processing, and optic nerve head end detection processing. In the reference position setting processing, the controller sets a reference position in a region of the optic nerve head identified in the site identification processing in a two-dimensional measurement region in which a three-dimensional tomographic image is captured. In the radial pattern setting processing, the controller sets a radial pattern that is a line pattern extending radially around the reference position, with respect to the two-dimensional measurement region. In the image extraction processing, the controller extracts a two-dimensional tomographic image (a two-dimensional tomographic image that intersects each of the plurality of lines of the radial pattern) in each of the plurality of lines of the set radial pattern, from the three-dimensional tomographic image. In the optic nerve head end detection processing, the controller detects a position of the end of the optic nerve head captured in the three-dimensional tomographic image based on the plurality of extracted two-dimensional tomographic images.
In a case where the reference position is correctly set in the region of the optic nerve head in the reference position setting processing, the optic nerve head will always be included in all of the plurality of two-dimensional tomographic images extracted according to the radial pattern in the image extraction processing. Therefore, by detecting the position of the end of the optic nerve head based on the plurality of extracted two-dimensional tomographic images, a probability that the end of the optic nerve head is erroneously detected from the two-dimensional tomographic image in which the optic nerve head is not captured, is reduced. It is possible to suppress an excessive increase in an amount of image processing compared with a case of processing all of a plurality of two-dimensional tomographic images configuring a three-dimensional tomographic image. Therefore, the end of the optic nerve head is also detected with high accuracy, by using a result of identification of the optic nerve head site performed based on the degree of deviation.
In a case where the first site is a layer/boundary, the controller may identify the fovea in the fundus as the second site, based on the degree of deviation, in the site identification processing. A state of at least any one of the layers and boundaries of the fundus differs between a position where the fovea is present and a position where the fovea is not present. Therefore, by setting a layer/boundary as the first site and the fovea as the second site, the fovea can be appropriately detected based on the degree of deviation.
In a case where the layer/boundary is the first site and the fovea is the second site, in the deviation degree acquisition processing, the degree of deviation may be acquired with at least any one of layers and boundaries nearer to a surface side of the retina than the retinal pigment epithelium (RPE), among the plurality of layers and boundaries of the fundus captured in the fundus image, as the first site. At the position where the fovea is present, the RPE, the Bruch's membrane, and the like are present, and the layers and boundaries near to the surface side of the retina than the RPE are missing. That is, at the position where the fovea is present, the degree of deviation related to identification of the layer/boundary nearer to the surface side than the RPE is larger than that at the position where the fovea is not present. Therefore, by setting at least any one of the layers and the boundaries nearer to the surface side of the retina than the RPE as the first site, the identification accuracy of the fovea is further improved.
In the deviation degree acquisition processing, the degree of deviation may be acquired with both of at least one of the RPE and Bruch's membrane (hereinafter, simply referred to as “RPE/Bruch's membrane”), and at least any one of layers and boundaries nearer to the surface side than the RPE, as the first site. In the site detection processing, a site may be detected, as the fovea, in which the degree of deviation related to identification of the layer/boundary nearer to the surface side than the RPE is more than the first threshold value and the degree of deviation related to identification of the RPE/Bruch's membrane is less than the second threshold value. In this case, a position where a plurality of layers/boundaries including the RPE/Bruch's membrane are missing due to the influence of a disease or the like and a position where the fovea is present are appropriately distinguished. Therefore, the identification accuracy of the fovea is further improved.
The second site to be identified based on the degree of deviation is not limited to the optic nerve head and the fovea. The second site may be a site other than the optic nerve head and fovca in the fundus (for example, a macula or a fundus blood vessel). For example, at a position where the fundus blood vessel (second site) is present, measurement light is blocked by the fundus blood vessel, and an imaging state of a layer/boundary (first site) at a position deeper than the fundus blood vessel deteriorates. Therefore, at the position where the fundus blood vessel is present, the degree of deviation related to identification of the layer/boundary at the position deeper than the fundus blood vessel is larger than that at a position where the fundus blood vessel is not present. Therefore, the controller may identify a site in which the degree of deviation related to identification of at least any one of layers/boundaries at positions deeper than the fundus blood vessel is more than the threshold value, as a site in which the fundus blood vessel is present. The fundus image processing apparatus may identify a site of a disease existing in the fundus as the second site.
In the deviation degree acquisition processing, the controller may input a three-dimensional tomographic image of the fundus into the mathematical model to acquire a two-dimensional distribution of the degree of deviation in a case where the fundus is viewed from the front (that is, in a case where the fundus is viewed along an optical axis of imaging light of the fundus image). In the site identification processing, a position of the second site in a case where the fundus is viewed from the front may be identified based on the two-dimensional distribution of the degree of deviation. In this case, the second site is identified based on more data than in a case of identifying a two-dimensional position of the second site from the two-dimensional fundus image. Therefore, the identification accuracy of the second site is further improved.
A specific method of acquiring a two-dimensional distribution of the degree of deviation from a three-dimensional tomographic image may also be selected as appropriate. For example, the controller may input each of a plurality of two-dimensional tomographic images configuring the three-dimensional tomographic image into the mathematical model and arrange the degree of deviation acquired for each two-dimensional tomographic image in two dimensions, to acquire the two-dimensional distribution of the degree of deviation. The controller may input the entire three-dimensional tomographic image into the mathematical model to acquire a two-dimensional distribution of the degree of deviation. The tomographic images (three-dimensional tomographic image and two-dimensional tomographic image) may be captured by various devices such as an OCT apparatus or a Scheimpflug camera.
The controller may input a two-dimensional fundus image into the mathematical model to identify the second site in the fundus. For example, the controller may input a two-dimensional front image in a case where the fundus is viewed from the front into the mathematical model to identify a fundus blood vessel as the first site. The controller may detect the second site (for example, an optic nerve head) based on the acquired two-dimensional distribution of the degree of deviation. The two-dimensional front image may be an image captured by a fundus camera, an image captured by a scanning laser ophthalmoscope (SLO), or the like. The two-dimensional front image may be an Enface image generated based on data of a three-dimensional tomographic image captured by the OCT apparatus. The two-dimensional front image may be an image generated from motion contrast data obtained by processing a plurality of pieces of OCT data acquired from the same position at different times (so-called “motion contrast image”).
The controller may further perform front image acquisition processing and auxiliary identification result acquisition processing. In the front image acquisition processing, the controller acquires a two-dimensional front image in a case where the fundus of which the three-dimensional tomographic image is captured is viewed from the front. In the auxiliary identification result acquisition processing, the controller acquires an auxiliary identification result that is an identification result of the second site, which is performed based on the two-dimensional front image. The second site may be identified based on the degree of deviation and the auxiliary identification result. In this case, in addition to the degree of deviation obtained from the three-dimensional tomographic image, the auxiliary identification result based on the two-dimensional front image is also taken into consideration, and thus the second site is more appropriately identified.
A specific method of acquiring the auxiliary identification result may be selected as appropriate. For example, the auxiliary identification result may be a result of identifying the second site by performing image processing on the two-dimensional front image. In this case, the image processing may be performed by the controller of the fundus image processing apparatus, or may be performed by another device. The controller may acquire the auxiliary identification result by inputting the two-dimensional front image front image acquired in the front image acquisition processing into a mathematical model that outputs an identification result of the second site in the two-dimensional front image.
A specific method of identifying the second site based on the auxiliary identification result and the degree of deviation may also be selected as appropriate. For example, the controller may extract a part that is likely to include the second site, from the entire three-dimensional tomographic image acquired in the image acquisition processing, based on the auxiliary identification result. The controller may acquire the degree of deviation by inputting the extracted three-dimensional tomographic image into the mathematical model, and may identify the second site based on the acquired degree of deviation. In this case, an amount of processing by the mathematical model is reduced, and thus the second site can be identified more efficiently. The controller may identify the second site by adding the identification result based on the degree of deviation and the auxiliary identification result after performing any weighting. The controller may notify a user of a warning, an error, or the like in a case where a difference between the identification result based on the degree of deviation and the auxiliary identification result does not satisfy conditions.
The mathematical model may output a distribution of scores indicating a possibility of the second site, together with an identification result of the first site of the fundus captured in the fundus image. In the site identification processing, the second site may be identified based on the degree of deviation and the distribution of the scores. In this case, the second site is identified based on the distribution of the score of the second site and the degree of deviation, which is not easily affected by the presence or absence of an eye disease or the like. Therefore, the identification accuracy of the second site is further improved.
A specific method of identifying the second site based on both the degree of deviation and the distribution of scores may also be selected as appropriate. For example, the controller may identify the second site by adding an identification result based on the degree of deviation and an identification result based on the distribution of scores. In this case, the controller may add the identification results after performing any weighting. However, the controller may also identify the second site without using the distribution of scores of the second site.
(Second Aspect)
The controller of the fundus image processing apparatus exemplified in the present disclosure performs image acquisition processing, reference position setting processing, radial pattern setting processing, image extraction processing, and optic nerve head end detection processing. In the image acquisition processing, the controller acquires a three-dimensional tomographic image of a fundus of a subject eye captured by irradiating a two-dimensional measurement region extending in a direction intersecting an optical axis of OCT measurement light with the measurement light. In the reference position setting processing, the controller sets a reference position in a region of the optic nerve head in the two-dimensional measurement region in which the three-dimensional tomographic image is captured. In the radial pattern setting processing, the controller sets a radial pattern that is a line pattern extending radially around the reference position, with respect to the two-dimensional measurement region. In the image extraction processing, the controller extracts a two-dimensional tomographic image (a two-dimensional tomographic image that intersects each of the plurality of lines of the radial pattern) in each of the plurality of lines of the set radial pattern from the three-dimensional tomographic image. In the optic nerve head end detection processing, the controller detects a position of the end of the optic nerve head captured in the three-dimensional tomographic image based on the plurality of extracted two-dimensional tomographic images.
In a case where the reference position is correctly set in the region of the optic nerve head in the reference position setting processing, the optic nerve head will always be included in all of the plurality of two-dimensional tomographic images extracted according to the radial pattern in the image extraction processing. Therefore, by detecting the position of the end of the optic nerve head based on the plurality of extracted two-dimensional tomographic images, a probability that the end of the optic nerve head is erroneously detected from the two-dimensional tomographic image in which the optic nerve head is not captured is reduced. It is possible to suppress an excessive increase in an amount of image processing compared with a case of processing all of a plurality of two-dimensional tomographic images configuring a three-dimensional tomographic image. Therefore, the end of the optic nerve head is appropriately detected with high accuracy.
In a case where the tomographic image captured by the OCT apparatus is used for diagnosis, it is desirable that not only information regarding the optic nerve head but also various types of information such as a retina thickness can be obtained based on the tomographic image. Here, it is also conceivable to capture a plurality of two-dimensional tomographic images by actually scanning the fundus with measurement light along the radial pattern after setting the center of the radial pattern in the region of the optic nerve head. Even in this case, it seems that a position of the end of the optic nerve head can be detected from a plurality of captured two-dimensional tomographic images. However, it is difficult to obtain various types of information such as the retina thickness from a plurality of two-dimensional tomographic images captured according to the radial pattern. In contrast, the controller of the fundus image processing apparatus of the present disclosure may perform fundus analysis processing (for example, analysis of a thickness of a specific layer of the retina) on the three-dimensional tomographic image acquired in the image acquisition processing, in addition to the optic nerve head end detection processing. That is, according to the fundus image processing apparatus of the present disclosure, by using the three-dimensional tomographic image, it is possible not only to detect a position of the end of the optic nerve head with high accuracy but also to obtain an analysis result of the fundus.
The details of “the end of the optic nerve head” to be detected may be selected as appropriate. At least any one of, for example, a Bruch's Membrane Opening (BMO), the margin of the optic disk, and parapapillary atrophy (PPA) may be detected as the end of the optic nerve head.
Any of various apparatuses may function as the fundus image processing apparatus. For example, an OCT apparatus itself may function as the fundus image processing apparatus in the present disclosure. A device (for example, a personal computer or the like) capable of exchanging data with the OCT apparatus may function as the fundus image processing apparatus. Controllers of a plurality of devices may cooperate to perform processing.
The OCT apparatus may include a scanning unit. The scanning unit performs scanning, with measurement light applied to the tissue by an irradiation optical system, in a two-dimensional direction intersecting the optical axis. The three-dimensional tomographic image may be obtained by the scanning unit performing scanning, with a spot of the measurement light, in a measurement region, in the two-dimensional direction. In this case, a three-dimensional tomographic image is appropriately obtained by the OCT apparatus.
However, a configuration of the OCT apparatus may be changed. For example, the irradiation optical system of the OCT apparatus may simultaneously irradiate a two-dimensional region on the tissue of a subject with the measurement light. In this case, a light receiving element may be a two-dimensional light receiving element that detects an interference signal in the two-dimensional region on the tissue. That is, the OCT apparatus may acquire OCT data according to the principle of so-called full-field OCT (FF-OCT). The OCT apparatus may simultaneously irradiate an irradiation line extending in the one-dimensional direction on the tissue with the measurement light and perform scanning, with the measurement light, in a direction intersecting the irradiation line. In this case, the light receiving element may be a one-dimensional light receiving element (for example, a line sensor) or a two-dimensional light receiving element. That is, the OCT apparatus may acquire a tomographic image according to the principle of so-called line field OCT (LF-OCT).
The controller may further perform alignment processing of performing image alignment, in the direction along the optical axis of the OCT measurement light, of the three-dimensional tomographic image or the two-dimensional tomographic image extracted in the image extraction processing. The controller may detect a position of the end of the optic nerve head based on the two-dimensional tomographic image for which the image alignment has been performed. In this case, by performing image alignment, the deviation of the annular optic nerve head end in the direction along the optical axis of the OCT measurement light (tissue depth direction), is reduced. Therefore, the end of the optic nerve head is detected with higher accuracy.
The controller may further perform optic nerve head position detection processing of automatically detecting a position of the optic nerve head in a two-dimensional region intersecting the optical axis of the OCT measurement light, based on the image of the fundus. The controller may set a reference position at the automatically detected position of the optic nerve head. In this case, even in a case where the accuracy of automatic detection of the position of the optic nerve head is low, if the detected position is within the actual optic nerve head region, the end of the optic nerve head is appropriately detected in the subsequent optic nerve head end detection processing. Therefore, the detection processing is performed more smoothly.
In the optic nerve head position detection processing, a center position of the optic nerve head may be detected. In this case, there is a higher probability that a reference position will be within the region of the optic nerve head than in a case where a position other than the center of the optic nerve head is detected and set as a reference position.
A specific method of automatically detecting a position of the optic nerve head based on the image of the fundus may be selected as appropriate. As an example, at the position where the optic nerve head is present, the NFL is present, and layers and boundaries at positions deeper than the NFL are missing. Therefore, in a case where at least anyone of the layers and boundaries of the fundus (hereinafter simply referred to as a “layer/boundary”) captured in the three-dimensional tomographic image is detected by a mathematical model trained by using a machine learning algorithm, the uncertainty of detection of a layer/boundary at a position deeper than the NFL is high, at the position of the optic nerve head. Therefore, the controller may automatically detect the position (center position) of the optic nerve head based on the uncertainty in a case where the layer/boundary at the position deeper than the NFL is detected by the mathematical model. For example, the controller may detect a region where the uncertainty is equal to or more than a threshold value as a region of the optic nerve head, and detect the center of the detected region (for example, the center of gravity) as a center position of the optic nerve head.
The controller may automatically detect a position of the optic nerve head based on the two-dimensional front image in a case where the three-dimensional tomographic image is viewed from the front (a direction along the optical axis of the OCT measurement light). For example, the controller may perform known image processing on the two-dimensional front image, detect a region of the optic nerve head, and detect the center of the detected region as the center position of the optic nerve head. The controller may input the two-dimensional front image into the mathematical model that detects and outputs the position of the optic nerve head captured in the two-dimensional front image, to automatically detect a position (center position) of the optic nerve head. The two-dimensional front image may be a front image (so-called “Enface image” or the like) generated based on the three-dimensional tomographic image acquired in the image acquisition processing. The two-dimensional front image may be an image (for example, a fundus camera image or an SLO image) captured according to a principle different from the imaging principle of the three-dimensional tomographic image.
However, a method of setting a reference position may be changed. For example, the controller may set a reference position at a position designated by a user, in the two-dimensional measurement region. That is, the user may set the reference position by himself/herself. In this case, by setting the reference position in the region of the optic nerve head based on the user's experience or the like, the end of the optic nerve head is appropriately detected in the subsequent optic nerve head end detection processing. The controller may set a reference position at a position designated by the user, for example, in a case where the automatic detection of the position of the optic nerve head described above fails. The user may be made to set the reference position without performing the automatic detection of the position of the optic nerve head. For example, in a case where a position of the optic nerve head detected in the past is stored, a reference position may be set at the stored position of the optic nerve head. In this case, the processing of automatically detecting the position of the optic nerve head may be omitted.
In the optic nerve head end detection processing, a mathematical model trained by using a machine learning algorithm may be used. The mathematical model may be trained to output a detection result of the end of the optic nerve head captured in an input two-dimensional tomographic image. The controller may input the plurality of two-dimensional tomographic images extracted in the image extraction processing into the mathematical model and acquiring the position of the end of the optic nerve head output from the mathematical model, to detect a position of the end of the optic nerve head. In this case, the position of the end of the optic nerve head is automatically and appropriately detected from the plurality of two-dimensional tomographic images extracted according to the radial pattern.
The position of the end of the optic nerve head automatically detected by using the machine learning algorithm may be corrected according to an instruction from the user. For example, the controller may display the position of the end of the optic nerve head output from the mathematical model, on a display device, together with the two-dimensional tomographic image input to the mathematical model. The controller may correct the position of the end of the optic nerve head according to an instruction from the user who has checked the displayed position of the end of the optic nerve head. In this case, even in a case where the accuracy of automatic detection of the end of the optic nerve head is low, the position is appropriately corrected by the user. Therefore, the end of the optic nerve head is detected with higher accuracy.
However, a specific method of detecting a position of the end of the optic nerve head may be changed. For example, the controller may accept input of an instruction from the user in a state in which the two-dimensional tomographic image extracted in the image extraction processing is displayed on the display device. The controller may detect the position designated by the user as a position of the end of the optic nerve head. As described above, the two-dimensional tomographic image appropriately extracted according to the radial pattern always includes the optic nerve head. Therefore, the user can appropriately input (give an instruction for) the position of the end of the optic nerve head by checking the displayed two-dimensional tomographic image. As a result, the end of the optic nerve head is detected with high accuracy. The controller may automatically detect a position of the end of the optic nerve head by performing known image processing on the plurality of two-dimensional tomographic images extracted in the image extraction processing.
In the optic nerve head end detection processing, the controller may perform a smoothing processing on the detection results of the plurality of positions detected based on the plurality of two-dimensional tomographic images, to detect a position of the annular end of the optic nerve head. For example, due to the presence of a fundus blood vessel or the like, a position of the end of the optic nerve head in some two-dimensional tomographic images may be erroneously detected. In this case, the erroneously detected position of the annular end of the optic nerve head is separated from the appropriately detected position. In contrast, by performing a smoothing processing on the detection results of the plurality of detected positions, the influence of some of the erroneously detected positions is reduced. Therefore, the position of the annular end of the optic nerve head is more appropriately detected.
The controllcr may further perform optic nerve head center specifying processing of specifying a center position of the optic nerve head, based on the position of the optic nerve head end detected in the optic nerve head end detection processing. In this case, the center position of the optic nerve head is specified based on the position of the end of the optic nerve head detected with high accuracy. Therefore, the center position of the optic nerve head is specified with high accuracy.
A specific method of specifying a center position of the optic nerve head based on the detected position of the end of the optic nerve head may be selected as appropriate. For example, the controller may specify a detected position of the center of gravity of the annular optic nerve head end as a center position of the optic nerve head. The controller may fit an ellipse to the detected end of the optic nerve head and specify the center position of the fitted ellipse as a center position of the optic nerve head.
The controller may set the center position of the optic nerve head specified in the optic nerve head center specifying processing as a setting position of the reference position in the reference position setting processing, and perform the reference position setting processing, the radial pattern setting processing, the image extraction processing, and the optic nerve head end detection processing again. As the reference position becomes closer to a center position of the optic nerve head, a position of the end of the optic nerve head in each of the plurality of two-dimensional tomographic images extracted according to the radial pattern becomes more approximate, and thus the detection accuracy of the annular end of the optic nerve head becomes higher. Therefore, the accuracy of detection is further improved by detecting a position of the end of the optic nerve head again with the center position of the optic nerve head specified in the optic nerve head center specifying processing as a reference position.
The controller may further perform annular shape extraction processing and output processing. In the annular shape extraction processing, the controller extracts a two-dimensional tomographic image in an annular line pattern centered on the center position of the optic nerve head specified in the optic nerve head center specifying processing (that is, an image into which a tomographic image that intersects the annular line pattern in a cylindrical shape, is deformed in two dimensions), from the three-dimensional tomographic image. In the output processing, the controller outputs information regarding the two-dimensional tomographic image extracted in the annular shape extraction processing. In this case, a state of the tissue in the vicinity of the optic nerve head is appropriately observed with reference to the center position of the optic nerve head detected with high accuracy.
In a case where the information regarding the two-dimensional tomographic image extracted according to the annular line pattern is output, a specific method of outputting the information may be selected as appropriate. For example, the controller may display the extracted two-dimensional tomographic image on the display device. The controller may display a graph representing a thickness of a specific layer of the retina in the extracted two-dimensional tomographic image (for example, a thickness of the NFL or a thickness from the ILM to the NFL), on the display device. The controller may display at least anyone of a two-dimensional tomographic image of a patient and a graph, in comparison with disease-free normal eye data.
At the position where the fundus blood vessel is present, the end of the optic nerve head is difficult to be captured in the tomographic image, and thus the end of the optic nerve head is more likely to be erroneously detected. The controller may acquire information regarding the position of the fundus blood vessel in the measurement region in which the three-dimensional tomographic image is captured. The controller may adjust at least any one of an angle of the overall radial pattern, an angle of at least any one of the lines included in the radial pattern, a length of the line, the number of lines, and the like, to reduce an amount of overlap between the lines of the radial pattern and the fundus blood vessels as much as possible. In this case, the influence of the fundus blood vessels is reduced, and thus the detection accuracy of the end of the optic nerve head is further improved.
A method of acquiring information regarding a position of the fundus blood vessel may be selected as appropriate. For example, the controller may perform known image processing on a two-dimensional front image of the fundus (for example, an Enface image, an SLO image, or a fundus camera image), to detect a position of the fundus blood vessel. The controller may input the fundus image (a two-dimensional front image, a three-dimensional tomographic image, or the like) into the mathematical model trained by using the machine learning algorithm, to acquire a detection result of the fundus blood vessel output from the mathematical model. The controller may input an instruction, of the user who has checked the fundus image, on the position of the fundus blood vessel, to acquire information regarding the position of the fundus blood vessel.
The controller may adjust at least any one of the angle of the overall radial pattern, an angle of at least any one of the lines included in the radial pattern, a length of the line, the number of lines, and the like according to an instruction input by the user who has checked the fundus image. In this case, the user can appropriately set the radial pattern to reduce an amount of overlap between the lines of the radial pattern and the fundus blood vessels as much as possible.
The controller may also detect various structures of the fundus based on a detection result of the end of the optic nerve head. For example, in a case where the BMO is detected as the end of the optic nerve head, the controller may detect a position of an optic disk recess (Cup) based on the detected BMO. As an example, the controller may set a straight line parallel to a reference straight line passing through a detected pair of BMOs and separated, by a predetermined distance, from the reference straight line toward the surface side of the retina. The controller may detect a position where the set straight line and the internal limiting membrane (ILM) in the fundus image intersect, as a position of the Cup. The controller may detect the shortest distance between the detected BMO and the ILM in the fundus image as the minimum thickness (minimum rim width) of the nerve fiber layer.
(Apparatus Configuration)
Hereinafter, one of the typical embodiments in the present disclosure will be described with reference to the drawings. As shown in
As an example, a personal computer (hereinafter, referred to as a “PC”) is used for the mathematical model construction apparatus 101 of the present embodiment. Although the details will be described later, the mathematical model construction apparatus 101 trains a mathematical model by using data of a fundus image of the subject eye (hereinafter, referred to as a “fundus image for training”) acquired from the OCT apparatus 10A and data indicating a first site (a site of the optic nerve head, in the present embodiment) of the subject eye of which the fundus image for training is captured. As a result, the mathematical model is constructed. However, a device that can function as the mathematical model construction apparatus 101 is not limited to a PC. For example, the OCT apparatus 10A may function as a mathematical model construction apparatus 101. Controllers of a plurality of devices (for example, a CPU of a PC and a CPU 13A of the OCT apparatus 10A) may cooperate to construct a mathematical model.
A PC is used for the fundus image processing apparatus 1 of the present embodiment. However, a device that can function as the fundus image processing apparatus 1 is not limited to a PC. For example, the OCT apparatus 10B or a server may function as the fundus image processing apparatus 1. Ina case where the OCT apparatus 10B functions as the fundus image processing apparatus 1, the OCT apparatus 10B can process a captured fundus image while capturing the fundus image. A portable terminal such as a tablet terminal or a smartphone may function as the fundus image processing apparatus 1. Controllers of a plurality of devices (for example, a CPU of a PC and a CPU 13B of the OCT apparatus 10B) may cooperate to perform various processing.
In the present embodiment, a case where a CPU is used as an example of a controller that performs various processing will be illustrated. However, it goes without saying that a controller other than the CPU may be used for at least some of various devices. For example, by employing a GPU as a controller, a processing speed may be increased.
The mathematical model construction apparatus 101 will be described. The mathematical model construction apparatus 101 is provided in, for example, a manufacturer that provides the fundus image processing apparatus 1 or a fundus image processing program to a user. The mathematical model construction apparatus 101 includes a controller 102 that performs various control processing and a communication I/F 105. The controller 102 includes a CPU 103 that is a controller that performs control, and a storage device 104 that can store programs, data, and the like. The storage device 104 stores a mathematical model construction program for performing a mathematical model construction processing (refer to
The mathematical model construction apparatus 101 is connected to an operation unit 107 and a display device 108. The operation unit 107 is operated by a user in order for the user to input various instructions to the mathematical model construction apparatus 101. For the operation unit 107, at least any one of, for example, a keyboard, a mouse, and a touch panel may be used. A microphone or the like for inputting various instructions may be used together with the operation unit 107 or instead of the operation unit 107. The display device 108 displays various images. As the display device 108, various devices (for example, at least any one of, for example, a monitor, a display, and a projector) capable of displaying an image may be used. The “image” in the present disclosure includes both a still image and a moving image.
The mathematical model construction apparatus 101 may acquire data of a fundus image (hereinafter, may be simply referred to as a “fundus image”) from the OCT apparatus 10A. The mathematical model construction apparatus 101 may acquire data of the fundus image from the OCT apparatus 10A by using at least any one of, for example, wired communication, wireless communication, and a detachable storage medium (for example, a USB memory).
The fundus image processing apparatus 1 will be described. The fundus image processing apparatus 1 is provided in, for example, a facility (for example, a hospital or a health examination facility) for diagnosing or examining an examinee. The fundus image processing apparatus 1 includes a controller 2 that performs various control processing and a communication I/F 5. The controller 2 includes a CPU 3 which is a controller that performs control, and a storage device 4 that can store programs, data, and the like. The storage device 4 stores a fundus image processing program for performing fundus image processing (refer to
The fundus image processing apparatus 1 is connected to an operation unit 7 and a display device 8. As the operation unit 7 and the display device 8, various devices may be used in the same manner as the operation unit 107 and the display device 108 described above.
The fundus image processing apparatus 1 may acquire a fundus image (in the present embodiment, a three-dimensional tomographic image of the fundus) from the OCT apparatus 10B. The fundus image processing apparatus 1 may acquire a fundus image from the OCT apparatus 10B by using at least any one of, for example, wired communication, wireless communication, and a detachable storage medium (for example, a USB memory). The fundus image processing apparatus 1 may acquire a program or the like for realizing the mathematical model constructed by the mathematical model construction apparatus 101, via communication or the like.
The OCT apparatus 10 (10A, 10B) will be described. As an example, in the present embodiment, a case where the OCT apparatus 10A providing a fundus image to the mathematical model construction apparatus 101, and the OCT apparatus 10B providing a fundus image to the fundus image processing apparatus 1 are used, will be described. However, the number of OCT apparatuses used is not limited to two. For example, the mathematical model construction apparatus 101 and the fundus image processing apparatus 1 may acquire fundus images from a plurality of OCT apparatuses. The mathematical model construction apparatus 101 and the fundus image processing apparatus 1 may acquire fundus images from one common OCT apparatus.
As shown in
The OCT light source 11 emits light (OCT light) for acquiring OCT data. The coupler 12 divides the OCT light emitted from the OCT light source 11 into measurement light and reference light. The coupler 12 of the present embodiment combines the measurement light reflected by a subject (in the present embodiment, the fundus of a subject eye E) and the reference light generated by the reference optical system 20, to interfere with each other. That is, the coupler 12 of the present embodiment serves as both a branch optical element that branches the OCT light into the measurement light and the reference light, and a multiplexing optical element that combines reflected light of the measurement light and the reference light.
The measurement optical system 13 guides the measurement light divided by the coupler 12 to the subject, and returns the measurement light reflected by the subject to the coupler 12. The measurement optical system 13 includes a scanning unit 14, an irradiation optical system 16, and a focus adjustment unit 17. By being driven by a drive unit 15, the scanning unit 14 can perform scanning with (deflect) the measurement light in a two-dimensional direction intersecting an optical axis of the measurement light. The irradiation optical system 16 is provided further toward the downstream side (that is, the subject side) of the optical path than the scanning unit 14, and irradiates the tissue of the subject with the measurement light. The focus adjustment unit 17 moves an optical member (for example, a lens) included in the irradiation optical system 16 in a direction along the optical axis of the measurement light, to adjust a focus of the measurement light.
The reference optical system 20 generates reference light and returns the reference light to the coupler 12. The reference optical system 20 of the present embodiment reflects the reference light divided by the coupler 12 by using a reflection optical system (for example, a reference mirror), to generate the reference light. However, a configuration of the reference optical system 20 may also be changed. For example, the reference optical system 20 may transmit the light incident from the coupler 12 without reflecting the incident light, to return the incident light to the coupler 12. The reference optical system 20 includes an optical path length difference adjustment unit 21 that changes an optical path length difference between the measurement light and the reference light. In the present embodiment, an optical path length difference is changed by moving the reference mirror in the optical axis direction. A configuration for changing an optical path length difference may be provided in the optical path of the measurement optical system 13.
The light receiving element 22 receives interference light between the measurement light and the reference light generated by the coupler 12, to detect an interference signal. In the present embodiment, the principle of Fourier domain OCT is employed. In the Fourier domain OCT, the spectral intensity (spectral interference signal) of the interference light is detected by the light receiving element 22, and a complex OCT signal is acquired by performing Fourier transform on the spectral intensity data. As an example of the Fourier domain OCT, any of spectral-domain-OCT (SD-OCT), swept-source-OCT (SS-OCT), and the like, may be employed. For example, time-domain-OCT (TD-OCT) may be employed.
In the present embodiment, the scanning unit 14 scans, with a spot of the measurement light, in a two-dimensional measurement region, and thus three-dimensional OCT data (three-dimensional tomographic image) is acquired. However, the principle of acquiring three-dimensional OCT data may also be changed. For example, three-dimensional OCT data may be acquired based on the principle of line field OCT (hereinafter, referred to as “LF-OCT”). In the LF-OCT, the measurement light is simultaneously applied on an irradiation line extending in the one-dimensional direction in the tissue, and the interference light between the reflected light of the measurement light and the reference light is received by a one-dimensional light receiving element (for example, a line sensor) or a two-dimensional light receiving element. In the two-dimensional measurement region, scanning with the measurement light is performed in a direction intersecting the irradiation line, and thus the three-dimensional OCT data is acquired. The three-dimensional OCT data may be acquired based on the principle of full-field OCT (hereinafter, referred to as “FF-OCT”). In the FF-OCT, the measurement light is applied to the two-dimensional measurement region on the tissue, and the interference light between the reflected light of the measurement light and the reference light is received by a two-dimensional light receiving element. In this case, the OCT apparatus 10 may not include the scanning unit 14.
The front observation optical system 23 is provided for capturing a two-dimensional front image of the tissue of the subject (in the present embodiment, the fundus of the subject eye E) in real time. The front observation image in the present embodiment is a two-dimensional front image in a case where the tissue is viewed from the direction (front direction) along the optical axis of the measurement light of the OCT. In the present embodiment, a scanning laser ophthalmoscope (SLO) is employed as the front observation optical system 23. However, for the configuration of the front observation optical system 23, a configuration other than an SLO (for example, an infrared camera that collectively irradiates a two-dimensional imaging range with infrared light to capture a front image), may be employed.
The controller 30 performs various types of control of the OCT apparatus 10. The controller 30 includes a CPU 31, a RAM 32, a ROM 33, and a nonvolatile memory (NVM) 34. The CPU 31 is a controller that performs various types of control. The RAM 32 temporarily stores various types of information. The ROM 33 stores a program executed by the CPU 31, various initial values, and the like. The NVM 34 is a non-transitory storage medium capable of storing storage contents even in a case where the power supply is cut off. The controller 30 is connected to an operation unit 37 and a display device 38. As the operation unit 37 and the display device 38, various devices may be used in the same manner as the operation unit 107 and the display device 108 described above.
A method of capturing a fundus image in the present embodiment will be described. As shown in
The OCT apparatus 10 may acquire (generate) an Enface image 45 that is a two-dimensional front image in a case where the tissue is viewed from the direction (front direction) along the optical axis of the measurement light, based on the captured three-dimensional tomographic image 43. In a case where the enface image 45 is acquired in real time, the front observation optical system 23 may be omitted. Data of the enface image 45 may be, for example, integrated image data in which luminance values are integrated in a depth direction (Z direction) at respective positions in the XY direction, integrated values of spectral data at respective positions in the XY direction, and luminance data at each position in the XY direction in a certain depth direction, or luminance data at each position in the XY direction in any layer of the retina (for example, the surface layer of the retina). The Enface image 45 may be obtained from a motion contrast image (for example, an OCT angiography image) obtained by acquiring a plurality of OCT signals from the same position in the tissue of the patient's eye at different times.
(Structure of Layer/Boundary of Fundus)
With reference to
The layers of the fundus will be described. In the fundus, from the surface side (upper side in
As boundaries that arc likely to appear in tomographic images, for example, NFL/GCL (a boundary between the NFL and the GCL), IPL/INL (a boundary between the IPL and the INL), and OPL/ONL (a boundary between the OPL and the ONL), RPE/BM (a boundary between the RPE and the BM), and BM/choroid (a boundary between the BM and the choroid), are present.
(Mathematical Model Construction Processing)
A mathematical model construction processing performed by the mathematical model construction apparatus 101 will be described with reference to
In the following description, as an example, a case will be exemplified in which a mathematical model that outputs an identification result of at least any one (a specific layer/boundary that is the first site in the present embodiment) of a plurality of layers/boundaries captured in a fundus image, by analyzing an input two-dimensional tomographic image, is constructed. However, in the mathematical model construction processing, a mathematical model that outputs a result different from the identification result of the layer/boundary may be constructed. For example, a mathematical model that outputs a detection result of the end of the optic nerve head captured in the input two-dimensional tomographic image (details thereof will be described later) is also constructed by the mathematical model construction processing.
The mathematical model exemplified in the present embodiment is trained to output a distribution of scores indicating a probability that each site (each A scan image) in a two-dimensional tomographic image is the second site (the optic nerve head in the present embodiment), together with an identification result of the first site (specific layer/boundary in the present embodiment) captured in the input fundus image.
In the mathematical model construction processing, the mathematical model is constructed by training the mathematical model with a training data set. The training data set includes input side data (input training data) and output side data (output training data).
As shown in
Next, the CPU 103 performs training of the mathematical model using a training data set according to a machine learning algorithm (S3). As the machine learning algorithm, for example, a neural network, a random forest, boosting, and a support vector machine (SVM), are generally known.
The neural network is a technique that mimics the behavior of biological nerve cell networks. The neural network includes, for example, a feedforward neural network, a radial basis function (RBF) network, a spiking neural network, a convolutional neural network, a recursive neural network (a recurrent neural network, a feedback neural network, or the like), and a probabilistic neural network (a Boltzmann machine, a Basian network, or the like).
The random forest is a method of performing learning based on randomly sampled training data to generate a large number of decision trees. In a case where the random forest is used, the branch of plurality of decision trees learned in advance as a discriminator are traced, and an average (or a majority) of results obtained from the respective decision trees is taken.
The boosting is a method of generating a strong discriminator by combining a plurality of weak discriminators. The strong discriminator is constructed by sequentially learning a simple and weak discriminator.
The SVM is a method of constructing two classes of pattern discriminators by using a linear input element. The SVM learns parameters of the linear input element on the basis of, for example, a criterion (hyperplane separation theorem) of obtaining the margin maximizing hyperplane in which a distance to each data point is maximized from the training data.
The mathematical model refers to, for example, a data structure for predicting a relationship between input data (in the present embodiment, data of a two-dimensional tomographic image similar to the input training data) and output data (in the present embodiment, data of an identification result of the first site). The mathematical model is constructed by being trained with a training data set. As described above, the training data set is a set including input training data and output training data. For example, each piece of correlation data (for example, weights) between input and output is updated through training.
In the present embodiment, a multi-layered neural network is used as a machine learning algorithm. The neural network includes an input layer for inputting data, an output layer for generating data of an analysis result desired to be predicted, and one or more hidden layers between the input layer and the output layer. A plurality of nodes (also called units) are disposed in each layer. Specifically, in the present embodiment, a convolutional neural network (CNN) that is a kind of multi-layered neural network is used.
Other machine learning algorithms may be used. For example, generative adversarial networks (GAN) that utilize two competing neural networks may be employed as a machine learning algorithm.
The processing in S1 to S3 are repeatedly performed until the construction of the mathematical model is completed (S4: NO). In a case where the construction of the mathematical model is completed (S4: YES), the mathematical model construction processing is ended. In the present embodiment, a program and data for realizing the constructed mathematical model are incorporated in the fundus image processing apparatus 1.
(Fundus Image Processing)
Fundus image processing performed by the fundus image processing apparatus 1 will be described with reference to
As shown in
The CPU 3 performs image alignment, in the direction along the optical axis of the OCT measurement light (the Z direction in the present embodiment), of the three-dimensional tomographic image acquired in S1 (S2). In
As an example, in S2 of the present embodiment, image alignment in the Z direction is performed between a plurality of two-dimensional tomographic images configuring the three-dimensional tomographic image. For each of the plurality of two-dimensional tomographic images configuring the three-dimensional tomographic image, image alignment is performed between a plurality of pixel arrays (in the present embodiment, a plurality of A-scan images extending in the Z direction) configuring the two-dimensional tomographic image.
Instead of the processing in S2, the image alignment in the Z direction may be performed for a plurality of two-dimensional tomographic images extracted in S11 that will be described later. In this case, the detection accuracy of the end of the optic nerve head is improved. The image alignment processing may be omitted.
Next, the CPU 3 performs the site identification processing (S3). The site identification processing is a processing of identifying a specific site in the two-dimensional measurement region 40 (refer to
In a case where the automatic optic nerve head detection is successful (S5: YES), a reference position is set at a position of the detected optic nerve head (in the present embodiment, the center position of the optic nerve head automatically detected in S3) (S6). As shown in
On the other hand, in a case where the automatic optic nerve head detection fails (S5: NO), the CPU 3 sets a reference position at a position designated by the user (S7). In the present embodiment, the CPU 3 receives input of an instruction from the user in a state in which the fundus image (for example, a two-dimensional front image) of the subject eye that is an examination target, is displayed on the display device 8. In a case where the user inputs an instruction for designating a position via the operation unit 7, the CPU 3 sets a reference position at the designated position. In some cases, a reference position may be set through the processing in S7 without performing the processing in S3, S5, and S6.
Next, the CPU 3 sets a radial pattern centered on the reference position in the two-dimensional measurement region 40 (S10). As shown in
Next, the CPU 3 extracts a two-dimensional tomographic image 64 (refer to
The CPU 3 acquires a position of the end of the optic nerve head (the BMO in the present embodiment) in each of the plurality of two-dimensional tomographic images 64 extracted in S11 (S12). In the present embodiment, the CPU 3 inputs the two-dimensional tomographic image 64 into the mathematical model. The mathematical model is trained by using a machine learning algorithm to output a detection result of a position of the BMO captured in the input two-dimensional tomographic image. Specifically, as shown in
In the present embodiment, the position of the BMO automatically detected by using the machine learning algorithm is corrected according to an instruction from the user. Specifically, as shown in
Next, the CPU 3 performs a smoothing processing on the detection results of the plurality of positions detected based on the plurality of two-dimensional tomographic images 64, to detect a position of the annular end of the optic nerve head (an annular BMO in the present embodiment) (S13). As a result, even in a case where a position of the end is erroneously detected for some of the two-dimensional tomographic images 64, the influence of the erroneous detection is suppressed. As an example, in the present embodiment, a smoothing processing using a one-dimensional Gaussian filter is performed, on each dimension of XYZ of the detection results of the plurality of positions detected based on the plurality of two-dimensional tomographic images 64. A smoothing processing using a three-dimensional Gaussian filter may be performed on the plurality of probability maps 65, before a position of the BMO is detected. Elliptical fitting or the like for a plurality of detection results may be used for smoothing.
The CPU 3 specifies a center position of the optic nerve head based on the position of the end of the optic nerve head detected in S12 and S13 (S14). As an example, in the present embodiment, the CPU 3 specifies the detected position of the center of gravity of the annular BMO in the XY plane as the center position of the optic nerve head in the XY plane.
The CPU 3 displays the detected position of the end of the optic nerve head on the display device 8 (S20). In the present embodiment, as shown in
Next, as shown in
The CPU 3 extracts, from the three-dimensional tomographic image acquired in S1, a two-dimensional tomographic image in the annular line pattern 71 set in S21 (that is, an image into which a tomographic image that intersects the annular line pattern 71 in a cylindrical shape deformed in two dimensions) (S22).
The CPU 3 processes the two-dimensional tomographic image extracted in S22 to generate a layer thickness graph representing a thickness of a specific layer of the retina (for example, a thickness of the NFL or a thickness from the ILM to the NFL) captured in the two-dimensional tomographic image (S23).
The CPU 3 displays the layer thickness graph generated in S23 on the display device 8 in a state of comparison with the data of the normal eye (S24).
(Site Detection Processing)
The site detection processing performed by the fundus image processing apparatus 1 will be described with reference to
In general, a plurality of layers and boundaries are normally present around the optic nerve head, but specific layers and boundaries are missing at the position of the optic nerve head. Specifically, at a position where the optic nerve head is present, the NFL is present, and layers and boundaries at positions deeper than the NFL are missing. Based on the above findings, in the site detection processing of the present embodiment, the optic nerve head is identified based on the degree of deviation of a probability distribution in a case where the mathematical model identifies a specific layer/boundary (layer/boundary at a position deeper than the NFL).
As shown in
Next, the CPU 3 acquires a two-dimensional front image in a case where the fundus of which the three-dimensional tomographic image 43 acquired in S31 (or S1) is captured is viewed from the front (that is, the direction along the OCT measurement light) (S32). As an example, in S32 of the present embodiment, the Enface image 45 (refer to
The CPU 3 acquires an auxiliary identification result of the second site (the optic nerve head in the present embodiment), based on the two-dimensional front image acquired in S32 (S33). A method of auxiliary identification of the second site for the two-dimensional front image may be selected as appropriate. In the present embodiment, the CPU 3 identifies the optic nerve head by performing known image processing on the two-dimensional front image.
The CPU 3 extracts a part in which the second site (the optic nerve head in the present embodiment) is included with a high probability, from the entire three-dimensional tomographic image 43 acquired in S31 (or S1), based on the auxiliary identification result acquired in S33 (S34). As a result, an amount of subsequent processing is reduced, and thus the second site is detected more appropriately.
The CPU 3 extracts a T-th two-dimensional tomographic image (where an initial value of T is “1”), from among the plurality of two-dimensional tomographic images configuring the three-dimensional tomographic image extracted in S34 (S36).
By inputting the T-th two-dimensional tomographic image into the mathematical model, the CPU 3 acquires a probability distribution of coordinates at which an M-th (where an initial value of M is “I”) layer/boundary is present in each of the plurality of one-dimensional regions A1 to AN, as a probability distribution for identifying the first site (specific layer/boundary) (S37). The CPU 3 acquires the degree of deviation of a probability distribution related to the M-th layer/boundary (S38). The degree of deviation is a difference in the probability distribution acquired in S37 with respect to the probability distribution in a case where the first site is accurately identified. In a one-dimensional region where the first site is present, the degree of deviation tends to be small. On the other hand, in a one-dimensional region where the first site is not present, the degree of deviation tends to be large. This tendency is likely to appear regardless of the presence or absence of an cyc disease or the like.
In the present embodiment, the entropy of the probability distribution P is calculated as the degree of deviation. The entropy is given by the following (Equation 1). The entropy H(P) takes a value of 0≤H(P)≤log (number of events), and becomes a smaller value as the probability distribution P is biased. That is, the smaller the entropy H(P), the higher the identification accuracy of the first site tends to be. The entropy of the probability distribution in a case where the first site is accurately identified is 0.
H(P)=−Σp log(p) (Equation 1)
Next, the CPU 3 determines whether or not the degree of deviation of all layers/boundaries to be identified in the T-th two-dimensional tomographic image has been acquired (S40). In a case where the degree of deviation of some layers/boundaries is not acquired (S40: NO), “1” is added to the order M of layers/boundaries (S41), the processing returns to S37, and the degree of deviation of the next layer/boundary is acquired (S37, S38). In a case where the degree of deviation of all layers/boundaries has been acquired (S40: YES), the CPU 3 stores the degree of deviation of the T-th two-dimensional tomographic image in the storage device 4 (S42).
Next, the CPU 3 determines whether or not the degree of deviation of all the two-dimensional tomographic images configuring the three-dimensional tomographic image has been acquired (S44). In a case where the degree of deviation of some two-dimensional tomographic images is not acquired yet (S44: NO), “1” is added to the order T of the two-dimensional tomographic images (S45), the processing returns to S36, and the degree of deviation of the next two-dimensional tomographic image is acquired (S36 to S42).
In a case where the degree of deviation of all the two-dimensional tomographic images has been acquired (S44: YES), the CPU 3 acquires a two-dimensional distribution of a magnitude of the degree of deviation (hereinafter, simply referred to as a “deviation degree distribution”) in a case where the fundus is viewed from the front (S47). In the present embodiment, as shown in
The CPU 3 acquires a distribution of scores indicating a probability that each site (each A-scan image) is the second site (hereinafter, referred to as a “score distribution of the second site”) (S48). As described above, the score distribution of the second site is output from the mathematical model together with the identification result of the first site.
Next, the CPU 3 generates an identification result of the second site based on the degree of deviation in a case where the mathematical model identifies the first site (S49). In the present embodiment, as shown in
The techniques disclosed in the above embodiment are merely examples. Therefore, the techniques exemplified in the above embodiment may be changed. For example, the CPU 3 may detect a structure other than the optic nerve head in the fundus, based on a detection result of the end of the optic nerve head detected through the fundus image processing (refer to
In S3 of the fundus image processing (refer to
In S10 in
In the fundus image processing of the above embodiment (refer to
In the above embodiment, the site identification processing shown in
At the position where the fundus blood vessel (second site) is present, the measurement light is blocked by the fundus blood vessel, and thus an imaging state of a layer/boundary (first site) at a position deeper than the fundus blood vessel tends to deteriorate. Therefore, at the position where the fundus blood vessel is present, the degree of deviation related to identification of the layer/boundary at the position deeper than the fundus blood vessel is larger than that at a position where the fundus blood vessel is not present. On the basis of the above findings, in S47, a deviation degree distribution of at least any one of layers/boundaries at positions deeper than the fundus blood vessel may be acquired, as the degree of deviation related to analysis of the first site. In S49, a site having the degree of deviation more than a threshold value may be identified as a site (second site) of the fundus blood vessel.
For example, only some of the plurality of techniques exemplified in the above embodiment may be performed. For example, in S33 and S34 (refer to
The processing of acquiring a three-dimensional tomographic image in S1 in
The processing of acquiring a fundus image in S31 in
Number | Date | Country | Kind |
---|---|---|---|
2021-160684 | Sep 2021 | JP | national |