This application discloses systems, computer-readable media, and methods for the automated analysis and display of images of an anatomical breast to assist a human physician in the inspection of such images.
According to “Recent Advances in Breast Imaging, Mammography, and Computer-Aided Diagnosis of Breast Cancer”, SPIE Press, Bellingham, Wash., 2006, breast cancer is the most common type of cancer in women worldwide. While mammography coupled with physical examination is the current standard for breast cancer screening, one in five breast cancers may be missed by mammography screening, and when suspicious lesions are found and referred to biopsy, about four in five biopsies may turn out to be benign (i.e., “false positives”) and were thus arguably unnecessary. Mammography is the process of using low-dose amplitude-X-rays to examine the human breast. Missed detections and false positives may be attributed to several factors including poor image quality, improper patient positioning, inaccurate interpretation, fibro-glandular tissue obscuration, subtle nature of radiographic findings, eye fatigue, and/ or oversight.
Breast density has been acknowledged to be a factor in effective mammogram interpretation. For example, there is a consideration that mammographic imaging techniques are less successful with denser breast tissue than with predominantly fat breast tissue. Fibro-glandular tissue in dense breasts tend to attenuate x-rays to a greater degree than does fat tissue, leading to increased difficulty in detection of cancer sites for denser breasts. There may also be a strong correlation between breast tissue density and the risk of developing breast cancer. Thus, accurate classification of breast density is required for effective mammogram interpretation. Automated techniques for classifying breast density may be beneficial to radiologists, not only as a means for interpreting mammograms, but as an aid in establishing a strategy for follow-up care (e.g., additional imaging exams, biopsies, etc.).
Accurately determining the density of an anatomical breast can be a challenging task for a computer system due to the wide range of types of breasts that may be encountered in a clinical setting. While some breasts are heterogeneous (i.e., exhibiting a mixture of dense and fat tissue), many other breasts are homogeneous (i.e., exhibiting predominantly dense or predominantly fat tissue). (An example of such a breast is shown in
In view of the foregoing, various embodiments of the present disclosure are directed to methods for classifying and displaying breast density.
In particular, in a computer system comprising at least one processor, at least one input device, and at least one output device, a method of computing and outputting a density estimate of a breast, comprises: obtaining, by means of at least one input device, at least two digital images of at least a portion of the breast, wherein each image represents a view of at least a portion of the breast from a specific angle; computing, in at least one processor, a breast density estimate using information from the at least two digital images; and outputting, by means of at least one output device, the computed density estimate.
Computing the breast density estimate may comprise in at least one processor, for each digital image, computing at least one feature value; in at least one processor, for each digital image, computing an image breast density estimate using computed image feature values; and in at least one processor, computing the breast density estimate using computed image breast density estimates. At least one digital image may be a two-dimensional CC digital image and at least one digital image may be a two-dimensional MLO digital image; and at least one image breast density estimate may be computed by means of a cranio-caudal (CC) computer-based classifier, using computed image feature values of a CC digital image, and at least one image breast density estimate may be computed by means of a medio-lateral oblique (MLO) computer-based classifier, using computed image feature values of a MLO digital image. The CC computer-based classifier may comprise feature values that distinguish breasts of different densities projected from a cranio-caudal angle and the MLO computer-based classifier may comprise feature values that distinguish breasts of different densities projected from a medio-lateral oblique angle. At least two digital images of the breast may be tomographic images and each image breast density estimate may be computed by means of a tomographic image computer-based classifier, using computed image feature values of a tomographic digital image. Each tomographic computer-based classifier may comprise feature values that distinguish breasts of different densities projected from a specific tomographic angle. Computing the breast density estimate may comprise, in at least one processor, for each digital image, computing at least one feature value; and in at least one processor, computing the breast density estimate using computed image feature values. Computing the breast density estimate may further comprise using information from at least one digital image of at least a portion of a breast opposite to the breast. At least one digital image may represent a two-dimensional CC view of at least a portion of the breast, and at least one digital image may represent a two-dimensional MLO view of at least a portion of the breast. The images may be a plurality of tomographic images of at least a portion of the breast. The computed density estimate may comprise an estimate of whether the breast belongs to at least one of four predetermined breast density categories of entirely fatty, scattered fibro-glandular dense, heterogeneously dense, and extremely dense breasts.
In a computer system having at least one input device, at least one processor and at least one output device, a method of computing and outputting a density estimate of a breast comprises: obtaining, by means of at least one input device, at least one digital image of at least a portion of the breast, wherein each image represents a view of at least a portion of the breast from a specific angle; obtaining, by means of at least one input device, at least one digital image of at least a portion of a breast opposite to the breast, wherein each image represents a view of at least a portion of the opposite breast from a specific angle; computing, in at least one processor, a breast density estimate using information from the at least one digital breast image and at least one digital opposite breast image; and outputting, by means of at least one output device, the computed density estimate.
Computing the breast density estimate may comprise: in at least one processor, for each digital breast image, computing at least one feature value using the said digital breast image and a digital opposite breast image; in at least one processor, for each digital breast image, computing an image breast density estimate using computed image feature values; and in at least one processor, computing the breast density estimate using computed image breast density estimates. Computing the breast density estimate may comprise: in at least one processor, for each digital breast image, computing at least one feature value using the said digital breast image and a digital opposite breast image; and in at least one processor, computing the breast density estimate using computed image feature values. An asymmetrical subtraction of information relating to the digital opposite breast image from information relating to the digital breast image may be performed. At least one digital breast image may represent a two-dimensional CC view of at least a portion of the breast, and at least one digital breast image may represent a two-dimensional MLO view of at least a portion of the breast. The digital breast images may be tomographic images of at least a portion of the breast. The computed density estimate may comprise an estimate of whether the breast belongs to at least one of four predetermined breast density categories of entirely fatty, scattered fibro-glandular dense, heterogeneously dense, and extremely dense breasts.
A medical imaging system comprises: a source configured to obtain digital images of breasts; a processor coupled with the source configured to compute a density estimate of a breast using information from at least two digital images, wherein a first digital image represents a view of at least a portion of the breast from a specific angle and wherein a second digital image is chosen from a group consisting of a further view of at least a portion of the breast from a second specific angle, and a view of at least a portion of an opposite breast from the specific angle; and an output device coupled with the processor configured to output the computed density estimate.
The source may be configured to obtain a plurality of tomographic images of at least a portion of the breast and the processor may be configured to compute the density estimate using tomographic images. The processor may be further configured to compute a plurality of reconstructed slices from the plurality of tomographic images and to compute the density estimate using reconstructed slices.
In a computer system having at least one input device, at least one processor and at least one output device, a method of computing and outputting a density estimate of a breast comprises: obtaining, by means of at least one input device, at least one digital image of at least a portion of the breast; computing, in at least one processor, parenchyma information relating to the breast using texture information and density information derived from the at least one digital image; computing, in at least one processor, a breast density estimate using computed parenchyma information; and outputting, by means of at least one output device, the computed density estimate.
The parenchyma information may be computed for individual pixels of the digital image. Parenchyma information for a specific area of the breast may be computed based in part on the location of the area in the breast. Density information may be given a stronger weighting than texture information in computing parenchyma information. Parenchyma information may be computed further using texture information and density information derived from at least one digital image of at least a portion of an opposite breast. A digital representation of at least a portion of the breast may be segmented into breast parenchyma and breast non-parenchyma using computed parenchyma information. The digital representation may be segmented by thresholding the computed parenchyma information. The breast density estimate may be computed using feature values of segmented breast parenchyma. The breast density estimate may be computed further using feature values of segmented breast non-parenchyma. The breast density estimate may be computed using feature values of computed parenchyma information. The computed density estimate may comprise an estimate of whether the breast belongs to at least one of four predetermined breast density categories of entirely fatty, scattered fibro-glandular dense, heterogeneously dense, and extremely dense breasts.
In a computer system having at least one input device, at least one processor and at least one output device, a method of computing and outputting a density estimate of a breast comprises obtaining, by means of at least one input device, at least one digital image of at least a portion of the breast; computing, in at least one processor, vessel line information from the at least one digital image; computing, in at least one processor, parenchyma information from the at least one digital image, using computed vessel line information; computing, in at least one processor, a breast density estimate using computed parenchyma information; and outputting, by means of at least one output device, the computed density estimate. Parenchyma information may be computed by means of treating computed vessel line information as non-parenchyma.
Other aspects of the present disclosure are computer-readable media having computer-readable signals stored thereon. The computer-readable signals define instructions which, as a result of being executed by a computer or computer system, instruct the computer or computer system to perform one or more of the methods disclosed herein. That is to say, the computer-readable medium has the said instructions stored therein.
Yet other aspects of the present disclosure are computers or computer systems having at least one processor, at least one input device, and at least one output device. The computers or computer systems may include or may facilitate the use of computer-readable media with instructions stored therein which, as a result of being executed by the computers or computer systems, instruct the computer or computer system to perform one or more of the methods disclosed herein.
It should be appreciated that all combinations of the foregoing concepts and additional concepts discussed in greater detail below are contemplated as being part of the inventive subject matter disclosed herein.
In the following detailed description of embodiments, reference is made to the accompanying drawings that form a part hereof, and in which are shown, by way of illustration and not by way of limitation, specific embodiments in which the methods and systems disclosed herein may be practiced. It is to be understood that other embodiments may be utilized and that logical, mechanical, and electrical changes may be made without departing from the scope of the methods and systems disclosed herein.
This disclosure is directed to computer systems, computer-readable media, and methods for the automated classification and display of breast density in digital mammographic imaging.
System 100 of
Image acquisition unit 115 is representative of a system capable of acquiring digital mammographic images of anatomical breasts and, in certain embodiments, capable of transmitting digital data representing such mammographic images to image viewing station 110 for further processing. For example, image acquisition unit 115 may be a computed radiographic (CR) mammography system such as those offered by the AGFA Healthcare of Ridgefield Park, N.J. (AGFA); or Fujifilm Medical Systems of Stamford, Conn. (Fuji); a digital radiographic (DR) mammography system such as those offered by the General Electric Company of Fairfield, Conn, (GE); or a tomographic mammography system, such as a digital breast tomosynthesis (DBT) imaging system offered by GE; Hologic, Inc. of Bedford, Mass. (Hologic); or Siemens AG of Munich, Germany (Siemens).
In certain embodiments, image acquisition unit 115 may connect to and communicate with a digitizer apparatus 120, such as a laser scanner with approximately 50 micron resolution, for digitizing developed x-ray mammograms of anatomical breasts. Such x-ray mammograms may be produced as films by image acquisition unit 115 and require digitizing prior to the execution of the methods disclosed hereinbelow by image viewing station 110. Image acquisition unit 115 and/or image viewing station 110 may connect to and communicate with digitizer apparatus 120 via any type of communication interface as described hereinabove. In such embodiments, the interface may function as the input device through which is obtained the digitized image for processing by the methods described herein.
Image viewing station 110 is representative of a general purpose computer system containing instructions for analyzing the medical breast imagery and outputting the medical imagery and/or breast density classification information that may be automatically derived from the medical breast imagery in the form of data. Image viewing station 110 may further comprise a processor unit 122, a memory unit 124, an input interface 126, an output interface 128, and program code 130 containing instructions that can be read and executed by the station. Input interface 126 may connect processor unit 122 to an input device such as a keyboard 136, a mouse 138, a means for acquiring the images for processing as described hereinabove, and/or another suitable device as will be known to a person of skill in the art, including for example and not by way of limitation a voice-activated system. Thus, input interface 126 may allow a user to communicate commands to the processor as well as to acquire images. One such exemplary command is the execution of program code 130 tangibly embodying the automated breast density classification methods disclosed herein. Output interface 128 may further be connected to processor unit 122 and an output device such as a graphical user interface (GUI) 140. Thus, output interface 128 may allow image viewing station 110 to transmit data from the processor to the output device, one such exemplary transmission including medical imagery and breast density classification data for display to a user on GUI 140.
Memory unit 124 may include conventional semiconductor random access memory (RAM) 142 or other forms of memory known in the art; and one or more computer readable-storage mediums 144, such as a hard drive, floppy drive, read/write CD-ROM, tape drive, flash drive, optical drive, etc. Stored in program code 130 may be an image reconstruction unit 146 for constructing additional imagery from the images acquired by image acquisition unit 110 in accordance with certain embodiments of the methods disclosed herein. While image reconstruction unit 146 is depicted as being a component within image viewing station 120, one skilled in the art will appreciate that image reconstruction unit 146 may also be deployed as part of one or more separate computers, computer processors, or computer systems. For example, image reconstruction unit 146 may be deployed as part of a review workstation system that constructs additional tomographic breast imagery from direct projections acquired by a DBT imaging system. One example of a review workstation system that performs such acts is the DexTop Breast Imaging Workstation, offered by Dexela Limited, London, United Kingdom.
While image viewing station 110, image acquisition unit 115, and digitizer apparatus 120 are depicted as being separate components within system 100, one skilled in the art will appreciate that any combination of such components may be deployed as parts of a single computer, computer processor, and/or computer system.
Referring now to
In certain embodiments, digital mammographic imagery acquired at step 210 may comprise a single digital mammographic image or “view” of an anatomical breast under study. For example,
In certain other embodiments, the digital mammographic imagery acquired at step 210 may comprise a plurality of digital mammographic images or views of an anatomical breast under study. As will be further discussed hereinbelow, the density estimation of the anatomical breast may be improved by introducing information regarding the breast tissue as projected from different angles. For example, digital mammographic imagery acquired at step 210 may comprise both a CC image and a MLO image of an anatomical breast under study. By way of another example, digital mammographic imagery acquired at step 210 may comprise a plurality of digital tomographic mammography images of an anatomical breast, such as those that are acquired using a DBT imaging system. For example, DBT imaging systems may acquire 10 digital tomographic mammography images (i.e., “direct projection images” or “slices”) of a single anatomical breast composing digital mammographic imagery by moving a source at 4 degree increments around a plane. By way of yet another example, digital mammographic imagery acquired at step 210 may comprise a synthetic CC and a synthetic MLO mammographic image, both of which may be computed from a plurality of digital tomographic mammography images of an anatomical breast. Such synthetic mammographic imagery may be an advantageous way to visually present both the anatomical breast under study and the breast density classification data to radiologists who are accustomed to reviewing conventional CC and MLO mammography images, yet wish to realize the advantages of tomographic imaging procedures. Of course, the multiple views and images described herein are exemplary only, and other multiple breast images may be utilized by the methods and systems disclosed herein.
In yet certain other embodiments, digital mammographic imagery acquired at step 210 may comprise at least one image of an anatomical breast under study and at least one image of the patient's opposite breast. As will be further discussed herein below, the density estimation of an anatomical breast may be improved by studying the tissue of both breasts. For example, and not by way of limitation, the acquired digital mammographic imagery may comprise a CC and an MLO image of a patient's right anatomical breast and a CC and an MLO image of the patient's left anatomical breast.
There may be many other suitable ways known to persons of skill in the art to acquire digital mammographic imagery not presented by way of example in which the advantages of the disclosure may be achieved, and those are intended to be encompassed within this disclosure.
Digital mammographic imagery acquired at step 210 is then processed by image viewing station 110 at step 220 to compute breast density classification information associated with the anatomical breast under study. According to an embodiment of this disclosure, prior to computing breast density classification information at step 220, the digital mammographic imagery acquired at step 210 may be inverted or transformed to make tissue that attenuates more x-rays appear bright and tissue which x-rays pass through appear dark. The transformed imagery may also be sub-sampled so as to gain computational efficiencies without losing critical diagnostic information.
Referring now to
An anatomical breast can be divided into two major tissue categories or components: fibro-glandular tissue and fat tissue. The density of an anatomical breast is described by the relative amount of fibro-glandular tissue present in the breast. Therefore, dense breasts will have more fibro-glandular tissue while fatty breasts will have less. The term “parenchyma” is used to describe the fibro-glandular tissue in digital mammographic imagery. The amount of parenchyma within a mammogram is directly proportional to the amount of fibro-glandular tissue present in the breast. Thus, highly accurate identification of the breast parenchyma is required in order to characterize the fibro-glandular breast tissue from fatty breast tissue.
Parenchyma will only appear within the actual anatomical breast, yet digital mammographic imagery acquired at step 210 will typically include areas outside of the anatomical breast region that are of diagnostic insignificance. Thus, according to an embodiment of the disclosure, image data representing the anatomical breast region may be first identified from other regions in the acquired digital mammographic imagery, such as but not limited to the foreground and background regions of the image. The anatomical breast region may be identified using any number of suitable automated image processing techniques, which are advantageous over requiring a radiologist or other user to manually identify such a region. One non-limiting example of a suitable image processing technique is a region growing method, such as, but not limited to, the acts disclosed in U.S. Pat. No. 6,091,841 “Method and system for segmenting desired regions in digital mammograms” assigned to Qualia Computing, Inc., which is incorporated herein by reference.
Some tissue of the anatomical breast may be obscured by the pectoral muscle in images taken at certain projection angles, such as in MLO images. Recognizing that it may be extremely difficult to distinguish the underlying nature of this breast tissue from the pectoral muscle due to the projection of the pectoral muscle tissue, according to an embodiment of the disclosure, image data representing the pectoral muscle may be identified and excluded from further analysis as potential breast parenchyma. The pectoral muscle may be identified using any number of suitable automated image processing techniques, which are advantageous over requiring a radiologist or other user to manually identify such a region. Several non-limiting examples of suitable image processing techniques include Hough transforms, self-organizing neural networks, Russ operators, and tunable parametric edge detection. Recognizing that such techniques may never be 100% accurate and that use of any of the pectoral muscle area in further analysis of breast parenchyma may be disadvantageous, according to a further embodiment of the disclosure, a boundary defining the areas between the breast and pectoral muscle may be thickened using, for example, a morphological operation. Other methods may be used.
Tissue at the margin of a breast may not be as thick as the surrounding, compressed tissue and thus, the pixels along the margin may tend to “roll-off' or dim in the imagery. A visual example of this can be seen in
In dense breasts, an example of which is illustrated in
Referring now to
Turning first to the texture map created at step 810, according to an embodiment of the disclosure, the texture or wispiness of the anatomical breast under study may be characterized by measuring local changes in pixel intensities across the breast region, or at least a statistically significant portion of the breast region. Local regions of the breast with a high range of intensities indicate an area with significant texture or wispiness, while regions with a low range of intensities indicate a homogenous or non-wispy region. In certain embodiments, a range filter may be applied to the breast region to measure local changes in pixel intensities. The kernel size of the range filter may be chosen empirically, such that the filter effectively characterizes the fibro-glandular tissue expected to exhibit such texture or wispiness. A 3×3 kernel may suitably output both a maximum intensity value under the kernel and a minimum intensity value under the kernel, which may then be differenced to compute a single local intensity range at each pixel. Kernels of other sizes may also be used. Pixels with a significantly high intensity value (e.g., greater than 2 standard deviations above the mean intensity of all pixels under the kernel) may be normalized (e.g., to 2 standard deviations above the mean intensity of all pixels under the kernel) to normalize noise that may be incorrectly evaluated as wispy tissue. Other cutoffs than 2 standard deviations may be used. In other embodiments, a wavelet filter, a Gabor filter, a spatial grey-level dependence (SGLD) method, a Laws method, or a gradient analysis method may be applied to the breast region to measure local changes in pixel intensities and suitably characterize texture. Other methods known to persons of skill in the art also may be utilized to measure local changes in pixel intensities and suitably characterize texture, and the methods may be used singly or in combination. The response of each local region to the texture characterization may be represented in the form of the texture map created at step 810.
Turning next to the density map created at step 820, independent from characterizing the texture or wispiness of the anatomical breast under study, the density of the anatomical breast under study may be characterized by, according to another embodiment of the disclosure, measuring pixel intensity differences inside the breast region, or at least a statistically significant portion of the breast region. Breast pixels that have a relatively high intensity value when compared against an estimate of the background (i.e., fatty tissue) intensity value indicate regions with significant density.
There are numerous ways of estimating the background intensity value inside the breast region. One way is to form a background value estimation region suspected to contain mostly fatty tissue and measure a global pixel intensity value from this region. According to an embodiment of the disclosure, a background value estimation region may be formed by thresholding a weighted distance map evaluated for pixels inside the breast region. Assuming a distance map is formed from the breast region, each pixel value may be normalized between 0 and 1 and then weighted by raising to 0.1 power. Then, pixel values may be thresholded at values greater than 0.9 to form a suitable background value estimation region containing mostly fatty breast tissue.
The background intensity value estimate may then be used to measure pixel intensity differences by forming an intensity difference image (i.e., an intensity difference map). According to an embodiment of the disclosure, a suitable intensity difference map may be formed by subtracting the computed background intensity value estimate from the value at each breast pixel. Negative values may be set to 0. To account for noise, breast pixels having difference values that are above two standard deviations above the mean may be set equal to two standard deviations above the mean. (Other values than two standard deviations may be used as cutoffs.)
According to an embodiment of the disclosure, a binary thresholding operation may be performed on the intensity difference map to set each pixel to either a dense tissue pixel or a non-dense tissue pixel, thus forming a density map. The threshold may be dynamically set equal to the background intensity value estimate plus the standard deviation of the intensity of the breast region. Other cutoffs may be used. The response of each pixel to the thresholding operation may be represented in the form of the density map created at step 820.
In other embodiments of the disclosure, a suitable density map may be created at step 820 by modeling a histogram of the distribution of the gray-level pixel intensities inside the breast region and computing a threshold that separates an estimate of the dense parenchyma pixels from fatty pixels. For example, a Mixture of Gaussians algorithm may be performed that assumes a Gaussian distribution of higher intensity values represents dense tissue pixels and a Gaussian distribution of lower intensity values represents fatty tissue pixels.
In accordance with an embodiment of the disclosure, the texture information computed at step 810 may be input into the density map calculation at step 820 to further refine density estimates. For example, in embodiments where a Mixture of Gaussians algorithm may be performed, texture information may be input as a free variable or “dimension” into the Gaussian distribution estimate.
In accordance with further embodiments of the disclosure, the computed threshold used to separate fatty pixels from dense pixels may be compared against a background intensity or density estimate (e.g., extracted from the segmented chest wall area of the image). If the computed threshold is similar to the background estimate, this may indicate that the range in pixel intensity differences between fatty and dense tissue is minimal and that the breast under study is more likely to be entirely fatty or entirely dense. The results of such a comparison may be stored for later use and introduced (e.g., as a feature) into density classification estimates described herein below.
According to an embodiment of the disclosure, to emphasize range in darker tissue regions of the breast, the texture map created at step 810 may be re-weighted prior to combining texture and density information. This may be achieved, for example, by re-weighting pixels in said texture map in accordance with the inverted pixel values of the intensity difference map that may be used, in certain embodiments, to form the density map created at step 820 as described hereinabove. Empirically, it has been found that by re-weighing pixels of the texture map from step 810 that appear within an intensity range of 0.25-0.5 in the intensity difference map by a power of 0.75 and by re-weighting pixels of that texture map that do not appear within this intensity range in the intensity difference map by a power of 0.25, darker tissue regions of the breast may be suitably emphasized.
According to an embodiment of the disclosure, to emphasize areas of extreme denseness in the breast, the density map created at step 820 may be re-weighted prior to combining texture and density information. This may be achieved, for example, by re-weighting pixels in that density map in accordance with the thresholded pixel values of the intensity difference map that may be used, in certain embodiments, to form the density map created in step 820 as described hereinabove. Empirically, it has been found that by thresholding the intensity difference map at 0.975 and setting the pixels of that density map that are both in the thresholded intensity difference map and have less than a 0.4 intensity value in the texture map created in step 810, areas of extreme denseness can be suitably emphasized.
Having separately characterized the texture and density of the anatomical breast under study, information concerning both the texture in the texture map and the density in the density map may be combined to characterize a probability or likelihood whether each pixel in the breast region is representative of breast parenchyma and form a tissue map at step 830. According to an embodiment of the disclosure, because density may be a better overall descriptor than texture as to the nature of breast tissue, a probability map may be formed by linearly combining pixels of the texture map with a weight of 40% and pixels of the density map with a weight of 60%.
Various rules may then be applied to characterize whether each pixel in the breast region is a parenchyma tissue pixel or a fatty (i.e., non-parenchyma) tissue pixel. According to an embodiment of the disclosure, a distance map may be used to set candidate parenchyma pixels towards the skin line to non-parenchyma (e.g., binary ‘0’) if the probability map also indicated a low likelihood (e.g., less than 0.4) that such pixels represent parenchyma tissue. This rule combines texture and density information with a distance to the breast skin line measurement to characterize fatty breast pixels. Of course, other likelihoods may be used as a cutoff, and other rules known to those of skill in the art may be applied.
A binary thresholding operation may be performed on the probability map to set each pixel to either a parenchyma tissue pixel or a non-parenchyma tissue pixel. Noting that the dynamic range of intensities in heterogeneous breasts will typically be greater than the dynamic range of intensities in homogenous breasts, according to an embodiment of the disclosure, the threshold value may be determined dynamically based on a measurement characterizing the area of denseness with respect to the total area of the breast. For example, if the area of density as represented by density map 820 is large when compared with the breast region (e.g., greater than 75% but less than 95%), there is less dynamic range in the image and a higher threshold value may be required for the system to characterize a pixel to be parenchyma tissue. For example, a threshold value of 0.65 may be used to label pixels in such breasts to either a parenchyma tissue pixel or a non-parenchyma tissue pixel; otherwise, a threshold value of 0.55 may be used. Of course, other values may be used as a threshold value, and other factors may be used to adjust the value. By using a dynamic threshold, a more accurate representation of the parenchyma and non-parenchyma tissue can be achieved.
According to an embodiment of the disclosure, pixels representing small objects (e.g. less than 250 mm2) may also be set to 0 to avoid misclassifying small, high-intensity structures that may be pockets of parenchyma that are outside of the breast parenchyma disk or cancerous lesions. The latter may be required when testing the methods disclosed herein on samples of anatomical breasts having known cancerous lesions. Of course, other small-size cutoffs may be used.
It has been found that in fatty breasts such as the breast illustrated in
It has also been found that when an anomaly such as a large cancer appears in the breast, these objects are also frequently segmented as part of the parenchyma tissue map. Such objects are also not statistically reflective of the characteristics of the breast parenchyma and thus, when introduced into the density classification, may contribute to misclassification. In accordance with one embodiment of the disclosure, this problem may be dealt with by comparing the parenchyma tissue map against a parenchyma tissue map extracted from another view of the patient's opposite breast at step 850. This process is typically called an asymmetric comparison. A region of parenchyma in one breast that exhibits substantially different characteristics from the corresponding parenchyma region in the other breast indicates a high likelihood that the region is likely an anomaly, not parenchyma. Such regions may be subtracted from the computed parenchyma tissue map before additional breast density classification processing is performed. In accordance with an embodiment of the disclosure, this step may be achieved by registering the single-view mammogram against a corresponding mammogram and performing a difference analysis (e.g., subtraction of one computed parenchyma tissue map from another map, followed by a thresholding operation) to compute asymmetric differences.
Again referencing
According to an embodiment of the disclosure, a histogram of intensity features of the tissue map formed at step 410 may be computed as part of calculating the plurality of feature values at step 420. For example, features describing the intensity of the non-parenchyma pixels may be computed such as, but not limited to, the dynamic range, the standard deviation, the skewness, and the kurtosis of the intensity of the non-parenchyma pixels. Ratio features describing the relationship between parenchyma and non-parenchyma pixel intensity features may be computed such as, but not limited to, the ratio of the median, the maximum, the minimum, and/or the standard deviation of the parenchyma pixel intensities to the ratio of the median, the maximum, the minimum, and/or the standard deviation of the non-parenchyma pixel intensities. Of course, values of other intensity characteristics may be utilized as well.
According to an embodiment of the disclosure, texture features of the tissue map formed at step 410 may be computed as part of determining the plurality of feature values at step 420. For example, features describing the intensity of the parenchyma pixels may be computed such as, but not limited to, the dynamic range, the standard deviation, the skewness, and the kurtosis of the intensity of the parenchyma pixels. A feature describing the number of “holes” in the image (i.e., regions of non-parenchyma that appear within the parenchyma region—may be set to a minimum number of pixels such as 5) may also be computed to characterize texture. Of course, values of other texture characteristics may be utilized as well.
According to an embodiment of the disclosure, shape or morphological features of the tissue map formed at step 410 may be computed as part of determining the plurality of feature values at step 420. For example, features describing the shape or morphology of the parenchyma pixel region may be computed such as, but not limited to, the number of parenchyma pixel region objects. Features describing the shape or morphology of the non-parenchyma pixels may be computed such as, but not limited to, the total area (i.e., size) of the non-parenchyma pixels. Ratio features describing the relationship between parenchyma and non-parenchyma pixel shape features may be computed such as, but not limited to, the ratio of the total area of the parenchyma to the non-parenchyma region and the ratio of the area of the parenchyma to the non-parenchyma region according to various quadrants (e.g., right, left, top, bottom) of the image. Of course, values of other morphological characteristics may be utilized as well.
Features known to a person of skill in the art, other than intensity features, texture features and shape or morphological features may be used.
At classification step 430, feature values calculated at step 420 are compared against classification parameters representing tissue characteristics of anatomical breasts of different densities. Such classification parameters may be stored in memory unit 124, for example. Based on this comparison, a breast density estimate is computed and assigned to the anatomical breast. The goal of classification is to group items that have similar feature values into groups and thus, the goal at classification step 430 is to group the anatomical breast under study into a density category based on feature values determined at step 420. According to embodiments of the present disclosure in which the anatomical breast may be classified to one of four breast density categories in accordance with ACR BI-RADS guidelines, the classification parameters stored in memory unit 124 represent feature characteristics of the tissue of entirely fatty breasts (0-25% glandular); scattered fibroglandular dense breasts (25-50% glandular); heterogeneously dense breasts (50-75% glandular); and extremely dense breasts (75-100% glandular). Of course, if only two classes are used for classification purposes, a suitable divide between the classes may be established, and classification systems utilizing other numbers of classes may also be used.
According to an embodiment of the disclosure, a plurality of different sets of classification parameters may be stored in memory unit 124. Each classification parameter set may correspond to the characteristics of anatomical breasts of various densities as imaged from a particular image angle. For example, and not by way of limitation, a set of classification parameters derived from CC images of anatomical breasts of different densities may be selected to classify feature values extracted from a CC image, while a set of classification parameters derived from MLO images of anatomical breasts of different densities may be selected to classify feature values extracted from a MLO image. By way of another example, again not by way of limitation, each classification parameter set may correspond to the characteristics of anatomical breasts of various densities as imaged from a specific digital tomographic mammography imaging angle (i.e., “a direct projection angle”) and/or by a specific imaging technique. Dynamic selection of classification parameters may be advantageous because certain fibro-glandular and/or fat tissue characteristics may be more descriptive based on the angle from which the anatomical breast is imaged, and/or the imaging technique used. For example, it has been found that the ratio of median intensity values between fibro-glandular and fat tissue may be a descriptive feature in characterizing breast tissue in CC images. It has also been found that the number of regions segmented as part of the parenchyma may be a descriptive feature in characterizing scattered fibroglandular dense breasts from heterogeneously dense breasts in MLO images. These feature characteristics are merely presented as examples. Different feature characteristics may be more descriptive if the anatomical breast is characterized at different angles, different projection depths, or different resolutions, for example, or if different types of images are analyzed.
According to certain embodiments, the image content may be analyzed to automatically derive information associated with the angle at which the anatomical breast is projected (e.g., whether the image is a CC or a MLO view). For example, and not by way of limitation, three features may be computed on the segmented breast region: the ratio of the width of the top of the segmented breast region to the overall width of the segmented breast region; an error estimate that indicates how well the breast region contour is fit by a parabola; and the fractional overlap of the top and bottom portions of the segmented breast region when a map or mask of the region is “folded over” at the widest row of the breast region. The first feature may be used because CC view masks are typically much narrower at the top than are MLO view masks. The second feature may be used because CC view masks tend to have a parabolic shape, while MLO view masks usually have a more complicated shape. The third feature may be used because CC view masks typically are much more symmetric about the horizontal line containing the widest part of the mask than are MLO view masks. After computing the three features, a two-class difference-of-discriminants classifier may be applied to the feature data to decide whether the image is a CC view or an MLO view. A CC is class 1, and has a non-negative difference of discriminants, and an MLO is class 0, and has a negative difference of discriminants. The algorithm may use the breast direction and the CC-MLO classification result to classify CC or MLO view and/or right or left breast. Alternatively, other techniques may be used to decide if the image is a CC or MLO, or other, view. Alternatively, parameters associated with the image may be read from a file header associated with the digital mammography imagery acquired at step 210 and used to select an appropriate set of classification parameters from memory at classification step 430.
According to an embodiment of the disclosure, a decision tree classifier may be employed at classification step 430 to compare the classification parameters retrieved from storage against feature values calculated at step 420 and, based on this comparison, compute and assign a breast density class for the anatomical breast. Decision tree classifiers may be advantageous in that they may correctly classify breast density with a substantially high accuracy while at the same time, the rules used to achieve such classification are simple to understand and interpret. Thus, a physician who may be interested in understanding the behavior of the breast density classification system and methods disclosed herein may benefit from such classifiers. However, any number of classification algorithms or combination of classification algorithms (e.g., committees) known in the art such as, but not limited to, a linear classifier, a quadratic classifier, a neural network classifier, a decision-tree classifier, a fuzzy logic classifier, a support vector machine (SVM) classifier, a Bayesian classifier, a k-nearest neighbor classifier, or a syntactical classifier may also be used to perform classification 430. (See Pattern Classification, Duda et al., John Wiley & Sons, New York, October 2000.)
According to an embodiment of the disclosure, the classification rules presented below may be performed on CC digital mammographic images. However, these are provided by way of non-limiting example only. Other rules may be sued for such images, and for other images the actual classification rules that may be developed will depend on the combination of sample digital mammographic images, feature values, and classification algorithm or algorithm(s) implemented for performing classification step 430.
By way of non-limiting example, for CC digital mammographic images a ratio feature value describing the median intensity values of parenchyma to non-parenchyma breast pixels may be used in combination with, for example, features describing the distribution (e.g., the standard deviation or the kurtosis) of intensity values of the parenchyma pixels alone and/or non-parenchyma pixels alone in distinguishing an entirely fatty breast (BI-RADS density category 1) from a scattered fibro-glandular dense breast (BI-RADS density category 2). Entirely fatty breasts will exhibit either a low distribution of parenchyma pixel intensity values or a high distribution of non-parenchyma pixel intensity values, while scattered fibro-glandular dense breasts will typically exhibit the opposite behavior. A high median intensity ratio feature used in combination with, for example, low intensity skewness features of both parenchyma and non-parenchyma breast pixels may be useful in distinguishing an extremely dense breast (BI-RAD density category 4) from other types of breasts.
According to an embodiment of the disclosure, the classification rules presented below may be performed on MLO digital mammographic images. However, these are also provided by way of non-limiting example only. Other rules may be used for such images. If the kurtosis distribution of intensity features of both the parenchyma and non-parenchyma pixels are both lower than predetermined classification thresholds, the breast may be classified as a dense breast (BI-RADS density categories 3 and 4). A skewness distribution of intensity features of the non-parenchyma pixels may further be compared against a predetermined threshold in which if the threshold is not met, the breast may be classified as an extremely dense breast (BI-RADS density category 4) and if the threshold is met, the breast may be classified as a heterogeneously dense breast (BI-RADS density category 3). If the kurtosis distribution of intensity feature of the non-parenchyma pixels is lower than the predetermined classification threshold and the kurtosis distribution of intensity feature of the parenchyma pixels is higher than the predetermined classification threshold, additional classification may be performed to determine breast density. For example, a skewness distribution of intensity features of the non-parenchyma pixels may further be compared against a predetermined threshold in which if the threshold is not met, the breast may be classified as extremely dense (BI-RADS density category 4) and if the threshold is met, the breast may be classified as a fatty breast (BI-RADS density categories 1 and 2). The ratio of the total area of the non-parenchyma pixels to parenchyma pixels may be compared against a predetermined threshold in which if the threshold is not met, the breast may be classified as entirely fatty (BI-RADS density category 1) and if the threshold is met, the breast may be classified as a scattered fibro-glandular density (BI-RADS density category 2).
According to an embodiment of the disclosure, the classifier may be designed such that both a “hard” and “soft” breast density classification decision is computed at classification step 430, or such that only a “hard” or a “soft” classification decision is made. The “hard” breast density classification decision may indicate one breast density category in accordance with the BI-RADS guidelines described hereinabove). The “soft” breast density classification decision may indicate a probability (e.g., on a 0-100% scale) or degree to which the breast under study exhibits the characteristics of a dense or fatty breast. This probability may be expressed as a percentage describing the amount of fibro-glandular or glandular tissue, which may be determined based on the percentage area of parenchyma with respect to the percentage area of the breast. Alternatively, the amount of fibro-glandular or glandular tissue may be further based upon the “hard” breast density classification decision output by the classifier. For example, if the computed percentage describing the amount of fibro-glandular or glandular tissue is 50% and the classifier density output is a scattered fibro-glandular dense breast (BI-RADS density category 2), a final percentage glandular tissue output may be 39%, which is half-way between the percentage ranges encompassed within this BI-RADS density category. Alternatively, if the classifier density output is a scattered fibro-glandular dense breast (BI-RADS density category 2), a distance measurement (e.g., the Mahalanobis distance) may be used to measure the similarity between the features of the anatomical breast under study and the features of anatomical breasts labeled to the same BI-RADS density category to compute a final percentage glandular tissue estimate.
Referring now to
A plurality of feature sets may then be calculated that characterize the fibro-glandular tissue and fat tissue in each tissue map. This may be advantageous because certain fibro-glandular tissue and/or fat tissue characteristics may be more descriptive based on the angle from which the anatomical breast is imaged. For example, by way of a non-limiting example presented in
At multi-view classification step 1730, the plurality of feature sets are then compared against a single set of classification parameters stored in memory unit 124 and based on this comparison, a breast density estimate is computed and assigned to the anatomical breast. Note that at multi-view classification step 1730, in contrast to classification step 430, the classification parameters in the single set represent characteristics of the fibro-glandular and fat tissue of anatomical breasts of different densities as imaged from a plurality of angles or image sources. For example, features and classification parameters describing the intensity of the fibro-glandular breast tissue from both the CC and MLO views, when taken together, may be more descriptive of the actual breast tissue than features and classification parameters describing such tissue from a single image view alone. Features and classification parameters describing the intensity of the fibro-glandular breast tissue using a plurality of projection views acquired using tomography may be more descriptive of the actual breast tissue than features acquired from a single image view.
Then, a case-based breast density estimate may be computed at step 1820 for the anatomical breast by statistically combining the individual, single-view breast density estimates. According to an embodiment of the disclosure, the mode of the single-view breast density estimates may be computed and assigned as the case-based breast density estimate at step 1820. If the modes are equal, the mean of the breast density estimates may be computed and assigned as the case-based breast density estimate. In further embodiments of the disclosure, a suitable case-based breast density estimate may be assigned by selecting the minimum or the maximum single-view breast density estimates, or by utilizing any technique known to persons of skill in the art to obtain a best estimate from a group of estimates. Any such exemplary computations may be made to integrate the breast density classification information from multiple images to arrive at a breast density estimate that is more accurate than an estimate that is solely derived from a single mammographic image.
Thus, using additional mammographic images of the same anatomical breast improves the accuracy in which density can be estimated. Breast density classification information may also be computed at step 220 by performing the acts described in
At output step 230, breast density classification information determined at step 220 may be transferred in the form of data to memory unit 124 for storage so that it can be retrieved at a later time by a radiologist or other user of system 100. Alternatively, breast density classification information determined at step 220 may be automatically output via output interface 128 to GUI 140 in the form of a report, as part of an image, or other visual depiction means that enables a radiologist or other user of system 100 to understand that the breast density classification information represents the density of the anatomical breast under study.
Having described the systems, computer-readable media, and methods disclosed herein in detail and by reference to specific embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of this disclosure. More specifically, although some aspects of this disclosure may be identified herein as preferred or particularly advantageous, it is contemplated that the methods and systems disclosed herein are not necessarily limited to these preferred aspects.
This application is a continuation-in-part of U.S. Ser. No. 12/511,701 filed on Jul. 29, 2009, the entire contents of which are herein incorporated by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
Parent | 12511701 | Jul 2009 | US |
Child | 12533952 | US |