1. Field of the Invention
The present invention relates to a digital image processing technique, and more particularly to a method and apparatus for processing medical images and detecting malignant areas in a medical image.
2. Description of the Related Art
Mammography images and identification of abnormal structures in mammography images are important tools for diagnosis of medical problems of breasts. For example, identification of cancer structures in mammography images is important and useful for prompt treatment and prognosis.
Reliable cancer detection, however, is difficult to achieve because of variations in anatomical shapes of breasts and medical imaging conditions. Such variations include: 1) anatomical shape variations between breasts of various people or breasts of the same person; 2) lighting variations for medical images of breasts, taken at different times; 3) pose and view changes in mammograms; 4) change in anatomical structure of breasts due to aging of people; etc. Moreover, malignant lesions are often difficult to distinguish from benign lesions. Such imaging situations pose challenges for both manual identification and computer-aided detection of cancer in breasts.
One way to detect cancer in breasts is to look for signs of malignant masses. Signs of malignant masses may often be very subtle, especially in the early stages of cancer. However, such early cancer stages are also the most treatable, hence detecting subtle signs of malignant masses offers great benefits to patients. Malignant areas in breasts (and in other organs) often appear as irregular regions surrounded by patterns of star-like, radiating linear structures, also called spicules, or spiculated structures. Since malignant masses are often accompanied by spiculated structures, spiculated margin is a strong indicator of malignant masses. Moreover, spiculated structures indicate a much higher risk of malignancy, than calcifications and other types of structural mass changes.
Typical/conventional spiculation detection methods employ two-step approaches, which first extract line structures, and then compute features based on the detected lines. Thus, the performance of spicular detection for typical/conventional spiculation detection methods depends heavily on the performance of line structure extraction. Hence, spiculation detection is often difficult or ineffective if the extracted line structures are not good enough for spiculation detection. One spiculation detection technique is described in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619. In the technique described in this work, a line structure map is constructed using a steerable filter approach, which is a method to efficiently compute maximum filtering response of breast images with a line-like kernel. A statistical analysis is then applied to detect spiculated structures based on the line structure map. This technique, however, still poses computational problems for automated detection of spiculated structures, because of the large computational load required by the technique. Moreover, the statistical analysis used to detect spiculated structures in this technique does not characterize well aspects of spiculated structures, such as the directional diversity of spiculated line structures. Hence, this technique encounters challenges when used for automated detection of spiculated structures in breasts.
Disclosed embodiments of this application address these and other issues by implementing methods and apparatuses for spiculation detection using separable steerable filters, to extract line structure maps for mammography images. The computational load is significantly reduced by using separable steerable filters. Line structure maps for mammography images are then characterized using multiple features related to spiculation of line structures, and overall spiculation scores are extracted for pixels inside mammography images. Spiculated structures are then identified using extracted spiculation scores. Methods and apparatuses described in this application can implement spicule detection algorithms that enhance mass detection capability of Computer Aided Design (CAD) systems. The methods and apparatuses described in this application can also be used for detection of spiculated structures in other anatomical parts besides breasts.
The present invention is directed to a method and an apparatus for identifying a spicule candidate in a medical image. According to a first aspect of the present invention, an image processing method for identifying a spicule candidate in a medical image comprises: accessing digital image data representing an image including a tissue region; processing the digital image data by generating at least one line orientation map and at least one line strength map for the tissue region for at least one scale, using separable filters; calculating spicule feature values based on the at least one line orientation map and the at least one line strength map; and identifying a spicule candidate based on the calculated features values.
According to a second aspect of the present invention, an image processing apparatus for identifying a spicule candidate in a medical image comprises: an image data input unit for accessing digital image data representing an image including a tissue region; a line structure extraction unit for processing the digital image data, the line structure extraction unit generating at least one line orientation map and at least one line strength map for the tissue region for at least one scale, using separable filters; a feature map generator unit for calculating spicule feature values based on the at least one line orientation map and the at least one line strength map; and a spicule candidate determining unit for identifying a spicule candidate based on the calculated features values.
Further aspects and advantages of the present invention will become apparent upon reading the following detailed description in conjunction with the accompanying drawings, in which:
Aspects of the invention are more specifically set forth in the accompanying description with reference to the appended figures.
The image input unit 25 provides digital image data representing medical images. Medical images may be mammograms, X-ray images of various parts of the body, etc. Image input unit 25 may be one or more of any number of devices providing digital image data derived from a radiological film, a diagnostic image, a digital system, etc. Such an input device may be, for example, a scanner for scanning images recorded on a film; a digital camera; a digital mammography machine; a recording medium such as a CD-R, a floppy disk, a USB drive, etc.; a database system which stores images; a network connection; an image processing system that outputs digital data, such as a computer application that processes images; etc.
The image processing unit 100 receives digital image data from the image input unit 25 and performs spiculation detection in a manner discussed in detail below. A user, e.g., a radiology specialist at a medical facility, may view the output of image processing unit 100, via display 65 and may input commands to the image processing unit 100 via the user input unit 75. In the embodiment illustrated in
In addition to performing spiculation detection in accordance with embodiments of the present invention, the image processing unit 100 may perform additional image processing functions in accordance with commands received from the user input unit 75. The printing unit 45 receives the output of the image processing unit 100 and generates a hard copy of the processed image data. In addition or as an alternative to generating a hard copy of the output of the image processing unit 100, the processed image data may be returned as an image file, e.g., via a portable recording medium or via a network (not shown). The output of image processing unit 100 may also be sent to image output unit 56 that collects or stores image data, and/or to processing unit 55 that performs further operations on image data, or uses image data for various purposes. The processing unit 55 may be another image processing unit; a pattern recognition unit; an artificial intelligence unit; a classifier module; an application that compares images; etc.
Operation of image processing unit 100 will be next described in the context of mammography images, for detecting spiculation structures, which are a strong indicator of malignant masses. However, the principles of the current invention apply equally to other areas of medical image processing, for spiculation detection in other types of anatomical objects besides breasts.
Generally, the arrangement of elements for the image processing unit 100 illustrated in
Image operations unit 110 sends the preprocessed breast image to line structure extraction unit 120, which identifies line structures in the breast image. Feature map generator unit 180 receives at least two images, including a line intensity image and a line orientation image, and calculates spicule features for areas including line structures. Spicule candidate determining unit 190 uses the spicule features to determine if various line structures are spiculation structures. Finally, spicule candidate determining unit 190 outputs a breast image with identified spiculation structures, including spicule candidates and their spicule feature values.
Spicule candidate and spicule feature maps output from spicule candidate determining unit 190 may be sent to processing unit 55, image output unit 56, printing unit 45, and/or display 65. Processing unit 55 may be another image processing unit, or a pattern recognition or artificial intelligence unit, used for further processing. For example, in one implementation, spiculation candidate locations and their spiculation feature values, output from spicule candidate determining unit 19, are passed to a final mass classifier that uses the spicule feature values along with other features for characterizing masses. Operation of the components included in the image processing unit 100 illustrated in
Image operations unit 110, line structure extraction unit 120, feature map generator unit 180, and spicule candidate determining unit 190 are software systems/applications. Image operations unit 110, line structure extraction unit 120, feature map generator unit 180, and spicule candidate determining unit 190 may also be purpose built hardware such as FPGA, ASIC, etc.
The breast image ROI is sent to line structure extraction unit 120. Line structure extraction unit 120 determines line structures in the breast image ROI (S208), and outputs a line intensity map for the breast image ROI (S212) and a line orientation map for the breast image ROI (S218). Feature map generator unit 180 receives the line intensity map and the line orientation map for the breast image ROI, and obtains a feature map for the breast image ROI (S222). Spicule candidate determining unit 190 receives the feature map for the breast image ROI and determines spicule candidates in the breast image ROI (S228). Finally, spicule candidate determining unit 190 outputs a breast image with identified spiculation structures (S232).
Spiculated lines typically have various widths. The width of spiculated lines can be characterized by a scale parameter σ. For each scale parameter σ, line structure extraction unit 120A extracts a line strength and a line orientation at pixels inside a breast image ROI. Line structures are extracted by examining the maximum filtering response with a line-like kernel. High correlation with a target shape pattern—a line, in the current case—is considered to indicate high probability for the presence of the target (i.e., a line).
A filter method described in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619, the entire contents of which are hereby incorporated by reference, may be used. For this purpose, the second directional derivative of Gaussian distribution is used as a line kernel. Other line kernels and distributions may also be used.
A line strength map Lσ and a line orientation map Θσ are computed by taking the maximum of the convolution of a breast image ROI and the kernel, over various angles. σ is the scale parameter that controls the line width to be detected and hence, sets the scale for line structure analysis.
Using techniques described in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619, the entire contents of which are hereby incorporated by reference, the line strength and orientation at pixels inside a mammography image can be computed as illustrated below. More precisely, at a pixel at position (x,y) in the breast image ROI, the line strength and line orientation at scale σ are obtained by formulas (1) and (2):
where I is the original breast image ROI, * is the 2-D convolution operator, and Kσ,θ is the second directional derivative at angle θ of Gaussian kernel, given by:
Spiculated lines typically have various widths and orientations. Line strength and line orientation may be calculated for various σ and θ using formulas (1) and (2), in order to detect various line structures forming spiculation structures. However, applying the convolution in formulas (1) and (2) for various σ and θ requires large volumes of data processing, and creates serious computational problems for practical CAD systems. To circumvent this difficulty, a multiscale approach related to the scale parameter σ is used by line structure extraction unit 120A to detect spiculated line structures.
As illustrated in
To obtain the final line strength map in step S324, each pixel in the final line strength map is set to the maximum of line strength over scales. For example, if pixel P1 has line strength Lσ0(x, y) in the line strength map at scale 0 at step S305_0 (where σ0 is the scale parameter associated with scale 0), line strength Lσ1(x, y) in the line strength map at scale 1 at step S305_1 (where σ1 is the scale parameter associated with scale 1), . . . , line strength LσQ(x, y) in the line strength map at scale Q at step S305_Q (where σQ is the scale parameter associated with scale Q), then pixel P1 is set to have line strength=max(Lσ0(x, y), Lσ1(x, y), . . . , LσQ(x, y)) in the final line strength map at step S324. The line strength map is also called line intensity map in the current application.
To obtain the final line orientation map in step S326, the line orientation of each pixel in the final line orientation map is set to the line orientation at the scale where the line strength yields maximum over scales. For example, if pixel P1 has line strength Lσ0(x, y) in the line strength map at scale 0 at step S305_0, line strength Lσ1(x, y) in the line strength map at scale 1 at step S305_1, . . . , line strength LσQ(x, y) in the line strength map at scale Q at step S305_Q, and maximum pixel line strength for pixel P1, equal to max(Lσ0(x, y), Lσ1(x, y), . . . , LσQ(x, y)), occurs at scale j, 0≦j≦Q, then the line orientation of pixel P1 in the final line orientation map in step S326 is set to be the line orientation Θσj(x, y) at the scale j, as obtained in a step S305_j.
Line and direction merging unit 260 next obtains a line structure image, using data from the intensity line map and the line orientation map (S329). The line structure image obtained in step S329 may be a breast image with identified lines throughout the breast image. Selective line removal unit 270 receives the line structure image. Since the line extraction algorithm often yields strong response on the breast pectoral edge, false spicule candidates may be identified on the pectoral edge. To reduce this type of false spicule candidates, selective line removal unit 270 removes lines on pectoral edge (S340). The pectoral edge of breasts may be determined manually, or using automated pectoral edge detection techniques.
Pectoral edge line removal may be performed before or after the line intensity and line orientation maps are combined to obtain the line structure image.
Before obtaining the final line strength map and final line orientation map in steps S324 and S326, from line strength maps Lσ(x, y) for scales 0, 1, . . . , Q, thinned line maps may be obtained. For example, a thinned line map Lσ0
In an exemplary implementation, thinning is performed by applying binarization and then a morphological method.
Instead of thinning, line strength maps may be only thresholded. Like thinning, thresholding makes line orientation estimation more accurate by excluding line pixels with low line intensity.
As shown by formulas (1) and (2), given a scale parameter σ, the convolution I*Kσ,θ for various angles θ needs to be computed, to obtain Θσ(x, y) and Lσ(x, y). Calculating the convolution I*Kσ,θ for various angles θ is computationally burdensome, and hard to implement in a practical real-time automated system.
A computationally efficient alternative is to apply a steerable filter method, so that the convolution I*Kσ,θ is not computed for multiple angles θ by brute force. Steerable filters are “a class of filters in which a filter of arbitrary orientation is synthesized as a linear combination of a set of ‘basic filters’ ”, as described in “The Design and Use of Steerable Filters”, by W. T. Freeman and E. H. Adelson, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 13, No. 9, September 1991, p. 891-906, the entire contents of which are hereby incorporated by reference. The basic filters can be selected to be independent of orientation θ, while the weight functions used in the linear combination of basic filters depend on θ only. It is known that Kσ,θ can be represented with only three basic filters, as demonstrated in “Generic Neighborhood Operators”, by J. J. Koenderink and A. J. van Doom, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol 14, pp. 597-605, 1992, “The Design and Use of Steerable Filters”, by W. T. Freeman and E. H. Adelson, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 13, No. 9, September 1991, p. 891-906, and/or “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619, the entire contents of these three publications being hereby incorporated by reference. Hence, Kσ,θ can be represented with only three basic filters as shown in equation (4) below:
where wσ,i(.,.) is the i-th basic filter and ki(θ) is its corresponding weight function.
Then, by the linearity of convolution, I*Kσ,θ can be computed by:
from formulas (1) and (2), only the θ and Lσ for the extremum filtering response I*Kσ,θ are needed. Hence, I*Kσ,θ does not need to be computed for all angles, as it is enough to examine only the orientations (i.e. angles θ) that satisfy
becomes:
Hence, to find the line strength
and the line orientation
for scale parameter σ, given basic filters wσ,i and corresponding weight functions ki(θ), i=1, 2, 3 (S380), line strength map generation unit 250 computes I*wσ,i, i=1, 2, 3 (S382); finds roots θ1 and θ2 for
compares the values of k1(θ)(I*wσ,1)+k2(θ)(I*wσ,2)+k3(θ)(I*wσ,3) at the two roots θ1 and θ2 (S386); selects the root yielding the higher value for k1(θ)(I*wσ,1)+k2(θ)(I*wσ,2)+k3(θ)(I*wσ,3) to be the line orientation Θσ (S307A); and obtains the line strength Lσ=I*Kσ,Θ
There are many choices for the basic filters wσ,i and weight functions ki(θ) for Kσ,θ expressed in formula (4). For example, basic filters wσ,i=Kσ,iπ/3, i=1, 2, 3 were used in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619, the entire contents of which are hereby incorporated by reference. Using these basic filters wσ,i is still computationally burdensome, because the 2-D convolution operations in I*wσ,i are computationally expensive, since Kσ,π/3 and Kσ,2π/3 are not separable for the directions π/3 and 2π/3. Moreover, the 2-D convolution operations in I*wσ,i are even more computationally expensive for larger σ.
To decrease computational complexity, the following basic filters and weight functions can be chosen to represent Kσ,θ:
k
1(θ)=cos2 θ (10)
k
2(θ)=−2 cos θsin θ (11)
k
3(θ)=sin2 θ (12)
The basic filters wσ,i can then be represented as separable filters as illustrated in equations (13), (14), (15), (16), (17), (18) below:
w
σ,1(x,y)=s1(x)s3(y) (13)
w
σ,2(x,y)=s2(x)s2(y) (14)
w
σ,3(x,y)=s3(x)s1(y) (15)
where
This choice of basic filters has been described in “The Design and Use of Steerable Filters”, by W. T. Freeman and E. H. Adelson, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 13, No. 9, September 1991, p. 891-906, the entire contents of which are hereby incorporated by reference. By using the separable filters described in equations (13), (14), (15), (16), (17), (18) the 2-D filtering step S382 in the flow diagram of
For this purpose, for a scale parameter σ (S378A), separable filters wσ,i and corresponding weight functions ki(θ), i=1, 2, 3 are selected based on equations (13), (14), (15), (16), (17), (18) (S380A), as illustrated in the flow diagram of
For angle Θσ obtained in step S307A in
as described in equation (5).
In this manner, the line strength map at angle θ=Θσ(x, y) can be computed efficiently for each pixel position (x,y). The separability of the representations using the basic filters reduces computation per pixel from N̂2 to 2N, hence providing a N/2 speedup, where N is the width or height of 2-D filtering kernel.
In one exemplary embodiment, 10 spicule feature values are calculated for line structure images including line intensity images and line orientation images.
The line concentration feature measure calculated by feature map generator unit 180 is similar to a feature developed in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619, the entire contents of which are hereby incorporated by reference. For each pixel i of interest in a breast image ROI, let Ai be the set of the line pixels in the gray ring R556, and S, be the set of pixels that belong to Ai and point to the center circle C552. The gray ring R556 has pixel i as center, inner radius R_in, and outer radius R_out, and the center circle C552 has pixel i as center, and radius R. Let Ni be the number of pixels in Si. Suppose that Ki is its corresponding random variable, which is the number of pixels that belong to Ai and point to the center circle C552. That is, Ni is a realization of the random variable Ki. Assuming that line orientation is uniformly distributed, the expectation and standard deviation of Ki are computed. The line concentration feature measure is a statistic defined by equation (19):
To compute a directional entropy feature for pixel i, the line structure maps in Si are examined. Suppose a directional entropy feature is computed based on the number of line pixels for pixel i. A histogram is generated for the angle of all radiating line pixels in Si. Let Li,k be the number of pixels in the k-th bin from the angle histogram:
The feature described by equation (20) is similar to the directional entropy in patent application US2004/0151357 titled “Abnormal Pattern Detecting Apparatus”, by inventors Shi Chao and Takeo Hideya, the entire contents of which are hereby incorporated by reference, but there are two important differences between the feature described by equation (20) in this disclosure and the directional entropy in patent application US2004/0151357. First, fi,2 is computed based on pixels only Si, while the directional entropy in US2004/0151357 was computed for Ai, since there was no motivation in US2004/0151357 to consider diversity of the pixels that are not oriented to spicule center. Second, the angle histogram is computed from Θσ in a pixel-based manner, while in US2004/0151357 line directions are computed by labeling one direction for each connected line.
Another feature calculated by feature map generator unit 180 for spiculation detection is the line orientation diversity feature. To calculate the line orientation diversity feature, let n+,i be the number of bins where Li,k is larger than the median value calculated based on an assumption that line orientation is random. Suppose that N+,i is a random variable, which is the number of bins where Li,k is larger than the median value calculated based on an assumption that line orientation is random. That is, n+,i is a realization of N+,i. The line orientation diversity feature is then defined as
The line orientation diversity feature is similar to a feature developed in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619, the entire contents of which are hereby incorporated by reference.
In many cases where only one or two strong lines exist in a spiculation structure, the spicule features (especially the line concentration feature and the line orientation diversity feature) have high responses and little specificity. To aid in spiculation detection in such cases, two additional linearity features are calculated by feature map generator unit 180 for spiculation detection. The two additional spicule linearity features employed are described by equations (22) and (23) below:
To obtain the bins used in equations (22) and (23), directions are divided into a number of bins, and the radiating line pixels in each direction bin are then counted. The three bins that have most radiating line pixels are then chosen, to obtain the “top 3 direction bins” used in equations (22) and (23).
Feature map images are typically rough, as they are sensitive to noise and small pixel positions differences. To reduce the effect of noise, smoothed spicule features for smoothed feature images are extracted by feature map generator unit 180 for spiculation detection. The smoothed spicule features may be extracted by filtering with a smoothening filter. Five smoothed spicule features fi,6, fi,7, fi,8, fi,9, fi,10 are calculated. The five smoothed spicule features are obtained by smoothing the five features fi,1, fi,2, fi,3, fi,4, fi,5 described by equations (19), (20), (21), (22), and (23).
More or fewer than 10 features may also be used to detect spiculated structures.
In an exemplary embodiment using the 10 spicule feature values fi,1, fi,2, fi,3, fi,4, fi,5, fi,6, fi,7, fi,8, fi,9, fi,10, for pixels i belonging to breast image ROIs, an SVM predictor uses 10 features fi,2 fi,3, fi,4, fi,5, fi,6, fi,7, fi,8, fi,9, fi,10 (or more or less features, depending on availability) as an input vector, and calculates an SVM prediction score as a final spiculation feature.
In one exemplary embodiment, the SVM predictor picks two spicule candidates based on SVM scores. In order to avoid duplicated candidates in one breast region, two spicule candidates may be chosen so that they are spatially sufficiently apart from each other.
Using SVM prediction scores as final features for spicule identification improves performance by offering flexibility in the decision boundary. The spiculation detection methods and apparatuses presented in this application are more accurate than the typical/conventional spiculation detection algorithms, and computationally more efficient compared to spiculation detection algorithms presented in “Detection of Stellate Distortions in Mammograms”, by N. Karssemeijer and G. M. te Brake, IEEE Transactions on Medical Imaging, Vol. 15, No. 5, October 1996, p. 611-619. Thus, the advantages of the present invention are readily apparent.
The methods and apparatuses described in the current application detect spiculation structures and enable detection of cancer masses. The methods and apparatuses described in the current application are automatic and can be used in computer-aided detection of cancer in breasts. The methods and apparatuses described in the current application can be used to detect spiculation structures and malignant masses in other anatomical parts besides breasts. For example, methods and apparatuses described in the current application can be used to detect spiculation structures and malignant masses in lung cancer.
Although detailed embodiments and implementations of the present invention have been described above, it should be apparent that various modifications are possible without departing from the spirit and scope of the present invention.