1. Field of the Invention
The present invention relates to anatomic region-based medical image processing and relates to automated detection of human diseases. It more specifically relates to computer-aided detection (CAD) methods for automated detection of lung nodules in chest images, such as segmentation of anatomic regions in chest radiographic images and identification of orientation of postero-anterior (PA) chest images using fuzzy clustering techniques.
2. Background Art
Lung cancer is the leading type of cancer in both men and women worldwide. Early detection and treatment of localized lung cancer at a potentially curable stage can significantly increase the patient survival rate.
Among the common detection techniques for lung cancer, such as chest X-ray, analysis of the types of cells in sputum specimens, and fiber optic examination of bronchial passages, chest radiography remains the most effective and widely used method. Although skilled pulmonary radiologists can achieve a high degree of accuracy in diagnosis, problems remain in the detection of the lung nodules in chest radiography due to errors that cannot be corrected by current methods of training, even with a high level of clinical skill and experience.
Studies have shown that approximately 68% of retrospectively detected lung cancers were detected by one reader and approximately 82% were detected with an additional reader as a “second-reader.” A long-term lung cancer screening program conducted at the Mayo Clinic found that 90% of peripheral lung cancers were visible in small sizes in retrospect, in earlier radiographs.
An analysis of human error in the diagnosis of lung cancer revealed that about 30% of the missed detections were due to search errors, about 25% were due to recognition errors, and about 45% were due to decision-making errors. (Reference is made to Kundel, H. L, et al., “Visual Scanning, Pattern Recognition and Decision-Making in Pulmonary Nodule Detection”, Investigative Radiology, May-June 1978, pp 175-181, and Kundel, H. L., et al., “Visual Dwell Indicates Locations of False-Positive and False-Negative Decisions”, Investigative Radiology, June 1989, Vol. 24, pp 472-478, which are incorporated herein by reference.) The analysis suggests that the miss rates for the detection of small lung nodules could be reduced by about 55% with a computerized method. According to the article by Stitik, F. P., “Radiographic Screening in the Early Detection of Lung Cancer”, Radiologic Clinics of North America, Vol. XVI, No. 3, December 1978, pp 347-366, which is incorporated herein by reference, many of the missed lesions would be classified as T1M0 lesions, the stage of non-small cell lung cancer that Mountain, C. F., “Value of the New TNM Staging System for Lung Cancer”, 5th World Conference in Lung Cancer Chest, 1989 Vol. 96/1, pp 47-49, which is incorporated herein by reference, indicates has the best prognosis (42%, 5 year survival). It is this stage of lung cancer, with lesions smaller than 1.5 cm in diameter, and located outside the hilum region, that needs to be detected by a radiologist in order to improve survival rates.
Computerized techniques, such as computer-aided detection (CAD), have been introduced to assist in the detection of lung nodules during the early stage of non-small cell lung cancer. The CAD technique requires the computer system to function as a second reader to double-check the films that a primary physician has examined. An exemplary automated system for the detection of lung nodules may include five functional units. They are:
Obviously, identification of the lung field location is a top priority in lung nodule detection. Although a number of computer algorithms have been developed to automatically identify the lung regions in a digitized postero-anterior chest radiograph (DCR), they can be generally described as either edge-based or area-based in terms of methodology. Typically, the edge-based lung region segmentation approach has been described in the following references: J. Duryea and M. Boone, “A fully automated algorithm for the segmentation of lung fields on digital chest radiographic images,” Med. Phys. 22, pp 183-191, 1995; X.-W. Xu and K. Doi, “Image features analysis for computer-aided diagnosis: Accuracy determination of ribcage boundary in chest radiographs,” Med. Phys. 22, pp 617-626, 1995; and F. M. Carrascal, J. M. Carreira, M. Souto, P. G. Tahoces, L. Gornez, and J. J. Vidal, “Automatic calculation of total lung capacity from automatically traced lung boundaries in postero-anterior and lateral digital chest radiographs,” Med. Phys. 25, pp 11 18-1131, 1998.
This edge-based approach detects the edge lines in profiles, signatures, or kernels of its original two-dimensional image. Duryea et al. presented an automated algorithm to identify both lungs on digital chest radiographs. Starting points for the edge tracing process, which provided four edges, i.e., upper-medial, upper-lateral, lower-medial, and lower-lateral edges, were extracted based on horizontal profiles. The algorithm was evaluated with 802 images. The average accuracies were 95.7% for the right lung and 96.0% for the left lung.
Xu et al. developed a computerized method for automated determination of ribcage boundaries in digital chest radiographs. The average position of the top of the lung was determined based on the vertical profile and its first derivative in the upper central area of the chest image. Top lung edges and rib cage edges were determined within search ROIs (regions of interest), which were selected over top lung cages and rib cage. The complete rib cage boundary was obtained by smoothly connecting three curves. Xu et al. used a subjective evaluation to examine the accuracy of the results for 1000 images. The overall accuracy of the method was 96% based on the evaluations of five observers.
Carrascal et al. developed an automated computer-based method for the calculation of total lung capacity (TLC), by determining the pulmonary contours from digital PA and lateral radiographs of the thorax. This method consists of four main steps: 1) determining a group of reference lines in each radiograph; 2) defining a family of rectangular ROIs, which include the pulmonary borders, in each of which the pulmonary border is identified using edge enhancement and thresholding techniques; 3) removing outlying points from the preliminary boundary set; and 4) correcting and completing the pulmonary border by means of interpolation, extrapolation, and arc fitting. The method was applied to 65 PA chest images. Three radiologists carried out a subjective evaluation of the automatic tracing of the pulmonary borders with use of a five-point rating scale. The results were 44.1% with a score of 5, 23.6% with a score of 4, 7.2% with a score of 3, 19.0% with a score of 2, and 6.1% with a score of 1.
On the other hand, the approach of area-based lung region segmentation usually uses image features, such as density (pixel gray level), histogram, entropy, gradients, and co-occurrence matrix to perform classification. In the existing methods, the techniques used include neural network techniques and discriminant analysis, examples of which follow, and which are incorporated herein by reference.
M. F. McNitt-Gray, H. K. Huang, and J. W. Sayre, “Feature selection in the pattern classification problem of digital chest radiograph segmentation”, IEEE Trans. Med. Images 14, pp 537-547, 1995, employed a linear discriminator and a feed-forward neural network to classify pixels in a digital chest image into five areas using a set of image features selected. The five areas represent five different anatomic regions: 1) includes heart, subdiaphragm, and upper mediastinum; 2) includes right and left lungs; 3) includes two side axillas; 4) includes base of head/neck; and 5) is the background, which includes the area outside the patient projection but within the radiation field. McNitt-Gray et al. introduced a list of candidate features that include gray-level-based features, measures of local differences, and measures of local texture. A feature selection step was used to choose a subset of features from the list of candidate features. The number of nodes in the input layer was determined by the number of features in the subset. The neural network classifier was trained by using back-propagation learning.
Hasegawa, S.-C. Lo, J.-S. Lin, M. T. Freedman, and S. K. Mun, “A shift invariant neural network for the lung field segmentation in chest radiography”, J. of VLSI Signal Processing 18, pp 241-250, 1998, developed a computerized method using a shift-invariant neural network for the segmentation of lung fields in chest radiography. Only pixel gray levels served as inputs of the neural network. The lung fields were extracted by employing a shift-invariant neural network that used an error back-propagation training method. In order to train the neural network, Hasegawa et al. generalized the corresponding reference image in advance for each of the training cases. In their study, a set of computer algorithms was also developed for smoothing the initially-detected edges of lung fields. The results indicated that 86% of the segmented lung fields globally matched the original chest radiographs for 21 testing images.
Tsujii, M. T. Freedman, and S. K. Mun, “Automated segmentation of anatomic regions in chest radiographs using an adaptive-sized hybrid neural network”, Med. Phys. 25, pp 998-1007,1998, developed an automated computerized method for lung segmentation. In contrast with the method of Hasegawa et al., Tsujii et al. chose four image features as inputs of the neural network; these are relative addresses (Rx, Ry), normalized density, and histogram equalized entropy. The network was trained using 14 images. The trained neural network classified lung regions with 92% accuracy when compared against the 71 test images following the same rules used for the training images.
N. F. Vittitoe, R. Vargas-Voracek, and Carey E. Floyd, Jr., “Markov random field modeling in postero-anterior chest radiograph segmentation”, Med. Phys. 26, pp 1670-1677, 1999, presented an algorithm to identify multiple anatomical regions in a digitized PA chest radiograph utilizing Markov random field (MRF) modeling. The MRF model was developed using 115 chest radiographs. An additional 115 chest radiographs served as a test set. On average for the test set, the MRF technique correctly classified 93.3% of the lung pixels, 89.8% of the subdiaphragm pixels, 78.3% of the heart pixels, 86.1% of the mediastinum pixels, 90.1% of the body pixels, and 88.4% of the background pixels.
Unfortunately, direct comparison of the performance of these various techniques can not be made because of the differences in the data sets. A common point of uncertainty with these experiments is universality of the specific data set used. If a method is tested using 1000 “similar” images, the meaning of the calculated accuracy is limited. Generally, different digitizers, different patients and different film-makers should affect the accuracy of a method of analysis. Additionally, the existing methods do not deal with identification of orientation of PA chest images. None of these methods simultaneously considers how to provide useful information for classification of lung nodules while segmenting lung regions.
Accordingly, one object of this invention is to provide a novel segmentation method, based on fuzzy clustering, and a set of specified post-processing techniques, which include noise reduction, determination of top and bottom points of lung, border detection, boundary smoothing, and modification of regions, for automated identification of anatomic regions in chest radiographs.
Another object of this invention is to provide a novel identification method for the detection of orientation of PA chest radiographs that may be oriented in either portrait or landscape view.
The invention further enables the detection of indicators of lung diseases, such as lung nodules. The invention also can be used for other areas, including but not limited to (1) breast tumor detection, (2) brain MRI segmentation, (3) interstitial lung disease classification, (4) CT image segmentation, (5) microcalcification identification, and (6) anatomic-region-based image processing.
Additionally, the invention may be embodied as a computer programmed to carry out the inventive method, as a storage medium storing a program for implementing the inventive method, and as a system for implementing the method.
These and other objects are achieved according to an embodiment of the present invention by providing a new method for segmenting anatomic regions in a digitized PA chest image and identifying orientation of a digitized PA chest image, including (a) subsampling image data obtained in order to speed up the computational process; (b) performing a fuzzy clustering algorithm for the subsampled image data and thus generating a rough image; (c) subjecting the rough image to a filter that is designed to assimilate isolated points in each region; (d) identifying the orientation of original chest image based on the rough image after step (c); (e) determining lung's top points and bottom points in the rough image; (f) detecting the border points of each region; (g) smoothing the boundaries of each region; and (h) adjusting the boundaries of each region based on human experience.
According to an embodiment of the invention, step (a) includes setting a reduction factor of image size to two to obtain a 263×319 image from an original 525×637 image.
According to an embodiment of the invention, step (b) includes performing a Gaussian clustering algorithm for the subsampled image data to generate a rough image in which pixels are classified into several classes based on pixel gray level.
Preferably, the employed Gaussian clustering method of step (b) includes performing self-organizing classification for pixels under a predetermined number of classes, where training or prior knowledge is unnecessary. Moreover, the process maybe fully automatic, and parameters need not be problem-specific.
According to a preferred embodiment of the invention, step (c) includes using a 3×3 table filter to assimilate isolated points in each class.
According to an embodiment of the invention, step (d) includes identifying the orientation of the original chest image based on the rough image generated by step (c). Preferably, the orientation identification method of step (d) further includes detecting a midline landmark of the chest image and determining a boundary of the central zone which includes most of the superior mediastinum (Sms), most of the heart (Hrt) area, and part of the subdiaphragm (Sub).
According to an embodiment of the invention, step (e) includes detecting the outer point of top lung (OTL), inner point of top lung (ITP), outer point of bottom lung (OBL), and inner point of bottom lung (IBL) for both right lung and left lung, based on the obtained rough image.
According to an embodiment of the invention, step (f) includes detecting the border points of the lung region, using information of top lung points and bottom lung points that were detected in step (e), based on the rough image, from the top of the lung to the bottom of the lung.
According to an embodiment of the invention, step (g) includes using heuristic rules based on spatial information to smooth the boundaries of the lung zones. In this step, two processes are preferably used. These are top-down trimming and bottom-up trimming. The former serves (1) to cut the connection between the top lung and the shoulder, and (2) to cut the connection between the bottom lung and the background, if applicable. The latter is designed to refill any part of the lung region that is misclassified. Preferably, the boundary obtained by using this bidirectional trimming method is not only smooth but also natural.
Preferably, step (h) of an embodiment of the inventive method includes using a set of empirical parameters that are determined by testing an entire training image data set to adjust the area of each of the regions, through extension and/or shrinking of the boundaries.
Definitions
In describing the invention, the following definitions are applicable throughout (including above).
A “computer” refers to any apparatus that is capable of accepting a structured input, processing the structured input according to prescribed rules, and producing results of the processing as output. Examples of a computer include: a computer; a general purpose computer; a supercomputer; a mainframe; a super mini-computer; a mini-computer; a workstation; a microcomputer; a server; an interactive television; a hybrid combination of a computer and an interactive television; and application-specific hardware to emulate a computer and/or software. A computer can have a single processor or multiple processors, which can operate in parallel and/or not in parallel. A computer also refers to two or more computers connected together via a network for transmitting or receiving information between the computers. An example of such a computer includes a distributed computer system for processing information via computers linked by a network.
A “computer-readable medium” refers to any storage device used for storing data accessible by a computer. Examples of a computer-readable medium include: a magnetic hard disk; a floppy disk; an optical disk, like a CD-ROM or a DVD; a magnetic tape; a memory chip; and a carrier wave used to carry computer-readable electronic data, such as those used in transmitting and receiving e-mail or in accessing a network.
“Software” refers to prescribed rules to operate a computer. Examples of software include: code segments; instructions; computer programs; and programmed logic.
A “computer system” refers to a system having a computer, where the computer comprises a computer-readable medium embodying software to operate the computer.
A “network” refers to a number of computers and associated devices that are connected by communication facilities. A network involves permanent connections such as cables or temporary connections such as those made through telephone or other communication links. Examples of a network include: an internet, such as the Internet; an intranet; a local area network (LAN); a wide area network (WAN); and a combination of networks, such as an internet and an intranet.
The features of the present invention and the manner of attaining them will become apparent, and the invention itself will be understood, by reference to the following description and the accompanying drawings, wherein:
Next is unit 200, the fuzzy clustering unit. According to a preferred embodiment of the invention, in this unit, a Gaussian clustering method (GCM) is employed. Fuzzy clustering is an unsupervised learning technique by which a group of objects is split up into some subgroups based on a measure function. GCM is one of the most commonly used clustering methods. It has a complete Gaussian membership function derived by using a maximum-fuzzy-entropy interpretation.
Here, xk represents the k-th input, i.e., k-th pixel, vi represents the center vector of cluster i. uik represents membership assignment, that is the degree to which the input k belongs to cluster i. σ is a real constant greater than zero, which represents the “fuzziness” of classification. T represents the maximum number of iterations, ε is a small positive number that determines the termination criterion of the algorithm. N and c represent the number of inputs and number of clusters, respectively. Note that in
The third unit (300) serves to identify the orientation of a PA chest image. According to the method of the invention, this task is designed to find the orientation of the “spinal” area of a PA chest image. Preferably, the inventive orientation identification method is based on the rough image instead of the original image. Obviously, the difference between portrait and landscape images is that for a portrait image there is a rectangle located in the middle section of the horizontal direction and oriented in the vertical direction, whereas, for a landscape image such a rectangle is located in middle section of the vertical direction and oriented in the horizontal direction. In this rectangle almost all the pixels are of the white gray value. The length of the long side of the rectangle is close to the image's height for the portrait case or close to the image's width for the landscape case.
The method of identifying orientation of a chest image based on the rough image is simple but effective. The default assumption for the method is that the image is landscape. To judge whether an image is in portrait orientation or not, two conjunctive conditions are used. First, in a portrait image, there is a rectangle as defined above that is located in the middle section of the horizontal direction and oriented in the vertical direction. Further, in a portrait image, gray level value must be black at point (width/4, height/2) and point (3width/4, height/2). Here, “width” represents image width in pixels, and “height” represents image height in pixels. If an image is portrait, it can be passed to post-processing unit (400) directly. Otherwise, a landscape image would be rotated to become a portrait image first, and then passed to the next processing unit. As will be noted below, according to an embodiment of the method of the invention, this rectangle can be used in determining the central zone of a PA chest image.
According to an embodiment of the inventive method, the purpose of the isolated-point assimilation part 1350 is to assimilate isolated white points in a black cluster and isolated black points in a white cluster.
To segment lung regions based on a rough image, the first step is to locate landmark points. Landmark points here include top lung edge points and bottom lung edge points. To determine top lung edge points, rough images are classified into two types. Type 1 images are those in which the boundary of the top lung is clearly separated, as shown in
For Type 1, as shown in
The search process for the top lung edge for Type 2 images is divided into four (4) steps, which will be described in terms of the right lung (i.e., the left side of
Similarly, for determination of bottom lung edge points, rough images are classified into two (2) cases. Case 1 refers to those in which the boundary of bottom lung is clearly separated as shown in
Top-down trimming part (3350) of the post-processing unit (400) in
According to an embodiment of the invention, the bottom-up trimming part (4350) of post-processing unit (400) in
Extension/shrink fitting part (5350) of the post-processing unit (400) in
Table 1 illustrates the chest image orientation identification performance of the method for 3459 images. Of them, 519 images were landscape. images, and the rest were portrait images.
Table 2 illustrates the rib-cage detection performance of the method for 3459 chest images.
It should be noted that, as in any ill-defined problem, the evaluation criterion used here is very subjective. The “quit” case indicates that the method as embodied for these trials was unable to deal with a given image.
The same concept has been expanded to lung segmentation in a CT image.
Obviously, numerous modifications to and variations of the present invention are possible in light of the above technique. It is, therefore, to be understood that within the scope of the appended claims, the invention may be implemented in situations other than as specifically described herein. Although the present application is focused on chest image and CT image, the concept can be expanded to other medical images and other object segmentation problems, such as MRI, brain and vessel segmentation, and the like. The invention is thus of broad application and not limited to the specifically disclosed embodiment.
This application claims the benefit of U.S. Provisional Application No. 60/394,238, filed Jul. 9, 2002, and incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
60394238 | Jul 2002 | US |