N/A
The subject matter described herein relates to systems and methods for processing optical images, and, more particularly, to automatic detection of polyps in optical images.
Colorectal cancer (CRC) is the second highest cause of cancer-related deaths in the United States with 50,830 estimated deaths in 2013. More than 80% of CRC cases arise from adenomatous polyps, which are precancerous abnormal growths of the colon wall. The preferred screening method for polyp detection and removal is an optical colonoscopy (OC) procedure, during which a colonoscopist meticulously examines the colon wall using a tiny camera that is inserted and guided through the colon. The goal of an OC is to detect and remove colorectal polyps, which may be precursors to CRC. Thus, it has been shown that timely removal of polyps can significantly reduce the mortality of CRC.
However, polyp detection with OC remains a challenging task and, as evidenced by several clinical studies, a significant portion of flat and pedunculated polyps remain undetected during colon screening with OC. High polyp detection rate requires a high level of attentiveness, alertness, and sensitivity to visual characteristics of polyps from colonoscopists and such qualities may only be procured after years of practice and experience. It is therefore important to reduce polyp miss-rate as it decreases the incidence and mortality of CRC.
Computer-aided polyp detection has recently been considered as a tool for reducing polyp miss-rate. For example, during an OC procedure, regions with suspected polyps can be highlighted for further examination. Existing approaches for polyp detection primarily rely on the shape or texture of polyps. However, shape information is susceptible to partial and fragmented image segmentation and can mislead a detector towards irrelevant objects in the complex endoluminal scene. Texture may also be unreliable because its visibility depends on camera-polyp distance. Thus, the texture of a polyp becomes fully visible only when the camera captures close shots of the surface of a polyp. This condition is often met when polyps have already been detected by operators. On the other hand, shape information cannot be considered as a reliable measure because polyps appear in a variety of forms ranging from sessile to peduncular shapes. Therefore, texture-based and shape-based polyp detectors offer limited practical value.
Consequently, considering such limitations of previous technological approaches, it would be desirable to have a system and method for accurate and reliable polyp detection in optical colonoscopy images that is shape-based and can compensate for the concomitant drawbacks of shape-based detection.
In accordance with one aspect, a system for automated polyp detection in optical colonoscopy images is disclosed. The system includes an input configured to acquire a series of optical images, a processor, and a memory. The memory contains instructions that, when executed by the processor, causes the processor to perform a process on the series of optical images. The processor applies a color filter to create a plurality of color filtered images for each optical image of the series of optical images, and locates a series of edge pixels on at least one of the plurality of color filtered images. The process then obtains a plurality of oriented image patches corresponding to each of the series of the edge pixels. The plurality of oriented image patches are represented by an intensity signal characterized by at least one of rotation invariance or illumination invariance. An edge normal for each edge pixel is then estimated. At least one classification system is constructed corresponding to the plurality of color filtered images. The at least one classification system is configured to enhance low level features of the plurality of color filtered images prior to classification, and appearance features are generated from a series of feature vectors selected from the classification system. An edge classification threshold analysis is then performed on each of the plurality of oriented image patches. Based on the edge classification threshold analysis, the processor determines that a given image patch is consistent with an edge threshold. A report is generated indicating potential polyps with a vote accumulation greater than a threshold. The vote accumulation includes a probabilistic output for each of the potential polyps in the absence of a predefined parametric model.
In accordance with another aspect, a method for automated polyp detection in optical colonoscopy images is disclosed. The method includes applying a color filter to create a plurality of color filtered images for each of a plurality of optical images. A series of edge pixels are located on at least one of the plurality of color filtered images, and a plurality of oriented image patches are obtained corresponding to each of the series of the edge pixels. The plurality of oriented image patches are represented by an intensity signal characterized by at least one of rotation invariance or illumination invariance. An edge normal is estimated for each edge pixel, and at least one classification system is constructed corresponding to the plurality of color filtered images. The at least one classification system is configured to enhance low level features of the plurality of color filtered images prior to classification. Appearance features are generated from a series of feature vectors selected from the at least one classification system, and an edge classification threshold analysis is performed on each of the plurality of oriented image patches. Based on the edge classification threshold analysis, a given image patch is determined to be consistent with an edge threshold. A report is generated indicating potential polyps with a vote accumulation greater than a threshold. The vote accumulation includes a probabilistic output for each of the potential polyps in the absence of a predefined parametric model.
The foregoing and other aspects and advantages of the disclosure will appear from the following description. In the description, reference is made to the accompanying drawings which form a part hereof, and in which there is shown by way of illustration one embodiment. Such embodiment does not necessarily represent the full scope of the disclosure, however, and reference is made therefore to the claims and herein for interpreting the scope of the disclosure.
The present disclosure describes embodiments that overcome the aforementioned drawbacks by providing a system and method for polyp detection based on an observation that image appearance around the polyp boundaries differs from that of other boundaries in colonoscopy images. To reduce vulnerability against misleading objects, the imaging processing method localizes polyps by detecting polyp boundaries, while filtering out irrelevant boundaries, with classification and feature extraction methods. To filter out irrelevant boundaries, a boundary removal mechanism is provided that captures changes in image appearance across polyp boundaries. Thus, the boundary removal mechanism is minimally affected by texture visibility limitations. In addition, the present disclosure describes embodiments that overcome the challenges posed by partial polyp segmentation by applying a vote accumulation scheme that enables polyp localization from fragmented edge segmentation maps without requiring perfect identification of whole polyp boundaries and knowledge about shapes and size of polyps. Therefore, the systems and methods as described herein assist both experienced, and inexperienced, colonoscopists with accurately detecting and locating polyps.
In addition, the present disclosure describes embodiments that overcome the aforementioned drawbacks by providing a system and method for polyp detection that combines image context with shape information to minimize the misleading effect of irrelevant objects with polyp-like boundaries. Given an input image, the present method begins with collecting a crude set of boundary pixels that are refined by a patch descriptor and classification scheme, before feeding a voting scheme for polyp localization. The patch descriptor quickly and efficiently characterizes image appearance across object boundaries, and is both rotation invariant and robust against linear illumination changes. A two-stage classification framework is also provided that is able to enhance low level image features prior to classification. Unlike traditional image classification where a single patch undergoes the processing pipeline, the present system fuses the information extracted from a pair of patches for more accurate edge classification. In addition, a vote accumulation scheme that robustly detects objects with curvy boundaries in fragmented edge maps is provided. The voting scheme produces a probabilistic output for each polyp candidate, but does not require any predefined parametric model of polyps (e.g., circle and ellipse).
Thus, the disclosed polyp detection system and method are based on two key observations. First, polyps, irrespective of their morphology, feature a curvy segment in their boundaries. The polyp detection system uses this property to localize polyps by detecting objects with curvy boundaries. Second, image appearance across the polyp boundaries is highly distinct from that of vessels, lumen, and specular reflections. Thus, present patch descriptor and classification scheme is able to distinguish polyp boundaries from the boundaries of other colonic objects, producing a refined edge map for the vote accumulation scheme.
The methodology of the present disclosure is based on image appearance variation between polyps and their surrounding tissue. The rationale takes into account that local patterns of color variation across the boundary of polyps differ from the patterns of color variation that occur across the boundary of folds, lumen, and vessels.
Turning to
The image acquisition hardware 104 may be designed to acquire optical image data continuously or intermittently, for example, during a medical procedure, such as a colonoscopy, and relay optical image data for processing. The image acquisition hardware 104 may require operator direction, input or feedback, or may be designed to operate autonomously.
The processor 106 may be configured to process optical image data, including image data obtained during a medical procedure, such as a colonoscopy. In one embodiment, the processor 106 may be designed to process optical images, generated from optical image data, by applying a plurality of color filters. One non-limiting example of a plurality of filters may include a red (R), green (G) and blue (B) filter, often referred to as an RGB filter. Within this example, an hue-saturation-lightness (HSL) or hue-saturation-value (HSV) coordinate representation of the RGB model may be used. Other non-limiting examples of color maps include La*b* (or Lab color space). In addition, it is possible to use more than one color map, for instance, RGB+La*b*. Regardless of the filter, map, color space, particular combination of filters, maps, or color spaces, the present disclosure provides a system and method for polyp detection that leverages the appearance of color variation between polyps and surrounding tissues.
The input 108 may take any suitable shape or form, as desired, for operation of the polyp detection system 100, including the ability for selecting, entering or otherwise specifying parameters consistent with detecting polyps of a requisite or desired size or shape.
The output 110 may take any suitable shape or form, as desired, and may include a visual and/or audio system, configured for displaying, for example, acquired optical images as a result of a medical procedure, such as a colonoscopy, and also configured, for example, to highlight and/or alert an operator of the polyp detection system 100 upon identification of a polyp with the requisite or desired features.
The memory 112 may contain software 114 and data 116, and may be configured for storage and retrieval of image processing information and data to be processed by the processor 106. In one embodiment, the software 114 includes instructions directed to performing optical image processing for polyp detection. In another embodiment, the data 116 includes optical image data.
Turning to
Illustrating the general steps associated with performing the polyp detection of process block 204, is a flow diagram shown in
Next, at process block 306, image patches 305 of polyp boundaries are captured by extracting oriented sub-images 307 along the edge normals that correspond to the R, G, and B color channels. The feature extraction method is then applied on the oriented sub-images 307 to generate appearance features. Next, at process block 308, the crude edge map is refined by means of a classification scheme that operates on the extracted features. More specifically, a two-stage classification system is used to filter out irrelevant, non-polyp edges (e.g., those edges lying on folds and vessels), while retaining edges around and on the polyp boundary. By filtering out the irrelevant boundaries, the change in image appearance across the boundary of polyps is revealed, which differs from what is observed across other boundaries. Thus, image patches 305 that are extracted from polyp boundaries may be referred to as “positive patches” and the remaining image patches as “negative patches.” Similarly, the edges that lie on the boundary of polyps may be referred to as “polyp edges” and the remaining edges may be referred to as “non-polyp edges.” Once the crude edge map is refined, at the next process block 310, a vote accumulation scheme is applied to the refined map to localize polyps.
Then, at process block 404, a crude set of edge pixels are detected in order to characterize image appearance across polyp boundaries. To do so, any suitable edge detection method, such as a Canny's method, is applied on the three color channels of the input images to extract as many edges as possible. The goal is to obtain an overcomplete edge map which will further be refined through a classification scheme.
Next, at process block 406, oriented image patches are extracted along the edge direction. Particularly, as shown in
Next, at process block 408, a robust and accurate estimation of edge normal may be computed. The normal directions are used to extract oriented patches around the edges. In certain embodiments, an accurate estimate of edge normal is important to collecting sound oriented patches 504. In some cases, edge normals computed using a gradient-based approach are often inaccurate, resulting in a non-smooth map of edge normal and poorly aligned image patches. For example, as shown in
To capture the unique image appearance of polyps along the boundaries of the polyps, oriented sub-images, for example of size 64×64, may be extracted along the edge normals, as shown at process block 410. As a result, in the extracted sub-images, the edges appear vertically in the middle. For example, an edge pixel at angle θ, may include two possible normals (i.e., θ−π/2 and θ+π/2) that give two horizontally mirrored sub-images.
Patches, for example of size 8×16, are formed all over each sub-image with about 50% overlap along horizontal and vertical directions. Each patch may then be averaged vertically, resulting in a 1-dimensional (1D) intensity signal which presents intensity variation along the horizontal axis. A 1D discrete cosine transform (DCT) may then be applied to obtain a compact and informative presentation of the signal. To achieve invariance against constant illumination changes, the DC component (i.e., the average patch intensity) is discarded. To achieve invariance against linear illumination scaling, the AC coefficients may be divided by the norm of the DCT coefficients vector. Thus, the descriptor can partially tolerate nonlinear illumination change over the whole sub-image, particularly if the nonlinear change can be decomposed to a set of linear illumination changes on the local patches. Then, the first few normalized AC coefficients may be selected from each patch corresponding to low frequency intensity changes and concatenated to form a feature vector for a sub-image, as shown at process block 412.
The above described image descriptor provides rotation invariance, which is important because features that can consistently represent image appearance across edges lying at arbitrary directions are needed. In addition, the image descriptor provides illumination invariance, which is important in colonoscopy since the source of light moves along with the camera, thereby causing the same segment of a polyp boundary to appear with varying contrast in different frames.
In an alternative embodiment, image patches of polyp boundaries are obtained and oriented in order to construct principal component analysis (PCA) models that correspond to the R, G, and B color channels. To fully utilize the information content of color patches, three PCA models corresponding to R, G, and B color channels may be constructed. Eigenvectors are then selected from each of the three PCA models such that approximately 90% of the total eigenvalue sum is covered, resulting in a 60-dimensional feature space. To generate appearance features, each image patch is projected along the selected eigenvectors and the resultant projection coefficients are concatenated to form the feature vectors. Compared to general purpose features, such as Haar and steerable filters, PCA eigen images 600, as shown in
Returning to
To train the first layer, N1 oriented patches may be collected around boundaries of polyps and four sub-negative classes: vessels, lumen areas, specular reflections, and around edges at random locations in training images. A five-class classifier may be trained using the patch descriptor. The output of the classifier may be an array of probabilities for the five object classes. Compared with the low level input features, which encode local image variation, the generated output array contains mid-level features that measure the global similarity between the underlying input sub-image and the general appearance of the predefined structures.
To train the second layer, N2 pairs of oriented patches may be collected from polyp boundaries and other random locations in training images. In one non-limiting example, let {pi1, pi2} be the extracted pair of patches around ith edge with the corresponding normals, {ni1, ni2}, where ∠ni1ε[0,π) and ∠ni2=∠ni1+π. Based on the state of the ith edge, a label, yiε{0, 1, 2} may be assigned to each pair of patches, where “0” is for a non-polyp edge, “1” is for a polyp edge with normal being ni1, and “2” is for a polyp edge with normal being ni2. Such labeling is possible given the ground truth for polyps. Next, low level features may be extracted from each pair of patches and the classifier trained may be applied in the first layer, resulting in two arrays of mid-level features per pair that are further concatenated to form a feature vector, {fi, yi}. Once the feature vectors are collected, a three-class classifier may be trained to learn both edge label and edge normal direction. In the test stage, the label with maximum probability may be assigned to the underlying edge pixel.
More generally, the second layer of classification fuses knowledge captured from a pair of patches because combining the two sources of information can yield more accurate edge classification. For example, if the patch pi1 around the ith edge resembles the appearance of a polyp, and the counterpart patch pi2 looks very similar to the average appearance of specular reflections or lumen areas, it may be difficult to determine the underlying edge pixel. In one embodiment, the first patch may be relied on and a polyp edge with edge normal being ni1 may be declared. In another embodiment, information from the counterpart patch may be considered and a non-polyp edge may be declared. In yet another embodiment, to determine the underlying pixel, a second classifier in the mid-level feature space may be trained to fully utilize such relationships.
At the next process block 416, a polyp vote accumulation analysis is performed. In the ideal classification scenario, all non-polyp edge pixels are removed and the arrangement of positive pixels indicates the locations of polyps. However, in practice, a portion of non-polyp edges may pass the classification stage and induce false positives. On the knowledge that false positive edges often appear on elongated and low-curvature edge segments, a vote accumulator scheme that will mitigate the effect of false positive edges, but also enable polyp localization from fragmented edges, may be utilized. The vote accumulation scheme assigns high values to the regions that are partially surrounded by curvy edges, but gives low values to the regions that are surrounded by elongated low curvature edges.
Turning to
Such edge grouping prior to vote casting minimizes vote accumulation in the regions that are surrounded by low curvature boundaries. The voters in each category then cast votes at their surrounding pixels according to their voting directions and classifications confidence. This results in K voting maps that are further multiplied to form the final voting map whose maximum vote accumulation (MVA) indicates the location of a polyp candidate, as shown the equation below:
where Mv(x, y) is the vote cast by the voter v at a receiver pixel r=[x, y], which is computed as follows:
where σ controls the size of the voting field.
However, in the accumulation scheme described above, votes received at each pixel are accumulated irrespective of the voters' orientation. This implies that the proposed accumulator may be undesirably sensitive to any accumulation of votes, no matter whether they are received from the edge pixels forming a circle or from the edge pixels arranged on parallel lines. Thus, regions delineated by parallel edge segments or by low curvature counters, in general, may not represent polyps. High responses in such regions may result in false positive detections.
As shown in
To overcome this problem, a constraint may be imposed on the voting scheme, such that regions can attain high accumulated votes if they receive adequate votes from voters whose normal directions span a wide range of [0, 2π), for example. Thus, regions surrounded by low curvature boundary receive low accumulation, because the edge normal of their surrounding voters can only partially cover the range of [0, 2π). Despite the proper handling of low curvature boundaries, this constraint undesirably weakens vote accumulation for the polyps that partially appear behind the folds or close to image borders. In such cases, polyps are segmented incompletely and, thus, the angular range of interest is only covered partially.
To alleviate the above problem, normal directions from [0, 2π) to [0, π), for example, can be mapped and a constraint can be placed on its maximum coverage. This allows for detecting semi-elliptical structures in edge maps. To measure the coverage, normal directions can be discretized into four values, for example. Mathematically, this may be expressed as:
In one non-limiting example, the edge normal in the ranger of
is mapped to
the edge normal in the range of
is mapped to
and so on. The positively classified edges may then be categorized into four categories according to the quantized values, and the voting process can be performed four times. Each time the voting process is performed, the edge pixels of one specific category are allowed to vote. As illustrated by process block 310 of
Turning now to
Returning to
where Rθ is an indicator variable that takes 1 if the ray at angle θ hits at least 1 voter, and 0 otherwise.
In an alternative embodiment, the post polyp voting mechanism performed at process block 418 may include defining isocontours of a voting map and then using the isocontours to estimate the unknown parameters of the bands. The isocontour Φc of the voting map V may be defined as Φc={(x, y)|V(x, y)=cM} where M denotes the maximum of the voting map and c is a constant between 0 and 1. The isocontours of a voting map can predict where actual object boundary is through a regression model. Since isocontours may get corrupted by other nearby voters in the scene, several isocontours may be obtained, as shown in
Referring again to
Specific examples are provided below, illustrative of the above-described polyp detection method. These examples are offered for illustrative purposes only, and are not intended to limit the scope of the present disclosure in any way. Indeed, various modifications of the disclosure in addition to those shown and described herein will become apparent to those skilled in the art from the foregoing description and the following example and fall within the scope of the appended claims.
As a non-limiting example, a CVC-ColonDB was used to evaluate the methodology described in the current disclosure, where CVC-ColonDB is the only publicly available polyp database, consisting of 300 colonoscopy images with 20 pedunculated polyps, 180 sessile polyps, and 100 flat polyps. The patch descriptor was evaluated first and then the whole polyp detection system was compared with the state-of-the-art methods.
For feature evaluation, 50,000 oriented images around polyp and other boundaries in colonoscopy images were collected. Half of the images corresponding to the first 150 images were selected for training a random forest classifier and used the rest for testing.
For system evaluation, a five-fold cross validation was implemented. To train the two-stage classifier, N1=100,000 oriented image patches were collected from training images with approximately 20,000 samples for each of the five predefined classes, and N2=100,000 oriented image patches were collected with 50% of images being extracted around polyps and the rest around edges in random location in training images. For classification, the random forest classifier was chosen because of its high quality probabilistic output that is utilized in the voting scheme. The trained classification system, followed by the voting scheme, was then applied to the five test folds. The voting scheme detected 267 out of 300 polyps. Thus, outperforming the state-of-the-art, where only 252 candidates were cast inside the polyps.
Polyps appear in different sizes in colonoscopy images. Because large polyps can often be effortlessly detected by colonoscopists, σv may be adjusted for detecting polyps of small and moderate sizes. Considering that the area of missed polyps are usually 9 to 16 times smaller than images, σv may vary between 70 and 90. A detection may be considered as a “true detection” if the maximum of the voting map falls inside the ground truth contour provided in the database. The accuracy of the polyp detection method can be measured by adding the true and false detections generated in the five test folds. For σvε[70, 90], the detection point was placed inside 267 polyps and produced 33 false detections which outperforms SA-DOVA descriptor with 252 true and 48 false detections (p<0.05, Z=2.01). Furthermore, while SA-DOVA descriptor requires five parameters to be tuned, the present disclosure has only two parameters, namely, σv whose range of variation can be automatically set, and σg, which as shown in
Examples of polyp localization are shown in
To obtain precision and recall rates, a threshold was changed on the probabilities assigned to the generated polyp candidates. A non-limiting exemplary experiment was conducted to compare one embodiment of the system of the present disclosure with conventional systems. As shown in the table of
In a different system evaluation, a five-fold cross validation was implemented to evaluate the polyp detection system using CVC-ColonDB. A detection was “true” if it fell inside the ground truth. As shown in the table of
Examples of polyp localization are shown in
The system trained on the entire CVC-ColonDB was further evaluated using eight short colonoscopy videos. The free-response ROC curve, as shown in
In summary, colorectal cancer most often begins as abnormal growth of the colon wall, commonly referred to as polyps. It has been shown that the timely removal of polyps with optical colonoscopy can reduce the incidence and mortality of colorectal cancer. However, polyp detection with optical colonoscopy is a challenging task and as reported, many polyps remain undetected. Computer-aided detection may offer promises of reducing polyp miss-rate.
The current disclosure describes a system and method that systematically exploits the unique appearance of polyp boundaries to suppress non-polyp edges, yielding a cleaner edge map for the above-described vote accumulation scheme. This approach can accommodate a large variation of polyp shapes, eliminate parallel edge configurations, and enable polyp detection from partially identified boundaries of polyps. Thus, the system and methods disclosed can improve the classification stage by enhancing the feature space with classification systems trained in consistent subspaces of the negative class (i.e., the boundaries of lumen, vessels, and folds, etc.), as well as evaluate the suggested methodology on a significantly larger polyp database.
In addition, the present polyp detection system and method provides feedback to enhance colonoscopists' diagnostic capabilities, especially during long and back-to-back colonoscopies, where human factors, such as insufficient attentiveness and fatigue, result in misdetection of polyps. In addition, the current disclosure describes a method that is not limited to optical colonoscopy and have provided effectiveness for polyp detection in capsule endoscopy.
The present disclosure has been described in terms of one or more exemplary embodiments, and it should be appreciated that many equivalents, alternatives, variations, and modifications, aside from those expressly stated, are possible and within the scope of the disclosure.
This application claims priority from U.S. Patent Application No. 61/983,868 filed Apr. 24, 2014.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2015/027552 | 4/24/2015 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
61983868 | Apr 2014 | US |