The present disclosure relates generally to a computer aided diagnosis of diseases, and more specifically, relates to the detection and diagnosis of colonic polyps using in vivo imaging.
Colorectal carcinoma is the third most common cancer in both men and women worldwide. According to the American Cancer Society, an estimated 101,340 cases of colon cancer and 39,870 cases of rectal cancer are expected to occur in 2011. Colorectal cancer incidence rates have been decreasing for the past two decades, from 66.3 cases per 100,000 persons in 1985 to 45.3 in 2007. The declining rate accelerated from 1998 to 2007 (e.g., 2.9% per year in men and 2.2% per year in women), which can be attributed to the increase in the use of colorectal cancer screening tests that allow the detection and removal of colorectal polyps before they can progress to cancer. In contrast to the overall decline, among younger adults less than 50 years old who are not at the average risk and for which the screening is not recommended, the colorectal cancer incidence rate has been increasing by 1.6% per year since 1998. This indicates that it may be beneficial to perform regular screening examinations by a healthcare professional, which can result in the detection and removal of precancerous growths at an early stage when the growths are most treatable. Fiber-optical colonoscopy (“FOC”) is currently the preferred method for colon polyp detection. However, the perceived discomforts associated with preparing the colon, and the potential proliferation risk, have impeded the usage of FOC. As a potential minimally-invasive screening technique, computed tomographic colonography (“CTC”), or CT-based virtual colonoscopy, has shown several advantages over a FOC. Computer-aided detection (“CADe”) of polyps has been proposed to improve the consistency and sensitivity of CTC interpretation, and to reduce interpretation burden.
A typical CADe pipeline for CTC starts from a three-dimensional (“3D”) model of the colon generated from the 2D CT data. From the 3D model, a segmented colon wall can be derived. Based on the segmented colon wall, initial polyp candidates (“IPCs”), each of which can be represented by a group of image voxels, namely a patch, can be localized on the colon wall. Unfortunately, due to the complexity of the colon structure (e.g., folds, cecal valve, etc.), and the presence of colonic materials such as fluid and fecal residuals which can mimic the structure of polyps, there can be a substantial number of false positives (“FP”) in the IPC pool. Therefore the reduction of the FP rate remains a challenge for the current CADe pipelines.
To achieve a high FP reduction rate, a large number of features have been developed to classify the IPCs. Combined with some empirical constraints, these features have worked well in differentiating some spherical IPCs from colon folds, which are a major source of FP findings in the colon. These features can be divided into two categories. The first category is geometry-related features that consider the global or local shape variations of the IPCs, such as shape index (“SI”), curvedness, sphericity ratio, convexity or concavity and surface normal overlap. These geometric features can generally be based on the assumption that there exists an iso-surface between a polyp and its neighboring tissues. Accurately detecting geometric features generally requires good quality image segmentation before the feature extraction procedure. The second category of features is texture-related features which typically consider the internal structure pattern of each IPC volume, such as gradient concentration, growth ratio, density projection, and statistical indices of the aforementioned features. For example, Yoshida and Nappi computed the values of some volumetric features (e.g., SI, curvedness, CT density value, gradient, gradient concentration (“GC”), and two GC-evolved features of dGC and mGC from all the image voxels in each IPC volume). (Nappi J. and Yoshida H., “Feature-guided analysis for reduction of false positives in CAD of polyps for computed tomographic colonography”, Medical Physics, 30(7): 1592-1601, 2002). The distribution of the values of each feature over the 3D space can form an internal pattern (e.g., volumetric texture) for each IPC. These volumetric textures can be depicted by some statistic indices of the distribution. Their results can indicate that these internal textural features have improved the performance of the CADe for CTC. The advantage of volumetric textural features over geometry-related features is that they can make full use of the image voxels inside an IPC volume to identify some internal patterns.
Based on the above, it may be beneficial to depict the texture features of an IPC volume and utilize the model for FP reduction in CADe. It may further be beneficial to perform computer aided diagnosis (“CADx”) for differentiating among pathology types of the various polyps detected by CADe.
These and other deficiencies can be addressed with the exemplary systems, methods, and computer-accessible mediums for the detection and diagnosis of polyps set forth in the present disclosure.
These and other objects of the present disclosure can be achieved by provision of systems, methods and computer-accessible mediums for diagnosing a region of interest (“ROI”) within an anatomical structure which can include receiving a 3D volumetric representation of the anatomical structure, identifying at least one volume of interest (“VOI”) of the anatomical structure, generating a first feature set based on a density, a gradient and a curvature of the volume of interest, and comparing the first feature set to a second feature set to diagnose the region of interest to at least one of a plurality of pathology types.
In some exemplary embodiments, the generation of the first feature set can include determining a gradient of the volume of interest, determining a curvature of the volume of interest, and combining the gradient and the curvature with the original density to produce the first feature set. In some exemplary embodiments, at least one of the gradient or the curvature or the original density can be determined using a 3D Haralick model. In some exemplary embodiments, the gradient can be determined using a gray-level gradient co-occurrence matrix. In certain exemplary embodiments, the curvature can be determined using a gray-level curvature co-occurrence matrix. In certain exemplary embodiments, the original density can be determined using a gray-level co-occurrence matrix. In certain exemplary embodiments, the second feature set can be generated by manually analyzing a plurality of regions of interest. In certain exemplary embodiments, the anatomical structure can be a polyp. In some exemplary embodiments, the region of interest can be detected. In some exemplary embodiments, the region of interest is diagnosed only if the anatomical structure is detected to be a polyp. In certain exemplary embodiments, the detection can include comparing the volume of interest to the volume of normal (“VON”) in the 3D volumetric representation of the anatomical structure.
In certain exemplary embodiments, the 3D volumetric representation of the anatomical structure can be generated using an in-vivo imaging method. In certain exemplary embodiments, the first feature set can be compared to the second feature set using a support vector machine (“SVM”). In certain exemplary embodiments, 2D imaging information of the anatomical structure can be received and converted into the 3D volumetric representation of the anatomical structure. In certain exemplary embodiments, the 2D imaging information can be generated using computed tomography. In a preferred embodiment, the first feature set and the second feature set have at least 50 features.
In accordance with a further exemplary embodiment are systems, methods and computer-accessible mediums for diagnosing a region of interest within an anatomical structure which can include receiving first information related to a gradient and a curvature of a volume of interest and a volume of normal of the anatomical structure, and comparing the first information to the second information to diagnose the region of interest to be at least one of a plurality of pathology types.
Further objects, features and advantages of the present disclosure will become apparent from the following detailed description taken in conjunction with the accompanying Figures showing illustrative embodiments of the present disclosure, in which:
a)-(c) are exemplary images with manually-drawn outlines of the boundaries of the VOI and VON on the image slices;
Throughout the drawings, the same reference numerals and characters, unless otherwise stated, are used to denote like features, elements, components, or portions of the illustrated embodiments. Moreover, while the present disclosure will now be described in detail with reference to the Figures, it is done so in connection with the illustrative embodiments and is not limited by the particular embodiments illustrated in the Figures.
The exemplary embodiments of the present disclosure may be further understood with reference to the following description and the related appended drawings. The exemplary embodiments of the present disclosure relate to exemplary systems, methods and computer-accessible mediums for the computer-aided detection and diagnosis of polyps in a virtual colonoscopy. Specifically, the exemplary system, method and computer-accessible medium can extract features of a region of interest, classify the features, and compare the features to a known feature set in order to detect and diagnose a polyp.
After the 3D imaging information is generated and/or received, one or more volume(s) of interest within the anatomical structure being analyzed can be selected and corresponding volume(s) of normal can also be selected at step 115. A VOI can be selected based on IPCs in the 3D volumetric information, each of which can be represented by a group of image voxels. VONs can be selected based on each VOI, at a certain local distance away from the VOI. Based on the VOIs and the VONs, a feature set of the VOIs and VONs can be extracted at step 120, and compared to a further, known, feature set at step 125. The further feature set can be stored in a database, and can be generated based on a prior automatic or manual classification of real-world, non-virtual, biopsies.
The CT images can be acquired using the different image slice thickness from patient to patient. In order to facilitate the feature extraction, it can be beneficial to interpolate the CT images into the same image slice thickness for all patients. t For that purpose, each volumetric data of a CTC scan can undergo a known Monotonic Cubic Interpolation procedure to transform the image elements into isotropic cubic voxels, as described in Fritsch. (Fritsch F and Carlson R. “Monotone piecewise cubic interpolation”, SIAM Journal on Numerical Analysis (SIAM), 17(2): 238-246, 1980, the disclosure of which is hereby incorporated by reference in its entirety.) Because the data can have uniform voxel spacing within the transverse plane (e.g., the image slices), and a larger voxel spacing between image slices, the interpolation can only be performed along the axial direction. Additionally, as the data can be acquired utilizing fecal tagging, electronic colon cleansing, which is known in the art, can be performed to remove the tagged colonic materials via an exemplary statistical image segmentation and post-segmentation operation. After the electronic colon cleansing is performed, a clean virtual colon lumen, and a gradient or partial volume (“PV”) layer representing the mucosa or the inner border of the colon wall in a volumetric shell form, can be achieved.
From the segmented PV layer representing the mucosa or the inner border of the colon wall in a volumetric shell form, computer-aided detection of initial polyp candidates (“IPCs”) (S. Wang, H. Zhu, H. Lu, and Z. Liang (2008), “Volume-based Feature Analysis of Mucosa for Automatic Initial Polyp Detection in Virtual Colonoscopy”, International Journal of Computer Assisted Radiology and Surgery, vol. 3, no. 1-2, 131-142); (H. Zhu, Y. Fan, H. Lu, and Z. Liang, “Improving Initial Polyp Candidate Extraction for CT Colonography”, Physics in Biology and Medicine, vol. 55, no. 3, 2087-2102, (2010), the disclosure of each publication is hereby incorporated by reference in its entirety), can be applied to the volumetric shell to identify suspicious patches or a group of image voxels in the shell. The patch can then be called IPC. From a group of image voxels in the shell, a volume can be extracted which includes image voxels for the associated IPC. The image voxels in the volume associated with an IPC can be used in the feature computation procedure, and in the selection in the volume. The selected features can then be used to reduce the number of FPs in the IPC pool (e.g., the CADe) and further, to differentiate the TPs as malignance or benign (e.g., CADx).
For each IPC identified, its location can be determined on the 2D CT image slices. The IPC borders can be manually outlined on each image slice by repeatedly reviewing it in different window positions and window widths. For example,
For the borders between the IPC and the lumen (e.g., the ROI-air border), the image contrast is typically very high, and the drawing can be a relatively easy task. Further, the erroneous inclusion of air pixels can be corrected by a known computerized procedure. For the borders between the IPC and the wall (e.g., ROI-tissue border), the image contrast can be limited, and the drawing task is more challenging. According to exemplary embodiments of the present disclosure, the ROI-tissue border can be determined and drawn according to an observed gray value variation by repeated review of an IPC at different window positions and window widths. Some prior knowledge can also be taken into account in the drawing procedure. For example, a ROI-tissue border can often be recognized as having a convex shape and being confined in the mucosa layer.
A VOI of normal tissue, or Volume of Normal (“VON”) can be obtained by a similar method of accumulating a number of ROIs of normal tissue or regions of normal (“RON”) on a few 2D CT image slices. A VON can be drawn at a distance proximate to the VOI in the same CTC dataset. The criterion for drawing the RON-tissue borders can be such that they can include normal tissues, and be confined in the mucosa layer and, can have a convex shape. The RON-air borders can have a shape that is either convex or concave depending on its location. For example, if a RON-air border is located on a colon fold, it can have a concave shape. If it is located on the colon wall, it can have a convex shape. The exemplary system, method and computer-accessible medium can also aid in the drawing of RONs such that all the RONs of a VON are consistent to their neighbor RONs. For example,
By stacking the ROIs and RONs together, initial VOIs and VONs pairs can be obtained. A computerized procedure can be applied to the initial VOIs and VONs to remove the included air voxels. For example, a segmented PV layer, such as disclosed in Wang (Wang S, Li L, Cohen H, Mankes S, Chen J, and Liang Z, “An EM approach to MAP solution of segmenting tissue mixture percentages with application to CT-based virtual colonoscopy”, Medical Physics, 3512: 5787-5798, 2008,) the disclosure of which is hereby incorporated by reference in its entirety, can be used where the air percentage in each of the included voxels has been computed. In order to consider the whole volume of each initial VOI or VON, a global threshold strategy, such as described in Gonzalez (Gonzalez R and Woods R. Digital Image Processing, 2nd ed., Pearson Education, Delhi, India, (2002), the disclosure of which is hereby incorporated by reference in its entirety,) can be employed to remove the included air voxels.
The procedure for determining the VOI from the group of voxels of a patch, which is initially detected as IPC, can be automated by dilation and erosion operations under constraints of convexity and concavity as described in Zhu et al. (H. Zhu, Z. Liang, M. Barish, P. Pickhardt, J. You, S. Wang, Y. Fan, H. Lu, R. Richards, E. Posniak, and H. Cohen (2010), “Increasing CAD Specificity by Projection Features for CT Colonography”, Medical Physics, vol. 37, no. 4, 1468-1481).
From the obtained volumes of either the VOI or the VON, various features can be extracted for the purposes of CADe and CADx.
A VOI or VON can be treated as a 3D image I, and then texture features can be extracted from the 3D image to form a feature vector for I. For 3D image I, a corresponding 3D gradient image Ig, can be computed using, for example, a modified Sobel operator to compute Ig, such that it can be applied in a 3D mode. The implemented 3D Sobel kernel in the z-direction can be shown as, for example:
For a voxel in image I which can be denoted as I(i, j, k), the derivatives Gx(i, j, k), Gy(i, j, k) and Gz(i, j, k) can be computed in the three orthogonal directions, respectively. The corresponding voxel value in the gradient image Ig can be computed according to, for example:
I
g(i,j,k)=√{square root over (Gx2(i,j,k)+Gy2(i,j,k)+Gz2(i,j,k))}{square root over (Gx2(i,j,k)+Gy2(i,j,k)+Gz2(i,j,k))}{square root over (Gx2(i,j,k)+Gy2(i,j,k)+Gz2(i,j,k))}. (2)
In order to analyze the texture of the volume image I in 3D mode, a model that elaborates the concept of the 3D textures can be used. Utilizing a texture analysis method, such as an extension of the analysis proposed by Haralick et al. (Haralick R and Shanmugam K. “Textural features for image classification”, IEEE Transactions on Systems, Man, and Cybernetics, SMC-3(6): 610-621 (1973), the disclosure of which is hereby incorporated by reference in its entirety), a model can be developed that accounts for the frequency of gray level co-occurrence pairs for a certain distance d along one direction in the 3D space, and records them in a 2D matrix Md,θ, (e.g., gray level co-occurrence matrix (“GLCM”)). The term (x, y, z) can denote the coordinate of a voxel in I, and (x′, y′, z′) can denote another voxel whose Euclidian distance is d along direction θ from (x, y, z). The element (i, j) in the 2D co-occurrence matrix, GLCM, can be computed by, for example:
where x, y, z represents the dimensions in three axes of I. If the distance d is fixed, the number of co-occurrence matrices that can be generated can be equal to the number of directions used for image I. According to exemplary embodiments of the present disclosure, the distance d can be measured by voxel units, along direction +θ and −θ; the details of which are described below.
Exemplary 3D texture features can include a gradient, a curvature, and a combination of the gradient and the curvature. These exemplary 3D texture features can be created by applying the method proposed by Haralick, but extended into the 3D space, as described in more detail below.
The Haralick model can be used to analyze texture patterns in a 2D gray-level image (e.g., the original density). The basis of Haralick features is the GLCM (e.g., the GLCM shown below). This matrix can be square with dimension Ng, where Ng can be the number of gray-levels of the density image. Element p(i, j) can be the normalized frequency pixel with value i adjacent to a pixel with value j in a specific direction θ with distance d, which can usually be set to 1. As each pixel has 8 nearest neighbors, there can be four directions, (e.g., 0° (180°), 45° (225°), 90° (270°) and 135° (315°)). The direction 0° can be considered to be the same as direction 180° for the feature calculations. Therefore, only four directions need to be considered. Fourteen initial features are computed from each direction, resulting in a total of 4×14 initial features in each 2D case. For each of the 14 initial features of a direction, the average value and range value over the four directions can be computed, resulting in a total of 2×14 the final features, 14 for the average and 14 for the range. The GLCM representing the so called Haralick features can be written as, for example:
The Haralick method described above can be extended to 3D gray-level image (e.g., the original density). In a 3D image, each voxel can have 26 distance 1 (d=1) neighbors, which can result in 13 directions in the 3D model, such as shown in
The GLCM can capture the correlation information between a pixel and its neighbors in the 2D gray-level density image or in the 3D gray-level density image above. Useful pattern information can also be depicted by the density-gradient/curvature pair, which can produce improved information regarding the lesions because of the higher order representations of the texture patterns, similar to the amplification in microscopy. In addition to the expansion from 2D to 3D, and the amplification, another advantage of the feature calculation above is that the parameter selection step for θ and d, as is known in the art, can be avoided because only one matrix (e.g., the GLCM) will be generated, and no parameter needs to be optimized (both θ and d are determined as described above).
As the orientation of each VOI or VON is unknown, and it is also unknown if the texture features of each VOI or VON are directionally invariant, it is preferable to use a model that is isotropic. In such a case, the derived volumetric features can be invariant regardless of which direction a VOI or VON is oriented. To achieve this, the directions can be uniformly distributed on a unit sphere. In this way, the isotropic trait can be obtained if each feature over all directions is averaged. The resulting directions can be shown in exemplary Table 1 below, which can be represented by vector directions. For each direction in Table 1, directions +θ and −θ can be used to compute a GLCM. Thus, 26 directions, or 13 pairs of directions, can be elaborated which are uniformly distributed on a unit sphere for the GLCM model.
Based on the concept of GLCM derivation in the 3D image, the model can be extended to include a 3D gray level and gradient co-occurrence matrix (“GLGCM”), which can be a derivation of the 3D image (Step 1005 of
The gradient image has a similar size to the original density image. This GLGCM can have a dimension Ng×Ngra where Ngra can be for the number of gradient levels in Ig. Element pgra (i,j) can be the normalized frequency of a voxel with value i in density gray-level image and j in corresponding gradient image in the same position. For example:
Similar to the GLCM, the GLGCM can capture the second-order statistics as the value of a voxel in Ig can be computed from a local neighborhood, which can reflect the inter-voxel relationship in the gradient space. As shown in
The dimensions of these co-occurrence matrices of the GLCM and the GLGCM can be very large. According to equation (3), the dimension of the GLCM can reach the maximum gray level in I. The dimension of the GLGCM can be even larger because its dimension can reach the maximum gray level in I, or the maximum gradient in Ig, whichever is larger. The CT density value in a VOI or VON can typically range from −1024 HU to 3071HU. Preferably, the original CT value range can be shifted to the range of 1˜4096 in order to guarantee positive i and j in equation (3). Although there is usually only one type of structure in a VOI or VON, and the gray level range can be relatively narrow, the range can still be too wide as compared to the limited dimension of a VOI or VON. A co-occurrence matrix derived based on the above can have the character of a singularity (e.g., contains a lot of zero elements), and can be difficult to manipulate. Therefore, after the shifting operation on the gray values in I, a scaling operation on the gray values of I and Ig can be performed to map the values into the same range which can result in two normalized images I′ and I′g, respectively. The scaling operation can have two parameters for the mapping, named rescaling factors S and Sg. The rescaling factors S and Sg can be determined for an adequate scale from which the textures can be viewed, and can be similar to the magnifying scale of a microscope in biopsy.
After the normalization via the shifting and scaling operations, GLCMs in 13 chosen direction pairs according to equation (3) can be computed (e.g., see Table 1), and 14 square-form GLCMs (corresponding the 14 initial features) from I′ can be obtained. Similarly, a GLGCM from I′ and I′g can be computed according to equation (5).
During In the implementation of extracting the features from GLGCM in the gradient space, there can be a simplification procedure in order to save computation time. For example, the recorded frequency in the GLCM can be employed for an image voxel in the density image to determine the corresponding frequency in the GLGCM for that voxel in the gradient image. Therefore, 14 features can be captured which reflect essentially the same patterns of the 28 features from the GLCM.
In addition to the gradient information, the curvature can be considered to reflect the higher order differentiation of the texture patterns, by which a geometric object deviates from being flat (e.g., a 3D surface). If a surface-like pattern exists inside the 3D volume image, the curvature information can help to improve diagnosis performance. As with the GLGCM, a gray-level curvature co-occurrence matrix (“GLCCM”) can be built (step 1010 of
where I.. can be the partial derivatives of the grey-level image function I(x,y,z). The 3D Deriche filters can be used to compute the partial derivatives of the image data, where, for example:
f
0(x)=c0(1+α|x|)e−α|x|
f
1(x)=−c1xα2e−α|x|
f
2(x)=c2(1−c3α|x|)e−α|x| (8)
The normalization coefficients c0, c1, c2, c3 can be set to, for example:
where α can control the degree of smoothing. The partial derivative can be determined to be, for example:
I
xx=(f2(x)f0(y)f0(z))*I
I
xy=(f1(x)f1(y)f0(z))*I (10)
Iyy, Iyz, Ixz, Izz can be determined by substituting the variables in the above equations. Two principal curvatures can be calculated using H, and the Gaussian curvature and the GLCCM can be built up, for example, as:
where Ncur is the number of curvature levels in the curvature image. By the same description using
During the extraction of the features from GLCCM in the curvature space, there can be a simplification procedure in to save computation time. The recorded frequency in the GLCM can be recorded for an image voxel in the density image to find the corresponding frequency in the GLCCM for that voxel in the curvature image. Therefore, 14 features can be captured, which reflect essentially the same patterns of the 28 features from the GLCM.
Both the GLGCM and the GLCCM can be based on the relationship of the high order images and the base grey level image. However, the co-occurrence matrix is constructed as a high order, which can be done by building a gradient curvature co-occurrence matrix (“GCCM”) (step 1015 of
where Ngra and Ncur are the number of gradient and curvature levels in Igra and Icur respectively. In this case, similar to the simplification implementation for the models of the GLGCM and the GLCCM, a total of 14 texture features can be calculated based on the GCCM. The recorded frequency in the GLGCM can be recorded for an image voxel in the gradient image to find the corresponding frequency in the GCCM for that voxel in the curvature image. Therefore, 14 features can be captured which reflect essentially the same patterns of the 28 features from the GLGCM.
The GLCM, GLGCM, GLCCM, and GCCM, for each VOI or VON, can represent our texture model of the volume. From this texture model, texture features can be derived to perform the CADe and CADx tasks. For example, from the original GLCM of Haralick et al, 14 features can be computed, as suggested by Haralick et al. (Haralick R and Shanmugam K., “Textural features for image classification”, IEEE Transactions on Systems, Man, and Cybernetics, SMC-3(6): 610-621 (1973)). In one preferred embodiment, the 12 features can include Angular Second Moment (e.g., Energy), Contrast, Correlation, Sum of Squares (e.g., Variance), Inverse Difference Moment, Sum Average, Sum Variance, Sum Entropy, Entropy, Difference Variance, Difference Entropy, Information Measures of Correlation in two forms, and Maximal Correlation Coefficient. Equations for the computation of these features can be seen in the following equations, and in Table 2, where p(i, j) can be the (i, j)th entry in a normalized gray-tone spatial-dependence matrix. Ng can be the number of distinct gray levels in the quantized image. Σi and Σj are Σi=1N
p
x(i)=Σi=1N
p
y(j)=Σj=1N
μx, μy, σx and σy can be the means and standard deviations of px and py.
HX and HY are entropies of px and py.
The above 14 features can be the initial features on a direction. For 13 directions in 3D space, there can be a total of 13×14 initial features. The initial features' average and range over the 13 directions can be computed, resulting in 2×14 final features; 14 for the average and 14 for the range. As previously discussed, the averaging over uniformly distributed 3D directions can produce the isotropic trait for each feature from the GLCMs. Two first-order statistical features can also be computed (e.g., mean and variance), directly from the volume data of VOI or VON. Thus 2×14+2=30 features can be computed from the GLCM.
Similarly 2×14 final features from the GLGCM, 2×14 final features from the GLCCM, and 2×14 final features from the GCCM can be computed. Concatenating these features together, a 114-feature vector can be formed for each VOI or VON (step 1020 of
The exemplary texture model with derived volumetric features was tested in a CTC database that includes 67 patients who have each undergone CTC screening with FOC follow-ups. Each patient followed a one-day low-residue diet with oral contrast for fecal tagging, and underwent two CTC scans in both supine and prone positions. Multi-detector (e.g., 4- and 8-MDCT) scanners were used and 134 CTC scans were produced. The scanning protocols included mAs modulation in the range of 120-216 mA with kVp of 120-140 values, 1.25-2.5 mm collimation, and reconstruction interval of 1 mm. The scanners rotated at a speed of 0.5 seconds of rotation. A total of 119 polyps and masses, sized in the range of 4-30 mm, were confirmed by both FOC and CTC. The scans were considered at two positions for each patient as two different scans, which resulted in 134 CTC scans and 238 polypoid cases. To avoid the interference from the oral contrast tagging, 47 cases that were sized less than 5 mm and almost totally buried in the contrasted materials were excluded from the study. Thus a database with 191 lesions (e.g., polyps and masses) was created from the 134 CTC scans. As described above, both manual and automated procedures outlined 191 VOI cases corresponding to the true polyps and masses or true positives in the 134 CTC scans. The additional 191 VONs from normal tissue regions were extracted for comparison purpose, resulted in 382 cases in total to evaluate the proposed volumetric features. Given the biopsy results in the FOC screening, the detailed information of the database of 382 cases is shown in Table 2. The numbers of hyperplastic (“H”), tubular adenoma (“Ta”), tubulovillous adenoma (“Va”) and adenocarcinoma (“A”) polyps or masses was 56, 94, 34 and 7, respectively. The trend of the risk rate is also shown by the right column in Table 3.
To ensure a high sensitivity rate, a typical CADe pipeline for CTC can generate a large number of IPCs, which can be a mixture of TPs and FPs. A variety of filters which utilize some geometric or textural features, have been designed to reduce the FPs. To test the ability of the proposed volumetric texture features in FP reduction, the 191 VOIs and 191 VONs were treated as IPCs from the initial operation of a CADe pipeline or CAD of IPCs (S. Wang, H. Zhu, H. Lu, and Z. Liang (2008), “Volume-based Feature Analysis of Mucosa for Automatic Initial Polyp Detection in Virtual Colonoscopy”, International Journal of Computer Assisted Radiology and Surgery, vol. 3, no. 1-2, 131-142] [H. Zhu, Y. Fan, H. Lu, and Z. Liang, “Improving Initial Polyp Candidate Extraction for CT Colonography”, Physics in Biology and Medicine, vol. 55, no. 3, 2087-2102, (2010)), where VOIs can reflect the TPs and VONs can reflect the FPs. Two classifiers of support vector machine (“SVM”) and linear discriminant analysis (“LDA”) were used to perform the task of FP reduction in the IPC pool. All the derived feature vectors were randomly assembled, and were trained and tested under a two-fold cross-validation strategy. For each fold, all features were randomly assigned to two sets d0 and d1, such that both sets were equally sized. Then training was performed on do and testing was performed on d1, followed by training on d1 and testing on d0. This testing scheme has the advantage in that the sample size of the training and test sets is large, and each feature vector is used for both training and testing on each fold. The two-fold cross-validation was repeated 50 times. The average sensitivity and specificity of the classification can indicate the ability of the proposed method to reduce the number of FPs.
In order to discriminate among the pathology types of the polyps using the proposed model (
After the interpolation procedure, an average of 725±91 slices was obtained for all CTC scans, and a cubic voxel size of 0.7174±0.0744 min was used. In the extraction of VOIs and VONs, the parameter s used to stop the iteration was 10−8. The global threshold T was set at 452.83±59.09 HU for the entire database.
A suitable SVM package for use in the present systems and methods is described in Chang. (Chang C and Lin C. “LIBSVM: A library for support vector machines”, ACM Transactions on Intelligent Systems and Technology, 2(27):1-27, 2011), the disclosure of which is hereby incorporated by reference in its entirety). A guide was followed to use a grid search to select the best-fit parameters. The LDA used was implemented by the R CARN package [e.g., R CARN package Online]. For the purpose of testing the proposed model in a CADe pipeline for FP reduction, the 382 vectors extracted from VOIs and VONs were randomly sorted, and then randomly split in half. The two-fold cross-validation was implemented to obtain the sensitivity and the corresponding specificity. The random grouping and two-fold cross-validation procedures were iteratively repeated 50 times. The final results were drawn on the average of the 50 iterations, as shown in Table 4 below.
According to the evaluation method of
Referring to
According to the accumulative variance percentage curve shown in
As is evident from the above, there is no significant difference between groups H and Ta. However, significant difference does exist between groups of A and Va (p<0.05), A and H/Ta (p<0.001), and Va and H/Ta (p<0.001). The Norm group is significantly different from each of the four polypoid groups. This can also be seen in the CADe results in Table 3 above.
To further explore the above case of no significant difference between groups H and Ta, more Hotelling T-square tests were performed as the number of principal components in
As shown in
Further, the exemplary processing arrangement 905 can be provided with or include an input/output arrangement 930, which can include, for example, a wired network, a wireless network, the internet, an intranet, a data collection probe, a sensor, etc. As shown in
The term “about,” as used herein, should generally be understood to refer to both the corresponding number and a range of numbers. Moreover, all numerical ranges herein should be understood to include each whole integer within the range.
While illustrative embodiments of the disclosure are disclosed herein, it will be appreciated that numerous modifications and other embodiments may be devised by those skilled in the art. For example, the features for the various embodiments can be used in other embodiments. Therefore, it will be understood that the appended claims are intended to cover all such modifications and embodiments that come within the spirit and scope of the present disclosure.
The present application claims the benefit under 35 U.S.C. §119(e) of U.S. Provisional Patent Application Ser. No. 61/619,208 filed on Apr. 2, 2012, the content of which is hereby incorporated by reference in its entirety.
This invention was made with government support under grant number CA082402 awarded by the National Institute of Health. The government has certain rights in the invention.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US13/32110 | 3/15/2013 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
61619208 | Apr 2012 | US |