The present invention relates to medical imaging of the prostate, and more particularly, to fully automatic segmentation of the prostate in 3D magnetic resonance images.
Detection and delineation of the prostate in magnetic resonance (MR) and computed tomography (CT) images is an important problem in medical imaging analysis. For example, prostate cancer is often treated by radiation therapy where precise targeting of the prostate is important. The exposure of neighboring organs, such as the bladder or the rectum, to ionizing radiation should be reduced to an absolute minimum. For this purpose, accurate localization of the prostate by exact organ delineation is critical. In addition, prostate delineation in medical images can be used for MR spectroscopy analysis (MRSI) to obtain metabolic information about prostatic tissue. Further, recent imaging techniques such as MR perfusion may also benefit from automatic prostate delineation in order to compute and normalize model parameters such as blood flow within the segmented area. Fully automatic detection of the prostate may also be useful for automatic scan range planning of subsequent scans, for example, to automatically position saturation bands.
Manual delineation of the prostate in 3D MR data, sometimes in combination with 3D CT data, is a challenging and labor intensive task, even for expert radiologists. The manual delineation of the prostate is a time-consuming process, and there is significant inter- and intra-user variability among manual segmentations of the prostate. Accordingly, a method for fully automatically segmenting the prostate in MR volumes is desirable.
The present invention provides a method and system for automatically segmenting the prostate in multi-spectral 3D MR images. Embodiments of the present invention provide a fully automatic segmentation of the prostate with a relatively short processing time that works with several types of data including pathological cases, and provides a segmentation accuracy that allows the segmentation results to be suitable for inverse radiation therapy planning.
In one embodiment, a prostate boundary is detected in the 3D MR image data using marginal space learning (MSL). The detected prostate boundary is refined using one or more trained boundary detectors. In order to refine the detected prostate boundary, the detected prostate boundary may be split into a plurality of patches corresponding to a plurality anatomical regions, and the detected prostate boundary may be refined using a plurality of trained boundary detectors corresponding to the plurality of patches.
These and other advantages of the invention will be apparent to those of ordinary skill in the art by reference to the following detailed description and the accompanying drawings.
The present invention is directed to a method and system for fully automatic segmentation of the prostate in multi-spectral 3D magnetic resonance (MR) images. A multi-spectral MR image is composed of vector-valued intensity values, i.e., one or more scalar intensity values per voxel. Different MR channels (T1-weighted, T2-weighted, etc.) may emphasize on different tissue characteristics providing additional information about the depicted image content. Embodiments of the present invention are described herein to give a visual understanding of the prostate segmentation method. A digital image is often composed of digital representations of one or more objects (or shapes). The digital representation of an object is often described herein in terms of identifying and manipulating the objects. Such manipulations are virtual manipulations accomplished in the memory or other circuitry/hardware of a computer system. Accordingly, is to be understood that embodiments of the present invention may be performed within a computer system using data stored within the computer system.
Embodiments of the present invention are directed to fully automated prostate segmentation in multi-spectral 3D MR images. Embodiments of the present invention utilize a top-down segmentation approach based on Marginal Space Learning (MSL) to detect the prostate in a multi-spectral 3D MR image. MSL decomposes the parameter space of the prostate along decreasing levels of geometrical abstraction into subspaces of increasing dimensionality by exploiting parameter invariance. At each level of abstraction, i.e., in each subspace, strong discriminative models are trained from annotated training data, and these models are used to narrow the range of possible solutions until the final shape of the prostate can be inferred. The basic MSL framework is described in greater detail in Zheng et al., “Four-Chamber Heart Modeling and Automatic Segmentation for 3D Cardiac CT Volumes Using Marginal Space Learning and Steerable Features”, IEEE T. Med. Imag. 27(11) (November 2008), pgs. 1668-1681, which is incorporated herein by reference. Contextual shape information for the prostate is introduced by representing candidate shape parameters with high-dimensional vectors of 3D generalized Haar features and steerable features derived from observed volume intensities in an MR image.
Accordingly to an embodiment of the present invention, artificially generated training data can be generated from real-world training images by randomly perturbing (12 parameter affine perturbance) the real-world training images. This dramatically reduces the amount of annotated training data needed in order to generate discriminative models of a sufficiently high accuracy.
For combined 3D rigid prostate detection and shape inference, an extended MSL-based framework is used. A structure of interest's (i.e., the prostate) center is estimated as c=(c1, c2, c3)εIR3, orientation as RεSO(3), scale as s=(s1, s2, s3)ε{SεIR3|si>0, i=1, 2, 3}, and shape as x=(x1, y1, z1, . . . xn, yn, zn)TεIR3n. The shape parameter comprises canonically sampled 3D points xi=(xi, yi, zi)T, iε{1, . . . , n}, on the surface of the object to be segmented. Note that R is relative to c, s is relative to c and R, and x is relative to c, R, and s. Let V={1, 2, . . . , N}, NεIN, be a set of indices of image voxels, (yv)vεV, yvε{−1,1}, a binary segmentation of the image voxels into object and non-object voxels, and ƒ be a function with Y=ƒ(I,Θ) that provides a binary segmentation of volume I using segmentation parameters Θ=(c, R, s, x). Let Z=(zΘ) be a family of high dimensional feature vectors extracted from a given multi-spectral input volume I=(iv)vεV and associated with different discretized configurations of Θ. In embodiments of the present invention, Z can include voxel-wise context encoding multi-spectral 3D generalized Haar-like features to characterize possible object centers and multi-spectral steerable features that are capable of representing hypothetical orientations and optionally scaling relative to a given object center or shape surface point.
In order to detect the prostate, we search for an optimal parameter vector:
maximizing the posterior probability of the presence, i.e., y=1, of the prostate given the discriminative model M(Θ) and the features Z extracted from the input volume I using a certain set of values for the parameters Θ.
Let π(c)(Z), π(c,R)(Z), π(c,R,s)(Z), and π(c,R,s,x)(Z) denote the vectors of components of Z associated with individual groups of elements (c), (c,R), (c, R, s), and (c, R, s, x) of the parameter vector Θ. The MSL method avoids exhaustively searching the high-dimensional parameter space spanned by all possible Θ by exploiting the fact that ideally for any discriminative model for center detection with parameters M(c) working on a restricted amount of possible features,
holds, as the object center c is invariant under relative reorientation, relative rescaling, and relative shape positioning. Similarly, we have
for combined position-orientation detection with model parameters M(c,R), where only features π(c*,R)(Z) with c=c* are considered. This is due to the fact that position and orientation are invariant under relative rescaling and relative shape positioning. Analogous considerations yield
for the target object's scaling, and
for the target object's shape where M(c,R,s,x,y,z) are the parameters of a local shape model with respect to individual surface points (x, y, z)T, and parameters M(c,R,s,x) represent a global shape model. Equations (2)-(5) set up a chain of discriminative models exploiting search space parameter invariance for combined 3D shape detection and shape inference. This allows different discriminative models to be applied in descending order of geometrical abstraction as, in embodiments of the present invention, the object center c alone is the most geometrically abstract and the complete set of parameters Θ is the least abstract shape representation. Therefore, MSL establishes a hierarchical decomposition of the search space along decreasing levels of geometrical abstraction with increasing dimensionality of the considered parameter subspace.
Let Z be the set of annotated multi-spectral image volumes in their transformed feature representation as described above. Z is referred to herein as the training data. In order to detect the nine parameter similarity transformation of the optimal parameter Θ*, i.e., c*, R*, and s*, discriminative models P(y=1|π(c*)(Z)), P(y=1|π(c*,R)(Z)), and P(y=1|π(c*,R*,s)(Z)), are learned (trained) based on the training data. Following the concept of MSL, a set of positive and negative training examples C={(π(c)(Z),y)|ZεZ} are generated from the training data to train a probabilistic boosting tree (PBT) classifier for position detection. The feature vectors π(c)(Z) can be multi-spectral 3D generalized Haar-like features encoding voxel context of candidate object centers based on observed intensity values. Decreasing the level of geometric abstraction, a PBT classifier is analogously trained for combined position-orientation detection based on an extended set of training examples G={(π(c,R)(Z),y)|ZεZ}, where the feature vectors π(c,R)(Z), associated with (c,R) and a multi-spectral image volume, are multi-spectral steerable features. Steerable features allow varying orientation and scaling to be encoded in terms of aligned and scaled intensity sampling patterns. According to an advantageous implementation, multi-spectral steerable features are also used to train a PBT classifier for full nine parameter similarly transformation detection based on an extended set of training examples S={(π(c,R,s)(Z),y)|ZεZ}, where π(c,R,s)(Z) is derived from (c, R, s) and the associated multi-spectral image volume.
In order to detect the final object shape for each individual brain structure, we further decompose
π(c,R,s,x)(Z)=(π(c,R,s,x
where π(c,R,s,x
in an iterative manner. The term p(yi=1|π(c,R,s,x
The global shape model can be implemented as an active shape model (ASM), which can be used to incorporate prior shape during segmentation. Active shape models are described in detail in Cootes et al. “Active Shape Models-Their Training and Application” Comp. Vis. Image Understand. 61(1) (January 1995), pgs. 38-59, which is incorporated herein by reference. In an ASM, the shape of a target structure is represented as a cloud of points, which are either manually or automatically placed at certain characteristic locations within the class of images to be processed. Once these sets of labeled point features, or landmarks, are established for each image, they are linearly aligned to each other in order to remove translation, rotation, and scaling as far as possible. This can be done using the generalized Procrustes analysis (GPA), which is well known and described in detail in Gower “Generalized Procrustes Analysis” Psychmetrika 40(1) (March 1975), pgs. 33-50, which is incorporated herein by reference. After the GPA all the shapes are transformed to a common coordinate system—the model space of the ASM. The remaining variability can be described as a prior model using a Point Distribution Model (PDM).
As described above, MSL-based detection can be used for segmenting target anatomic objects.
Referring to
At step 204, the voxel intensities of the received multi-spectral 3D MR image data are standardized. This intensity standardization allows the segmentation method of
Steps 206-212 of the method of
At step 206, center position candidates of the prostate are detected in multi-spectral MR image data using a trained position detector. The trained position detector is a trained discriminative model, such as a PBT, trained based on the training data including the artificially generated training data. The position detector is used to scan voxels of the multi-spectral 3D MR image data to search for the center position of the prostate. The position detector may scan all voxels of the multi-spectral 3D MR image data or may scan voxels within a constrained reasonable search range. The position detector returns a confidence value for every voxel that represents the probability that the voxel is the center position of the prostate. For the purposes of distinguishing more likely center candidates from less likely center candidates, a 3D context surrounding each voxel of interest is represented by 3D Haar-like features that may have been computed on all multi-spectral input sequences. These features are input to the position detectors, and a certain number of center position candidates with the highest confidence scores are kept.
At step 208, position-orientation candidates are detected based on the position candidates using a trained position-orientation detector. For each of the position candidates detected at step 206, a reasonable range of orientations is checked resulting in a number of position-orientation hypotheses. The position-orientation detector is a discriminative model, such as a PBT, trained based on the training data including the artificially generated training data. The position-orientation detector relies on multi-spectral 3D steerable features to encode contextual information derived from the MR image data and returns a confidence value for every position-orientation hypothesis. A certain number of position-orientation hypotheses with the highest confidence value are kept as the detected position-orientation candidates.
At step 210, a full similarity transformation (position-orientation-scale) is detected based on the position-orientation candidates using a trained full similarity transformation detector. For each of the position-orientation candidates detected at step 208, a reasonable range of scales is checked resulting in a number of full similarity transformation (position-orientation-scale) hypotheses. The full similarity transformation detector is a discriminative model, such as a PBT, trained based on the training data including the artificially generated training data. The full similarity transformation detector relies on 3D steerable features to encode contextual information derived from the MR image data and returns a confidence value for every full similarity transformation hypothesis. The full similarity transformation is detected based on one or more of the full similarity transformation hypotheses having the highest scores. For example, the full similarity transformation can be detected as the full similarity transformation hypothesis having the highest confidence score. Alternatively, the full similarity transformation can be detected by clustering a certain number of full similarity transformation candidates having the highest scores. The detected full similarity transformation is used to transform a prostate model, such as a mean prostate model learned from the training data, to the MR image data.
According to an exemplary embodiment, it is also possible that the full transformation detection of step 210 can be made more robust by taking into account 3D landmarks related to the prostate and detected with separate trained landmark detectors. For example, the base and the apex of the prostate can be detected using a separate trained landmark detector for each one. The positions of the detected base and apex can then be taken into account, for example using a discriminative anatomical network (DAN), when detecting the full similarity transformation of the prostate.
At step 212, the boundary of the detected prostate is refined using one or more trained boundary detectors. On the lowest level of geometrical abstraction, the shape representation of the prostate can be enriched by a triangulated mesh surface that is fit to the MR image data using the detected full similarity transformation, which can be thought of as a local coordinate system. By using a generative shape model built from the ground truth annotations, and several discriminative models (boundary detectors) relying on multi-spectral 3D steerable features for object boundary/non-boundary distinction, the final prostate shape can be inferred using an iterative search, in which the global shape's and the boundary candidates' posterior probability are maximized in turn.
According to an advantageous implementation of the present invention, in order to improve segmentation accuracy, the prostate's surface can be split into several patches for boundary detection. The patches correspond to different regions of the prostate's surface that may have different characteristics. For example, patches corresponding to the following boundary zones of the prostate may be used: anterior fibromuscular stroma, peripheral, apex, and base. By splitting the prostate's surface into several patches, patch-specific discriminative models (boundary detectors) can be trained to more precisely capture the local characteristics of each boundary patch of interest. For example, the prostate's shared boundary with the bladder largely differs in appearance from the boundary shared with the rectum. The final prostate shape can be determined using the iterative search in which the following steps are iterated: a) the global shape's posterior probability is maximized using an active shape model (ASM); b) the posterior probability of the boundary points in each patch are maximized using the trained boundary detector for that patch.
At step 214, the prostate segmentation results are output. For example, the prostate segmentation results can be output by displaying the prostate segmentation results on a display of a computer system, or other display device. It is also possible that the prostate segmentation results can be output by storing the prostate segmentation results, for example, on a storage or memory of a computer system or on a computer readable medium. Once the prostate is segmented in the 3D MR image data, volume quantification can be performed to determine the volume of the prostate. The prostate segmentation results may be used for inverse radiotherapy planning, accurate does estimation based on the volumetric measurement, and optimizing clinical workflows. The prostate segmentation can also be used in brachytherapy planning.
As described above, the method of
According to an advantageous embodiment of the present invention, in addition to segmenting the prostate, anatomical landmarks associated with the prostate can also be detected using trained anatomical landmark detectors. This anatomical landmark detection can be done as part of the prostate segmentation, as described above, to achieve more robust segmentation results. Alternatively, the anatomical landmarks can be detected by constraining the anatomical landmark detectors to the segmented prostate. Anatomical landmarks associated with the prostate may include, but are not limited to the base and apex of the prostate. The detection of such anatomical landmarks associated with the prostate allows saturation bands to be placed for subsequent scans relative to the detected anatomy. In addition, the detected anatomic landmarks can be used for image registration, for example to compare the prostate in prior and follow-up scans.
The above-described methods for segmenting the prostate in multi-spectral 3D MR images may be implemented on a computer using well-known computer processors, memory units, storage devices, computer software, and other components. A high level block diagram of such a computer is illustrated in
The foregoing Detailed Description is to be understood as being in every respect illustrative and exemplary, but not restrictive, and the scope of the invention disclosed herein is not to be determined from the Detailed Description, but rather from the claims as interpreted according to the full breadth permitted by the patent laws. It is to be understood that the embodiments shown and described herein are only illustrative of the principles of the present invention and that various modifications may be implemented by those skilled in the art without departing from the scope and spirit of the invention. Those skilled in the art could implement various other feature combinations without departing from the scope and spirit of the invention.
This application claims the benefit of U.S. Provisional Application No. 61/262,330, filed Nov. 18, 2009, the disclosure of which is herein incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
5751838 | Cox et al. | May 1998 | A |
7136516 | Alyassin | Nov 2006 | B2 |
7916919 | Zheng et al. | Mar 2011 | B2 |
8068654 | Barbu et al. | Nov 2011 | B2 |
8073252 | Florin et al. | Dec 2011 | B2 |
8098918 | Zheng et al. | Jan 2012 | B2 |
20050283054 | Reiman | Dec 2005 | A1 |
20070014462 | Rousson et al. | Jan 2007 | A1 |
20080085050 | Barbu et al. | Apr 2008 | A1 |
20080101676 | Zheng et al. | May 2008 | A1 |
20080298659 | Spence et al. | Dec 2008 | A1 |
20080306379 | Ikuma et al. | Dec 2008 | A1 |
20090030304 | Feiweier et al. | Jan 2009 | A1 |
20090316988 | Xu et al. | Dec 2009 | A1 |
20100020208 | Barbu | Jan 2010 | A1 |
20100074499 | Wels et al. | Mar 2010 | A1 |
20100080434 | Seifert et al. | Apr 2010 | A1 |
20100134517 | Saikaly et al. | Jun 2010 | A1 |
20100286517 | Kamen et al. | Nov 2010 | A1 |
20110058720 | Lu et al. | Mar 2011 | A1 |
Entry |
---|
Deformable Segmentation of 3-D Ultrasound Prostate Images Using Statistical Texture Matching Method Yiqiang Zhan and Dinggang Shen; Mar. 2006. |
Four-Chamber Heart Modeling and Automatic Segmentation for 3D Cardiac CT Volumes Using Marginal Space Learning and Steerable Features Yefeng Zheng; Nov. 2008. |
Schmidt et al“Spine detection and labeling using a parts based graphical model” pp. 122-133 Oct. 2007. |
Cootes T.F. et al., “Active Shape Models—Their Training and Applicaton”, Computer Vsion and Image Understanding, 61(1), Jan. 1995, pp. 38-59. |
Cox, I.J. et al., “Dynamic Histogram Warping of Images Pairs for Constant Image Brightness”, In IEEE International Conference on Image Processing, vol. II, Washington D.C., USA, Oct. 1995. |
Georgescu, B. et al., “Database-Guided Segmentation of Anatomical Structures with Complex Appearance”, In IEEE Comp. Soc. Conf. Comp. Vis. Pat. Recog., San Diego, CA, USA, Jun. 2005. |
Gower, J.C., Generalized Procrustes Analysis, Psychometrika 40(1), Mar. 1975, pp. 33-50. |
Tu, Z., “Probabilistic Boosting-Tree: Learning and Discriminative Models for Classification, Recognition, and Clustering”, In IEEE Int'l. Conf. Comp. Vis., Beijing, China, Oct. 2005, pp. 1589-1596. |
Zheng, Y., et al., “Four-Chamber Heart Modeling and Automatic Segmentation for 3D Cardiac CT Volumes Using Marginal Space Learning and Steerable Features”, IEEE Transactions on Medical Imaging, 27(11), Nov. 2008, pp. 1668-1681. |
Bharatha, A., et al. “Evaluation of Three-Dimensional Finite Element-Based Deformable Registration of Pre-and Intraoperative Prostate Imaging”, Medical Physics 28(12) pp. 2551-2560, (2001). |
Gubern-Merida, A., et al. “Atlas Based Segmentation of the Prostate in MR Images”, MICCAI 2009 Prostate Segmentation Challenge, London UK. (2009). |
Heimann T., et al., “3D Segmentation in the Clinic: A Grand Challenge”, pp. 7-15. (2007). |
Klein, S., et al. “Segmentation of the Prostate in MR Images by Atlas Matching”, Proceeding of IEEE International Symposium on Biomedical Imaging, pp. 13001303. (2007). |
Number | Date | Country | |
---|---|---|---|
20110116698 A1 | May 2011 | US |
Number | Date | Country | |
---|---|---|---|
61262330 | Nov 2009 | US |