This disclosure relates to magnetic resonance imaging (MRI), and in particular, to image reconstruction.
One way to generate an image of the interior of an opaque structure makes use of a phenomenon known as nuclear magnetic resonance (NMR). Generally, NMR is a phenomenon by which atoms absorb energy provided from an external magnetic field, and subsequently radiate the absorbed energy as photons. By controlling the magnetic field throughout a region, one can control the frequency and phase of the emitted photons to vary as a function of the emitting atom's position in the region. Therefore, by measuring the emitted photons' frequency and phase, one can tell where the atoms are present inside the region.
In
An image 102 can be reconstructed from the k-space dataset 100 by various mathematical techniques. These techniques are useful for imaging the internal structure of a human being. In this context, the hydrogen atoms that make up the human's body are caused to undergo nuclear magnetic resonance. In the context of imaging the hydrogen in a human or other animal, this technique is sometimes referred to as magnetic resonance imaging (MRI). Since hydrogen is present in nearly every organ, tissue, fluid, or other part of a human being. MRI typically provides a relatively detailed image of the human being through non-invasive means.
In some MRI contexts, it is desirable to quickly acquire the k-space data necessary to produce an image. For example, when imaging a patient's heart, the image quality is often enhanced if the patient suppresses respiratory-induced motion by holding his breath while the k-space data is acquired. Some patients experience discomfort or difficulty holding their breath for extended periods of time, particularly patients who are in need of cardiac imaging.
One way to reduce the time required to acquire the k-space data is to employ multiple detectors, with each detector configured to detect photons from different spatial regions. This approach is referred to as “parallel” MRI, or pMRI.
In general, in one aspect, reconstructing an image from MRI data provided by an MRI machine includes: selecting a system transformation based on the MRI machine; selecting a subspace based on the system transformation; obtaining a solution vector that solves a minimization problem, the minimization problem being formulated on the subspace based on the MRI data; and displaying the image reconstructed from the solution vector.
Reconstructing an image also includes estimating a sensitivity of a receiving coil in the MRI machine, and selecting the system transformation includes selecting the system transformation based on the sensitivity. Selecting the system transformation also includes selecting the system transformation based on a sampling pattern used by the MRI machine. Obtaining a solution vector includes selecting a regularization parameter, and the subspace is selected independently of the regularization parameter. Reconstructing an image also includes selecting a basis of the subspace. The basis is selected using a conjugate-gradient least-square technique. The basis is selected using an least-squares QR technique. Obtaining a solution vector includes selecting a regularization parameter, and the basis is selected independently of the regularization parameter. The subspace consists of a Krylov subspace. Obtaining a solution vector includes using an LSQR algorithm. Obtaining a solution vector includes using a conjugate-gradient algorithm. Reconstructing an image also includes selecting a regularization parameter corresponding to a substantially vertical portion of an L-curve. Reconstructing an image also includes selecting a first regularization parameter and a second regularization parameter. The first regularization parameter corresponds to a real part of the solution vector, and the second regularization parameter corresponds to an imaginary part of the solution vector. The first regularization parameter and the second regularization parameter are selected as maximum parameters yielding a pre-determined error. The first and second regularization parameters are selected from an interval, the interval being based on a singular value spectrum of the system transformation. The interval consists of numbers between 10^2 and 10^{circumflex over (6)}. The MRI data is acquired by uniformly sampling a k-space. The MRI data is acquired by non-uniformly sampling a k-space. The MRI data is acquired by sampling over an entire k-space. The MRI data is acquired by sampling over half a k-space.
Other aspects include other combinations of the features recited above and other features, expressed as methods, apparatus, systems, program products, and in other ways. Other features and advantages will be apparent from the description and from the claims.
Although only two detector coils 204 are shown in
The computer 206 receives data provided by the MRI machine 200, and includes instruction to perform image reconstruction described below (see
a shows a schematic k-space dataset 300 constructed by irregular or non-uniform sampling. Each line 302 represents the amounts and relative phases of photons detected (or “sampled”) at a particular frequency by a detector coil 204. Typically, a relatively high quality image can be constructed from a central region 304 of k-space. Thus, in some approaches, the k-space dataset 300 is sampled along lines 302 that are concentrated in the central region 304. Using non-uniform sampling allows for comparatively greater flexibility in manipulating artifacts inherent with subsampling, and provides easily self-referenced coil sensitivity data. However, some mathematical techniques for image reconstruction rely on uniform sampling, therefore those techniques would be inapplicable to this k-space dataset 300.
By contrast,
Additionally, k-space datasets may have certain symmetries. For example, in certain cases, k-space datasets tend to have conjugate symmetry. In general, if a k-space dataset possesses such symmetry, then only a portion of the dataset need be acquired in order to recover the entire dataset. In the case of conjugate symmetry, for example, the entire k-space dataset may be reconstructed from a suitable half of the k-space dataset.
The techniques described below do not rely on any particular sampling strategy, and therefore can be used with any k-space dataset. Moreover, the techniques described below can be used on an entire k-space dataset, or half of a k-space dataset.
Theoretical Background
One can represent a k-space dataset as a vector or collection of vectors. Each sampled frequency is represented by a separate vector, whose dimension is equal to the phase resolution of the detector coil 204 that gathers data at that frequency. Each component of the vector represents the complex signal strength detected at the corresponding phase and frequency encoding. As a notational convention, all vectors are assumed to be column vectors, unless otherwise specified.
Reconstructing an image from a k-space dataset is equivalent to finding a vector ρ that solves the minimization problem:
minρ∥s−Pρ∥2, (1)
where s is an n-dimensional vector in the k-space dataset, P is an n×m matrix describing the phase and frequency encoding, along with estimates of the detector coils' sensitivities, ρ is an m-dimensional vector containing the reconstructed image (or a portion thereof), and ∥-∥2 is the L2 norm. The maxtrix P is sometimes referred to as the system matrix. The dimension of s equals the number of distinct phases that can be detected by each detector coil 204, multiplied by the number of coils employed. The dimension of ρ represents the resolution of the reconstructed image.
Typically, (e.g. at high sub-sampling rates) the dimensions of s, ρ, and P are such that the minimization problem (1) is ill-conditioned, and therefore solutions of (1) can be extremely sensitive to error or noise. For example, a solution ρ of (1) may have the property that relatively slight changes in the acquired data results in relatively large changes in other solutions. One way to mitigate this sensitivity is to solve a “regularized” minimization problem:
minρ{∥s−Pρ∥22+λ2∥Lρ∥22}. (2)
Here, L is a linear operator that is selected to impose a desired constraint on the solution ρ, and λ is a real or complex scalar. Often, the choice of L depends on how the k-space data is acquired. If data over the full range of k-space is acquired, then typically, L=I, the identity operator.
In some embodiments, data over only half of k-space is acquired. Nevertheless, a full image can be recovered, based on symmetries that typically exist in k-space data, or based on known properties of solutions ρ. If data over only half of k-space is acquired, on can use a regularization parameter λ and regularization operator L to separately constraint the real and imaginary part of the solution. On can then rewrite the minimization problem (2) as
where real {ρ} denotes the real portion of ρ, and image {ρ} denotes the imaginary portion of ρ. In some embodiments, L can be defined by
where I is an identify operator of dimension equal to that of ρ, and c=λ/λ. This operation L acts nontrivially on the imaginary part of ρ. An operator L=(1/c)L can also be used to act on the real part of ρ. In this case, λr should be replaced by λ in equation (2′). With the choice of L as in equation (2″), the regularized minimization problem (2′) has two parameters: λr and c.
In general, other choices of L can be accommodated. If L is invertible, then the techniques described below can be applied by a transformation of variables. For example, one transformation of variables is to introduce the variable η=Lρ, or ρ=L−1η. Reformulating minimization problem (2) or (2′) in terms of η eliminates the appearance of L, and the techniques described below can be used. In some implementations the system matrix can be considered to the PL−1, as opposed to P.
The regularized minimization problem (2) can be thought of as seeking a solution to the original minimization problem (1), while simultaneously attempting to constrain the solution ρ. In this context, the “regularization parameter” λ (or λr) represents the relative importance of finding a ρ that minimizes (1) and controlling the norm of ρ. For example, if λ is small, then a solution of (2) is close to a solution of (1). If used, the regularization parameter c represents the relative importance of adhering to the assumed symmetries in k-space.
Krylov Subspaces and Orthonormal Bases
Without any additional constraints, solving the minimization problem (2) is computationally intensive, in part due to the dimensions of s and ρ. The Krylov subspace techniques described below provide a way to project the minimization problem (2) to a smaller-dimensional subspace. This results in a more tractable minimization problem. To illustrate this process, we note that the minimization problem (2) is equivalent to solving:
(PHP+λ2LHL)ρ=PHs, (3)
wherein the exponent “H” denotes the Hermitian (or complex-conjugate) of a matrix. In what follows, we shall assume L=I, although the other case L≠1 follows by employing a change of variables; e.g., η=Lρ, or ρ=L−1η, to make the following observations applicable. Equation (3) is sometimes referred to as the “normal equations” corresponding to the minimization problem (2).
In general, given an n-dimensional vector v in a vector space V, and a linear operator T on V, the kth-order Krylov subspace of V generated by v and T is defined to be the linear vector space spanned by v and k−1 successive images of v under T; that is, Kk(v,T)=span{v, Tv, T2v, . . . , Tk−1v}. Where v and T are understood, the k-th order Krylov subspace is simply denoted Kk.
A Krylov subspace can be constructed based on s and P. Although the n×m matrix P is generally not a square matrix, the matrix PHP is always n×n. Thus, Krylov subspaces can be formed from the vector PHs and the operator PHP. The kth-order Krylov subspace associated with the normal equations (3) (for λ=0) is given by:
Kk=span{PHs, (PHP)(PHs), . . . , (PHP)k−1(PHs)}. (4)
With the k-th Krylov subspace identified, the minimization problem (1) and (2), or the normal equations (3) can be restricted to the k-th Krylov subspace. For example, minimization problem (1) restricts to:
where ρk denotes the solution obtained in the k-th Krylov subspace. One class to strategies for solving the restricted problems involves constructing an orthonormal basis of the Krylov subspace Kk. As described more fully below, this can be accomplished iteratively; that is, the orthonormal basis for Kk−1 can be used to construct the orthonormal basis of Kk.
Projecting to a Krylov Subspace with a CGLS-Constructed Basis.
One way to construct such an orthonormal basis involves conjugate-gradient least squares (“CGLS”) techniques. Using these techniques, a matrix Vk with columns {v, . . . , vk} can be constructed such that: the columns of Vk form an orthonormal basis of Kk; βv=PHs for some scalar β; and
(PHP)Vk=Vk+3{tilde over (H)}k (5)
for a (k+1)×k tridiagonal matrix {tilde over (H)}k. Because {tilde over (H)}k is tridiagonal, equation (5) yields a three-term recurrence that can be used to construct the next column vk from the previous columns vk−1, vk−2. Thus, the columns of Vk of need not be stored for all k, as k varies from iteration to iteration (see
If ρk is a solution of equation (3) restricted to be in a Krylov subspace Kk, then ρk factors as:
ρk=Vkyk (6)
for some coefficient vector yk, because the columns of Vk form an orthonormal basis of Kk. Inserting equation (6) into equation (3) and applying equation (5) yields:
((PHP)+λ2I)Vkyk=(Vk−1{tilde over (H)}k+λ2Vk)yk=βv1. (7)
To enforce the condition that the error between the vectors ρk and ρ is orthogonal to Kk, one multiplies (7) by VH to obtain:
(Hk+λ2I)yk=βe1. (8)
where Hk is the k×k leading tridiagonal submatrix of {tilde over (H)}k, and e1=Vk+1Hv1 is the first canonical unit vector. Equation (8) is referred to as the “projection” of equation (3) onto the kth-order Krylov subspace. Because Hk is tri-diagonal, a three-term recurrence relation can be used to obtain solutions to equation (8) using only the solutions and orthonormal basis vectors from the previous three iterations. Obtaining the three-term recurrence relation is explained in Appendix A.
Projecting to a Krylov Subspace with LSQR-Constructed Basis.
One can also use least-square QR (“LSQR”) techniques to construct a basis of the Krylov subspace Kk. As in the CGLS case described above, the vectors in this orthogonal basis are also denoted {v1, . . . , vk}. It can be arranged that if these vectors are placed as columns in a matrix Vk, then the matrix Vk satisfies the recurrence relation
PVk=Uk+1Bk, (9)
where Bk is a k+1×k bidiagonal matrix and Uk+1=[u1, . . . , uk+1] is a matrix with orthonormal columns such that u1=s/∥s∥2. Similarly to the CGLS case, since the matrix Bk is bidiagonal, the columns of Vk (and Uk+1) can be generated from a two-term recurrence. Thus, the entire matrix Vk (or Uk+1) is not needed to generated subsequent columns of the matrix as k increases—only the most recent two columns are needed. Not storing the entire matrices Vk or Uk+1 can save computational resources.
Since Vk describes a basis of Kk, the solution to equation (3) above is given by ρk=Vkyk for some appropriate coefficient vector yk. Instead, for example, a solution ρk may be constructed by using the bidiagonal nature of the Bk matrix. In particular, using equation (4) and the equation ρk=Vkyk, equation (3) can be written as
where e1 is the first canonical unit vector of length k+1 in Bk, β=∥s∥2, and the last equality follows from the fact that the L2-norm is invariant under left multiplication by a matrix with orthonormal columns. Therefore, solving the minimization problem (3) has been reduced solving the minimization problem (12), which is a problem of smaller dimension.
Using this approach, ρk is not formed from the product Vkyk explicitly: rather, the bidiagonal nature of Bk is exploited to obtain ρk from only the most recent iterates. Further, based on reasoning explained in Appendix A, one can also obtain values of the norms ∥s−Pρk∥2 and ∥ρk∥2 from short-term recurrences. As discussed more fully below, these norms can be used to select regularization parameters.
Image Reconstruction
Referring to
data is acquired from the MRI machine (step 404). To reconstruct the image, initially, a particular Krylov subspace of order k0 is selected (step 406). With k=k0, a basis Vk according to equation (5) is constructed, for example by using conjugate-gradient least squares of LSQR techniques (step 408). Using the basis Vk, the “projected” image reconstruction minimization problem is formulated as in equation (8).
This minimization problem is solved, for example using the truncated singular value decomposition approach described in connection with equation (10) (step 410). The result is a solution of equation (3) that depends on λ, and possibly c. A particular value of λ (and c, if necessary) is selected (step 412), as described below (see
Selection of Regularization Parameter(s)
Referring to
The L-curve 500 is obtained by plotting ∥ρ∥2 against ∥s−Pρ∥2 on a log-log scale, parameterized by λ. The L-curve 500 shown in
Typically, the L-curve 500 for parallel MRI reconstruction problems has a region 502 in which it is nearly vertical, and an inflection point 504. As λ varies within region 502, ∥s−Pρ∥2 changes relatively slowly. Consequently, the resulting solutions to equation (3) are relatively stable. Thus, for image reconstruction, λ is selected within region 502. Specifically, λ is chosen by partitioning the λ-interval corresponding to the substantially vertical region 502 into subintervals [λn−1, λn], and selecting λ to be the largest λn such that the condition:
|∥s−Pρλ
is satisfied for some threshold C, where ρλ
In some examples, the logarithmic error threshold is chosen to be 0.01. The point λ=λ+ schematically illustrates the location of a typical value satisfying the condition (13). A value of λ that satisfies the condition (13) may be near the inflection point 504, but need not be coincident with it.
Once a value of λ is fixed, then ρ is an approximate solution of equations (3). This solution may subsequently be used to generate an image according to known techniques. When the iterations have converged to produce an image of sufficient quality, it is displayed (step 416 in
Referring to
Referring to
In some implementations, the maximum value in the regularization range is based on the greatest singular value of the matrix P. For example, the maximum value in the regularization range may be equal to the greatest singular value of P, or may be a scalar multiple of the greatest singular value of P, the scalar multiple being based on the process by which P is determined from the detector coils 204.
The singular values of P often are distributed in two clusters, with each singular value in the first cluster being larger than each singular value in the second cluster. In some implementations, the minimum value in the regularization range is selected based on the greatest singular value in the second cluster. For example, the minimum value in the regularization range may be equal to the greatest singular value in the second cluster, or may be a scalar multiple of the greatest singular value in the second cluster, the scalar multiple being based on the process by which P is determined from the detector coils 204.
An error threshold is identified (step 702). This error threshold is analogous to the value C in equation (13). In some implementations, the logarithmic error threshold is equal to 0.01. In step 704, each axis in the regularization range (e.g., the λi and λr axes) is partitioned into intervals 604 (see
An endpoint of an interval is selected (step 706). For example, the largest partition endpoint on the λr axis. Starting from the endpoint selected in step 706, the error terms ∥s−Pρ∥2 are computed along a diagonal trajectory 606 (step 708). In
It is determined whether interval endpoints exist that have not been used in the above steps (step 712). If such endpoints exist, steps 706-712 are repeated. In repeating steps 706-712, several values of λr and λi are recorded. Once all the interval endpoints have been identified, the maximum λr and λi of all those recorded in the various iterations of step 710 are used as regularization parameters.
Other embodiments are within the scope of the following claims.
In this appendix, certain recurrence relations and algorithms are described. The notation used in this appendix follows the MATLAB conventions, and follows that of Gene Golub and Charles Van Loan, “Matrix Computations,” second edition, Johns Hopkins University Press, 1989. To the extent the MATLAB notation MATLAB and Matrix Computations is inconsistent, MATLAB notation governs.
LSQR Recurrence and Algorithm
The matrix recursions
PVk=Uk+1Bk,
PHUk+1=VkBkH+αk+1vk+1ek+1r
translate into the following matrix-vector (coupled two-term) recursions:
βu1=s;
βi+1ui+1=Avi−αivi;
α1v1=PHu1;
αi+1vi+1=PHui+1−βi+1v1
where αi and βi indicate scalar normalization constants.
for any λ. To minimize the functional on the right of the equation (1), one can use a QR factorization. However, the stacked matrix on the right is sparse, and because Bk differs from Bk−1 only in the addition of a new row and column, the factorization at step k requires only a few additional Givens (plane) rotations. For notational convenience in the QR factorization described below, the dependence on λ will be suppressed.
To solved equation (1), the QR factorization is applied to the augmented matrix as follows:
where Qk is an orthonormal k×k matrix, Rk is an upper bidiagonal k×k matrix, and otherwise the notation follows LSQR: An Algorithm for Sparse Linear Equations and Sparse Least Squares, ACM Transactions on Mathematical Software, Vol. 8, No. 1, Mar. 1982, pp. 43-71; Vol. 8, No. 2 June 1982, pp. 195-209. By the relationship of Bk to Bk−1, the matrix Qk is a product of Qk−1 (augmented by ek in the last row) and plane rotations that “zero-out” the new last row in Bk, as well as the extra λ in the last row of the stacked matrix.
This yields Rkyk=fk as the solution to the minimization problem on the right in equation (1). However, the recursion for ρk can be obtained by the fact that ρk=Vky=Dkfk, where Rkyk=fk is the solution to the minimization problem. The columns of Dk are obtained from the fact that RkTDkT=VkT, and that RkT is lower bidiagonal. Thus, solving for the last row of DkT depends only on dk−1 and vk. This allows ρk to be obtained from a linear combination of ρk−1 and dk.
By way of example, in the case k=2, the augmented matrix is transformed through four plane rotations as:
For k=3, the augmented matrix transform as:
To obtain a recurrence for ∥ρk∥2, write
ρk=VkRk−1fk=Vk
where Rk
For the norms of the residual, ∥Pρk−s∥2, note that
However, an estimate of the first term on the right is given by
where the vector qk differs from the previous qk−1 only in the last component. Thus, ∥qk∥22 can be computed from ∥qk−1∥22+ψk2.
Algorithm LSQR for Complex λ
= norm([
= ĉ*
Conjugate-Gradient Recurrence and Algorithm:
The matrix recurrence relations
PHPVk=Vk+1Hk=Vk
where gk=[0, . . . , 0, βk] and ek is the kth canonical unit vector, can be re-written as
(PHP+λ2I)Vk=Vk({tilde over (H)}k+λ2I)+gkekT.
since the matrix Tk=(
Tk=LkDkLkH
where Lk and Dk are defined by:
The entries μ1 and di in Lk and Dk depend on λ, but that dependence has been suppressed for notational convenience.
Note that the entries of Lk, Dk differ from the entries in Lk−1, Dk−1 respectively only in the last row and column of each matrix. Thus, the matrix recurrence above can be written as a matrix-vector recurrence by equating like columns, obtaining:
βk−1vk=−conj(βk−2)vk−2+PH(Pv2)−α2v2,
where the βi terms are the normalization constants located on the subdiagonal of Tk, their conjugates are on the superdiagonal of Tk, and the αi terms are on the main diagonal of Tk. Therefore, the recursion for the entries of Lk and Dk is given by:
μt−1=βt−1/dt−1,
di=αi−|μi−1|2*dt−1,
with initial condition d1=α1.
Define Ck=VkLk−H and let pk be the solution to LkDkpj=βe1. Now
ρk=Vkyk=VkTk−1βe1=(VkLk−H)Dk−1Lk−1βe1=Ckpk.
From the definition of Ck, we have
LkCkT=VkT,
so the lower bidiagonal structure of Lk gives the recurrence
ck=vk−μk−1ck−1.
Writing pk=[φ1, . . . , φk]T, we have
However, because Lk−1Dk−1pj=βe1, we have pk=[pk−1; φk], so that
φ1=β/d1
φk=−(μk−1dk−1φk−1)/dk−1
from which it follows that
ρk=Ckpk=Ck−1pk−1+φkck=ρk−1+φkck.
Using the relations above, the residual norm of the regularized equations can be written as:
∥(PHP+λ2I)ρk−PHs∥2=|βk∥ekTy|=|βk∥φk|,
since yk is a solution to LkHyk=pk.
Conjugate Gradient Algorithm (Complex λ)
This invention was made with Government support under Grant No. RRO19703 awarded by the National Institutes of Health. The U.S. Government has certain rights in this invention.
Number | Name | Date | Kind |
---|---|---|---|
5546472 | Levin | Aug 1996 | A |
6208139 | Foo et al. | Mar 2001 | B1 |
6680610 | Kyriakos et al. | Jan 2004 | B1 |
6748096 | Chuang | Jun 2004 | B2 |
6975900 | Rudy et al. | Dec 2005 | B2 |
20010038285 | Zhu | Nov 2001 | A1 |
20030120163 | Rudy et al. | Jun 2003 | A1 |
20040082870 | Rudy et al. | Apr 2004 | A1 |
20050270024 | Lin | Dec 2005 | A1 |
20070297656 | DeMan et al. | Dec 2007 | A1 |
20080012564 | Lin | Jan 2008 | A1 |
20080107319 | Chang et al. | May 2008 | A1 |
Number | Date | Country | |
---|---|---|---|
20080175451 A1 | Jul 2008 | US |