1. Field of the Invention
The present invention relates to reducing noise in computed tomography (CT) images during CT-fluoroscopy.
2. Discussion of the Background
CT-fluoroscopy involves continuous scanning of a slice or volume of a subject for monitoring in real time, such as monitoring interventions. If a regular dose of x-rays is used, the subject will be exposed to a large x-ray dose. If a lower dose is used, then image noise is increased. In CT, image noise is inversely proportional to the square root of the x-ray tube current. As the tube current is decreased to reduce dose, the image noise increases, resulting in poor image quality. One method used to reduce image noise is to average the image slices at the same location, but this produces blurring of the edges since there is bound to be movement of the subject, voluntary or involuntary, during the scan. For example, involuntary motion can be due to breathing or beating of the heart.
One aspect of the present invention is a computed-tomography method including exposing an object with x-rays at a plurality of scans at a position of the object to obtain projection data at a plurality of views, defining a group of views, where each scan includes a first number of the groups, generating first images respectively using projection data from each group of views, generating second images from plural ones of the first images, generating third images by averaging respective pluralities of the second images, generating a gradient image using at least one of the second and third images, and generating a display image by weighting one the of second images and one of the third images using the gradient image.
In another aspect of the invention, a computed-tomography apparatus includes an x-ray source to expose an object with x-rays at a plurality of scans at a position of the object to obtain projection data at a plurality of views, an x-ray detector, a data collection unit, a data processing unit connected to the data collection unit, and a display. The data processing unit includes a memory storing x-ray projection data for a plurality of scans at a position of an object to obtain projection data at a plurality of views, and the data processing unit generates first images respectively using projection data from each group of views, generates second images from plural ones of the first images, generates third images by averaging respective pluralities of the second images, generates a gradient image using at least one of the second and third images, and generates a display image on the display by weighting one the of second images and one of the third images using the gradient image.
A more complete appreciation of the invention and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
X-ray controller 8 supplies a trigger signal to high voltage generator 7. High voltage generator 7 applies high voltage to x-ray source 3 with the timing with which the trigger signal is received. This causes x-rays to be emitted from x-ray source 3. Gantry/bed controller 9 synchronously controls the revolution of rotating ring 2 of gantry 1 and the sliding of the sliding sheet of bed 6. System controller 10 constitutes the control center of the entire system and controls x-ray controller 8 and gantry/bed controller 9 such that, as seen from the subject, x-ray source 3 executes so-called helical scanning, in which it moves along a helical path. Specifically, rotating ring 2 is continuously rotated with fixed angular speed while the sliding plate is displaced with fixed speed, and x-rays are emitted continuously or intermittently at fixed angular intervals from x-ray source 3. The source may also be scanned circularly.
The output signal of two-dimensional array type x-ray detector 5 is amplified by a data collection unit 11 for each channel and converted to a digital signal, to produce projection data. The projection data output from data collection unit 11 is fed to processing unit 12. Processing unit 12 performs various processing using the projection data. Unit 12 performs interpolation, backprojection and reconstruction. Unit 12 determines backprojection data reflecting the x-ray absorption in each voxel. In the helical scanning system using a cone-beam of x-rays, the imaging region (effective field of view) is of cylindrical shape with radius o) (is there a word missing here?) centered on the axis of evolution. Unit 12 defines a plurality of voxels (three-dimensional pixels) in this imaging region, and finds the backprojection data for each voxel. The three-dimensional image data or tomographic image data compiled by using this backprojection data is sent to display device 14, where it is displayed visually as a three-dimensional image or tomographic image.
In typical CT operation, projection data is collected over one rotation of the x-ray source (full scan). The number of views collected per rotation in time (TRot) is NVPR, and during each view, data is collected from a set of detectors Nd. There may be one or more rows of detectors. For ease of explanation, a detector with one row is considered. The views collected over one rotation can be represented as a matrix shown in
A more detailed view of collection unit 11 and processing unit 12 is shown in
For CT fluoroscopy, the same slice position is scanned repeatedly for more than one rotation (NRot). The total number of views collected is given by NRot|NVPR compared with just NVPR in the case of typical CT operation. Since there is a continuous feed of the views, it is not necessary to wait until the end of an integral number of TRot to reconstruct an image. A real time image may be reconstructed using data views equal to NVPR at any given time (the views are counted backwards from any point in time). Preferably, real-time images are reconstructed at a desired fraction of the rotation, such as every ¼ or ⅙ rotation.
The upper limit is determined by hardware speed and memory needed to reconstruct images. For example, having four partial images per second implies four displayed images per second. In an extreme limit, in a mathematical sense, a partial image after every view may be created, that is 900 partial images per second or 900 displayed images per second (in this example). However, for the human eye, anything beyond 25-30 images per second is not significant. Hence, in practice no more than about 20 or 25 partial images per second (900 views) may be computed to provide good quality partial images. Note that for example purposes, 900 views per second are used but this number can take on other values as needed.
As an example, assume that a total of 1800 views are collected and 900 views are required to reconstruct 1 image (Full-Scan). Then, in theory an image can be reconstructed using the view ranges (1 . . . 900), (2 . . . 901), (3 . . . 902) & so on. However, in practice, the ability of the hardware to keep up with the pace of reconstruction may be limited.
In another example, if NVPR=900, each view block contains 225 (900/4) views. There will be a significant overlap in terms of views when reconstructing consecutive images. It is therefore not necessary to backproject NVPR views to reconstruction every single image. Partial images, shown in
In the example of
First image=PI(0)+PI(1)+PI(2)+PI(3)=FS(3)
Second image=first image−PI(0)+PI(4)=FS(4)
Third image=second image−PI(1)+PI(5)=FS(5)
Using one adding and one subtracting operation to create the images reduces the number of operations as opposed to three additions. Here, a partial image (PI) can be computed from as small as one view. In the example, 900 consecutive (in time) partial images may be added added to give one full scan image. Computationally, using larger number of views (such as 225 in the example) to create partial images is more practical. Further, partial images may be computed using a partial scan, such as a half-scan image.
According to the invention, the images may be averaged before being displayed. This is illustrated in
In this example:
First Display Image OTA(5)=average (FS(3)+FS(4)+FS(5))
Second Display Image OTA(6)=average (FS(4)+FS(5)+FS(6))
The above OTA approach works ideally when the object being scanned is stationary. However, when there is voluntary or involuntary motion, edges in the displayed image may be blurred. In a second approach to noise reduction, non-overlapping time images (NTA) are averaged. These images are smooth (less noise). This is illustrated in
NNTA is defined as the number of non-overlapping time average images. For example, NTA(11)=FS(3)+FS(7)+FS(11).
A gradient image, described in more detail below, is used to determine the contribution to each pixel in the display image from the NTA image and from the FS image. For pixels in the gradient image that have a high value (indicating an edge), the pixels in the display image will have a significantly larger contribution from the FS image (sharp image) and pixels in the display image that have a low value (indicating smooth regions) will have a larger contribution from the NTA image (smooth image).
The gradient image may be obtained as follows:
In a first approach, a difference of consecutive FS images is found, and there is (NSPR−1)/NSPR rotation overlap.
Grad1k=abs(FS(k)−FS(k−1)), where k≧NSPR.
In a second approach, a difference of FS images is found, with no overlap between the images.
Grad2m=abs(FS(m)−FS(m−NSPR)), where m≧2·NSPR−1
In a third approach, a difference between FS and NTA images is found
Grad3p=abs(FS−NTAp), where p≧NNTA·NSPR−1
If there is object motion (as is usually the case), scheme 1 is a better approach than scheme 2.
Once the gradient image is obtained, the gradient, FS and NTA images are blended.
Here, x0 and w are parameters, where x0 represents the “center” of the curve and w controls the “width” of the curve.
The parameters may be chosen by an operator or can be set automatically depending on the scan conditions and the slice position in the object being imaged.
The gradient curve remains fixed for each pixel. In other words, the ‘shape’ of the curve does NOT depend on ‘x’ value, which would the gradient value at any voxel. Therefore, going from one voxel to another is tantamount to moving along the x-axis which would in turn yield a corresponding value (alpha) on the y-axis. However, the value of α for each pixel is different and this value is determined by the value of the gradient at that pixel, and is given by:
For each pixel in the gradient image, a new value of α is determined based on the gradient value.
At any given pixel, if the gradient value is high, a higher value of α is used such that a higher contribution to the displayed image comes from the FS (sharp) image and, on the other hand, if the gradient value is low, this means that the pixel belongs to a low frequency region and a higher contribution to the displayed image comes from the NTA (smooth) image. The following equation describes the blending to obtain the blended image BI.
Blp(n)=(1−a)NTAp(n)+α·FSp(n),
where 0<p<Number of pixels.
The gradient curve may be automatically selected. When the gradient image is computed, the statistics (mean, median and standard deviation) of the noise values in a soft-tissue region may be computed. On the x-axis, which represents the gradient value, the soft-tissue region and the high gradient regions will be segregated as shown in
An example of an image obtained according to the invention is illustrated in
A more detailed view of processing unit 12 is shown in
The invention may also be embodied in the form a computer-readable medium containing a stored program to cause a computer to carry out the various operations and functions described above.
Numerous other modifications and variations of the present invention are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the invention may be practiced otherwise than as specifically described herein.
This application is a Continuation of and claims the benefit of priority under 35 U.S.C. §120 from U.S. Ser. No. 12/539,674, filed Aug. 12, 2009, the contents of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5594767 | Hsieh | Jan 1997 | A |
5907593 | Hsieh et al. | May 1999 | A |
6400789 | Dafni | Jun 2002 | B1 |
6970585 | Dafni et al. | Nov 2005 | B1 |
7006592 | Ali et al. | Feb 2006 | B2 |
7164745 | Tsuyuki | Jan 2007 | B2 |
20090052727 | Eusemann et al. | Feb 2009 | A1 |
Number | Date | Country |
---|---|---|
11-009589 | Jan 1999 | JP |
Entry |
---|
Office Action issued Feb. 4, 2014, in Japanese Patent Application 2010-181051 (with English-language translation). |
Number | Date | Country | |
---|---|---|---|
20120294417 A1 | Nov 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12539674 | Aug 2009 | US |
Child | 13557467 | US |