This application claims the banefit of Japanese Application No. 2002 -198645 Filed Jul. 8, 2002.
The present invention relates to an image processing method and apparatus, and more particularly, to a method and apparatus for extracting an image, which meets a predetermined condition, from among a plurality of images.
Medical images are helpful in diagnosing a patient condition. Medical images provide unique information depending on what modality is used to produce the medical images. Medical images produced by an appropriate modality are employed according to a purpose of diagnosis.
In order to given more appropriate diagnosis on the basis of diverse information, a plurality of kinds of medical images produced by a plurality of modalities, such as, an X-ray computed tomography (CT) apparatus, a magnetic resonance imaging (MRI) apparatus, and a positron emission tomography (PET) apparatus or a single photon emission computed tomography (SPECT) apparatus are employed.
For example, the X-ray CT apparatus and PET apparatus are used to image the same region of the same patient. Images produced by the X-ray CT apparatus and showing the structure of the cerebral parenchyma, and images produced by the PET apparatus and showing the active state of the brain are used to diagnose a lesion in terms of the structure of the lesion and the function thereof. At this time, a synthetic image produced by superposing two different kinds of images in order to facilitate simultaneous grasping of the structure and function.
In order to produce such a synthetic image, two images are aligned with each other. The alignment may be referred to as registration. The registration is performed by a computer.
For the registration, a technique described in, for example, “Multi-modal Volume Registration by Maximization of Mutual Information” written by W. M. Wells III et al. is adopted.
Generally speaking, the registration is achieved through transformation through which the amount of mutual information and a candidate image v(x) shares with a reference image u(x) is maximized. Namely,
where T denotes the transformation from the apparatus of coordinates for the candidate image to the apparatus of coordinates for the reference image.
I(u(x), v(T(x))) (2)
The formula 2 denotes mutual information.
I(u(x), v(T(x)))≡h(u(x))+h(v(T(x)))−h(u(x), v(T(x))) (3)
where h(•) denotes an entropy.
h(x)≡−∫p(x)In p(x)dx (4)
h(x,y)≡−∫p(x,y)In p(x,y)dxdy (5)
The above formulae provide an entropy.
The foregoing registration proves effective when it is previously apparent that two images result from visualization of the same region. Otherwise, even if transformation provides an image that shares the largest amount of mutual information with a reference image, the transformation does not always lead to significant registration.
Therefore, an object of the present invention is to provide a method and apparatus for extracting an image closest to a reference image from among a plurality of candidate images.
(1) According to one aspect of the present invention attempting to solve the aforesaid problem, there is provided an image processing method for: extracting a plurality of candidate images similar to a reference image from among a plurality of images by utilizing granulometry; transforming the plurality of candidate images on the basis of the reference image; calculating mutual information shared by each of the transformed candidate images and the reference image; and selecting a candidate image, which shares the largest amount of mutual information with the reference image, from among the plurality of candidate images.
(2) According to another aspect of the present invention attempting to solve the aforesaid problem, there is provided an image processing apparatus comprising: a extracting means for extracting a plurality of candidate images similar to a reference image from among a plurality of images by utilizing granulometry; a transforming means for transforming the plurality of candidate images on the basis of the reference image; a calculating means for calculating mutual information shared by each of the transformed candidate images and the reference image; and a selecting means for selecting a candidate image, which shares the largest amount of mutual information with the reference image, from among the plurality of candidate images.
According to the aspects (1) and (2) of the present invention, a plurality of candidate images similar to a reference image are extracted from among a plurality of images by utilizing granulometry. The candidate images are transformed based on the reference image. Mutual information shared by each of the transformed candidate images and the reference image is calculated. A candidate image that shares the largest amount of mutual information with the reference image is selected from among the plurality of candidate images. Consequently, an image closest to the reference image can be extracted from among the plurality of candidate images.
Preferably, the transformation includes matching of magnifications so that the transformation can be achieved effectively.
Preferably, the transformation includes alignment of barycenters so that the transformation can be achieved effectively.
Preferably, both the reference image and candidate images are medical images so that diagnosis can be achieved based on a synthetic image.
Therefore, the present invention provides a method and a apparatus capable of extracting an image closest to a reference image from among a plurality of candidate images.
Further objects and advantages of the present invention will be apparent from the following description of the preferred embodiments of the invention as illustrated in the accompanying drawings.
Referring to drawings, an embodiment of the present invention will be described below.
As shown in
The computer 100 also includes a display unit 104 and an operating unit 106. An image and other information provided by the computer 100 are displayed on the display unit 104. A user handles the operating unit 106, and various instructions and information are transferred from the operating unit 106 to the computer 100. The user uses the display unit 104 and operating unit 106 to operate the image processing apparatus interactively.
Actions to be performed by the present apparatus will be described below.
As illustrated, at step 202, a reference image is read. Consequently, for example, an image frame containing an image A shown in
At step 204, received images are read. Consequently, for example, an image frame containing an image B shown in
A plurality of images preserved in an external image file or the like are read. The plural images are different from one another in terms of, for example, a position of slicing.
At step 206, candidate images similar to the reference image are extracted from among the received images. For the extraction of the candidate images, granulometry that is a kind of morphological operation is utilized.
The granulometry is known as described in, for example, “Granulometry; Mathematical Morphology in Statistical Sense,” Information statistics II, Lecture 5 (Nov. 10, 2000).
The granulometry is adapted to the reference image and the plurality of received images. As a result of the granulometry, a pattern spectrum is provided for each of the reference images and plurality of received images. The similarity between the pattern spectrum of each received image and the pattern spectrum of the reference image is checked. Received images whose pattern spectra are similar to the pattern spectrum of the reference image are extracted as candidate images. Thus, only the received images whose pattern spectra are similar to the pattern spectrum of the reference image are extracted in advance, whereby succeeding image processing is performed efficiently.
At step 208, the candidate images are transformed. The transformation is geometric transformation from the apparatus of coordinates for the candidate images to the apparatus of coordinates for the reference image. The transformation includes matching of magnifications and alignment of barycenters.
The matching of magnifications is the processing of matching the size of a candidate image with the size of the reference image as described below. First, the reference image and candidate image are binary-coded. The binary coding is such that as for the reference image that is the image A, among all pixels that constitute data of the image frame containing the image A, pixels whose values are equal to or larger than a predetermined threshold are set to 1s and pixels whose values fall below the threshold are set to 0s. As for the candidate image, among all pixels that constitute data of the image frame containing the image B, pixels whose values are equal to or larger than a predetermined threshold are set to 1s, and pixels whose values fall below the threshold are set to 0s.
This results in, for example, image frames shown in
In general, the moment of an image is expressed as follows:
A magnification required to match the size of the image b with the size of the image a is given as follows:
where M(0,0) denotes the 0-order moment. The magnification k is calculated for each of the plurality of candidate images. The magnifications k are adapted to the candidate images, whereby the candidate images have the same size as the reference image.
In order to align barycenters with each other, the barycenters of the images a and b in the respective image frames are calculated. The barycentric coordinates (m,n) are calculated as follows:
m=M(1,0)/M(0,0) (8)
n=M(0,1)/M(0,0) (9)
where M(1,0) and M(0,1) denote first-order moments.
The above formulae are adapted to the image frames shown in
Assuming that the coordinates representing the barycenters a0 and b0 of the images A and B are, for example, (i0,j0) and (i0′,j0′) as shown in
Δi=i0−i0′ (10)
Δj=j0−j0′ (11)
where Δi denotes the magnitude of correction for the position of a barycenter in the direction of an I axis, and Δj denotes the magnitude of correction therefor in the direction of a J axis. The magnitudes of correction for the position of a barycenter are calculated for each of the plurality of candidate images.
The foregoing magnification k and the magnitudes of correction for the position of a barycenter (Δi,Δj) which are calculated for each candidate image are adapted to each candidate image. Consequently, the plurality of candidate images is transformed into images defined in the same apparatus of coordinates as the reference image. Thus, the transformation is achieved.
At step 210, mutual information shared by each candidate image and the reference image is calculated. Each candidate image to be subjected to this calculation is a transformed image.
In order to calculate mutual information, the entropy h(a) of the reference image is calculated. The entropy h(b) of each candidate image is calculated. A joint entropy h(a,b) that is a combination of the entropy of each candidate image and the entropy of the reference image is determined. Using the entropies, mutual information is calculated as follows:
MI(a, b)=h(a)+h(b)−h(a,b) (12)
The mutual information is determined for each of the plurality of candidate images. Since the plurality of candidate images is transformed images, mutual information shared by each candidate image and the reference image can be determined under the same geometric conditions.
Whether the amount of mutual information is large or small indicates a degree of approximation to which each candidate image approximates the reference image. Since the plurality of candidate images meets the same geometric conditions as the reference image, whether the amount of mutual information is large or small indicates a degree of approximation to which each candidate image approximates the reference image. Consequently, a candidate image sharing the largest amount of mutual information with the reference image is an image closest to the reference image.
At step 212, a candidate image sharing the largest amount of mutual information is selected. Consequently, an image closest to the reference image is selected from among the candidate images. The image has the highest possibility that it may result from visualization of the same section of a patient body as the reference image does. Since the selected candidate image has already been transformed, registration has also been completed.
Generally, the received images include an image that results from visualization of the same section of a patient body as the reference image does. Therefore, the desired image is automatically extracted through the aforesaid processing, and registration is automatically completed.
The candidate extracting part 302 extracts a candidate image similar to a reference image from among a plurality of received images. This functional block performs the processing equivalent to step 206. The candidate extracting part 302 is an example of an extracting means included in the present invention.
The transforming part 304 transforms each candidate image. This functional block performs the processing equivalent to step 208. The transforming part 304 is an example of a transforming means included in the present invention.
The mutual information calculating part 306 calculates mutual information shared by each candidate image and the reference image. This functional block performs the processing equivalent to step 210. The mutual information calculating part 306 is an example of a calculating means included in the present invention.
The image selecting part 308 selects a candidate image that shares the largest amount of mutual information with the reference image. This functional block performs the processing equivalent to step 212. The image selecting part 308 is an example of an image selecting means included in the present invention.
Described above is a case where an image closest to an image produced by a PET apparatus is extracted from among a plurality of images produced by an X-ray CT apparatus. The present invention is not limited to the images produced by the X-ray CT apparatus and PET apparatus but also applies to medical images produced by other modalities. Even in the case where the medical images produced by the other modalities are employed, an image produced by one modality and closest to an image produced by the other modality can be extracted in the same manner as mentioned above. Moreover, the present invention can apply not only to two-dimensional images but also to three-dimensional images. Moreover, needless to say, the images are not limited to the medical images.
Many widely different embodiments of the invention may be configured without departing from the spirit and the scope of the present invention. It should be understood that the present invention is not limited to the specific embodiments described in the specification, except as defined in the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
2002-198645 | Jul 2002 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
5623560 | Nakajima et al. | Apr 1997 | A |
6026142 | Gueziec et al. | Feb 2000 | A |
6201888 | Kalvin | Mar 2001 | B1 |
6462342 | Stearns | Oct 2002 | B1 |
6560354 | Maurer et al. | May 2003 | B1 |
6574356 | Lee et al. | Jun 2003 | B1 |
7106891 | Wyman et al. | Sep 2006 | B2 |
7142700 | Van Vaals | Nov 2006 | B2 |
20010031082 | Numaoka | Oct 2001 | A1 |
20010038386 | Numaoka | Nov 2001 | A1 |
20020076106 | Kondo et al. | Jun 2002 | A1 |
20020081008 | Wollenweber | Jun 2002 | A1 |
20020081009 | Licato et al. | Jun 2002 | A1 |
20020122575 | Vaisburd et al. | Sep 2002 | A1 |
20030004405 | Townsend et al. | Jan 2003 | A1 |
Number | Date | Country |
---|---|---|
06-165036 | Jun 1994 | JP |
Number | Date | Country | |
---|---|---|---|
20040028275 A1 | Feb 2004 | US |