This application claims priority to and the benefit of Korean Patent Application No. 10-2020-0037922 filed in the Korean Intellectual Property Office on Mar. 30, 2020, the entire contents of which are incorporated herein by reference.
The present invention relates to a method and an apparatus for generating a synthetic 2D image. More particularly, the present invention relates to a method and an apparatus for generating a synthetic 2D image, which generate and provide a synthetic 2D image by using a semantic segmentation technique.
Tomosynthesis is a technique for generating a 3D image based on 2D image data captured at a limited angle. The generated 3D image is advantageous in that it is possible to know the structure of a subject that is lost as data is superimposed in the conventional 2D image. In addition, compared to computed tomography (CT), the tomosynthesis has an advantage of taking less time to capture and effectively obtaining an image of a desired cross section with only a relatively small number of images.
However, there is an inconvenience that in order to perform diagnosis by extracting an image of a desired cross section from the generated 3D image, a user should perform the diagnosis while directly adjusting a depth. For this, a synthetic 2D technology that generates and provides a new 2D image to the user by using only valid information in 3D data or technologies that display a valid slice among 3D image data are developed. However, the existing method is difficult to be flexibly applied to other application technologies, and has limitations, such as being limited to the purpose of simply displaying an image.
The present invention has been made in an effort to provide a method and an apparatus for generating a synthetic 2D image, which generate a synthetic 2D image by using a semantic segmentation technique in order to facilitate diagnosis and analysis.
An exemplary embodiment of the present invention provides an apparatus for generating a synthetic 2D image, which includes: an image input unit receiving a 2D image at a plurality of angles or locations for a target object from a detector of a radiographic image acquisition apparatus; a tomosynthesis unit generating a reconstructed 3D image by using the 2D image input into the image input unit; a segmentation map generation unit generating a 3D segmentation map including segmentation data indicating characteristics or types of voxels constituting the 3D image; and a synthetic 2D image synthesis unit generating a synthetic 2D image by using the reconstructed 3D image and the 3D segmentation map.
In an exemplary embodiment, the synthetic 2D image synthesis unit may generate a synthetic 2D image in which the intensity of at least one material of the target object is adjusted by using the segmentation data.
In an exemplary embodiment, the segmentation map generation unit may assign a class label as the segmentation data to the voxel.
The segmentation map generation unit may receive the 2D image from the image input unit and acquire a plurality of 2D segmentation data by performing semantic segmentation in the 2D image, and generate the 3D segmentation map by using the plurality of 2D segmentation data.
The segmentation map generation unit may generate the 3D segmentation map by back-projecting the plurality of 2D segmentation data.
The segmentation map generation unit may perform the semantic segmentation for each class label for the 2D image to acquire 2D segmentation data for each class label.
In an exemplary embodiment, the segmentation map generation unit may generate the 3D segmentation map by performing the semantic segmentation for each voxel of the reconstructed 3D image.
In an exemplary embodiment, the apparatus may further include a semantic filter unit generating a semantic filter as a weight to be applied to the segmentation data of the voxel included in the 3D segmentation map.
The weight may be set separately for each segmentation data or for each voxel.
The weight may be generated through a normalization process.
The synthetic 2D image synthesis unit may generate the synthetic 2D image by multiplying the intensity of the voxel by the semantic filter.
Another exemplary embodiment of the present invention provides a method for generating a synthetic 2D image, which includes: (a) receiving a 2D image at a plurality of angles or locations for a target object from a detector of a radiographic image acquisition apparatus; (b) generating, by a tomosynthesis unit, a reconstructed 3D image by using the 2D images; (c) generating, by a segmentation map generation unit, a 3D segmentation map including segmentation data indicating characteristics or types of voxels of the reconstructed 3D image; and (d) generating, by a synthetic 2D image synthesis unit, a synthetic 2D image by using the reconstructed 3D image and the 3D segmentation map.
In an exemplary embodiment, in step (d), a synthetic 2D image may be generated in which the intensity of at least one material of the target object is adjusted by using the segmentation data.
In step (c), the segmentation map generation unit may receive the 2D image from the image input unit and acquire a plurality of 2D segmentation data by performing semantic segmentation in the 2D image, and generate the 3D segmentation map by using the plurality of 2D segmentation data.
In step (c), the 3D segmentation map may be generated by performing the semantic segmentation for each voxel of the reconstructed 3D image. Further, in step (d), a semantic filter as a weight to be applied to segmentation data of each voxel of the 3D segmentation map may be generated and the synthetic 2D image may be generated by using the semantic filter.
The weight may be set separately for each segmentation data or for each voxel.
The weight may be generated through a normalization process.
In an exemplary embodiment, in step (d), the synthetic 2D image may be generated by multiplying the intensity of the voxel by the semantic filter.
According to an exemplary embodiment of the present invention, when a synthetic 2D image is generated from a 3D image generated by tomosynthesis, intensity of a specific material or a part having a characteristic in an image is adjusted according to a request or setting of a user to provide an adaptive synthetic 2D image.
The foregoing summary is illustrative only and is not intended to be in any way limiting. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features will become apparent by reference to the drawings and the following detailed description.
It should be understood that the appended drawings are not necessarily to scale, presenting a somewhat simplified representation of various features illustrative of the basic principles of the invention. The specific design features of the present invention as disclosed herein, including, for example, specific dimensions, orientations, locations, and shapes will be determined in part by the particular intended application and use environment.
In the figures, reference numbers refer to the same or equivalent parts of the present invention throughout the several figures of the drawing.
Hereinafter, a preferred embodiment of the present invention will be described in detail with reference to the accompanying drawings. First, when reference numerals refer to components of each drawing, it is to be noted that although the same components are illustrated in different drawings, the same components are denoted by the same reference numerals as possible. Further, in describing the present invention, a detailed description of known related configurations and functions may be omitted to avoid unnecessarily obscuring the subject matter of the present invention. Further, hereinafter, the preferred embodiment of the present invention will be described, but the technical spirit of the present invention is not limited thereto or restricted thereby and the embodiments can be modified and variously executed by those skilled in the art.
The radiographic image acquisition apparatus 10 includes a radiation source 12 irradiating radiation R to a target object 18, and a detector 20 detecting the radiation R that is emitted from the radiation source 12 and transmitted through the target object 18. In an exemplary embodiment, the radiation R may be an X-ray, but the radiation R is not limited to the X-ray and may include other radiation such as a gamma ray that may be applied to a medical image. Further, in an exemplary embodiment, the target object 18 may be a breast, and the radiographic image acquisition apparatus 10 may further include a support plate 16 supporting the breast and a compression paddle 14 compressing the breast from the top (for reference, in
At least one of the radiation source 12 or the detector 20 of the radiographic image acquisition apparatus 10 may be rotated or repositioned at a predetermined angle with respect to the target object 18 and the detector 20 may acquire a 2D image for the target object 18 at a plurality of angles or locations. In an exemplary embodiment, the detector 20 is implemented in a digital scheme to acquire 2D projection data for the target object 18. In an exemplary embodiment, image data acquired by the radiographic image acquisition apparatus 10 may be used for Digital Breast Tomosynthesis (DBT).
The apparatus 100 for generating a synthetic 2D image according to an exemplary embodiment of the present invention may include an image input unit 110 receiving a plurality of 2D images, which are 2D projection data, from the detector 20 of the radiographic image acquisition apparatus 10, a tomosynthesis unit 120 reconstructing a 3D image using the image input through the image input unit 110, a segmentation map generation unit 130 generating a 3D segmentation map by using the image input into the image input unit 110, and a synthetic 2D image synthesis unit 150 generating a synthetic 2D image by using the 3D image reconstructed by the tomosynthesis unit 120 and the 3D segmentation map generated by the segmentation map generation unit 130. Further, the apparatus 100 for generating a synthetic 2D image may further include a semantic filter unit 140 generating a semantic filter for generating the synthetic 2D image according to the setting or request of the user. The semantic filter may be used for adaptively applying the 3D segmentation map generated by the segmentation map generation unit 130.
Meanwhile, the synthetic 2D image generated by the synthetic 2D image synthesis unit 150 may be presented to a user through a display 160. Further, the user may input setting or request for generating the synthetic 2D image through a user interface (not illustrated) such as a mouse, a keyboard, or a touch pad.
The image input unit 110 receives the plurality of 2D images which are the 2D projection data from the detector 20 of the radiographic image acquisition apparatus 10. The image input unit 110 may be configured to perform preprocessing such as noise removal or signal amplification for the 2D image transferred from the detector 20.
The tomosynthesis unit 120 applies a tomosynthesis technique to the plurality of 2D images transferred from the image input unit 110 and reconstructs the 3D image including the target object 18.
The segmentation map generation unit 130 generates the 3D segmentation map for a region including the target object 18. The 3D segmentation map may be appreciated as a set of voxel information on 3D classified according to characteristics or physical properties of the target object 18. Specifically, a class label for classifying the characteristic or type of each voxel of the reconstructed 3D image may be granted for each voxel and a set of class labels of the respective voxels may be the 3D segmentation map.
In an exemplary embodiment, the segmentation map generation unit 130 performs semantic segmentation for each of the plurality of 2D images transferred from the image input unit 110 to acquire 2D segmentation data and generates the 3D segmentation map from the 2D segmentation data.
Of course, it may be possible for the segmentation map generation unit 130 to directly generate the 3D segmentation data by using 3D segmentation, but it may be difficult to implementation the synthesis due to various problems. In the 3D data reconstructed by using the tomosynthesis, there is a ghost image phenomenon that a voxel in which data is deficient is influenced by surrounding voxels as a 2D image captured at a limited angle is used. A ghost image refers to a phenomenon that the corresponding voxel has incorrect information due to insufficient data for reconstruction and due to effect of the surrounding voxels. Since the 3D data shows a different pattern from general 3D data due to the ghost image, a segmentation result may not be good. Further, since the size of the 3D data may be several tens of times or hundreds of times larger than that of the 2D data, there may be a physical limitation in performing the semantic segmentation requiring a large amount of computation.
In an exemplary embodiment of the present invention, there is no limitation on the method for generating the 3D segmentation map by the segmentation map generation unit 130. However, when the limitation due to such a reason is expected, as described above, the segmentation map generation unit 130 performs the semantic segmentation for the plurality of 2D images to acquire 2D segmentation data and reconstructs the 2D segmentation data in 3D to generate the 3D segmentation map.
In order to facilitate the description of the present invention, the symbols of data or images in each step are defined as shown in Table 1.
The segmentation map generation unit 130 performs semantic segmentation for each of the plurality of 2D images Tp transferred from the image input unit 110 to acquire 2D segmentation data 1 to n (Sp1, Sp2, . . . , Spn). The semantic segmentation means that the class label is granted to each pixel in the 2D image Tp. The class label in the 2D image Tp including the target object 18 may be a bone, tissue, and air. Further, when the target object 18 is the breast, the class label may include a mammary gland. In an exemplary embodiment, the semantic segmentation for the 2D image Tp may be performed by using an artificial neural network such as a deep learning based algorithm. Further, it may also be possible to perform the semantic segmentation by using a technique such as K-means clustering, otsu thresholding, or fuzzy clustering.
Meanwhile, in the case of the 2D image acquired by the detector 20 of the radiographic image acquisition apparatus 10, various class labels may be granted to one pixel unlike a general image acquired by a camera, etc., due to characteristics of the radiation that passes through an object. That is, at least two class labels among class labels such as the bone, the tissue, the air, and the mammary gland may be granted to one pixel. As a result, multi-label pixelwise classification may be performed for each of the plurality of 2D images and the plurality of 2D segmentation data Sp1, Sp2, . . . , Spn may be acquired.
The segmentation map generation unit 130 generates the 3D segmentation map Srep by using the plurality of 2D segmentation data Sp1, Sp2, . . . , Spn. In an exemplary embodiment, the segmentation map generation unit 130 generates the 3D segmentation map Srep by using a back-projection algorithm for the plurality of 2D segmentation data Sp1, Sp2, . . . , Spn.
The semantic segmentation is performed for the plurality of 2D images Tp transferred from the image input unit 110 and the 2D segmentation data Sp is acquired. In the case of the 2D image Tp, a plurality of class labels may be granted to one pixel due to characteristics of a radiographic image and the multi-label pixelwise classification is performed for the plurality of 2D images Tp. In the example of
In other words, the segmentation map generation unit 130 may obtain 2D segmentation data Sp for each class label for each 2D image Tp obtained at various angles, and then generate the 3D segmentation data Sr for each class label by using the plurality of 2D segmentation data Sp. In this case, each voxel of the 3D segmentation data Sr for each class label may include a value (for example, a probability of a specific label) with respect to the degree to which the class label may be a specific class label. In an exemplary embodiment, in generating the 3D segmentation data Sr, a back-projection technique may be used.
Individual pixels constituting the 2D image Tp may include several class labels due to superimposition, but individual voxels constituting the reconstructed 3D image Tr may have one class label. For this reason, in the process of generating the 3D segmentation map Srep using the 3D segmentation data Sr for each class label, class labels for individual voxels are specified. In an exemplary embodiment, when a plurality of class labels exist in individual voxels of the 3D segmentation map Srep, the class label of the corresponding voxel may be determined as the class label with the highest probability. For example, when there are M 3D segmentation data (Srk, k=1 . . . M), the class label (Srep(x, y, z)) of a specific voxel of the 3D segmentation map (Srep) may be obtained by Equation 1.
Referring back to
The semantic filter unit 140 may generate the semantic filter from the 3D segmentation map Srep according to the setting or request of the user. The semantic filter may be described as a weight to be described below, and a specific material may be emphasized or suppressed by assigning the weight for each material in the synthetic 2D image synthesis process. That is, the intensity of an image pixel by a specific material in the generated synthetic 2D image may be adjusted.
In
By combining the weights for each label, a weight ωk which may be applied as the semantic filter may be acquired.
Numerical values illustrated in
As such, a semantic filter is generated in which the weight for the specific class label is increased or decreased, and applied to the synthetic 2D image synthesis to acquire an image in which a body tissue desired by the user is emphasized or weakened.
In
In
F
syn(r)=∫−∞∞∫−∞∞Tr(x,y)δ(x cos φ+y sin φ−r)dxdy [Equation 2]
A semantic filter for giving the weight for a specific material, i.e., a weight function ωn is applied to Equation 2, which is shown in Equation 3.
F
syn(r)=∫−∞∞∫−∞∞Tr(x,y)ωn(x,y)δ(x cos φ+y sin φ−r)dxdy [Equation 3]
In an exemplary embodiment, when a weight set by the user for each voxel in the 3D segmentation map Srep is ωk, ωn may be acquired by normalizing ωk according to a voxel group in a ray formed by the pixel of the image to be generated and the radiation source. Normalization may be performed by dividing a voxel to be normalized by the sum of weights of voxels in the ray formed by the radiation source.
When the weight of a specific voxel in the segmentation map Srep is represented by ωk(x, y), the sum of ωk(x, y) of voxels included in a specific ray ωk′(r) may be expressed as in Equation 4 and the relationship between ωn, ωk, and ωk′ may be expressed as in Equation 5, and accordingly, Equation 3 may be expressed as Equation 6 again.
Referring to
The image input unit 110 receives the plurality of 2D images which are the 2D projection data from the detector 20 of the radiographic image acquisition apparatus 10 (S100).
The tomosynthesis unit 120 applies a tomosynthesis technique to the plurality of 2D images transferred from the image input unit 110 and reconstructs the 2D images in 3D to generate the reconstructed 3D image (S110).
The segmentation map generation unit 130 generates the 3D segmentation map for a region including the target object 18. Specifically, the segmentation map generation unit 130 performs semantic segmentation for each of the plurality of 2D images transferred from the image input unit 110 to acquire 2D segmentation data (S120).
Next, the segmentation map generation unit 130 generates a 3D segmentation map by reconstructing the 2D segmentation data in 3D (S130).
The semantic filter unit 140 generates a semantic filter as a weight to be applied to segmentation data (e.g., class label information) of each voxel of the 3D segmentation map according to the user's request or setting (S140).
The synthetic 2D image synthesis unit 150 generates a synthetic 2D image by using the reconstructed 3D image and the semantic filter (S150).
Through the above process, a final synthetic 2D image may be acquired (S160) and the final synthetic 2D image may be presented to a user through a display 160.
The image input unit 110 receives the plurality of 2D images which are the 2D projection data from the detector 20 of the radiographic image acquisition apparatus 10 (S200).
The tomosynthesis unit 120 applies a tomosynthesis technique to the plurality of 2D images transferred from the image input unit 110 and reconstructs the 2D images in 3D to generate the reconstructed 3D image (S210).
The segmentation map generation unit 130 generates the 3D segmentation map by using the reconstructed 3D image (S220). In step S220, the 3D segmentation map may be generated by directly performing the semantic segmentation for each voxel of the reconstructed 3D image. As a result of performing step S220, the class label for each voxel may be assigned.
The semantic filter unit 140 generates a semantic filter as a weight to be applied to segmentation data of each voxel of the 3D segmentation map according to the user's request or setting (S230).
The synthetic 2D image synthesis unit 150 generates a synthetic 2D image by using the reconstructed 3D image and the semantic filter (S240).
Through the above process, a final synthetic 2D image may be acquired (S250) and the final synthetic 2D image may be presented to a user through a display 160.
A plurality of 2D images Tp which are 2D projection data are input from the detector 20 of the radiographic image acquisition apparatus 10 and the reconstructed 3D image Tr is generated (S300).
The segmentation map generation unit 130 generates the 3D segmentation map Srep (S310).
The synthetic 2D image synthesis unit 150 generates the synthetic 2D image Fsyn by using the reconstructed 3D image Tr and the 3D segmentation map Srep.
According to the present invention, a filter for emphasizing information requested by the user may be dynamically generated by using semantic segmentation data. In addition, by using semantic segmentation data, even voxels having similar pixel values (pixel intensity) are classified and different weights are used for each material or characteristic are used to obtain an adaptive synthetic 2D image.
As described above, the exemplary embodiments have been described and illustrated in the drawings and the specification. The exemplary embodiments were chosen and described in order to explain certain principles of the invention and their practical application, to thereby enable others skilled in the art to make and utilize various exemplary embodiments of the present invention, as well as various alternatives and modifications thereof. As is evident from the foregoing description, certain aspects of the present invention are not limited by the particular details of the examples illustrated herein, and it is therefore contemplated that other modifications and applications, or equivalents thereof, will occur to those skilled in the art. Many changes, modifications, variations and other uses and applications of the present construction will, however, become apparent to those skilled in the art after considering the specification and the accompanying drawings. All such changes, modifications, variations and other uses and applications which do not depart from the spirit and scope of the invention are deemed to be covered by the invention which is limited only by the claims which follow.
Number | Date | Country | Kind |
---|---|---|---|
10-2020-0037922 | Mar 2020 | KR | national |