The present application claims priority from Japanese Patent Application No. 2022-050635, filed on Mar. 25, 2022 and Japanese Patent Application No. 2022-150250, filed on Sep. 21, 2022, the entire disclosures of which are incorporated herein by reference.
The present disclosure relates to an image generation apparatus, a method, and a program, a learning apparatus, a method, and a program, a segmentation model, and an image processing apparatus, a method, and a program.
As a machine learning model that handles images, a convolutional neural network (hereinafter, abbreviated as CNN) that performs semantic segmentation for identifying a target object included in an image in units of pixels is known. For example, U-Net: Convolutional Networks for Biomedical Image Segmentation, Olaf Ronneberger, et al., 2015 has suggested segmentation using a U-shaped convolutional neural network (U-Net; U-Shaped Neural Network).
In the medical field, a medical image is segmented using a machine learning model, and determination of a progress of an illness on a segmented region is performed.
On the other hand, a lot of training data is required for learning of the machine learning model. Note that, in the medical field, since collection of data of a scarce disease is difficult, it is difficult to provide a machine learning model capable of accurately performing segmentation.
For this reason, for learning of a machine learning model for detecting a skin cancer, a technique that generates a pseudo image including skin cancers of various sizes using an existing medical image and correct answer data in which a region of a skin cancer in the medical image is masked has been suggested (see Mask2Lesion: Mask-Constrained Adversarial Skin Lesion Image Synthesis, Kumar Abhishek, et al., 2019). For example, a technique that learns a spatial distribution of a three-dimensional object having an existing shape, such as a chair, and generates an image of a chair having an unknown shape has also been suggested (see The shape variational autoencoder: A deep generative model of part-segmented 3D objects, C. Nash, et al., 2017).
Note that, like the technique described in Mask2Lesion: Mask-Constrained Adversarial Skin Lesion Image Synthesis, Kumar Abhishek, et al., 2019, it is not possible to generate data of features that are rarely included in existing learning data, only using correct answer data on an existing image. For this reason, it is difficult to construct a machine learning model capable of accurately segmenting a target object that is rarely included in an existing image, even though a generated image is added to learning data. Like the technique described in The shape variational autoencoder: A deep generative model of part-segmented 3D objects, C. Nash, et al., 2017, it is difficult to construct a machine learning model capable of accurately segmenting a target object different from an existing target object only by changing a shape of the existing target object. In particular, on a scarce disease, such as a progressive cancer, since a cancer tissue is often infiltrated into a part in the vicinity thereof, there is a demand for segmenting such a progressive cancer with excellent accuracy.
The present disclosure has been accomplished in view of the above-described situation, and an object of the present disclosure is to provide a machine learning model capable of accurately performing segmentation.
An image generation apparatus according to a first aspect of the present disclosure comprises at least one processor, in which the processor is configured to acquire an original image and a mask image in which masks are applied to one or more regions respectively representing one or more objects including a target object in the original image, derive a pseudo mask image by processing the mask in the mask image, and derive a pseudo image that has a region based on a mask included in the pseudo mask image and has the same representation format as the original image, based on the original image and the pseudo mask image.
An image generation apparatus according to a second aspect of the present disclosure is the image generation apparatus according to the first aspect of the present disclosure, in which the pseudo mask image and the pseudo image may be used as training data for learning a segmentation model that segments the object included in an image.
An image generation apparatus according to a third aspect of the present disclosure is the image generation apparatus according to the second aspect of the present disclosure, in which the processor may be configured to accumulate the pseudo mask image and the pseudo image as training data.
An image generation apparatus according to a fourth aspect of the present disclosure is the image generation apparatus according to any one of the first to third aspects of the present disclosure, in which the processor may be configured to derive the pseudo mask image that is able to generate the pseudo image including a target object of a class different from a class indicated by the target object.
The term “class is different” means that a type of a shape of the target object is different, that, in a case where the target object is a lesion included in a medical image, a progress of the lesion is different. The term “different class” means a class that has a small frequency of appearance or is not at all compared to other classes, in stored training data. For this reason, “the pseudo mask image that is able to generate the pseudo image including the target object of the class different from the class indicated by the target object” is derived, whereby it is possible to prepare training data of a class that is small or is not at all in existing training data. Accordingly, learning of the segmentation model is performed using such training data along with existing training data, whereby it is possible to construct a segmentation model that can segment a target object on an image including a target object having a small frequency of appearance.
An image generation apparatus according to a fifth aspect of the present disclosure is the image generation apparatus according to any one of the first to fourth aspects of the present disclosure, in which the processor may be configured to derive the pseudo mask image by processing the mask such that at least one of a shape or a progress of a lesion is different from that of a lesion included in the original image, based on a lesion shape evaluation index used as an evaluation index in medical practice for a medical image.
An image generation apparatus according to a sixth aspect of the present disclosure is the image generation apparatus according to any one of the first to fifth aspects of the present disclosure, in which the processor may be configured to derive the pseudo mask image by processing the mask until a normal organ has a shape to be evaluated as a lesion based on a measurement index in medical practice for a medical image.
An image generation apparatus according to a seventh aspect of the present disclosure is the image generation apparatus according to any one of the first to sixth aspects of the present disclosure, in which the processor may be configured to refer to at least one style image having predetermined density, color, or texture and generate the pseudo image having density, color, or texture depending on the style image.
The term “style image” is an image that represents an object of the same type as the target object having possible density, color, and texture of the target object.
An image generation apparatus according to an eighth aspect of the present disclosure is the image generation apparatus according to any one of the first to seventh aspects of the present disclosure, in which the processor may be configured to receive an instruction for a degree of processing of the mask and derive the pseudo mask image by processing the mask based on the instruction.
An image generation apparatus according to a ninth aspect of the present disclosure is the image generation apparatus according to the eighth aspect of the present disclosure, in which the processor may be configured to receive designation of a position of an end point of the mask after processing and designation of a processing amount as the instruction for the degree of processing.
An image generation apparatus according to a tenth aspect of the present disclosure is the image generation apparatus according to the eighth or ninth aspect of the present disclosure, in which the processor may be configured to receive the instruction for the degree of processing of the mask under a constraint condition set in advance.
An image generation apparatus according to an eleventh aspect of the present disclosure is the image generation apparatus according to any one of the first to tenth aspects of the present disclosure, in which, in a case where the original image includes a plurality of the objects, and the target object and a partial region of another object other than the target object have an inclusion relation, in the mask image, a region having the inclusion relation may be applied with a mask different from a region having no inclusion relation.
An image generation apparatus according to a twelfth aspect of the present disclosure is the image generation apparatus according to the eleventh aspect of the present disclosure, in which the processor may be configured to, in a case where the other object having the inclusion relation is an object fixed in the original image, derive the pseudo mask image by processing the mask applied to the target object conforming to a shape of a mask applied to the fixed object.
An image generation apparatus according to a thirteenth aspect of the present disclosure is the image generation apparatus according to any one of the first to twelfth aspects of the present disclosure, in which the processor may be configured to, in a case where the original image is a three-dimensional image, derive the pseudo mask image by processing the mask while maintaining three-dimensional continuity of the mask applied to the region of the target object.
An image generation apparatus according to a fourteenth aspect of the present disclosure is the image generation apparatus according to any one of the first to thirteenth aspects of the present disclosure, in which the original image may be a three-dimensional medical image, and the target object may be a lesion included in the medical image.
An image generation apparatus according to a fifteenth aspect of the present disclosure is the image generation apparatus according to the fourteenth aspect of the present disclosure, in which the medical image may include a rectum of a human body, and the target object may be a rectal cancer, and another object other than the target object may be at least one of a mucous membrane layer of the rectum, a submucosal layer of the rectum, a muscularis propria of the rectum, a subserous layer of the rectum, or a background other than the layers.
An image generation apparatus according to a sixteenth aspect of the present disclosure is the image generation apparatus according to the fourteenth aspect of the present disclosure, in which the medical image may include a joint of a human body, and the target object may be a bone composing the joint, and another object other than the target object may be a background other than the bone composing the joint.
A learning apparatus according to a seventeenth aspect of the present disclosure comprises at least one processor, in which the processor is configured to construct a segmentation model that segments a region of one or more objects including a target object included in an input image, by performing machine learning using a plurality of sets of pseudo images and pseudo mask images generated by the image generation apparatus according to any one of the first to sixteenth aspects of the present disclosure as training data.
A learning apparatus according to an eighteenth aspect of the present disclosure is the learning apparatus according to the seventeenth aspect of the present disclosure, in which the processor may be configured to construct the segmentation model by performing machine learning using a plurality of sets of original images and mask images as training data.
A segmentation model according to a nineteenth aspect of the present disclosure is constructed by the learning apparatus according to the seventeenth or eighteenth aspect of the present disclosure.
An image processing apparatus according to a twentieth aspect of the present disclosure comprises at least one processor, in which the processor is configured to derive a mask image in which one or more objects included in a target image to be processed are masked, by segmenting a region of one or more objects including a target object included in the target image using the segmentation model according to the nineteenth aspect of the present disclosure.
An image processing apparatus according to the twenty-first aspect of the present disclosure, in the image processing apparatus according to the twentieth aspect of the present disclosure, in which the processor may be configured to discriminate a class of the target object masked in the mask image using a discrimination model that discriminates a class of a target object included in a mask image.
An image generation method according to a twenty-second aspect of the present disclosure comprises acquiring an original image and a mask image in which masks are applied to one or more regions respectively representing one or more objects including a target object in the original image, deriving a pseudo mask image by processing the mask in the mask image, and deriving a pseudo image that has a region based on a mask included in the pseudo mask image and has the same representation format as the original image, based on the original image and the pseudo mask image.
A learning method according to a twenty-third aspect of the present disclosure constructs a segmentation model that segments a region of one or more objects including a target object included in an input image, by performing machine learning using a plurality of sets of pseudo images and pseudo mask images generated by the image generation method according to the twenty-second aspect of the present disclosure as training data.
An image processing method according to a twenty-fourth aspect of the present disclosure comprises deriving a mask image in which one or more objects included in a target image to be processed are masked, by segmenting a region of one or more objects including a target object included in the target image using the segmentation model according to the nineteenth aspect of the present disclosure.
An image generation program according to a twenty-fifth aspect of the present disclosure causes a computer to execute procedure of acquiring an original image and a mask image in which masks are applied to one or more regions respectively representing one or more objects including a target object in the original image, a procedure of deriving a pseudo mask image by processing the mask in the mask image, and a procedure of deriving a pseudo image that has a region based on a mask included in the pseudo mask image and has the same representation format as the original image, based on the original image and the pseudo mask image.
A learning program according to a twenty-sixth aspect of the present disclosure causes a computer to execute a procedure of constructing a segmentation model that segments a region of one or more objects including a target object included in an input image, by performing machine learning using a plurality of sets of pseudo images and pseudo mask images generated by the image generation method according to the twenty-second aspect of the present disclosure as training data.
An image processing program according to a twenty-aspect of the present disclosure causes a computer to execute a procedure of deriving a mask image in which one or more objects included in a target image to be processed are masked, by segmenting a region of one or more objects including a target object included in the target image using the segmentation model according to the nineteenth aspect of the present disclosure.
According to the present disclosure, it is possible to provide a machine learning model capable of accurately segmentation.
Hereinafter, an embodiment of the present disclosure will be described referring to the drawings. First, the configuration of a medical information system to which an image generation apparatus, a learning apparatus, and an image processing apparatus according to the present embodiment are applied will be described.
The computer 1 includes the image generation apparatus and the learning apparatus according to the present embodiment, and an image generation program and a learning program according to the present embodiment are installed thereon. The computer 1 may be a workstation or a personal computer or may be a server computer connected to the workstation or the personal computer through a network. The image generation program and the learning program are stored in a storage device of the server computer connected to the network or a network storage in a state of being accessible from the outside, and are downloaded to and installed on the computer 1 on demand. The image generation program and the learning program are recorded on a recording medium, such as a digital versatile disc (DVD) or a compact disc read only memory (CD-ROM), are distributed, and are installed on the computer 1 from the recording medium.
The computer 2 includes the image processing apparatus according to the present embodiment, and the image processing program of the present embodiment is installed thereon. The computer 2 may be a workstation or a personal computer or may be a server computer connected to the workstation or the personal computer through a network. The image processing program is stored in a storage device of the server computer connected to the network or a network storage in a state of being accessible from the outside, and is downloaded to and installed on the computer 2 on demand. The image processing program is recorded on a recording medium, such as a DVD or a CD-ROM, is distributed, and is installed on the computer 2 from the recording medium.
The imaging apparatus 3 is an apparatus that images a part to be diagnosed of a subject to generate a three-dimensional image representing the part, and specifically, is a CT apparatus, an MM apparatus, a positron emission tomography (PET) apparatus, or the like. The three-dimensional image generated by the imaging apparatus 3 and composed of a plurality of tomographic images is transmitted to and stored in the image storage server 4. In the present embodiment, the imaging apparatus 3 is an MRI apparatus, and generates an MRI image of a human body as the subject as a three-dimensional image. In the present embodiment, it is assumed that the three-dimensional image is a three-dimensional image including the vicinity of a rectum of the human body. For this reason, in a case where a patient with a rectal cancer is imaged, the rectal cancer is included in the three-dimensional image.
The image storage server 4 is a computer that stores and manages various kinds of data, and comprises a large capacity external storage device and software for database management. The image storage server 4 performs communication with other apparatuses through the network 5 in a wired or wireless manner and transmits and receives image data and the like. Specifically, the image storage server 4 acquires various kinds of data including image data of the three-dimensional image generated by the imaging apparatus 3 by way of the network, and stores and manages the acquired data in a recording medium, such as a large capacity external storage device. In the image storage server 4, training data for constructing a machine learning model for deriving a pseudo image, detecting an abnormal part, or discriminating a class of the abnormal part as described below is also stored. A storage format of image data and communication between the respective apparatuses by way of the network 5 are based on a protocol, such as Digital Imaging and Communication in Medicine (DICOM).
Next, the image generation apparatus and the learning apparatus according to the present embodiment will be described.
The storage 13 is realized by a hard disk drive (HDD), a solid state drive (SSD), a flash memory, or the like. In the storage 13 as a storage medium, an image generation program 12A and a learning program 12B are stored. The CPU 11 reads out the image generation program 12A and the learning program 12B from the storage 13, develops the image generation program 12A and the learning program 12B to the memory 16, and executes the developed image generation program 12A and learning program 12B.
Next, the functional configuration of the image generation apparatus and the learning apparatus according to the present embodiment will be described.
The information acquisition unit 21 acquires an original image G0 that is used to derive a pseudo image described below, from the image storage server 4. The information acquisition unit 21 acquires training data for constructing a trained model described below from the image storage server 4.
Here, in the present embodiment, the original image G0 is stored in conjunction with a mask image M0 in which a mask is applied to a region of an object included in the original image G0. In a case where a rectal cancer is included in the original image G0, information representing a stage of the rectal cancer is applied to the original image G0.
The application of the mask may be performed by a manual operation using the input device 15 to the original image G0 or may be performed by segmenting the original image G0. As the segmentation, semantic segmentation that performs class classification by labeling all pixels of an image in units of pixels is used. The semantic segmentation is performed by a semantic segmentation model that is a machine learning model constructed by performing machine learning to extract a region of an object included in an image. The semantic segmentation model will be described below. The original image G0 and the mask image M0 that are stored in the image storage server 4 may be images derived by the image processing apparatus of the present embodiment described below.
In the present embodiment, a rectal cancer is detected from a target image to be processed in the image processing apparatus described below. For this reason, a mask for identifying each region of the rectal cancer included in the original image G0, a mucous membrane layer of a rectum, a submucosal layer of the rectum, a muscularis propria of the rectum, a subserous layer of the rectum, and a background other than the layers is applied as the mask image M0 to the original image G0. The rectal cancer included in the original image G0 is an example of a target object, and the mucous membrane layer of the rectum, the submucosal layer of the rectum, the muscularis propria of the rectum, the subserous layer of the rectum, and the background other than the layers are an example of objects other than the target object.
Here, while an initial rectal cancer is present only in the mucous membrane layer, in a case where the rectal cancer progresses, the rectal cancer spreads outward from the mucous membrane layer, is infiltrated into the submucosal layer and the muscularis propria, and has an inclusion relation with the submucosal layer and the muscularis propria.
In the present embodiment, a region having an inclusion relation in the original image G0 is applied with a mask different from a region having no inclusion relation. For example, in the mask image M0 shown in
In a case of deriving the mask image M0 by the segmentation model, to segment a region having an inclusion relation by the semantic segmentation model, the segmentation model may be constructed by performing machine learning while preparing training data including correct answer data with a region having an inclusion relation and a region having no inclusion relation applied with different masks.
The pseudo mask derivation unit 22 derives a pseudo mask image by processing the masks included in the mask image M0 of the original image G0. To this end, a display control unit 74 displays a mask processing screen on the display 14.
The mask image M0 and the pseudo mask image Mf0 shown in
The three-dimensional model 41 is a three-dimensional image derived by extracting only the region of the rectal cancer in the original image G0 and performing volume rendering. An operator can change omnidirectional orientations of the three-dimensional model 41 by operating the input device 15.
In the pull-down menu 26, the stage of the rectal cancer can be selected. That is, in the pull-down menu 26, the stages T1, T2, T3ab, T3cd, T3MRF+, T4a, and T4b of the rectal cancer shown in
In the present embodiment, the pseudo mask derivation unit 22 derives a pseudo mask image capable of generating a pseudo image including a target object of a class different from the class of the target object. That is, in deriving the pseudo mask image Mf0, the pseudo mask derivation unit 22 processes a mask while deforming the three-dimensional model 41 such that a pseudo image including a rectal cancer of a stage different from the stage of the rectal cancer included in the original image G0 can be generated. For example, in a case where the stage of the rectal cancer included in the original image G0 is the stage T1 where the rectal cancer is present only in the mucous membrane layer, the three-dimensional model 41 corresponds to, for example, the rectal cancer of the stage T3ab progressed from the stage T1 by deforming the three-dimensional model 41 of the rectal cancer to extend to the muscularis propria. The three-dimensional model 41 corresponds to the rectal cancer of the stage T4a by deforming the three-dimensional model 41 of the rectal cancer to break through the muscularis propria.
For this reason, the operator views the displayed mask image M0 and selects a stage of a rectal cancer to be generated as a pseudo image from the pull-down menu 26.
In deriving the pseudo mask image Mf0, the pseudo mask derivation unit 22 deforms the three-dimensional model 41 such that a shape and/or a progress of a lesion is different from that of a lesion included in the original image G0 based on a lesion shape evaluation index used as an evaluation index in medical practice for a medical image. That is, the shape of the rectal cancer is deformed such that the shape and/or the progress of the rectal cancer included in the original image G0 is turned from the stage T3ab to the stage T3MRF+, for example. Alternatively, the three-dimensional model 41 is deformed until a normal organ has a shape to be evaluated as a lesion based on a measurement index in medical practice for a medical image.
The pseudo mask derivation unit 22 processes the mask while maintaining the three-dimensional continuity of the mask applied to the rectal cancer. For example, in deforming the three-dimensional model 41 to extend toward the outside of the rectum, a degree of deformation decreases as far from the center of the rectum in the three-dimensional model 41. With this, it is possible to deform the three-dimensional model 41 while maintaining the three-dimensional continuity in the original three-dimensional model 41, and as a result, it is possible to process the mask while maintaining the three-dimensional continuity of the mask applied to the rectal cancer.
The three-dimensional model 41 corresponds to the rectal cancer, and deforming the three-dimensional model 41 to make the stage of the rectal cancer progress is making the rectal cancer be infiltrated into the submucosal layer and the muscularis propria of the rectum. Here, while the rectal cancer is enlarged or deformed with the progress, a way of deformation depends on the shape of the rectum. That is, the rectal cancer is enlarged or deformed to match the shape of the rectum. Here, the rectum is fixed in the original image G0 without movement and deformation. For this reason, the pseudo mask derivation unit 22 deforms the three-dimensional model 41 matching the shapes of the masks applied to the fixed submucosal layer and muscularis propria in the vicinity of the rectal cancer. With this, it is possible to derive the pseudo mask image Mf0 representing the rectal cancer having a natural shape matching the shape of the rectum.
The operator may instruct the degree of deformation of the three-dimensional model 41.
On the mask processing screen 40A shown in
The derivation of the pseudo mask image Mf0 is not limited to that depending on the stage of the rectal cancer designated as described above. For example, in addition to or instead of the pull-down menu 26 for selecting the stage of the rectal cancer, a pull-down menu for selecting the presence or absence of application of a spinous protrusion may be displayed on the mask processing screen 40 shown in
In addition to or instead of the pull-down menu 26 for selecting the stage of the rectal cancer, a pull-down menu for selecting the presence or absence of application of a protrusion may be displayed on the mask processing screen 40 shown in
The pseudo mask derivation unit 22 derives information representing the stage of the rectal cancer on the derived pseudo mask image Mf0. Here, since the stage of the rectal cancer is selected from the pull-down menu 26 by the operator on the mask processing screen 40 shown in
In a case where the CONVERT button 28 is selected on the mask processing screen 40, the pseudo image derivation unit 23 derives a pseudo image having a region based on the mask included in the pseudo mask image Mf0, based on the original image G0 and the pseudo mask image Mf0. To this end, the pseudo image derivation unit 23 has a generator 50 that performs learning using generative adversarial networks (GAN). The generator 50 is constructed by performing learning to output a pseudo image having a region based on a mask included in the mask image in a case where the original image G0 including the rectal cancer and the mask image are input.
In the present embodiment, the pseudo image means an image having the same representation format as an image acquired by a modality that acquires the original image G0. That is, in a case where the original image G0 is an MRI image acquired by an MRI imaging apparatus, the pseudo image means an image having the same representation format as the MRI image. Here, the same representation format means that structures having the same composition are represented with the same density or brightness.
Here, as the generator 50, for example, a generator constructed by a technique of SPADE for generating a pseudo image from a mask image, described in “Semantic Image Synthesis with Spatially-Adaptive Normalization, Park, et al., arXiv:1903.07291v2 [cs.CV] 5 Nov. 2019” can be used.
The encoder 51 that configures the generator 50 is composed of a convolutional neural network (CNN) that is one multilayered neural network in which a plurality of processing layers are connected hierarchically, and in the present embodiment, outputs a latent representation z0 representing a feature quantity of the MRI image including the rectum in a case where the image 51 for learning is input.
The decoder 52 applies a mask of an individual region included in the mask image S2 for learning to generate a region represented by each mask while decoding the latent representation z0 output from the encoder 51, and outputs a pseudo image S3 having a region based on each of the masks included in the mask image and has the same representation format as the image 51 for leaning.
The discriminator 53 discriminates whether the input image is a real image or the pseudo image generated by the generator 50, and outputs a discrimination result TF0. Here, the real image is not an image generated by the generator 50, but an original image acquired by imaging a subject by the imaging apparatus 3. In contrast, the pseudo image is an image having the same representation format as the original image, generated from the mask image by the generator 50.
In the present embodiment, learning of the discriminator 53 is performed to correctly discriminate the discrimination result TF0 regarding whether the input image is a real image or the pseudo image generated by the generator 50. Learning of the generator 50 is performed to derive a pseudo image similar to the real image from the input mask image and is performed such that the discriminator 53 incorrectly discriminates the discrimination result TF0. With this, the generator 50 can generate a pseudo image having the same representation format as an MRI image of a real thing, not identified by the discriminator 53.
The pseudo image derivation unit 23 derives a pseudo image from the original image G0 and the pseudo mask image Mf0 derived by the pseudo mask derivation unit 22 with the generator 50 constructed in this manner. For example, in a case where the original image G0 and the pseudo mask image Mf0 shown in
The learning unit 24 learns the segmentation model that segments the MM image including the rectum into a plurality of regions. In the present embodiment, learning of a semantic segmentation model that segments the MM image into the regions of the rectal cancer, the mucous membrane layer of the rectum, the submucosal layer of the rectum, the muscularis propria of the rectum, the subserous layer of the rectum, and the background other than the layers is performed, and a trained semantic segmentation model is constructed. The semantic segmentation model (hereinafter, referred to as an SS model) is a machine learning model that outputs an output image with a mask representing an extraction target object (class) applied to each pixel of an input image as well known in the art. In the present embodiment, the input image to the SS model is the MRI image including the region of the rectum, and the output image is the mask image in which the regions of the rectal cancer, the mucous membrane layer of the rectum, the submucosal layer of the rectum, the muscularis propria of the rectum, the subserous layer of the rectum, and the background other than the layers in the MRI image are masked. The SS model is constructed by a convolutional neural network (CNN), such as Residual Networks (ResNet) or U-shaped Networks (U-Net).
In regard to the learning of the SS model, in addition to existing training data, that is, training data composed of a combination of the original image G0 and the mask image M0 of the original image G0, training data composed of a combination of the pseudo mask image Mf0 derived by the pseudo mask derivation unit 22 and the pseudo image Gf0 derived by the pseudo image derivation unit 23 is used. In the existing training data, the original image G0 and the pseudo image Gf0 are data for learning, and the mask image M0 and the pseudo mask image Mf0 are correct answer data. In the training data including the pseudo image Gf0 and the pseudo mask image Mf0, the pseudo image Gf0 is data for learning, and the pseudo mask image Mf0 is correct answer data.
The original image G0 and the pseudo image Gf0 are input to the SS model at learning, and a mask image in which an object included in the images is segmented is output. Next, a difference between the mask image output from the SS model, and the mask image M0 and the pseudo mask image Mf0 as correct answer data is derived as a loss. Then, learning of the SS model is repeated using a plurality of kinds of training data such that the loss decreases, and the SS model is constructed.
The learning unit 24 performs learning of a discrimination model that discriminates the stage of the rectal cancer on the MRI image including the rectum, and a trained discrimination model is constructed. In the present embodiment, an input image to the discrimination model is the MRI image including the regions of the rectum and a mask image obtained by segmenting the MRI image, and an output is the stage of the rectal cancer included in the MM image. The discrimination model is also constructed by a convolutional neural network, such as ResNet or U-Net.
At learning of the discrimination model, in addition to existing training data, that is, training data composed of a combination of the existing training data, that is, the original image G0, the mask image M0 of the original image G0, and information representing the stage of the rectal cancer included in the original image G0, training data composed of a combination of the pseudo mask image Mf0 derived by the pseudo mask derivation unit 22, the pseudo image Gf0 derived by the pseudo image derivation unit 23, and the information representing the stage of the rectal cancer included in the pseudo image Gf0 is used. In the existing training data, the original image G0 and the mask image M0 are data for learning, and information representing the stage of the rectal cancer of the original image G0 is correct answer data. In the training data including the pseudo image Gf0 and the pseudo mask image Mf0, the pseudo image Gf0 and the pseudo mask image Mf0 are data for learning, and information representing the stage of the rectal cancer is correct answer data.
The original image G0 and the mask image M0, and the pseudo image Gf0 and the pseudo mask image Mf0 are input to the discrimination model at learning, and information representing the stage of the rectal cancer included in such images is output. As information representing the stage of the rectal cancer, a probability of each stage of the rectal cancer is used. The probability has a value of 0 to 1. Next, a difference between the probability of each stage of the rectal cancer and the stage of the rectal cancer of correct answer data is derived as a loss. Here, in a case where it is assumed that the stage of the rectal cancer output from the discrimination model is (T1, T2, T3, T4)=(0.1, 0.1, 0.7, 0.1), and correct answer data is (0, 0, 1, 0), the difference between the probability of each stage of the rectal cancer output from the discrimination model and the probability of each stage of the rectal cancer in correct answer data is derived as a loss. Then, learning of the discrimination model is repeated using a plurality of kinds of training data such that the loss decreases, and the discrimination model is constructed.
The discrimination model may be constructed to output information representing the stage of the rectal cancer included in the input image in a case where only the mask image on the input image is input. In this case, in learning of the discrimination model, training data including the mask image M0 on the original image G0 as data for learning and information representing the stage of the rectal cancer included in the original image G0 as correct answer data, and training data including the pseudo mask image Mf0 on the pseudo image Gf0 as data for learning and information representing the stage of the rectal cancer included in the pseudo image Gf0 as correct answer data are used.
Next, the image processing apparatus according to the present embodiment will be described.
An image processing program 62 is stored in the storage 63. The CPU 61 reads out the image processing program 62 from the storage 63, develops the image processing program 62 to the memory 66, and executes the developed image processing program 62.
Next, the functional configuration of the image processing apparatus according to the present embodiment will be described.
The image acquisition unit 71 acquires a target image T0 to be a target of processing from the image storage server 4. The target image T0 is an MRI image including a rectum of a patient.
The segmentation unit 72 segments regions of an object included in the target image T0 to derive a mask image TM0 in which the regions of the object included in the target image T0 are masked. In the present embodiment, the mask image TM0 in which regions of a rectal cancer, a mucous membrane layer of a rectum, a submucosal layer of the rectum, a muscularis propria of the rectum, a subserous layer of the rectum, and a background other than the layers included in the target image T0 are segmented, and a mask is applied to each region is derived. To this end, an SS model 72A constructed by the learning apparatus according to the present embodiment is applied to the segmentation unit 72.
The discrimination unit 73 discriminates a stage of the rectal cancer included in the target image T0 and outputs a discrimination result. To this end, a discrimination model 73A constructed by the learning apparatus according to the present embodiment is applied to the discrimination unit 73. The target image T0 and the mask image TM0 of the target image T0 derived by the segmentation unit 72 are input to the discrimination model 73A, and the discrimination result of the stage of the rectal cancer included in the target image T0 is output.
The display control unit 74 displays the mask image TM0 derived by the segmentation unit 72 and the discrimination result of the stage of the rectal cancer derived by the discrimination unit 73 on the display 64.
Next, processing that is executed in the present embodiment will be described.
Here, since a scarce disease having a small frequency of appearance, such as a progressive cancer, has a small number of cases, it is not possible to prepare a sufficient amount of training data for constructing a machine learning model for performing segmentation and stage discrimination. For this reason, it is difficult to provide a machine learning model capable of accurately segmenting a scarce disease or accurately discriminating a class of a scarce disease, such as a stage of a progressive cancer.
In the present embodiment, the pseudo mask image Mf0 is derived by processing the mask in the mask image M0 on the original image G0, and the pseudo image Gf0 having the region based on the pseudo mask image Mf0 is derived. With this, it is possible to prepare training data including a target object of a class that is not in existing training data for constructing a segmentation model or is not at all. For example, a pseudo image Gf0 including a progressed rectal cancer can be prepared as training data. For this reason, a pseudo image Gf0 on a scarce disease is derived, is accumulated along with existing training data, and is used for learning of a learning model for performing segmentation and stage discrimination, whereby a sufficient amount of training data to such an extent that segmentation can be accurately performed even on a scarce disease can be prepared. Accordingly, it is possible to provide a machine learning model capable of accurately segmenting a scarce disease or accurately discriminating a class of a scarce disease, such as a stage of a progressive cancer, on a target image to be processed.
In the above-described embodiment, although the target object is the rectal cancer, the present disclosure is not limited thereto. A lesion, such as a cancer or a tumor of an organ or a structure other than the rectum can be used as a target object. For example, derivation of a pseudo mask and a pseudo image and construction of an SS model with a bone spur in a joint as a target object can be performed. Hereinafter, this will be described as another embodiment.
Here, the bone spur refers to a disease that a cartilage of an articular facet hypertropically grows and is gradually hardened and ossified, and becomes like a “spur”, and is one of characteristic findings of osteoarthritis to be expected around an articular facet. In such a case, a pseudo mask may be derived to form a bone spur with a bone composing a joint as a target object, and a pseudo image in which the bone spur is formed on the bone composing the joint may be derived.
Here, for description, a mask Ms0 is applied to only a region of the shinbone to be processed in the mask image M0 shown in
The three-dimensional model 91 is a three-dimensional image derived by extracting only a region near the joint of the shinbone in the original image G0 and performing volume rendering. The operator can change omnidirectional orientations of the three-dimensional model 91 by operating the input device 15. The operator designates a processing place of the mask by designating a desired position in the three-dimensional model 91 using the input device 15. Then, the operator designates a degree of processing of the mask by moving a knob 92A of the scale 92 using the input device 15. With this, the pseudo mask derivation unit 22 processes the three-dimensional model 91 to extend with the designated position as a starting point as shown in
In this case, the pseudo mask derivation unit 22 processes the mask while maintaining three-dimensional continuity of the mask applied to the shinbone. For example, in deforming to extend the designated position in the three-dimensional model 91, the degree of deformation decreases as far from the designated position in the three-dimensional model 91. With this, it is possible to deform the three-dimensional model 91 while maintaining three-dimensional continuity in the original three-dimensional model 91, and as a result, it is possible to process the mask like the shinbone on which the bone spur is formed while maintaining the three-dimensional continuity of the mask applied to the shinbone.
The pseudo image derivation unit 23 derives the pseudo image Gf0 including the shinbone on which the bone spur is formed, from the pseudo mask image Mf0 in which the bone spur is formed. Then, the learning unit 24 learns the SS model using the pseudo mask image WO in which the bone spur is formed and the pseudo image Gf0 as training data. With this, it is possible to construct a trained SS model capable of accurately segmenting the bone spur in the MRI image including the knee joint. Accordingly, the SS model constructed in this manner is applied to the segmentation unit 72 of the image processing apparatus according to the present embodiment, whereby it is possible to accurately segment the region of the bone spur in the MRI image including the knee joint.
The pseudo image Gf0 including the shinbone on which the bone spur is formed in the joint is used as training data, whereby it is possible to construct a discrimination model that discriminates the presence or absence of the bone spur, on the MRI image including the joint of the shinbone. In constructing the discrimination model that discriminates the presence or absence of the bone spur, the original image including the shinbone on which no bone spur is formed is also used as training data.
In another embodiment described above, although the pseudo mask image Mf0 and the pseudo image Gf0 are derived by processing the original image G0 in which no bone spur is included, a lesion to be added is not limited to the bone spur. An organ to be a target of addition of any lesion may be included, and an original image G0 may be processed to add the lesion to the original image G0 including no lesion, thereby deriving a pseudo mask image Mf0 and a pseudo image Gf0.
In each embodiment described above, in a case where the pseudo mask derivation unit 22 processes the mask to derive the pseudo mask image WO, constraint conditions may be set with respect to the deformation of the mask, and designation of the degree of processing of the mask may be received under the constraint conditions.
In the pseudo mask image Mf0 displayed on the mask processing screen 100 shown in
In the condition list 101, constraint conditions regarding deformation of a mask to be processed are displayed to be selectable by checkboxes. As shown in
In a case where “ROTATE AROUND CENTER OF GRAVITY OF FIXED MASK” is selected as the constraint condition, a center C0 of gravity of the mask Msf1 of the rectum as the fixed mask is displayed in the pseudo mask image Mf0. With this, in a case where the operator deforms and processes the mask Mfs0 using the input device 15, the deformation of the mask Msf0 is constrained such that only rotation around the center C0 of gravity of the fixed mask Msf1 is possible.
Then, after the mask Msf0 is processed and the pseudo mask image Mf0 is derived, in a case where the CONVERT button 103 is selected, the pseudo image derivation unit 23 derives the pseudo image Gf0 depending on the processed mask Msf0 and the selected stage of the rectal cancer.
Then, after the mask Msf0 is processed and the pseudo mask image Mf0 is derived, in a case where the CONVERT button 103 is selected, the pseudo image derivation unit 23 derives the pseudo image Gf0 depending on the processed mask Msf0 and the selected stage of the rectal cancer.
Then, after the mask Msf0 is processed and the pseudo mask image Mf0 is derived, in a case where the CONVERT button 103 is selected, the pseudo image derivation unit 23 derives the pseudo image Gf0 depending on the processed mask Msf0 and the selected stage of the rectal cancer.
On the mask processing screen 107 shown in
In each embodiment described above, although the pseudo mask image Mf0 is derived by deforming the rectal cancer to be increased or adding the bone spur, the present disclosure is not limited thereto. The technique of the present disclosure can also be applied to a case of deriving a pseudo image Gf0 on a disease like constriction.
For example, in a case of deriving a pseudo image Gf0 including a disease of vascular constriction, a pseudo mask image Mf0 is derived by processing a mask of a blood vessel of an original image G0 including no vascular constriction to be thinned, whereby it is possible to derive the pseudo image Gf0 including vascular constriction.
In the above-described embodiment, although the pseudo image derivation unit 23 derives the pseudo image Gf0 having the same representation format as the original image G0, the present disclosure is not limited thereto. In a case where the pseudo image derivation unit 23 derives a pseudo image, at least one of density, color, or texture of the pseudo image Gf0 may be changed. In this case, a plurality of style images having predetermined density, colors, and texture may be stored in the storage 13, and the pseudo image Gf0 may be derived to have the density, color, or texture selected from among a plurality of style images.
The style image list 122 displayed on the mask processing screen 120 shown in
In the above-described embodiment, for example, as a hardware structure of processing units that execute various kinds of processing, such as the information acquisition unit 21, the pseudo mask derivation unit 22, the pseudo image derivation unit 23, and the learning unit 24 of the image generation apparatus 20, and the image acquisition unit 71, the segmentation unit 72, the discrimination unit 73, and the display control unit 74 of the image processing apparatus 60, various processors described below can be used. As described above, various processors include a programmable logic device (PLD) that is a processor capable of changing a circuit configuration after manufacturing, such as a field programmable gate array (FPGA), a dedicated electric circuit that is a processor having a circuit configuration dedicatedly designed for executing specific processing, such as an application specific integrated circuit (ASIC), and the like, in addition to a CPU that is a general-purpose processor configured to execute software (program) to function as various processing units.
One processing unit may be configured with one of various processors described above or may be configured with a combination of two or more processors (for example, a combination of a plurality of FPGAs or a combination of a CPU and an FPGA) of the same type or different types. A plurality of processing units may be configured with one processor.
As an example where a plurality of processing units are configured with one processor, first, as represented by a computer, such as a client or a server, there is a form in which one processor is configured with a combination of one or more CPUs and software, and the processor functions as a plurality of processing units. Second, as represented by system on chip (SoC) or the like, there is a form in which a processor that realizes all functions of a system including a plurality of processing units into one integrated circuit (IC) chip is used. In this way, various processing units are configured using one or more processors among various processors described above as a hardware structure.
In addition, as the hardware structure of various processors, more specifically, an electric circuit (circuitry), in which circuit elements, such as semiconductor elements, are combined can be used.
Number | Date | Country | Kind |
---|---|---|---|
2022-050635 | Mar 2022 | JP | national |
2022-150250 | Sep 2022 | JP | national |