The present disclosure generally relates to a program, an information processing method, an information processing apparatus, and a model generation method.
In a medical field, complications caused by medical practices such as surgery and examination can become a problem. Accordingly, various systems for supporting prevention of the complications have been proposed.
For example, Japanese Patent Application Publication No. 2014-200549A discloses a medical image processing apparatus that detects a region of surrounding tissue existing around an aortic valve of a heart based on a medical image captured by an X-ray computed tomography (CT) apparatus, arranges a model image of a prosthetic valve to be replaced with the aortic valve in the medical image, and evaluates a risk of a complication based on a distance between the region of the surrounding tissue and the model image of the prosthetic valve.
However, in the disclosure according to Japanese Patent Application Publication No. 2014-200549A, it cannot be said that an evaluation of the risk of the complication by simple pattern matching based on the distance in the image is necessarily accurate.
A non-transitory computer-readable medium, an information processing method, an information processing apparatus, and a model generation method, are disclosed that are capable of suitably evaluating a risk of a complication.
A non-transitory computer-readable medium (CRM) storing computer program code executed by a computer processor that executes a process for outputting complication information is disclosed. The process includes acquiring a medical image obtained by imaging a lumen organ of a patient before treatment, inputting the acquired medical image into a trained model that is trained to output complication information on a complication that is likely to occur after the treatment, and outputting the complication information.
An information processing apparatus is disclosed, which comprises: an acquisition unit configured to acquire a medical image obtained by imaging a lumen organ of a patient before treatment; and an output unit configured to input the acquired medical image into a trained model that is learned to output complication information on a complication that is likely to occur after the treatment, and output the complication information.
A model generation method executed by a computer is disclosed, the model generation method comprising: acquiring training data including a medical image obtained by imaging a lumen organ of a patient before treatment and complication information on a complication that occurs after the treatment; and generating, based on the training data, a trained model that outputs the complication information when the medical image is received.
In one aspect, a risk of a complication can be suitably evaluated.
Set forth below with reference to the accompanying drawings is a detailed description of embodiments of a program, an information processing method, an information processing device, and a model generating method. Note that since embodiments described below are preferred specific examples of the present disclosure, although various technically preferable limitations are given, the scope of the present disclosure is not limited to the embodiments unless otherwise specified in the following descriptions. For convenience of explanation, dimensions in the drawings may be exaggerated and may be different from actual dimensions.
In the present embodiment, the intravascular treatment is described as an example, but a lumen organ as a target is not limited to a blood vessel, and may be, for example, another lumen organ such as a bile duct, a pancreatic duct, a bronchus, or an intestine.
The diagnostic imaging system 2 can include an intravascular diagnostic imaging apparatus 21, a fluoroscopic image capturing apparatus 22, and a display apparatus 23. The intravascular diagnostic imaging apparatus 21 is an apparatus for imaging an intravascular tomographic image of the patient, and can be, for example, an intravascular ultrasound (IVUS) apparatus that performs an ultrasonic examination using a catheter 211. The catheter 211 is a medical instrument to be inserted into the blood vessel of the patient, and includes an imaging core that transmits ultrasound and receives a reflected wave from the inside of the blood vessel. The intravascular diagnostic imaging apparatus 21 generates an ultrasound tomographic image (transverse tomographic image) based on a signal of the reflected wave received by the catheter 211, and displays the ultrasound tomographic image on the display apparatus 23.
In the present embodiment, the intravascular diagnostic imaging apparatus 21 generates an ultrasound tomographic image, and may also generate an optical coherence tomographic image by an optical method such as optical coherence tomography (OCT) or optical frequency domain imaging (OFDI).
The fluoroscopic image capturing apparatus 22 is an apparatus unit for capturing a fluoroscopic image in which a body of the patient is seen through, and can be, for example, an angiography apparatus that performs an angiographic examination. The fluoroscopic image capturing apparatus 22 can include an X-ray source 221 and an X-ray sensor 222, and generates an X-ray fluoroscopic image by the X-ray sensor 222 receiving X-rays emitted from the X-ray source 221. For example, an X-ray opaque marker is attached to a distal end of the catheter 211 to enable position alignment with a tomographic image generated by the intravascular diagnostic imaging apparatus 21.
As described above, examples of a medical image include an ultrasound tomographic image, an optical coherence tomographic image and an angiographic image, and the medical image may also be a computed tomography (CT) image, a magnetic resonance imaging (MRI) image, or the like.
The information processing apparatus 1 is an information processing apparatus capable of processing various types of information and transmitting and receiving information, and can be, for example, a server computer, or a personal computer. In the present embodiment, the information processing apparatus 1 is a server computer, and hereinafter, the information processing apparatus 1 is read as a server 1 for the sake of conciseness. The server 1 may be a local server installed in the same facility (hospital or the like) as the diagnostic imaging system 2, or may be a cloud server communicably connected via the Internet or the like. Based on the medical image (tomographic image and fluoroscopic image) of the patient on whom the intravascular treatment is performed, the server 1 functions as a treatment support apparatus that outputs complication information on a complication that is likely to occur after the treatment. Specifically, as will be described later, the server 1 performs machine learning using predetermined training data, and prepares, in advance, a learning model 50 (see
The complication information is a prediction result on the complication that is likely to occur due to the intravascular treatment, and can be information for assisting a user (health care worker) who treats a patient from a viewpoint of preventing complications. In the present embodiment, the intravascular treatment is an example, and a prediction result of the complication that is likely to occur due to percutaneous coronary intervention (PCI) using the catheter 211 is output as the complication information. For example, the complication information can include, in addition to a type of the complication that is highly likely to occur and a probability value of the type of the complication that is highly likely to occur, a prediction result of occurrence conditions of the complication, such as a treatment device that may cause the complication and use conditions of the treatment device. The treatment device can be, for example, a stent indwelled in a blood vessel, a balloon used for inflating a blood vessel, and is a device to be inserted into a blood vessel of a patient.
In addition, the server 1 can detect a dangerous region in which the complication is likely to occur based on the medical image, generate a second medical image indicating the detected dangerous region, and cause the display apparatus 23 to display the second medical image. The dangerous region can be, for example, a corresponding portion of a blood vessel in which the complication is highly likely to occur, and is a specific image region in the medical image. The server 1 can detect the dangerous region based on the medical image using the training model 50, generate the second medical image in which the dangerous region can be identified by a method such as color display, and cause the display apparatus 23 to display the second medical image.
Although the server 1 can perform processing using the training model 50 in the present embodiment, the training model 50 constructed by the server 1 may also be installed in the diagnostic imaging system 2 such that the processing may be locally executed.
The auxiliary storage unit 14 can be a nonvolatile storage area such as a large-capacity memory or a hard disk, and stores the program P necessary for the control unit 11 to execute processing and other data. In addition, the auxiliary storage unit 14 stores the training model 50 and a countermeasure information database (DB) 141. The training model 50 is a model of machine learning that is trained using the training data as described above, and is a model that receives a medical image as input and outputs the complication information on the complication that is likely to occur after the treatment. The training model 50 can be used as a program module forming artificial intelligence software. The countermeasure information DB 141 is a database that stores countermeasure information for complications. In the present embodiment, the server 1 predicts the complication using the training model 50, and presents, for example, to the user, the countermeasure information for the complication that is highly likely to occur.
The auxiliary storage unit 14 may be an external storage device connected to the server 1. The server 1 may be a multi-computer including a plurality of computers, or may be a virtual machine virtually constructed by software.
In the present embodiment, the server 1 is not limited to the configuration described above, and may include, for example, an input unit that receives operation input, and a display unit that displays an image. In addition, the server 1 may include a reading unit that reads a portable storage medium 1a such as a compact disk (CD), a digital versatile disc (DVD), or a universal serial bus (USB) memory, and may read and execute the program P from the portable storage medium 1a. Alternatively, the server 1 may read the program P from a semiconductor memory 1b.
As shown in
For example, the training model 50 can output a probability value for each of a plurality of types of complications as one of estimation items. The server 1 performs learning using case data on the plurality of types of complications as training data, and prepares the training model 50 that estimates a probability value representing a probability of occurrence of each type of complication.
A complication to be estimated is not particularly limited, and representative examples of the complication associated with PCI include a hemorrhagic complication, side branch occlusion, and peripheral embolism. The hemorrhagic complication can include a rupture of a blood vessel, dissociation of a blood vessel wall, and the like. For example, when a balloon is inflated in a lesion where unevenly distributed calcified tissue is present, the blood vessel may be ruptured. In addition, when the stent is expanded at a portion having a plaque, the dissociation of the blood vessel wall or a rupture of the plaque is likely to occur due to an edge of the stent.
The side branch occlusion can be a complication that occludes a side branch blood vessel by treating a main trunk (main branch). When the balloon is inflated or the stent is indwelled at a bifurcated portion where the main trunk and the side branch diverge, an inlet portion of the side branch may be compressed and blocked.
The peripheral embolism can be a complication in which a thrombus or the like occurs due to mechanical stimulation of the intravascular treatment and a peripheral artery is embolized. For example, when the balloon is inflated or the stent is indwelled for an unstable plaque (plaque including lipid components or inflammatory cells), contents of the plaque may flow out due to the mechanical stimulation and the thrombus is likely to occur.
The training model 50 estimates the probability value representing the probability of occurrence of each type of complication based on the medical image, and outputs the probability value as the complication information. The above complications are examples, and the complication to be estimated is not limited to the complications based on the medical image, and the outputs of the probability value as the complication information as disclosed herein.
In addition, the complication information can include an estimation result related to occurrence conditions of the complication. Specifically, the training model 50 estimates the treatment device that may cause the complication or the use conditions of the treatment device. As described above, the treatment device can be, for example, a balloon used for inflating a blood vessel, and a stent indwelled in a blood vessel, and can be a device to be inserted into a blood vessel. The treatment device is not limited to the balloon and the stent, and may be, for example, a rotablator for resection of a lesion.
For example, as shown in the upper right part of
The training model 50 outputs the probability value, the occurrence conditions, and the like of each type of complication as the complication information, generates the second medical image by processing the dangerous region in which the complication is likely to occur with respect to the input medical image, and outputs the second medical image as one piece of complication information. The second medical image is a medical image showing an image region in which a risk of occurrence of the complication is high, and is an image in which a dangerous region is displayed in a display mode different from that of other image regions as shown by hatching on a right side of
It is preferable that the training model 50 changes the display mode (display color) of the dangerous region according to the type of complication corresponding to the dangerous region. Accordingly, the user can rather easily grasp which complication the dangerous region presented in the second medical image corresponds to.
For example, the training model 50 can be a neural network generated by deep learning, and is a convolutional neural network (CNN) that extracts feature data of an input image by a large number of convolutional layers. The training model 50 includes an intermediate layer (hidden layer) in which a convolution layer that convolutes pixel information of the input image and a pooling layer that maps the convoluted pixel information are alternately connected, and extracts the feature data (feature data map) of the input image. The training model 50 outputs the complication information based on the extracted feature data.
Although the training model 50 is described as a CNN in the present embodiment, the training model 50 may be, for example, another neural network such as a recurrent neural network (RNN), or another model of machine learning such as a generative adversarial network (GAN), a support vector machine (SVM), or a decision tree.
The training model 50 according to the present embodiment can include a first encoder 51, a second encoder 52, a classifier 53, a first decoder 54, and a second decoder 55. Each of the first encoder 51 and the second encoder 52 can be an encoder that receives input of the tomographic image acquired from the intravascular diagnostic imaging apparatus 21 and the fluoroscopic image acquired from the fluoroscopic image capturing apparatus 22, and is an extractor that extracts the feature data of the input image by a large number of convolutional layers. Each of the first encoder 51 and the second encoder 52 performs convolution of the tomographic image and the fluoroscopic image, and extracts feature data of each image.
The classifier 53 can be a classifier that performs classification based on the feature data extracted by the first encoder 51 and the second encoder 52, and is an estimator that estimates the complication information other than the second medical image, such as a probability value and occurrence conditions of the complication. For example, the server 1 combines the feature data of the images extracted by the first encoder 51 and the second encoder 52, inputs the combined feature data into the classifier 53, and causes the classifier 53 to estimate the probability value of each type of complication and the like.
Each of the first decoder 54 and the second decoder 55 is a detector that detects the dangerous region from the tomographic image and the fluoroscopic image, and can be, for example, a decoder according to semantic segmentation (U-net or the like). The first decoder 54 and the second decoder 55 detect the dangerous region in pixel units from the tomographic image and the fluoroscopic image based on the feature data extracted by the first encoder 51 and the second encoder 52.
The semantic segmentation is a type of CNN, and is a type of encoder-decoder model that generates output data based on input data. The semantic segmentation can include, in addition to the convolution layer that compresses data of the input image, a deconvolution layer that maps (enlarges) feature data obtained by compression to an original image size. The deconvolution layer identifies which object is present at which position in the image in pixel units based on the feature data extracted by the convolution layer, and generates a label image generated by binarizing which object the pixel corresponds to.
Each of the first decoder 54 and the second decoder 55 includes a large number of deconvolution layers, and generates the label image generated by binarizing the dangerous region for each of the tomographic image and the fluoroscopic image. For example, the server 1 combines the feature data of the tomographic image and the fluoroscopic image extracted by the first encoder 51 and the second encoder 52, inputs the combined feature data into the first decoder 54 and the second decoder 55 respectively, and causes the first decoder 54 and the second decoder 55 to generate the label image.
The semantic segmentation is used to detect the dangerous region in the present embodiment, and other object detection models such as Region CNN (R-CNN) and You Only Look Once (YOLO) may also be used.
The configuration of the training model 50 illustrated in
The server 1 uses the training data to which correct complication information is assigned as label data (metadata) to perform learning on the medical image for training. For example, the server 1 uses data of a plurality of patients in whom the complication occurs after the treatment as the training data. In the training data, the intravascular tomographic image and the fluoroscopic image of the patient in whom the complication occurs are assigned with label data. The label data indicates the type of the complication that occurs in the patient, the occurrence conditions (inflated diameter, inflated pressure, inflated time, and the like of balloon) when the complication occurs, and the region (dangerous region) in the medical image in which the complication occurs.
The training data may include not only the data of the patient in whom the complication occurs but also the data of the patient in whom the complication does not occur. In addition, the training data is not limited to data of an actual patient, and may also be virtual data increased by a data generation unit such as GAN.
The server 1 inputs the tomographic image and the fluoroscopic image for training into the first encoder 51 and the second encoder 52 to extract the feature data, and inputs the extracted feature data into the classifier 53, the first decoder 54, and the second decoder 55 to output the complication information. The server 1 compares the output complication information with the correct complication information, and optimizes a parameter such as a weight between neurons so that the two pieces of complication information are approximate to each other. Accordingly, the server 1 generates the training model 50.
When the complication information is actually output for the patient to be treated, the server 1 acquires the medical image of the patient to be treated from the diagnostic imaging system 2, inputs the medical image into the training model 50, and outputs the complication information. As shown in
The server 1 may also input and process the tomographic images of the plurality of frames one by one into the training model 50, but it is preferable that the tomographic images of the plurality of consecutive frames are collectively input and the complication information can be estimated based on the tomographic images of the plurality of frames. In this case, for example, the training model 50 can be configured as a 3D-CNN (for example, 3D-U-net), and the tomographic images of the plurality of frames are handled as three-dimensional data in which coordinates of two-dimensional tomographic images are set as two axes and time points (generation time points) at which the tomographic images of the frames are acquired are set as one axis. The server 1 inputs the tomographic images (for example, 16 frames) for a predetermined unit time into the training model 50 as a set, and predicts the complication information based on the tomographic images of the plurality of frames.
Alternatively, the server 1 may set the training model 50 as a model obtained by combining the CNN and the RNN, and insert a long-short term memory (LSTM) layer that stores the feature data of the tomographic images of the frames into a rear side of the encoder, so that the complication information may be predictable based on the feature data of the tomographic images of the plurality of frames. By collectively processing the tomographic images of the plurality of consecutive frames, it is possible to improve prediction accuracy in consideration of not only a tomographic image at a certain position but also tomographic images at positions before and after the certain position.
The complication list 61 can be a list showing probability values of types of complications output from the classifier 53. For example, the server 1 determines whether a probability value of each type of complication output from the classifier 53 is equal to or greater than a predetermined threshold. The server 1 outputs names (types) of the complications determined to be equal to or greater than the threshold and the probability values of the complications as the complication list 61.
The occurrence condition list 62 is a list showing occurrence conditions of a complication output from the classifier 53. For example, the display apparatus 23 receives specified input for specifying the names (types) of the complications whose occurrence conditions are displayed for the complication list 61. Then, the display apparatus 23 displays estimated values of the occurrence conditions of a specified complication in the occurrence condition list 62. Specifically, the display apparatus 23 can display the inflated diameter, the inflated pressure, the inflated time, and the like of the balloon in which the complication is highly likely to occur.
The countermeasure information column 63 is a display column that displays countermeasure information for the complications. For example, similarly to the occurrence condition list 62, when the specified input of the names of the complications is received in the complication list 61, the display apparatus 23 displays the countermeasure information for the specified complications in the countermeasure information column 63.
The countermeasure information is information indicating a method of preventing the complications, and can be, for example, information indicating a treatment device that is effective for preventing the complications or a method of using the treatment device. According to the complication information output from the classifier 53, the server 1 reads the countermeasure information on the complication having a high probability value from the countermeasure information DB 141, and outputs the countermeasure information to the display apparatus 23. For example, the server 1 outputs the countermeasure information indicating another treatment device effective for preventing the complications, and proposes the user to use the devices in combination or substitute for another device.
The countermeasure information may be not only a method for preventing the complications, but also a countermeasure method that can be taken after the complications occur.
Further, the display apparatus 23 displays the tomographic image 64 generated by the intravascular diagnostic imaging apparatus 21 and the fluoroscopic image 65 captured by the fluoroscopic image capturing apparatus 22. In this case, the display apparatus 23 displays the second medical image obtained by processing the dangerous region of the original medical image as the tomographic image 64 and the fluoroscopic image 65. Specifically, as described above, the display apparatus 23 displays the second medical image indicating the dangerous region in a display mode different from that of the other image regions by superimposing the semitransparent mask on the dangerous region.
Although a two-dimensional tomographic image and a two-dimensional fluoroscopic image are shown as the second medical image in
Based on the training data, the control unit 11 generates the training model 50 that outputs the complication information on the complication that is likely to occur after the treatment when the medical image is input (S12). For example, as described above, the control unit 11 generates a neural network related to the CNN as the training model 50. The control unit 11 inputs the medical image for training into the training model 50 and acquires the complication information as output. The control unit 11 compares the output complication information with the correct complication information, and optimizes a parameter such as the weight between the neurons so that the two pieces of complication information are approximate to each other to generate the training model 50. The control unit 11 ends the series of processing.
The control unit 11 detects the dangerous region in which the complication is likely to occur based on the medical image acquired in S11 (S33). The control unit 11 generates the second medical image indicating the dangerous region detected in S33 in addition to the probability value and the occurrence conditions of the complication estimated in S32, the countermeasure information on the complication having the probability value equal to or greater than the threshold, and the like, outputs them to the display apparatus 23, and causes the display apparatus 23 to display them (S34). The control unit 11 ends the series of processing.
Only the medical image is used as the input into the training model 50 in the above description, but the present embodiment is not limited to only using the medical image as the input into the training model 50. For example, the server 1 may use, in addition to the medical image, the information on the balloon, the stent, and the like as the treatment device (diameter and length of stent, type and length of balloon, and the like) for input. Accordingly, based on the information on the treatment device itself, it is possible to suitably estimate the use conditions having a high possibility that the complication is likely to occur when the device is used. In addition, for example, the server 1 may use medical care information (for example, age, gender, and past medical history) on a patient to be treated for input. In this way, the input into the training model 50 is not limited to images.
As described above, according to the first embodiment, it is possible to suitably evaluate the risk of the complication by using the training model 50 that completes learning the training data.
According to the first embodiment, it is possible to present the complication that is highly likely to occur to the user by outputting the type and the probability value of the complication that is likely to occur.
According to the first embodiment, it is possible to more suitably support performing the treatment by outputting the countermeasure information for the complication as well.
According to the first embodiment, it is possible to more suitably support prevention of the complication by outputting the occurrence conditions of the complication, such as the treatment device that may cause the complication and the use conditions (expansion conditions) of the treatment device. [0083] According to the first embodiment, by outputting the second medical image indicating the dangerous region in which the complication is likely to occur, it is possible to suitably grasp a portion in which the complication is likely to occur.
In the present embodiment, an embodiment will be described in which only an intravascular tomographic image is used as a medical image. The same contents as those of the first embodiment are denoted by the same reference numerals, and descriptions of the same contents as those of the first embodiment will be omitted.
The vertical tomographic image 66 is an image obtained by imaging a cross section of the blood vessel parallel to the longitudinal direction, and is a tomographic image in which a direction of the cross section is orthogonal to that of a transverse tomographic image. The server 1 reconstructs transverse tomographic images of a plurality of frames generated by the intravascular diagnostic imaging apparatus 21, generates the vertical tomographic image 66 along the longitudinal direction of the blood vessel, and displays the vertical tomographic image 66 on the display apparatus 23.
In this case, similarly to the tomographic image 64 (transverse tomographic image), the server 1 generates the vertical tomographic image 66 in which the dangerous region can be identified, and causes the display apparatus 23 to display the vertical tomographic image 66. For example, as shown by hatching in
For example, the server 1 identifies the dangerous region in the vertical tomographic image 66 based on the detection result of the dangerous region in each of the plurality of transverse tomographic images serving as a basis of the vertical tomographic image 66. That is, when the dangerous region is detected from the plurality of continuous transverse tomographic images, the server 1 identifies a region of the vertical tomographic image 66 corresponding to each transverse tomographic image as the dangerous region, and superimposes the semitransparent mask thereon. The server 1 outputs, to the display apparatus 23, the vertical tomographic image 66 processed so that the dangerous region can be identified, and causes the display apparatus 23 to display the vertical tomographic image 66.
Since the second embodiment is the same as the first embodiment except for the above points, a detailed description of a flowchart and the like is omitted in the present embodiment.
In the present embodiment, an embodiment will be described in which correction input for correcting complication information output from the training model 50 is received from a user, and relearning is performed based on the corrected complication information.
When the correction input of the complication information is received, the control unit 11 uses a medical image input into the training model 50 and the corrected complication information as training data to perform the relearning, and updates the training model 50 (S302). That is, the control unit 11 updates a parameter such as a weight between neurons so that the complication information output from the training model 50 approximates to the corrected complication information. The control unit 11 ends the series of processing.
As described above, according to the third embodiment, it is possible to update the training model 50 through operations of the present system and improve estimation accuracy of the complication information.
The detailed description above describes embodiments of a program, an information processing method, an information processing device, and a model generating method. The invention is not limited, however, to the precise embodiments and variations described. Various changes, modifications and equivalents may occur to one skilled in the art without departing from the spirit and scope of the invention as defined in the accompanying claims. It is expressly intended that all such changes, modifications and equivalents which fall within the scope of the claims are embraced by the claims.
Number | Date | Country | Kind |
---|---|---|---|
2020-058997 | Mar 2020 | JP | national |
This application is a continuation of International Application No. PCT/JP2021/009280 filed on Mar. 9, 2021, which claims priority to Japanese Application No. 2020-058997 filed on Mar. 27, 2020, the entire content of both of which is incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2021/009280 | Mar 2021 | US |
Child | 17953581 | US |