The present invention relates to an image processing device, an image processing method, an image processing program, and a diagnosis support system.
There is a system that photographs an observation target placed on a glass slide with a microscope, generates a digitized pathological image, and performs various types of image analyses on the pathological image. For example, the observation target is a tissue or a cell collected from a patient, and corresponds to a piece of meat of an organ, saliva, blood, or the like.
As a conventional technique related to image analysis, a technique of inputting a pathological image to a morphology detector, detecting a morphology or a state of a cell nucleus, a cell membrane, or the like included in the pathological image, and calculating a feature amount obtained by quantifying a feature of the morphology is known. A skilled person such as a pathologist or a researcher sets adjustment items of an identifier for classifying or extracting a morphology or a state having a specific feature based on the calculation result of the feature amount and the specialized knowledge.
Patent Literature 1: JP 2018-502279 W
It is difficult for a user with poor specialized knowledge to associate a feature amount obtained by quantifying a feature of a morphology and a state calculated by the prior art with a feature based on the specialized knowledge of the user, and there is room for improvement.
Therefore, the present disclosure proposes an image processing device, an image processing method, an image processing program, and a diagnosis support system capable of appropriately displaying a feature amount obtained by quantifying an appearance feature of a morphology and easily setting an adjustment item of an identifier.
To solve the problems described above, an image processing device according to an embodiment of the present disclosure includes: in a case where designation of a plurality of partial regions corresponding to a cell morphology is received, the plurality of partial regions being extracted from a pathological image, a generation unit that generates auxiliary information indicating information about a feature amount effective when a plurality of partial regions is classified or extracted with respect to a plurality of feature amounts calculated from the image; and in a case where setting information about an adjustment item according to the auxiliary information is received, an image processing unit that performs an image process on the image using the setting information.
Hereinafter, the embodiments of the present disclosure will be described in detail with reference to the drawings. In the following embodiments, the same parts are denoted by the same reference signs, and a duplicate description will be omitted.
Further, the present disclosure will be described in the order of the following items.
<Present Embodiment>
1. Configuration of system according to present embodiment
2. Various kinds of information
2-1. Pathological image
2-2. Browsing history information
2-3. Diagnostic information
3. Image processing device according to present embodiment
4. Processing procedure
5. Another process
6. Effects of image processing device according to present embodiment
7. Hardware configuration
8. Conclusion
First, a diagnosis support system 1 according to the present embodiment will be described with reference to
The pathology system 10 is a system mainly used by a pathologist, and is applied to, for example, a laboratory or a hospital. As illustrated in
The microscope 11 is an imaging device that has a function of an optical microscope, captures an image of an observation target placed on a glass slide, and acquires a pathological image that is a digital image. Note that the observation target is, for example, a tissue or a cell collected from a patient, and is a piece of meat of an organ, saliva, blood, or the like.
The server 12 is a device that stores and holds a pathological image captured by the microscope 11 in a storage unit (not illustrated). In a case of receiving the browsing request from the display control device 13, the server 12 searches for a pathological image from a storage unit (not illustrated) and transmits the searched pathological image to the display control device 13. In addition, in a case of receiving the acquisition request for the pathological image from the image processing device 100, the server 12 searches for the pathological image from the storage unit and transmits the searched pathological image to the image processing device 100.
The display control device 13 transmits a browsing request for the pathological image received from the user to the server 12. Then, the display control device 13 causes the display device 14 to display the pathological image received from the server 12.
The display device 14 has a screen using, for example, liquid crystal, electro-luminescence (EL), cathode ray tube (CRT), or the like. The display device 14 may be compatible with 4K or 8K, or may be formed by a plurality of display devices. The display device 14 displays the pathological image controlled by the display control device 13. Note that, although details will be described later, the server 12 stores the browsing history information about the region of the pathological image observed by the pathologist via the display device 14.
The image processing device 100 is a device that transmits an acquisition request for a pathological image to the server 12 and performs the image process on the pathological image received from the server 12.
[2-1. Pathological Image]
As described above, the pathological image is generated by imaging the observation target with the microscope 11. First, an imaging process by the microscope 11 will be described with reference to
In
Subsequently, after generating the entire image, the microscope 11 identifies the region where the observation target A10 exists from the entire image, and sequentially images, by the high-resolution imaging unit, each divided region obtained by dividing the region where the observation target A10 exists for each predetermined size. For example, as illustrated in
When the stage is moved, the glass slide G10 may move on the stage. When the glass slide G10 moves, an unimaged region of the observation target A10 may occur. As illustrated in
Note that the low-resolution imaging unit and the high-resolution imaging unit described above may be different optical systems or may be the same optical system. In a case of the same optical system, the microscope 11 changes the resolution according to the imaging target. Furthermore, in the above description, an example is described in which the imaging region is changed by moving the stage, but the imaging region may be changed by the microscope 11 moving an optical system (high-resolution imaging unit or the like). The imaging element provided in the high-resolution imaging unit may be a two-dimensional imaging element (area sensor) or a one-dimensional imaging element (line sensor). The light from the observation target may be condensed and imaged using the objective lens, or may be dispersed and imaged for each wavelength using the spectroscopic optical system. In addition,
Subsequently, each high-resolution image generated by the microscope 11 is divided into predetermined sizes. As a result, a partial image (hereinafter, it is referred to as a tile image.) is generated from the high-resolution image. This point will be described with reference to
In the example illustrated in
Note that, in the example of
In this manner, the server 12 generates the tile image that is the minimum unit of the captured image of the observation target A10. Then, the server 12 sequentially combines the tile images of the minimum unit to generate the tile images having different hierarchies. Specifically, the server 12 generates one tile image by combining a predetermined number of adjacent tile images. This point will be described with reference to
The upper part of
Furthermore, the server 12 generates a tile image obtained by further combining tile images adjacent to each other among the tile images obtained by combining the tile images of the minimum unit. In the example of
By repeating such a combining process, the server 12 finally generates one tile image having a resolution similar to the resolution of the tile image of the minimum unit. For example, as in the above example, in a case where the resolution of the tile image of the minimum unit is 256×256, the server 12 generates one tile image T1 having a resolution of 256×256 finally by repeating the above-described combining process.
Note that a region D illustrated in
The server 12 stores the tile images of the respective hierarchies as illustrated in
Note that the server 12 may store the tile images of the respective hierarchies as illustrated in
Furthermore, the server 12 may not store the tile images of all the hierarchies. For example, the server 12 may store only the tile images of the lowermost layer, may store only the tile images of the lowermost layer and the tile images of the uppermost layer, or may store only the tile images of a predetermined hierarchy (for example, odd-numbered hierarchies, even-numbered hierarchies, and the like.). At this time, in a case where the tile images of an unstored hierarchy is requested from another device, the server 12 generates a tile image requested from another device by dynamically combining the stored tile images. In this manner, the server 12 can prevent the storage capacity from being overloaded by thinning out the tile images to be stored.
Furthermore, although the imaging conditions are not mentioned in the above example, the server 12 may store the tile images of the respective hierarchies as illustrated in
In addition, another example of the imaging condition includes a staining condition for the observation target A10. Specifically, in the pathological diagnosis, a specific portion (for example, a cell nucleus or the like) of the observation target A10 may be stained using a fluorescent reagent. The fluorescent reagent is, for example, a substance that is excited and emits light when irradiated with light of a specific wavelength. Then, different luminescent substances may be stained for the same observation target A10. In this case, the server 12 may store tile images of respective hierarchies as illustrated in
Furthermore, the number and resolution of the tile images described above are merely examples, and can be appropriately changed depending on the system. For example, the number of tile images combined by the server 12 is not limited to four. For example, the server 12 may repeat a process of combining 3×3=9 tile images. In the above example, the resolution of the tile image is 256×256, but the resolution of the tile image may be other than 256×256.
The display control device 13 extracts a desired tile image from the tile image group having the hierarchical structure according to an input operation of the user via the display control device 13 using software of a system capable of handling the tile image group having the hierarchical structure described above to output the extracted tile image to the display device 14. Specifically, the display device 14 displays an image of an arbitrary portion selected by the user among images of aby resolution selected by the user. By such a process, the user can obtain a feeling as if the user is observing the observation target while changing the observation magnification. That is, the display control device 13 functions as a virtual microscope. The virtual observation magnification here actually corresponds to the resolution.
[2-2. Browsing History Information]
Next, the browsing history information about the pathological image stored in the server 12 will be described with reference to
In the example of
While the pathological image is browsed as described above, the display control device 13 acquires the browsing information at a predetermined sampling cycle. Specifically, the display control device 13 acquires the center coordinates and the display magnification of the browsed pathological image at each predetermined timing, and stores the acquired browsing information in the storage unit of the server 12.
This point will be described with reference to
In the example of
Furthermore, the number of times each region has been browsed can be extracted from the browsing history information. For example, it is assumed that the number of times of display of each pixel of the displayed pathological image is increased by one each time a display region changing operation (for example, an operation of moving the display region and an operation of changing the display size) is performed. For example, in the example illustrated in
In a case where the operation of changing the display position is not performed by the viewer for a predetermined time (for example, 5 minutes), the display control device 13 may suspend the storage process of the browsing information. Furthermore, in the above example, an example is described in which the browsed pathological image is stored as the browsing information by the center coordinates and the magnification, but the present invention is not limited to this example, and the browsing information may be any information as long as it can identify the region of the browsed pathological image. For example, the display control device 13 may store, as the browsing information about the pathological image, tile identification information for identifying the tile image corresponding to the browsed pathological image or information indicating the position of the tile image corresponding to the browsed pathological image. Furthermore, although not illustrated in
[2-3. Diagnostic Information]
Next, diagnostic information stored in a medical information system 30 will be described with reference to
A diagnostic information storage unit 30A illustrated in
A diagnostic information storage unit 30B illustrated in
In the examples of
Next, the image processing device 100 according to the present embodiment will be described.
The communication unit 110 is realized by, for example, a network interface card (NIC) or the like. The communication unit 110 is connected to a network (not illustrated) in a wired or wireless manner, to transmit and receives information to and from the pathology system 10 and the like via the network. The control unit 150 described later transmits and receives information to and from these devices via the communication unit 110.
The input unit 120 is an input device that inputs various types of information to the image processing device 100. An input unit 111 corresponds to a keyboard, a mouse, a touch panel, or the like.
The display unit 130 is a display device that displays information output from the control unit 150. The display unit 130 corresponds to a liquid crystal display, an organic electro luminescence (EL) display, a touch panel, or the like.
The storage unit 140 includes a pathological image data base (DB) 141 and a feature amount table 142. For example, the storage unit 140 is realized by a semiconductor memory device such as a random access memory (RAM) and a flash memory, or a storage device such as a hard disk and an optical disk.
The pathological image DB 141 is a database that stores a plurality of pathological images.
The feature amount table 142 is a table that holds data of feature amounts of partial regions corresponding to cell nuclei and cell membranes extracted from the pathological image.
The feature amount is obtained by quantifying characteristics of various patterns including a tissue morphology and a state existing in a pathological image calculated from a partial region. For example, the feature amount corresponds to a feature amount output from a neural network (NN) such as a convolutional neural network (CNN). In addition, the feature amount corresponds to a cell nucleus or a color feature (Luminance, saturation, wavelength, spectrum, and the like), a shape feature (circularity, circumferential length), a density, a distance from a specific morphology, a local feature amount, a structure extraction process (nucleus detection and the like), information obtained by aggregating them (cell density, orientation, and the like), and the like of a cell nucleus. Here, respective feature amounts are indicated by feature amounts f1 to f10. Note that the feature amount may further include a feature amount fn other than the feature amounts f1 to f10.
The description returns to
The acquisition unit 151 is a processing unit that transmits an acquisition request for the pathological image to the server 12 and acquires the pathological image from the server 12. The acquisition unit 151 registers the acquired pathological image in the pathological image DB 141. The user may operate the input unit 120 to instruct the acquisition unit 151 to acquire the pathological image to be acquired. In this case, the acquisition unit 151 transmits an acquisition request for the instructed pathological image to the server 12 to acquire the instructed pathological image.
The pathological image acquired by the acquisition unit 151 corresponds to a whole slide imaging (WSI). Annotation data indicating part of the pathological image may be attached to the pathological image. The annotation data indicates a tumor region or the like indicated by a pathologist or a researcher. The number of WSIs is not limited to one, and a plurality of WSIs such as serial sections may be included. In addition to the patient ID, information such as the “diagnosis result”, the “grade”, the “tissue type”, the “genetic testing”, the “ultrasonic testing”, and “medication” described in
The analysis unit 152 is a processing unit that analyzes the pathological image stored in the pathological image DB 141 and calculates a feature amount. The user may operate the input unit 120 to designate the pathological image to be analyzed.
The analysis unit 152 acquires a pathological image designated by the user from the pathological image DB 141, and extracts a plurality of partial regions (patterns) from the pathological image by performing segmentation on the acquired pathological image. The plurality of partial regions includes individual cells, cell organs (cell nucleus, cell membrane, etc.), and cell morphology by a cell or cell organ aggregation. In addition, the partial region may be a region corresponding to a specific feature possessed in a case where the cell morphology is normal or in a case of a specific disease. Here, the segmentation is a technique of assigning a label of an object of a site in units of pixels from an image. For example, a learned model is generated by causing a convolution neural network to learn an image data set having a correct answer label, and an image (pathological image) desired to be processed is input to the learned model, whereby a label image to which a label of an object class is allocated in units of pixels can be obtained as an output, and a partial region can be extracted for each pixel by referring to the label.
Subsequently, the analysis unit 152 calculates a feature amount from the partial region. For example, the analysis unit 152 calculates a feature amount by inputting the image of the partial region to the CNN. In addition, the analysis unit 152 calculates a color feature (luminance value, dyeing intensity, etc.), a shape feature (circularity, circumferential length, etc.), a density, a distance from a specific morphology, and a local feature amount based on the image of the partial region. Any conventional technique may be used for the process in which the analysis unit 152 calculates the color feature, the shape feature, the density, the distance from the specific morphology, and the local feature amount. The analysis unit 152 registers the feature amounts (for example, the feature amounts f1 to f10) of the partial region in the feature amount table 142 in association with the region ID.
The analysis unit 152 may perform the above process after receiving an instruction for the pathological image to be analyzed from the user, or may calculate the feature amount of the partial region from a result of analyzing the entire pathological image in advance. Furthermore, the pathology system 10 may analyze the entire pathological image, and the analysis result by the pathology system 10 may be attached to the pathological image, and the analysis unit 152 may calculate the feature amount of the partial region using the analysis result by the pathology system 10.
The display control unit 153 is a processing unit that causes the display unit 130 to display screen information about a pathological image indicating the partial region (various patterns including the tissue morphology) extracted by the analysis unit 152 and receives designation of the partial region. For example, the display control unit 153 acquires the coordinates of each partial region from the feature amount table 142 and reflects the coordinates in the screen information.
In the example illustrated in
In the example illustrated in
The display control unit 153 outputs the region ID of the designated partial region and the information about the category of the partial region to the generation unit 154. In the following description, it is assumed that a region ID of a designated partial region is appropriately referred to as a “designated region ID”, and information about a category designated by the user is associated with the designated region ID. Note that the display control unit 153 outputs the first to fourth auxiliary information generated by the generation unit 154 to be described later to the display unit 130 to display it on the display unit 130.
The generation unit 154 is a processing unit that acquires the feature amount corresponding to the designated region ID from the feature amount table 142 and generates auxiliary information about the feature amount of the pathological image. The auxiliary information includes information that enables identification of a feature amount important for expressing a feature of a region desired to be classified or extracted, a distribution of feature amounts, and the like. For example, the generation unit 154 generates the first to fourth auxiliary information as the auxiliary information.
The process in which the generation unit 154 generates the “first auxiliary information” will be described. The generation unit 154 calculates a contribution rate (or importance) when classifying or extracting a partial region of the designated region ID for each category with respect to a plurality of feature amounts (for example, the feature amounts f1 to f10) calculated from the pathological image, and generates first auxiliary information.
As illustrated in
The process in which the generation unit 154 generates the “second auxiliary information” will be described. The generation unit 154 compares each feature amount corresponding to the designated region ID with a threshold value set in advance for each feature amount, and performs the process for identifying a feature amount equal to or greater than the threshold value for each category, thereby generating second auxiliary information.
The generation unit 154 compares the feature amounts f1 to f10 of the designated region ID corresponding to the second category with the threshold values Th1 to Th10 of the respective feature amounts. For example, in a case where the feature amount f1 is equal to or more than the threshold value Th1 and the feature amount f3 is equal to or more than the threshold value Th3, the generation unit 154 sets the feature amounts f1 and f3 as the feature amounts representing the characteristics of the second category.
The generation unit 154 compares the feature amounts f1 to f10 of the designated region ID corresponding to the third category with the threshold values Th1 to Th10 of the respective feature amounts. For example, in a case where the feature amount f5 is equal to or more than the threshold value Th5, the feature amount f3 is equal to or more than the threshold value Th3, and the feature amount f2 is equal to or more than the threshold value Th2, the generation unit 154 sets the feature amounts f5, f3, and f2 as the feature amounts representing the characteristics of the third category.
The generation unit 154 performs the above process to generate the second auxiliary information illustrated in
The generation unit 154 outputs the second auxiliary information to the display control unit 153 to request the display of the second auxiliary information. The display control unit 153 displays the second auxiliary information on the display unit 130. Note that the display control unit 153 may display the second auxiliary information on the display unit 130 in the form of a table illustrated in
The process in which the generation unit 154 generates the “third auxiliary information” will be described. The generation unit 154 generates third auxiliary information in which the distribution of respective partial regions is disposed in a feature space having the first feature amount fi and the second feature amount fj as axes among the feature amounts f1 to f10. The first feature amount fi and the second feature amount fj may be set in advance, or the feature amounts corresponding to the higher contribution rate may be set as the first feature amount fi and the second feature amount fj based on the contribution rate calculated in the case of generating the first auxiliary information in
For example, in a case where the partial region PA1 of the first category designated in
In a case where the partial region PB1 of the second category designated in
The generation unit 154 outputs the third auxiliary information to the display control unit 153 to request the display of the third auxiliary information. The display control unit 153 displays the third auxiliary information on the display unit 130. Note that, in the multidimensional feature amount, the generation unit 154 may calculate the low-dimensional feature amount obtained using dimension reduction such as principal component analysis or TSNE, and plot points corresponding to respective partial regions in the feature space of the low-dimensional feature amount.
The process in which the generation unit 154 generates the “fourth auxiliary information” will be described. Based on the feature amount table 142, the generation unit 154 generates a histogram of each of the feature amounts f1 to f10 as the fourth auxiliary information.
The histogram h1-1 is a histogram corresponding to the feature amount f1. In a case where the feature amount f1 of the partial region PA1 of the first category corresponds to the class value cm1, the generation unit 154 sets the color of the frequency corresponding to the class value cm1 to the first color. In a case where the feature amount f1 of the partial region PB1 of the second category corresponds to the class value cm2, the generation unit 154 sets the color of the frequency corresponding to the class value cm2 to the second color. In a case where the feature amount f1 of the partial region PC1 of the third category corresponds to the class value cm3, the generation unit 154 sets the color of the frequency corresponding to the class value cm3 to the third color.
The histogram h2-1 is a histogram corresponding to the feature amount f2. In a case where the feature amount f2 of the partial region PA1 of the first category corresponds to the class value cm1, the generation unit 154 sets the color of the frequency corresponding to the class value cm1 to the first color. In a case where the feature amount f2 of the partial region PB1 of the second category corresponds to the class value cm2, the generation unit 154 sets the color of the frequency corresponding to the class value cm2 to the second color. In a case where the feature amount f2 of the partial region PC1 of the third category corresponds to the class value cm3, the generation unit 154 sets the color of the frequency corresponding to the class value cm3 to the third color.
The histogram h3-1 is a histogram corresponding to the feature amount f3. In a case where the feature amount f3 of the partial region PA1 of the first category corresponds to the class value cm1, the generation unit 154 sets the color of the frequency corresponding to the class value cm1 to the first color. In a case where the feature amount f3 of the partial region PB1 of the second category corresponds to the class value cm2, the generation unit 154 sets the color of the frequency corresponding to the class value cm2 to the second color. In a case where the feature amount f3 of the partial region PC1 of the third category corresponds to the class value cm3, the generation unit 154 sets the color of the frequency corresponding to the class value cm3 to the third color.
The histogram h4-1 is a histogram corresponding to the feature amount f4. In a case where the feature amount f4 of the partial region PA1 of the first category corresponds to the class value cm1, the generation unit 154 sets the color of the frequency corresponding to the class value cm1 to the first color. In a case where the feature amount f4 of the partial region PB1 of the second category corresponds to the class value cm2, the generation unit 154 sets the color of the frequency corresponding to the class value cm2 to the second color. In a case where the feature amount f4 of the partial region PC1 of the third category corresponds to the class value cm3, the generation unit 154 sets the color of the frequency corresponding to the class value cm3 to the third color.
Although not illustrated, the generation unit 154 similarly generates histograms corresponding to the feature amounts f5 to f10. The generation unit 154 outputs the fourth auxiliary information to the display control unit 153 to request the display of the fourth auxiliary information. The display control unit 153 displays the fourth auxiliary information on the display unit 130.
Here, the display control device 153 may display all the 1st to 4th auxiliary information on the display unit 130, or may display only part of the auxiliary information. Furthermore, the user may operate the input unit 120 to designate the auxiliary information to be displayed. In the following description, in a case where the first to fourth auxiliary information is not particularly distinguished, they are simply referred to as auxiliary information.
In addition, after referring to the auxiliary information, the user may operate an input unit 130, refer to screen information Dis1 illustrated in
The description returns to
In a case where the image processing unit 155 performs the image process of classifying the partial region included in the pathological image according to the feature amount, the user operates the input unit 120 to set, as parameters, the feature amount (some feature amounts of the feature amounts f1 to f10) to be used at the time of classification, the importance of the feature amount, and the like.
In a case where the image processing unit 155 performs the image process of extracting a partial region having a specific feature amount from partial regions included in the pathological image, the user operates the input unit 120 to set, as parameters, the feature amount (some feature amounts of the feature amounts f1 to f10) to be used at the time of extraction, the threshold value for each feature amount at the time of extraction, and the like.
Here, an example of a processing result by the image processing unit 155 will be described.
The image processing unit 155 classifies each partial region included in the pathological image Ima1-1 into one of the first category, the second category, and the third category based on the parameter set to the user. The classification result is illustrated in a pathological image Ima1-2. In the pathological image Ima1-2, each partial region indicated by the first color is a partial region classified into the first category. Each partial region indicated by the second color is a partial region classified into the second category. Each partial region indicated by the third color is a partial region classified into the third category. The image processing unit 155 may output the pathological image Ima1-2 as the classification result to the display unit 130 to display it on the display unit 130.
The histogram h1-2 is a histogram corresponding to the feature amount f1. In the histogram h1-2, a distribution 41a is a distribution of the feature amounts of the partial regions classified into the first category. A distribution 42a is a distribution of the feature amounts of the partial regions classified into the second category. A distribution 43a is a distribution of the feature amounts of the partial regions classified into the third category.
The histogram h2-2 is a histogram corresponding to the feature amount f2. In the histogram h2-2, the distribution 41b is a distribution of the feature amounts of the partial regions classified into the first category. The distribution 42b is a distribution of the feature amounts of the partial regions classified into the second category. The distribution 43b is a distribution of the feature amounts of the partial regions classified into the third category.
The histogram h3-2 is a histogram corresponding to the feature amount f3. In the histogram h3-2, the distribution 41c is a distribution of the feature amounts of the partial regions classified into the first category. The distribution 42c is a distribution of the feature amounts of the partial regions classified into the second category. The distribution 43c is a distribution of the feature amounts of the partial regions classified into the third category.
The histogram h4-2 is a histogram corresponding to the feature amount f4. In the histogram h4-2, the distribution 41d is a distribution of the feature amounts of the partial regions classified into the first category. The distribution 42d is a distribution of the feature amounts of the partial regions classified into the second category. The distribution 43d is a distribution of the feature amounts of the partial regions classified into the third category.
Although not illustrated, the image processing unit 155 similarly generates histograms corresponding to the feature amounts f5 to f10. The image processing unit 155 may output the information about the histograms h1-2 to h4-2 illustrated in
The analysis unit 152 calculates a feature amount of each partial region (step S103). The display control unit 153 displays the pathological image indicating the partial region on the display unit 130 (step S104). The display control unit 153 receives designation of the partial region (step S105).
The generation unit 154 of the image processing device 100 generates auxiliary information (step S106). The display control unit 153 displays the auxiliary information on the display unit 130 (step S107).
In a case of receiving a change or an addition of the partial region to be designated (step S108, Yes), the image processing device 100 advances the process to step S105. On the other hand, in a case of not receiving the change or addition of the partial region to be designated (step S108, No), the image processing device 100 advances the process to step S109.
The image processing unit 155 of the image processing device 100 receives adjustment of parameters (step S109). The image processing unit 155 performs the classification or extraction process based on the adjusted parameters (step S110).
In a case where the re-adjustment of the parameter is received (step S111, Yes), the image processing device 100 advances the process to step S109. In a case where re-adjustment of the parameters is not received (step S111, No), the image processing device 100 ends the process.
The image processing device 100 may generate, as the auxiliary information, information capable of grasping the situation of a plurality of partial regions in the pathological image, such as the situation of the entire pathological image, and display the auxiliary information.
The analysis unit 152 of the image processing device 100 extracts partial regions from the ROI 40a and calculates a feature amount of each partial region in the same manner as the above processing. The generation unit 154 of the image processing device 100 generates auxiliary information 42a based on the feature amounts of respective partial regions of the ROI 40a and sets the auxiliary information in the screen information 45. For example, the auxiliary information 42a may be the third auxiliary information described with reference to
The user can grasp the features of the entire pathological image by referring to the screen information 45, and can use the features for parameter adjustment in a case where the image process is performed.
The image processing device 100 according to the present embodiment extracts a plurality of partial regions from a pathological image, and generates auxiliary information indicating a feature amount effective in a case of classifying or extracting a partial image with respect to a plurality of feature amounts calculated from the pathological image when designation of the partial region is received. In a case of receiving the setting of the parameter from the user who has referred to the auxiliary information, the image processing device 100 performs the image process on the pathological image using the received parameter. As a result, the feature amount obtained by quantifying the appearance feature of the morphology can be appropriately displayed by the auxiliary information, and adjustment of the parameter of the image process can be facilitated. For example, the “macroscopic and visible feature” of a specialist such as a pathologist can be easily associated with the “calculated quantitative feature”.
The image processing device 100 calculates a contribution rate when classifying the plurality of designated partial regions, and generates and displays, as auxiliary information, information in which the feature amount and the contribution rate are associated with each other. By referring to such auxiliary information, the user can easily grasp which feature amount should be emphasized to set the parameter in a case of classifying the plurality of partial regions for each category.
The image processing device 100 selects some feature amounts based on the magnitudes of the plurality of feature amounts calculated from the plurality of designated partial regions, and generates the selected feature amount as auxiliary information. By referring to such auxiliary information, the user can easily grasp the feature amount to be used in a case of extracting a partial region having the category same as that of the designated partial region.
The image processing device 100 performs segmentation on the pathological image and extracts a plurality of partial regions. As a result, the user can easily designate the region corresponding to the cell morphology included in the pathological image.
The image processing device 100 displays all the partial regions included in the pathological image, and receives selection of a plurality of partial regions among all the partial regions. As a result, the user can easily select the partial region to be used in the creation of the auxiliary information.
The image processing device 100 performs the factor analysis, prediction analysis, or the like to calculate a contribution rate. As a result, it is possible to calculate the feature amount effective in a case where the partial region is appropriately classified for each of the designated different categories.
The image processing device 100 generates a feature space corresponding to some feature amounts, and identifies a position in the feature amount space corresponding to the partial region designation of which was received based on the feature amount of the partial region designation of which was received. As a result, the user can easily grasp the position in the feature space with respect to the designated partial region.
The image processing device 100 identifies a feature amount having a high contribution rate and generates a feature space of the identified feature amount. As a result, the user can grasp the distribution of the designated partial regions in the feature space of the feature amount with a high contribution rate.
In a case where a plurality of ROIs is designated for the entire pathological image, the image processing device 100 generates auxiliary information based on the feature amount of the partial region included in each ROI. As a result, it is possible to grasp the features of the entire pathological image and to use the features for parameter adjustment when the image process is performed.
The image processing device according to each embodiment described above is implemented by, for example, a computer 1000 having a configuration as illustrated in
The CPU 1100 operates based on a program stored in the ROM 1300 or the HDD 1400, and controls each unit. For example, the CPU 1100 develops a program stored in the ROM 1300 or the HDD 1400 in the RAM 1200, and executes processing corresponding to various programs.
The ROM 1300 stores a boot program such as a basic input output system (BIOS) executed by the CPU 1100 when the computer 1000 is activated, a program depending on hardware of the computer 1000, and the like.
The HDD 1400 is a computer-readable recording medium that non-transiently records programs executed by the CPU 1100, data used by the programs, and the like. Specifically, the HDD 1400 is a recording medium that records an information processing program according to the present disclosure which is an example of program data 1450.
The communication interface 1500 is an interface for the computer 1000 to be connected to an external network 1550 (for example, the Internet). For example, the CPU 1100 receives data from another device or transmits data generated by the CPU 1100 to another device via the communication interface 1500.
The input/output interface 1600 is an interface that connects an input/output device 1650 and the computer 1000. For example, the CPU 1100 receives data from an input device such as a keyboard and a mouse via the input/output interface 1600. In addition, the CPU 1100 transmits data to an output device such as a display, a speaker, or a printer via the input/output interface 1600. Furthermore, the input/output interface 1600 may function as a media interface that reads a program or the like recorded in a predetermined recording medium (medium). The medium is, for example, an optical recording medium such as a digital versatile disc (DVD) or a phase change rewritable disk (PD), a magneto-optical recording medium such as a magneto-optical disk (MO), a tape medium, a magnetic recording medium, a semiconductor memory, or the like.
Furthermore, the computer 1000 is connected to a millimeter wave radar or a camera module (corresponding to an image generation unit 107 or the like) via the input/output interface 1600.
For example, in a case where the computer 1000 functions as the image processing device 100 according to the embodiments, the CPU 1100 of the computer 1000 executes the image processing program loaded on the RAM 1200 to implement the functions of the acquisition unit 151, the analysis unit 152, the display control unit 153, the generation unit 154, the image processing unit 155, and the like. Further, the HDD 1400 stores an image processing program or the like according to the present disclosure. The CPU 1100 reads the program data 1450 from the HDD 1400 and executes the program data, but as another example, the program may be acquired from another device via the external network 1550.
The image processing device includes a generation unit and an image processing unit. In a case of receiving designation of a plurality of partial regions extracted from a pathological image and corresponding to a cell morphology, the generation unit generates auxiliary information indicating information about a feature amount effective when classifying or extracting the plurality of partial regions with respect to the plurality of feature amounts calculated from the image. In a case of receiving setting information about an adjustment item according to the auxiliary information, the image processing unit performs the image process on the image using the setting information. As a result, the feature amount obtained by quantifying the appearance feature of the morphology can be appropriately displayed by the auxiliary information, and adjustment of the parameter of the image process can be facilitated. For example, it is possible to easily associated the “feature based on knowledge of a specialist such as a pathologist” with the “calculated quantitative feature”.
The generation unit calculates a contribution rate when classifying the plurality of designated partial regions, and generates, as the auxiliary information, information in which the feature amount and the contribution rate are associated with each other. By referring to such auxiliary information, the user can easily grasp which feature amount should be emphasized to set the parameter in a case of classifying the plurality of partial regions for each category.
The generation unit selects some feature amounts based on the magnitudes of the plurality of feature amounts calculated from the plurality of designated partial regions, and generates information about the selected feature amounts as the auxiliary information. By referring to such auxiliary information, the user can easily grasp the feature amount to be used in a case of extracting a partial region having the category same as that of the designated partial region.
The image processing device performs segmentation on the image and extracts the plurality of partial regions. As a result, the user can easily designate the region corresponding to the cell morphology included in the pathological image.
The image processing device further includes a display control unit that displays all partial regions extracted by the analysis unit and receives designation of a plurality of partial regions among all the partial regions. The display control unit further displays the auxiliary information. As a result, the user can easily select the partial region to be used in the creation of the auxiliary information.
The generation unit performs the factor analysis or the prediction analysis to calculate the contribution rate. As a result, it is possible to calculate the feature amount effective in a case where the partial region is appropriately classified for each of the designated different categories.
The generation unit generates a feature space corresponding to some feature amounts, and identifies a position in the feature amount space corresponding to the partial region designation of which was received based on a feature amount of a partial region designation of which was received. As a result, the user can easily grasp the position in the feature space with respect to the designated partial region.
The generation unit identifies a feature amount having a high contribution rate and generates a feature space of the identified feature amount. As a result, the user can grasp the distribution of the designated partial regions in the feature space of the feature amount with a high contribution rate.
In a case where a plurality of regions is designated for the pathological image, the generation unit generates auxiliary information for each of the plurality of regions. As a result, it is possible to grasp the features of the entire pathological image and to use the features for parameter adjustment when the image process is performed.
Number | Date | Country | Kind |
---|---|---|---|
2020-110156 | Jun 2020 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2021/020921 | 6/2/2021 | WO |