This application is based upon and claims the benefit of priority of the prior Japanese Patent Application No. 2021-031192, filed Feb. 26, 2021, the entire contents of which are incorporated herein by reference.
The disclosure of this specification relates to a system, a method and a computer-readable medium for supporting annotation for an image.
Currently, in the field of cell culture, culture states are grasped non-invasively from images using trained models constructed through execution of supervised learning. Incidentally, construction of a trained model used for image processing requires a large amount of training data including tagged images (with correct labels). An operation of creating such training data is called annotation.
In annotation, each of a large amount of images are manually tagged by people. The amount of operation is enormous. A technology of reducing the operation load for annotation is required. A technique related to such a problem is described in Japanese Patent Laid-Open No. 2017-009314, for example. Japanese Patent Laid-Open No. 2017-009314 discloses user interfaces suitable for annotation.
An image annotation support system according to an aspect of the present invention includes: a processor and a memory, the processor being configured to perform the following steps: generating classification information that classifies a plurality of target regions constituting a target image, based on a feature represented in the target image, the target image being an image serving as a candidate to be annotated; and arranging, on a screen of a display device, a classified image that visualizes the classification information, in a manner capable of being compared with the target image. An image annotation support system according to another aspect of the present invention includes: a processor and a memory, the processor being configured to perform the following steps: generating classification information that classifies a plurality of target regions constituting a target image, based on a feature represented in the target image, the target image being an image serving as a candidate to be annotated; and arranging, on the screen of a display device, a plurality of the classified images in which a plurality of pieces of generated classification information are visualized and which correspond to a plurality of the target images respectively taken at times different from each other, based on the times of the plurality of target images being taken.
An image annotation support method according to an aspect of the present invention includes: generating classification information that classifies a plurality of target regions constituting a target image, based on a feature represented in the target image, the target image being an image serving as a candidate to be annotated; and arranging, on a screen of a display device, a classified image that visualizes the classification information, in a manner capable of being compared with the target image.
An image annotation support method according to another aspect of the present invention includes: generating classification information that classifies a plurality of target regions constituting a target image, based on a feature represented in the target image, the target image being an image serving as a candidate to be annotated; and arranging, on the screen of a display device, a plurality of the classified images in which a plurality of pieces of generated classification information are visualized and which correspond to a plurality of the target images respectively taken at times different from each other, based on the times of the plurality of target images being taken.
A non-transitory computer-readable medium storing an image annotation support program according to an aspect of the present invention, the program causing a computer to execute processes of: generating classification information that classifies a plurality of target regions constituting a target image, based on a feature represented in the target image, the target image being an image serving as a candidate to be annotated; and arranging, on a screen of a display device, a classified image that visualizes the classification information, in a manner capable of being compared with the target image.
A non-transitory computer-readable medium storing an image annotation support program according to another aspect of the present invention, the program causing a computer to execute processes of: generating classification information that classifies a plurality of target regions constituting a target image, based on a feature represented in the target image, the target image being an image serving as a candidate to be annotated; and arranging, on the screen of a display device, a plurality of the classified images in which a plurality of pieces of generated classification information are visualized and which correspond to a plurality of the target images respectively taken at times different from each other, based on the times of the plurality of target images being taken.
The present invention will be more apparent from the following detailed description when the accompanying drawings are referenced.
The technology described in Japanese Patent Laid-Open No. 2017-009314 is adopted, which allows operations of a user for annotation to be supported. As a result, the annotation efficiency can be expected to be improved. However, the technology described in Japanese Patent Laid-Open No. 2017-009314 supports the operations of the user, but does not support various types of determination required for the user for annotation. For example, determination about which region should be tagged remains to be up to the user. No technology of supporting such determination is described.
Hereinafter, embodiments of the present invention are described.
The system 200 shown in
The system 200 includes: one or more imaging devices 100 that image cultured cells contained in the vessel C from below the vessel C; and a control apparatus 130 that controls the imaging devices 100. It is only required that each of the imaging devices 100 and the control apparatus 130 can mutually exchange data. Consequently, each of the imaging devices 100 and the control apparatus 130 can be wiredly, communicably connected to each other, or wirelessly, communicably connected. The vessel C that contains the cultured cells is, for example, a flask. However, the vessel C is not limited to a flask, and may be another culture vessel, such as a dish, or a well plate.
To image the cultured cells without taking out of an incubator 120, the imaging device 100 is used in a state of being arranged in the incubator 120, for example. More specifically, as shown in
As shown in
As shown in
The stage 103 changes the relative position of the imaging unit 105 with respect to the vessel C. The stage 103 is movable in the X direction and the Y direction that are parallel to the transparent window 101 (mounting surface) and are orthogonal to each other. Note that the stage 103 may further move also in the Z direction (height direction) orthogonal to both the X direction and the Y direction.
Note that
As shown in
The diffuser panel 107 diffuses the light emitted from the light source 106. Although not specifically limited, the diffuser panel 107 is, for example, a frost-type diffuser panel on which irregularities are formed on the surface. Note that the diffuser panel 107 may be a surface-coated opal-type diffuser panel, or a diffuser panel of another type. Furthermore, masks 107a for limiting a diffusion light emission region may be formed on the diffuser panel 107. The light emitted from the diffuser panel 107 travels in various directions.
As shown in
The image pick-up element 109 is an optical sensor that converts the detected light into an electric signal. Although not specifically limited, for example, a CCD (charge-coupled device) image sensor, a CMOS (complementary MOS) image sensor or the like is used as the image pick-up element 109.
The imaging device 100 configured as described above adopts oblique illumination in order to visualize a specimen S (cultured cells), which is a phase object, in the vessel C. Specifically, the light emitted by the light source 106 is diffused by the diffuser panel 107, and is emitted to the outside of the housing 102. That is, the light source units 104 emit light, which is to travel in various directions, to the outside of the housing 102 without through the optical system 108. Subsequently, a part of light emitted to the outside of the housing 102 is, for example, reflected by the upper surface of the vessel C, and is deflected above the specimen S. Part of the light deflected above the specimen S is emitted to the specimen S, and passes through the specimen S and the transparent window 101, and enters the housing 102 accordingly. Part of the light having entered the housing 102 is condensed by the optical system 108, and an image of the specimen S is formed on the image pick-up element 109. Lastly, the imaging device 100 generates an image of the specimen S (cultured cells) on the basis of an electric signal output from the image pick-up element 109, and outputs the image to the control apparatus 130.
The control apparatus 130 is an apparatus that controls the imaging device 100. The control apparatus 130 transmits an imaging instruction to the imaging device 100 arranged in the incubator 120, and receives an image taken by the imaging device 100.
The control apparatus 130 is an image processing apparatus that processes the image taken by the imaging device 100. The control apparatus 130 generates classification information that classifies a region constituting the image into some classes on the basis of the feature of the image extracted from this image.
Furthermore, the control apparatus 130 is a display control apparatus that visualizes and displays the classification information. The control apparatus 130 visualizes the classification information and arranges the information on a screen, in response to a request issued by a user. Hereinafter, the image where the classification information is visualized is described as a classified image.
Note that the control apparatus 130 may be what includes one or more processors, and one or more non-transitory computer-readable media, and is a typical computer, for example. More specifically, as shown in
Each of the one or more processors 131 is, for example, hardware that includes a CPU (central processing unit), a GPU (graphics processing unit), and a DSP (digital signal processor), and executes a program 132a stored in the one or more storage devices 132, thereby performing programmed processes. The programed processes include, for example, a classification process of generating the classification information, and a display control process of arranging classified images on the screen. That is, the processor 131 is an example of a classification unit of the system 200, and is an example of a control unit of the system 200. The one or more processors 131 may include an ASIC (application specific integrated circuit), and an FPGA (field-programmable gate array).
Each of the one or more storage devices 132 includes, for example, one or more freely selected semiconductor memories, and may further include one or more other storage devices. The semiconductor memories include, for example, volatile memories such as RAMs (random access memories), and nonvolatile memories such as ROMs (read only memories), programmable ROMs and flash memories. The RAMs may include, for example, DRAMs (dynamic random access memories), and SRAMs (static random access memories). The other storage devices may include, for example, magnetic storage devices that include magnetic disks, and optical storage devices that include optical disks.
Note that the one or more storage devices 132 are non-transitory computer-readable media, and are examples of storage units of the system 200. At least one of the storage devices 132 stores trained data 132b, which is to be used for generating the classified image.
The input device 133 is a device that the user directly operates, and is, for example, a keyboard, a mouse, a touch panel, etc. The display device 134 may be, for example, a liquid crystal display, an organic EL display, a CRT (cathode ray tube) display, etc. The display may internally include a touch panel. The communication device 135 may be a wired communication module, or a wireless communication module.
Note that the configuration shown in
When the processes shown in
Next, the system 200 classifies individual regions of the target image on the basis of the feature represented in the target image (step S2). Here, the control apparatus 130 obtains the target image output from the imaging device 100. Furthermore, in the control apparatus 130, the processors 131, which are an example of the classification unit, use the trained data 132b stored in the storage devices 132 to thereby generate classification information that classifies a plurality of target regions constituting the target image on the basis of the feature represented in the target image.
Note that each of the target regions is, for example, one or more pixels included in the target image. One pixel may constitute one target region. A plurality of pixels (e.g., 3×3 pixels) may constitute one target region. The number of divisions of the target regions may be freely defined by the user or the system. For example, the case where the number of pixels of 3*3 is adopted as the division unit has been exemplified. However, the vertical and lateral numbers of pixels are not necessarily the same. The classification information includes at least pieces of class information that correspond to the respective target regions. The class information is information that indicates the classes into which the corresponding target regions are classified.
When the classification information is generated, the system 200 arranges classified image that visualizes the classification information on a screen of the display device 134 in a manner capable of being compared with the target image (step S3). Here, in the control apparatus 130, the processors 131, which are an example of a control unit, assign each region a color or a pattern associated with the class into which the corresponding region is classified, thereby generating the classified image that visualizes the classification information. Furthermore, the processors 131 arrange the classified image on the annotation screen in a manner capable of being compared with the target image.
Note that
After the classified image is arranged on the screen, the system 200 accepts an input by the user, and annotates the image (step S4). First, the user herein determines the target image displayed on the screen as an image to be annotated. Input for annotating the target image is performed. In response to this, the processors 131 annotate the target image according to the input by the user, and generates training data.
By the system 200 performing the processes shown in
First, the classified image is displayed, which allows the user to roughly determine a portion to be annotated. As described above, this is because the image visualizes classification information (including the class information) where the classified image is classified based on the feature of the target image.
More specifically, the classification information can be assumed to be generated by the classification unit grasping, as a pattern, cells taken in the target image and classifying the difference in texture in the target image caused by the difference in the size of cells and contrast. Accordingly, it can be regarded that cells and the like having forms similar in appearance are taken in regions classified into the same class among the regions of the target image. Conversely, it can be regarded that cells and the like having forms different in appearance are taken in regions classified into different classes among the regions of the target image. For example, it is assumed that for an image of iPS cells in culture, undifferentiated regions and differentiated regions are annotated. The undifferentiated regions and the differentiated regions have different forms, and are classified into different classes accordingly. Meanwhile, the undifferentiated regions (or differentiated regions) have similar forms, and are classified into the same class accordingly. Consequently, when a target image where undifferentiated regions and differentiated regions are mixed is taken, the user can expect that a region associated with any class in a classified image generated from the target image is an undifferentiated region, and they can roughly discriminate the undifferentiated region from the classified image on the basis of the assumption. Likewise, also for the differentiated region, the differentiated region can be roughly determined from the classified image.
Furthermore, the classified image is arranged in a manner capable of comparison with the target image, which allows the user to verify the target image with respect to the portion roughly determined using the classified image, and determine whether the portion is a portion to be annotated or not. Note that
As described above, the system 200 can display the classified image, which can provide an operator performing annotation with information for supporting various types of determination. Accordingly, the efficiency and reliability of annotation can be expected to be improved. The system 200 provides the user with the information required for annotation, which allows even a person having no specialized knowledge to assign an annotation correctly. Accordingly, the restriction on selection of the operator can be alleviated. Consequently, securement of the operator is facilitated while the operation efficiency is improved. Accordingly, in comparison with the conventional art, a large amount of images can be efficiently annotated.
Note that in this embodiment, the example is described where the control apparatus 130 generate the classified image used for controlling the imaging device 100 and for supporting annotation. Alternatively, these may be performed separately by different apparatuses. For example, the control apparatus 130 may control the imaging device 100, and an apparatus different from the control apparatus 130 may generate the classified image. The apparatus different from the control apparatus 130 may display the classified image in response to a request by the user. The control apparatus 130 controls the imaging device 100 and generates the classified image. The apparatus different from the control apparatus 130 may display the classified image in response to a request by the user. The apparatus different from the control apparatus 130 is, for example, any of client terminals, which include a tablet, a smartphone, and a computer. These client terminals may be configured to be capable of communicate with the control apparatus 130. Likewise with the control apparatus 130, the client terminal may be what includes one or more processors, and one or more non-transitory computer-readable media.
The classification process in step S2 of
The learning process shown in
The learning process shown in
As shown in
In step S12, the same images for learning are set for both input data and training data on the neural network, and the neural network is repetitively trained. That is, the neural network is trained using the multiple images for learning. An auto encoder that extracts features of images through the trained neural network is constructed.
After the control apparatus 130 performs the processes in steps S11 and S12 for all the images for learning, this apparatus determines whether to finish learning or not (step S13). Here, the determination may be made based on whether the loss function is equal to or less than a reference value or not. When it is determined that the loss function is equal to or less than the reference value, the control apparatus 130 finishes the learning (YES in step S13), and outputs, as trained data, the auto encoder (hereinafter, described as trained data A for feature extraction) constructed in step S12 (step S14). Note that the trained data A may be the entire or a part of neural network (e.g., information on an input layer to the fifth intermediate layer) constituting the auto encoder.
In step S20 of learning the normalization method, as shown in
In step S22, the control apparatus 130 inputs the images for learning into the model M1, and outputs, as intermediate-layer images, data on the intermediate layer where the images for learning are compressed and the feature is extracted, for example. Note that in a case where the data on the intermediate layer is not a two-dimensional array (image format) but is a one-dimensional array, the array is transformed into two-dimensional array (image format), which is output as the intermediate-layer images. For example, in a case where all the first to fifth intermediate layers of the model M1 are used, in step S22 total 368 (=16+32+64+128+128)-channel of intermediate-layer images are generated.
After the intermediate-layer images are generated in step S22, the control apparatus 130 applies a statistical process to the intermediate-layer images on a channel-by-channel basis, and generates a statistical image (step S23). Here, the statistical process is an image filtering process that performs a statistical operation. The statistical operation is an operation, such as of average or variance, for example. In other words, the statistical process is a spatial filtering process that uses information on an intended pixel and pixels adjacent thereto.
Note that the statistical operation performed in the statistical process in step S23 may be one type (e.g., only averaging), or two or more types (e.g., average and variance calculations). In a case where two or more types of statistical operations are performed, the control apparatus 130 may output the operation results respectively as different channels.
The control apparatus 130 repeats the processes in steps S21 and S23 for all the images for learning (NO step S24). After all the images for learning are processed (YES in step S24), the control apparatus 130 learns normalization on a channel-by-channel basis (step S25), and outputs the trained data (step S26).
In step S25, the control apparatus 130 extracts the maximum pixel value and the minimum pixel value as normalization parameters, from among the statistical images generated in step S23, on a channel-by-channel basis. In step S26, the control apparatus 130 outputs the extracted normalization parameters (hereinafter, described as trained data B for normalization) as the trained data.
Lastly, the control apparatus 130 normalizes the statistical images on a channel-by-channel basis using the trained data B for normalization, generates feature images, and outputs the generated feature images (step S27). Here, the control apparatus 130 transforms the statistical images using the trained data B, and generates feature images having pixel values ranging from zero to one.
In step S30 of learning the information amount reduction method, as shown in
In step S32, the control apparatus 130 may reduce the image size by thinning out pixels from the feature image on the basis of a predetermined rule. The control apparatus 130 may reduce the image to that of a freely selected size, using interpolation.
The control apparatus 130 repeats the processes in steps S31 and S32 for all the feature images (NO step S33). After all the feature images are processed (YES in step S33), the control apparatus 130 applies principal component analysis to the feature vector of the second feature image having a reduced image size (step S34). Note that the principal component analysis may be performed after transforming the second feature image into a one-dimensional array as required.
The control apparatus 130 outputs, as trained data, a transformation matrix (hereinafter, described as trained data C for reducing the amount of information) for outputting principal components identified by the principal component analysis in step S34 in response to the input of the second feature image (step S35). Preferably, the number of principal components output by the transformation matrix is appropriately set depending on the complexity of the classification target. Preferably, the more complex the classification target is, the more the principal components remain.
In step S40 of learning the classification method, as shown in
The control apparatus 130 repeats the processes in steps S41 and S42 for all the second feature images (NO step S43). After all the second feature images are processed (YES in step S43), the control apparatus 130 applies cluster analysis to the feature vector of the third feature image having a reduced number of dimensions (step S44).
For example, the K-means method or the like can be adopted as the cluster analysis applied in step S44. The K-means method is preferable in that the trained classification rule can be stored. However, the cluster analysis method is not limited to the K-means method. Note that the cluster analysis may be performed after transforming the third feature image into a one-dimensional array as required.
The control apparatus 130 outputs, as trained data, the classification rule (hereinafter described as trained data D for classification) created by the cluster analysis in step S44 (step S45). Preferably, the number of clusters (the number of classes) as a classification result is appropriately set depending on the classification target before the cluster analysis in step S44 is performed.
The system 200 stores, in the storage devices 132, the trained data 132b (trained data items A to D) obtained by the learning process described above. When the user performs annotation, the system 200 supports the user in annotation using these trained data items.
Next, referring to
First, the control apparatus 130 obtains the target image (step S101). Here, for example, in step S1, the control apparatus 130 obtains, as the target image, the image obtained by the imaging device 100 taking cells cultured in the vessel C. As shown in
After the control apparatus 130 obtains the target image, this apparatus generates an intermediate-layer image using the trained data A (step S102). Here, the processors 131 input the intermediate-layer image, as the trained data A, into the auto encoder stored in the storage devices 132, and obtain the intermediate-layer data as the intermediate-layer image. As shown in
After the intermediate-layer images are generated, the control apparatus 130 applies a statistical process to the intermediate-layer images on a channel-by-channel basis, and generates a statistical image (step S103). Here, the processors 131 apply a spatial filtering process to the intermediate-layer images, thereby generating a statistical image. Note that the spatial filtering process performed in this step is the similar to the process in step S23 of
After the statistical image is generated, the control apparatus 130 normalizes the statistical images on a channel-by-channel basis using the trained data B, and generates feature images (step S104). Here, the processors 131 transform the statistical images into feature images having pixel values ranging from zero to one, using the normalization parameters stored as the trained data B in the storage devices 132. Note that the normalization parameters are stored on a channel-by-channel basis. Accordingly, the normalization parameters associated with the channel of the statistical image in step S104 are used. As shown in
After the feature image is generated, the control apparatus 130 reduces the image size of the feature image, and generates the second feature image (step S105). Similar to the process in step S32 of
After the second feature images are generated, the control apparatus 130 generate third feature images having a limitedly adjusted number of dimensions of the feature vectors, from the second feature images, using trained data C (step S106). Here, the processors 131 transform the second feature images into the third feature images made up of the principal components of the second feature images, using the transformation matrix stored as the trained data C in the storage devices 132. As shown in
After the third feature images are generated, the control apparatus 130 classify the feature vectors (three dimensional in this example) of the third feature images using the trained data D, and generates an index image (classification information) (step S107). Here, the processors 131 cluster the feature vectors corresponding to multiple pixels constituting the third feature images, using the classification rule stored in the storage devices 132 as the trained data D. The processors 131 further two-dimensionally arrange the index (class information) indicated in the classified classes, and generates the index image. As shown in
The system 200 stores, in the storage devices 132, the index image obtained by the classification process described above. When the user performs annotation, the system 200 generates the classified image obtained by visualizing the index image according to the indices as shown in
As shown in
As shown in
As shown in
To achieve such a transmissivity depending on the reliability, as shown in
Since the classification information includes the score information, the control apparatus 130 may determine the transmissivities of the multiple classified regions constituting the classified image, on the basis of the score information of the corresponding target region among the target regions. More specifically, when the stochastic score is low, the transmissivity is set to be high, which allows the state of actual cells to be more finely verified on the target image with respect to regions having low classification reliabilities. On the other hand, for the regions on which classification having high reliabilities have been made, annotation can be performed on the basis mainly of information on the classified image. Note that
The example where the classification information is visualized using the difference in color has been described above. However, various methods can be adopted as the color assigning method. For example, the index may be assigned to H (hue) in the HLS color space. Alternatively, a color that the people can easily identify may be selected, and the selected color may be assigned the index. Further alternatively, without any limitation to the example described above, any color may be assigned an index in conformity with the purpose.
The classification information may be visualized using the shading of color instead of the difference in color. Furthermore, the classification information may be visualized using the pattern. For example, as shown in
The example where the classification information is visualized by filling the regions with specific colors or patterns has been described above. However, it is only required that the classification information can be divided such that the regions with different classifications can be discriminated from each other. For example, visualization may be achieved by drawing the contours of the regions, instead of filling of the regions. For example, as shown in
Note that the system according to this embodiment supports selection of images to be annotated, by performing the processes shown in
After the processes shown in
After the time lapse imaging is finished, the control apparatus 130 obtains multiple pieces of classification information obtained by the time lapse imaging (step S210). Note that the multiple pieces of classification information obtained here correspond to multiple target images taken at times different from each other.
Subsequently, the control apparatus 130 generates multiple classified images where the obtained pieces of classification information are visualized. A shown in
An arrow assigned “ME” indicates that media have been replaced. An arrow assigned “P” indicates passage has been performed. Information about the time when the medium replacement or passage is performed is recorded by the operator pressing an operation button provided on the imaging device 100. The control apparatus 130 may display content indicating timings of medium replacement and passage, together with the classified images, on the basis of the time information recorded in the imaging device 100. The medium replacement and passage largely change the culture environment. Accordingly, it is beneficial, for selection of an annotation image, to provide information so as to demonstrate when these are performed.
After multiple classified images are arranged on the screen, the system accepts an input by the user, and annotates the image (step S230). Note that the process of step S230 is similar to the process in step S4 of
By performing the processes shown in
To achieve a high learning efficiency with a relatively small number of images in machine learning, it is preferable that the training data created by annotation include a wide variety of images. The system according to this embodiment arranges, on the screen, multiple classified images on the basis of the imaging times. Consequently, the user can easily grasp the change occurred in the target image, by comparing the classified images with each other. For example, the user can select a wide variety of images only by selecting images according to a reference of prioritizing a largely changed image. Accordingly, the user can easily and appropriately select the image to be annotated. In particular, the user can grasp the change caused in the image at a glance, not by displaying the target image in an arranged manner, but by displaying the classified images in an arranged manner instead. Consequently, the image to be annotated can be selected in a short time period, which increases the operation efficiency of annotation. By combining the annotation support methods according to the first embodiment, consistent supports are provided for selection of the image to be annotated to the operation of annotation. Accordingly, the user can assign annotations further efficiently.
Note that the description has been made with the example where after the time lapse imaging is finished, the image to be annotated is selected. Alternatively, the selection of the image to be annotated may be performed during the time lapse imaging period. The classified images created from the image having already obtained by the control apparatus 130 may be displayed in a time-series manner, which may allow the user to select the image to be annotated as needed.
As shown in
After the culture data is selected, the control apparatus 130 arranges a vessel image (image CT) simulating a culture vessel on the screen of the display device 134. Furthermore, the control apparatus 130 may arrange, on the vessel image, the classified images based on the target image obtained by taking the specimen in the vessel. The classified images may be arranged on the vessel image on the basis of the imaging position of the target image on the culture vessel.
As shown in
When the culture region (e.g., the well image W6) is selected on the vessel image, the control apparatus 130 arranges, on the screen, the classified images generated based on the target images taken in the culture regions according to the imaging times, as shown in
Furthermore, when a specific classified image (pasted image) is selected from among the classified images (pasted images) arranged in a time-series manner, the control apparatus 130 displays the selected classified image (the image Pc that is a pasted image) in an enlarged manner, as shown in
Subsequently, when a region intended to be observed specifically in detail in the classified image (the image Pc) is selected, the control apparatus 130 displays the classified image (the image Pc9 that is a pasted element images) including the selected region, in an enlarged manner as shown in
As described above, the system according to this embodiment can support selection of the images performed by the user until transition to the annotation screen, using the time-series display of the classified images. Accordingly, the user can select an appropriate image. Accordingly, they can avoid image reselection and the like, and efficiently perform annotation.
Note that in this embodiment, the example of displaying only the classified image until transition to the annotation screen has been described. Alternatively, for selecting the image to be annotated, the target image may be used in addition to the classified image. For example, similar to the first embodiment, the classified image and the target images may be displayed in a superimposed manner, or the classified images and the target images may be displayed in a switchable manner. That is, the multiple classified images may be arranged on the screen in a time-series order in a manner capable of comparison with the target images.
In this embodiment, the example of listing the classified images at the imaging times of taking images in a time-series order has been described. However, not all the classified images at the imaging times of taking the images are necessarily displayed. For example, images that the user intend not to display may be selected on the screen shown in
After the processes shown in
Subsequently, the control apparatus 130 selects multiple classified images to be arranged on the screen on the basis of comparison between the pieces of classification information obtained in step S320 (step S330). Here, the control apparatus 130 is only required to select largely changed images with priority. There is no limitation to a specific selection method. For example, as shown in
When the multiple classified images are selected, the control apparatus 130 arranges, on the screen, the selected classified images on the basis of the imaging times of the target images corresponding to the respective classified images (step S340). Note that the process in step S340 is similar to the process in step S220 of
After the selected classified images are arranged on the screen, the system accepts an input by the user, and annotates the image (step S350). Note that the process of step S350 is similar to the process in step S4 of
By performing the processes shown in
In this embodiment, the example of automatically selecting the classified images to be arranged on the screen on the basis of the magnitude of change is described. This selection is not necessarily performed by automatic selection by the system. For example, even without automatic selection by the system, the user can visually grasp the amount of change in each time period only if multiple classified images are displayed in a time-series order. Accordingly, as described above in the third embodiment, the user is only required to only annotate images determined to have a large amount of change. Images that do not serve as candidates of images to be annotated may be manually selected to be hidden. According to this embodiment, the rate of change between images is calculated. Accordingly, information on the rates of change may be displayed in proximity to the classified images displayed in a time-series order. The user may manually select the images to be hidden on the basis of the displayed information on the rates of change. Images to be annotated may be selected based on the information on the rate of change.
As described above, similar to the system according to the second embodiment, the system according to this embodiment concerned can support selection of the images performed by the user until transition to the annotation screen, using the time-series display of the classified images.
Subsequently, the control apparatus 130 selects transmissivities for classified images to be arranged on the screen on the basis of comparison between the pieces of classification information obtained in step S420 (step S430). Here, the control apparatus 130 is only required to select high transmissivities for largely changed images. There is no limitation to a specific selection method. For example, as shown in
When the transmissivities are selected, the control apparatus 130 arranges, on the screen, the selected classified images with the selected transmissivities, on the basis of the imaging times of the target images corresponding to the respective classified images (step S440). Here, the control apparatus 130 arranges the classified images with the selected transmissivities, in a manner of being superimposed on the corresponding target images. Accordingly, the largely changed target images is viewed translucently through the classified image, which can be used as reference for selection of the images to be annotated.
After multiple classified images are arranged on the screen, the system accepts an input by the user, and annotates the image (step S450). Note that the process of step S450 is similar to the process in step S4 of
As described above, similar to the system according to the second embodiment, the system according to this embodiment concerned can support selection of the images performed by the user until transition to the annotation screen, using the time-series display of the classified images.
After the processes shown in
Subsequently, when the user selects the target image to be annotated based on the classified images arranged on the screen, the control apparatus 130 annotates the target image selected by the user (step S540). Here, the control apparatus 130 may automatically annotate the target image selected by the user, on the basis of the classified image corresponding to the target image selected by the user. Specifically, the control apparatus 130 may annotate regions that belong to predetermined classes. For example, the control apparatus 130 may identify the regions belonging to the predetermined classes, on the basis of the classified images corresponding to the target images selected by the user, and annotate the identified regions among the target images selected by the user. Note that the control apparatus 130 may annotate the regions that are regions belonging to the predetermined class and have been subjected to highly reliable classification. After the automatic annotation, the control apparatus 130 causes the display device 134 to display the relationship between the automatically assigned annotations and the classes (step S550). Here, for example, the control apparatus 130 may cause the display device 134 to display an annotation screen shown in
Note that an image obtained by annotating the target image is used as the training data. The annotated image 1003 displayed on the annotation screen is not limited to an image obtained by annotating the target image. As shown in
The user verifies the annotated image 1003 and the like on the annotation screen, and determines whether the annotation has been appropriately performed. After the necessity of correcting the annotation is recognized, the user performs operations, such as of changing of the automatic annotation setting, and activation of the manual annotation setting, and instructs the control apparatus 130 to correct annotation. The control apparatus 130 accepts an input by the user, and corrects the annotation (step S560).
Hereinafter, an example is described where an input by the user for correcting the annotation is performed by a click operation using a mouse. Alternatively, the input by the user may be performed through a direct input with a finger or a stylus. The information (annotation information) input by the user through the input device 133 is accepted by an acceptance unit of the control apparatus 130. The acceptance unit may be, for example, processors in the control apparatus 130. When the accepted annotation information is output to the display device, the acceptance unit may appropriately transform the annotation information into a form suitable to output.
The user may change the number of grids as required. For example, as shown in
As described above, the system according to this embodiment also can exert advantageous effects similar to those of the system according to the second embodiment. The system according to this embodiment automatically annotate the image selected by the user, thereby allowing the load of the annotation operation on the user to be significantly reduced. Furthermore, by the user verifying the automatically assigned annotations and correcting only required points, a high level equivalent to that in the case of manually maintaining the reliability of annotation can be maintained. Accordingly, appropriate annotations can be efficiently assigned. By allowing the user to freely set the units of annotations, the efficiency and correctness of the annotation operation can be achieved at a higher level in a compatible manner.
The embodiments described above is a specific example in order to facilitate understanding of the invention. The present invention is not limited to these embodiments. Modified embodiments obtained by modifying the embodiments described above, and alternative embodiments that substitute the embodiments described above can be encompassed. That is, each embodiment may allow the configuration elements to be modified in a range without departing from the spirit and scope. By appropriately combining multiple configuration elements disclosed in one or more embodiments, new embodiments can be implemented. Some configuration elements may be removed from the configuration elements indicated in each embodiment, or some configuration elements may be added to the configuration elements indicated in the embodiments. The order of the processing procedures shown in each embodiment may be replaced as long as there is no contradiction. That is, the annotation support system, method and computer-readable medium according to the present invention may be variously modified and changed in a range without departing from the description of the claims.
In the embodiments described above, the examples are described where by arranging the classified images in a time-series manner, the user can easily grasp the change occurring in the target image. However, the control apparatus 130 may sequentially display the classified images one by one in a time-series manner. As shown in
In the embodiments described above, the examples have been described where the imaging device 100 that obtains the target image images the cultured cells in the incubator. However, there is no limitation on the imaging apparatus. For example, the apparatus may be an endoscope, a biological microscope, or an industrial microscope. The target image is not limited to the image of the cultured cells. For example, the image may be an image used for pathological diagnostic, or an image where industrial products have been taken. The imaging device and the control apparatus may be connected to each other via the Internet. As shown in
Number | Date | Country | Kind |
---|---|---|---|
2021-031192 | Feb 2021 | JP | national |