The present invention relates to a medical image processing apparatus, a medical image processing method, and a program, and in particular, to techniques of a medical image processing apparatus, a medical image processing method, and a program for assisting a user who observes a medical image.
A general endoscope apparatus emits illumination light from a distal end of an insertion part of an endoscope, and captures an image of an observation target with a camera to acquire a medical image. The captured medical image is displayed on a monitor, and a user observes the medical image displayed on the monitor and performs an examination.
In recent years, by using a recognizer that is trained through machine learning, it has become possible to recognize a medical image with high accuracy (A. Krizhevsky, I. Sutskever, and G. Hinton. ImageNet classification with deep convolutional neural networks. In NIPS, 2012). Also in an endoscope apparatus, it is considered to automatically recognize a specific scene by using a recognizer that is trained through machine learning and to notify a user of the recognized scene.
For example, JP2020-146202A describes a technique of providing, in accordance with an operation of an endoscope operator, a notification indicating information of interest in a region of interest included in a medical image acquired by an endoscope apparatus.
Here, if a user (doctor) wants to observe a specific examination scene, in some cases, it is insufficient to be notified only that the specific examination scene is recognized. If an insertion part of an endoscope apparatus is near the specific examination scene, the user can cause the endoscope apparatus to recognize the specific examination scene relatively immediately. On the other hand, if the insertion part of the endoscope apparatus is away from the specific examination scene, the user has to adjust the imaging position, angle, distance, and the like without any assistance until the endoscope apparatus recognizes the specific examination scene, and in some cases, it takes time until the specific examination scene is recognized.
The present invention has been made in view of such circumstances, and an object thereof is to provide a medical image processing apparatus, a medical image processing method, and a program by which a user can efficiently observe a medical image.
A medical image processing apparatus according to one aspect of the present invention for achieving the above object is a medical image processing apparatus including a processor configured to perform: medical image acquisition processing of sequentially acquiring time-series medical images; first scene recognition processing of recognizing at least one first scene from one medical image of the medical images; second scene recognition processing of recognizing a second scene from the one medical image if the at least one first scene is recognized; first notification processing of providing a notification indicating that the at least one first scene is recognized; and second notification processing of providing a notification indicating that the second scene is recognized.
According to this aspect, the first scene is recognized from the medical image, and the user is notified that the first scene is recognized, the second scene is recognized from the medical image, and the user is notified that the second scene is recognized. Accordingly, since the user is notified that the first scene and the second scene are recognized, the user can grasp where a camera (e.g., an insertion part of an endoscope) that photographs the medical image is located, and can observe the medical image more efficiently.
Preferably, the at least one first scene contains the second scene.
According to this aspect, after being notified that the first scene is recognized, the user can expect the second scene to be recognized, and can observe the medical image more efficiently.
Preferably, the medical image processing apparatus includes a second scene recognizer configured to perform the second scene recognition processing for each of the at least one first scene. The first scene recognition processing recognizes two or more first scenes of the at least one first scene, and in accordance with the two or more first scenes recognized in the first scene recognition processing, the second scene recognizer is selected to recognize the second scene.
Preferably, after the second scene is determined to be recognized in the second scene recognition processing, the first notification processing is not performed.
According to this aspect, once the second scene is recognized and the user is notified, the first notification processing is not performed, which can prevent a number of notifications from being provided and can prevent the observation from being interrupted by repeated notifications.
Preferably, after an image of the second scene is captured, the first notification processing is not performed.
According to this aspect, after the image of the second scene is captured, the first notification processing is not performed, which can prevent a number of notifications from being provided and can prevent the observation from being interrupted by repeated notification.
Preferably, after the second scene is determined to be recognized, the second scene recognition processing is not performed.
According to this aspect, after the second scene is recognized, the recognition processing of the second scene is not performed, and calculation resources can be efficiently used. In addition, this can prevent the observation from being interrupted by repeated notifications as a result of repeated recognition of the same second scene.
Preferably, after an image of the second scene is captured, the second scene recognition processing is not performed.
According to this aspect, after the image of the second scene is captured, the recognition processing of the second scene is not performed, and calculation resources can be efficiently used. In addition, this can prevent the observation from being interrupted by repeated notifications as a result of repeated recognition of the same second scene.
Preferably, the second notification processing continuously provides a notification indicating that the second scene is recognized.
According to this aspect, if there are a plurality of sites to be observed, it is possible to assist the user to comprehensively observe the sites.
Preferably, the first notification processing provides a notification by an indication on a screen, and the second notification processing provides a notification by sound.
Preferably, the indication on the screen is a sample image of the at least one first scene.
Preferably, the first scene recognition processing and the second scene recognition processing are performed by using a Convolutional Neutral Network.
Preferably, the first scene recognition processing recognizes the at least one first scene, based on a classification score.
Preferably, the second scene recognition processing recognizes the second scene, based on a degree of similarity.
Preferably, the at least one first scene and the second scene are scenes in which an image of a site inside a stomach is captured.
A medical image processing method according to another aspect of the present invention is a medical image processing method using a medical image processing apparatus including a processor configured to perform: a medical image acquisition step of sequentially acquiring time-series medical images; a first scene recognition step of recognizing a first scene from one medical image of the medical images; a second scene recognition step of recognizing a second scene from the one medical image if the first scene is recognized; a first notification step of providing a notification indicating that the first scene is recognized; and a second notification step of providing a notification indicating that the second scene is recognized.
A program according to another aspect of the present invention is a program for causing a medical image processing apparatus including a processor to execute a medical image processing method. The program causes the processor to perform: a medical image acquisition step of sequentially acquiring time-series medical images; a first scene recognition step of recognizing a first scene from one medical image of the medical images; a second scene recognition step of recognizing a second scene from the one medical image if the first scene is recognized; a first notification step of providing a notification indicating that the first scene is recognized; and a second notification step of providing a notification indicating that the second scene is recognized.
According to the present invention, the first scene is recognized from the medical image, the user is notified that the first scene is recognized, the second scene is recognized from the medical image, and the user is notified that the second scene is recognized. Accordingly, it is possible to grasp where the camera that captures the medical image is located, and to observe the medical image more efficiently.
Hereinafter, preferred embodiments of a medical image processing apparatus, a medical image processing method, and a program according to the present invention will be described with reference to the accompanying drawings.
As illustrated in
The endoscope 10 captures time-series medical images including a subject image and is, for example, a lower or upper digestive tract endoscope. The endoscope 10 has an insertion part 20, a handheld operating unit 21, and a universal cord 22. The insertion part is to be inserted into a subject (e.g., a stomach) and has a distal end and a proximal end. The handheld operating unit 21 is provided continuously from the proximal end side of the insertion part 20 and is held by a doctor, who is a surgeon, to perform various operations. The universal cord 22 is provided continuously from the handheld operating unit 21.
The entirety of the insertion part 20 is formed to have a small diameter and an elongated shape. The insertion part 20 is constituted by continuously providing, in order from the proximal end side to the distal end side thereof, a soft part 25, a bending part 26, and a tip part 27. The soft part 25 has flexibility. The bending part 26 can be bent by an operation of the handheld operating unit 21. In the tip part 27, an imaging optical system (objective lens), an imaging element 28, and the like, which are not illustrated, are incorporated.
The imaging element 28 is an imaging element of a complementary metal oxide semiconductor (CMOS) type or a charge coupled device (CCD) type. Image light of a site to be observed is incident on an imaging surface of the imaging element 28 through an observation window and the objective lens. The observation window, which is not illustrated, is open on a distal end surface of the tip part 27, and the objective lens, which is not illustrated, is disposed behind the observation window. The imaging element 28 captures the image light of the site to be observed, which is incident on the imaging surface (converts the image light into an electric signal) and outputs an image signal. That is, the imaging element 28 sequentially captures medical images. Note that the medical images are acquired as a moving image 38 and a still image 39, which will be described later.
The handheld operating unit 21 is provided with various operating members operated by a doctor (user). Specifically, the handheld operating unit 21 is provided with two types of bending operation knobs 29, an air/water supply button 30, and a suction button 31. The bending operation knobs 29 are used for a bending operation of the bending part 26. The air/water supply button 30 is for air supply/water supply operations. The suction button 31 is for a suction operation. The handheld operating unit 21 is further provided with a still image capturing instruction unit 32 and a treatment tool introduction port 33. The still image capturing instruction unit 32 is for issuing an instruction for capturing the still image 39 of the site to be observed. The treatment tool introduction port 33 is for inserting a treatment tool (not illustrated) into a treatment tool insertion path (not illustrated) penetrating through the insertion part 20.
The universal cord 22 is a connection cord for connecting the endoscope 10 to the light source apparatus 11. The universal cord 22 contains a light guide 35, a signal cable 36, and a fluid tube (not illustrated). The light guide 35, the signal cable 36, and the fluid tube penetrate through the insertion part 20. In addition, an end portion of the universal cord 22 is provided with a connector 37a and a connector 37b. The connector 37a is to be connected to the light source apparatus 11. The connector 37b branches off from the connector 37a and is to be connected to the endoscope processor apparatus 12.
By the connector 37a being connected to the light source apparatus 11, the light guide 35 and the fluid tube (not illustrated) are inserted into the light source apparatus 11. Thus, through the light guide 35 and the fluid tube (not illustrated), necessary illumination light, water, and gas are supplied from the light source apparatus 11 to the endoscope 10. As a result, the site to be observed is irradiated with the illumination light from an illumination window (not illustrated) on the distal end surface of the tip part 27. In accordance with a pressing operation on the above-described air/water supply button 30, the gas or water is injected from an air/water supply nozzle (not illustrated) on the distal end surface of the tip part 27 to the observation window (not illustrated) on the distal end surface.
By the connector 37b being connected to the endoscope processor apparatus 12, the signal cable 36 is electrically connected to the endoscope processor apparatus 12. Thus, through the signal cable 36, an image signal of the site to be observed is output from the imaging element 28 of the endoscope 10 to the endoscope processor apparatus 12, and also, a control signal is output from the endoscope processor apparatus 12 to the endoscope 10.
The light source apparatus 11 supplies the illumination light through the connector 37a to the light guide 35 of the endoscope 10. As the illumination light, light in various wavelength ranges in accordance with an observation purpose, such as white light (light in a white wavelength range or light in a plurality of wavelength ranges), light in one or more specific wavelength ranges, or a combination thereof is selected.
The endoscope processor apparatus 12 controls operations of the endoscope 10 through the connector 37b and the signal cable 36. In addition, based on the image signal acquired from the imaging element 28 of the endoscope 10 through the connector 37b and the signal cable 36, the endoscope processor apparatus 12 generates an image (also referred to as “moving image 38”) formed of time-series frame images 38a including the subject image. Furthermore, if the still image capturing instruction unit 32 is operated in the handheld operating unit 21 of the endoscope 10, concurrently with the generation of the moving image 38, the endoscope processor apparatus 12 acquires one frame image 38a in the moving image 38 as the still image 39 in accordance with the timing of an imaging instruction.
The moving image 38 and the still image 39 are medical images obtained by capturing images of the inside of the subject, that is, a living body. In addition, if the moving image 38 and the still image 39 are images obtained with the above-described light in the specific wavelength range (special light), both are special-light images. In addition, the endoscope processor apparatus 12 outputs the generated moving image 38 and the still image 39 to each of the display apparatus 13 and the medical image processing apparatus 14.
Note that the endoscope processor apparatus 12 may generate (acquire) the special-light image having information on the above-described specific wavelength range, based on a usual-light image obtained with the above-described white light. In this case, the endoscope processor apparatus 12 functions as a special-light image acquisition unit. Then, the endoscope processor apparatus 12 obtains a signal in the specific wavelength range by performing calculation based on red, green, and blue (RGB) color information or cyan, magenta, and yellow (CMY) color information included in the usual-light image.
Based on, for example, at least one of the usual-light image obtained with the above-described white light or the special-light image obtained with the above-described light in the specific wavelength range (special light), the endoscope processor apparatus 12 may generate a feature quantity image such as a known oxygen saturation image. In this case, the endoscope processor apparatus 12 functions as a feature quantity image generating unit. Note that each of the moving image 38 and the still image 39 including the usual-light image, the special-light image, and the feature quantity image is a medical image obtained by converting results of imaging or measuring of a human body into an image for the purpose of image diagnosis or examination.
The display apparatus 13 is connected to the endoscope processor apparatus 12 and functions as a display unit that displays the moving image 38 and the still image 39 input from the endoscope processor apparatus 12. A doctor (user) operates the insertion part 20 back and forth, for example, while viewing the moving image 38 displayed on the display apparatus 13, and, if a lesion or the like is found at the site to be observed, the doctor (user) operates the still image capturing instruction unit 32 to capture a still image of the site to be observed and give treatment such as diagnosis or biopsy. Note that the moving image 38 and the still image 39 are similarly displayed on the display 16 connected to the medical image processing apparatus 14, which will be described later. In addition, if the moving image 38 and the still image 39 are displayed on the display 16, a notification indication, which will be described later, is also provided together. Accordingly, a user preferably performs diagnosis or the like by viewing what is displayed on the display 16.
The medical image processing apparatus 14 is constituted by a medical image acquisition unit 40, a central processing unit (CPU) 41, a first scene recognition unit 42, a second scene recognition unit 43, a first notification unit 44, a second notification unit 45, a display control unit 46, an audio control unit 47, and a memory 48. Processing of each unit is implemented by one or more processors. Herein, the processor may be constituted by the CPU 41 or may be constituted by one or more CPUs that are not illustrated.
The CPU 41 operates based on various programs including an operating system and a medical image processing program according to the present invention that are stored in the memory 48, generally controls the medical image acquisition unit 40, the first scene recognition unit 42, the second scene recognition unit 43, the first notification unit 44, the second notification unit 45, the display control unit 46, and the audio control unit 47, and functions as some of these units.
The medical image acquisition unit 40 performs medical image acquisition processing and sequentially acquires time-series medical images. The medical image acquisition unit 40 acquires, from the endoscope processor apparatus 12 (
The first scene recognition unit 42 performs first scene recognition processing. The first scene herein refers to a scene in a wider range than a second scene described below, and the first scene contains the second scene. For example, if the inside of a stomach is examined with the endoscope apparatus, the first scene is the cardia, the pylorus, the stomach corner, the fundus, the body of the stomach, the pyloric antrum, the lesser curvature, the greater curvature, and the rest. Thus, the first scene can be a scene of each region of an examination target.
The first scene recognition unit 42 recognizes the first scene from an input medical image by various methods. For example, the first scene recognition unit 42 is constituted by a recognizer constituted by a Convolutional Neural Network or the like. The recognizer of the first scene recognition unit 42 learns an image (medical image) in order to recognize the first scene in advance, and recognizes the first scene by using a trained parameter.
The second scene recognition unit 43 performs second scene recognition processing. The medical image recognized as the first scene by the first scene recognition unit 42 is input to the second scene recognition unit 43. The second scene herein refers to a scene that is suitable for observation or diagnosis in the first scene and is a scene in a narrower range than the first scene. For example, the first scene recognition unit 42 recognizes, as the first scene, the cardia inside the stomach, and the second scene recognition unit 43 recognizes, as the second scene, a medical image having a scene that is suitable for observation and in which the cardia is at the center of the image. For example, the first scene recognition unit 42 recognizes, as the first scene, a case where the medical image is blurred due to a movement of the camera or the like or a case where the medical image is dark due to a shielding object, but the second scene recognition unit 43 recognizes, as the second scene, only a case where the medical image is captured at appropriate brightness without blur and shake. The recognizer of the second scene recognition unit 43 learns images (medical images) in advance in order to recognize the second scene, and recognizes the second scene by using a trained parameter.
The first scene recognition unit 42 and the second scene recognition unit 43 may recognize the scenes by determining the input medical image, based on classification or a degree of similarity. If the first scene recognition unit 42 and the second scene recognition unit 43 recognize the scenes by classifying the medical image, the technology described in a literature (B. Zhou, A. Lapedriza, J. Xiao, A. Torralba, and A. Oliva. Learning deep features for scene recognition using places database. In Neural Information Processing Systems (NIPS), pages 487-495, 2014. 1, 4, 6, 8) can be used. In addition, if the first scene recognition unit 42 and the second scene recognition unit 43 recognize the scenes, based on the degree of similarity of a feature quantity of the medical image, the technology described in a literature (FaceNet: A Unified Embedding for Face Recognition and Clustering https://arxiv.org/abs/1503.03832)) can be used.
The first notification unit 44 performs first notification processing and notifies the user that the first scene is recognized. The first notification unit 44 notifies the user that the first scene is recognized, by various methods. For example, the first notification unit 44 provides a notification indicating that the first scene is recognized, on the display 16 via the display control unit 46. Specifically, in a model diagram of an organ drawn in a sub-region of the display 16, the first notification unit 44 displays a region (notification indication) corresponding to the recognized first scene by coloring, flashing, or illuminating the region to notify the user that the first scene is recognized. While the first scene is recognized, the first notification unit 44 may continuously provide the notification indicating that the first scene is recognized, may end providing the notification after providing the notification for a certain period, or may gradually end providing the notification (e.g., the color gradually disappears). Note that although an example in which the first notification unit 44 provides a notification by providing a notification indication on the display 16 has been described above, the notification manner is not limited to this. For example, the first notification unit 44 may provide a notification by using a speaker 17 via the audio control unit 47. In this case, the speaker 17 outputs a notification sound to notify the user that the first scene is recognized.
The second notification unit 45 performs second notification processing and provides a notification indicating that the second scene is recognized. The second notification unit 45 notifies the user that the second scene is recognized, by various methods. For example, the second notification unit 45 provides a notification indicating that the second scene is recognized, on the display 16 via the display control unit 46. Specifically, the second notification unit 45 displays a diagram of the organ drawn in the sub-region of the display 16 by coloring a local region of the diagram. Specifically, in a model diagram of the organ drawn in the sub-region of the display 16, the second notification unit 45 provides a circle (notification indication) in a region corresponding to the recognized second scene and displays the circle in color, or flashes or illuminates the circle, to notify the user that the second scene is recognized. While the second scene is recognized, the second notification unit 45 may continuously provide the notification indicating that the second scene is recognized, may end the notification after providing the notification for a certain period, or may gradually end the notification (e.g., the color gradually disappears). Note that although an example in which the second notification unit 45 provides a notification by providing a notification indication on the display 16 has been described above, the notification manner is not limited to this. For example, the second notification unit 45 may provide a notification by using the speaker 17 via the audio control unit 47. In this case, the speaker 17 outputs a notification sound to notify the user that the second scene is recognized.
The first notification unit 44 and the second notification unit 45 may provide notifications independently of each other, or the first notification unit 44 and the second notification unit 45 may provide notifications in association with each other. If the first notification unit 44 and the second notification unit 45 provide notifications in association with each other, while one of the notifications is being provided, the other of the notifications may be refrained from being provided. In addition, the first notification unit 44 and the second notification unit 45 may provide notifications in different notification manners. For example, the first notification unit 44 may provide a notification by an indication on a screen on the display 16, and the second notification unit 45 may provide a notification by sound output from the speaker 17. In addition, the second notification unit 45 may provide a notification by an indication on the screen on the display 16, and the first notification unit 44 may provide a notification by sound output from the speaker 17.
The display control unit 46 causes the first notification unit 44 or the second notification unit 45 to display the notification indication on the display 16. Specifically, under control of the first notification unit 44, the display control unit 46 causes the display 16 to display a notification indication for providing a notification indicating that the first scene is recognized. In addition, under control of the second notification unit 45, the display control unit 46 causes the display 16 to display a notification indication for providing a notification indicating that the second scene is recognized. In addition, the display control unit 46 generates image data to be displayed, based on the medical images (the moving image 38) acquired by the medical image acquisition unit 40 and outputs the image data to the display 16. Thus, the user is notified that the first scene and the second scene are recognized while observing the medical image.
The audio control unit 47 causes the first notification unit 44 or the second notification unit 45 to reproduce a notification sound from the speaker 17. Specifically, under control of the first notification unit 44, the audio control unit 47 causes the speaker 17 to reproduce a notification sound for providing a notification indicating that the first scene is recognized. In addition, under control of the second notification unit 45, the audio control unit 47 causes the speaker 17 to reproduce a notification sound for providing a notification indicating that the second scene is recognized.
The memory 48 includes a flash memory, a read-only memory (ROM), a random access memory (RAM), a hard disk device, and the like. The flash memory, the ROM, and the hard disk device are non-volatile memories that store an operating system, various programs such as the medical image processing program according to the present invention, the still image 39 that is captured, and the like. In addition, the RAM is a volatile memory from which data can be read and on which data can be written at high speed and that functions as an area for temporarily storing various programs stored in the non-volatile memory and as a work area for the CPU 41.
Next, a specific configuration example of the first scene recognition unit 42 and the second scene recognition unit 43 will be described.
In this example, a case will be described in which seven sites inside the stomach are each observed, and a series of observations are performed in which an image of a representative scene among the respective sites is captured. Specifically, each of the seven sites inside the stomach is recognized as a first scene, and a representative scene, an image of which is to be captured, is recognized as a second scene.
The first scene recognition unit 42 is constituted by a first scene recognizer 42a, and the second scene recognition unit 43 is constituted by second scene recognizers 43a to 43g. The first scene recognizer 42a and the second scene recognizers 43a to 43g are trained models constituted by a convolutional neural network (CNN), which are subjected to machine learning in advance. For example, the first scene recognizer 42a is subjected to learning using learning data constituted by medical images obtained by capturing images of the seven sites inside the stomach so as to recognize respective scenes at the seven sites inside the stomach (see
The first scene recognizer 42a receives the moving image 38, and recognizes the first scene in each frame image 38a. For example, the first scene recognition unit 42 recognizes the first scene in the frame image 38a, based on a classification score. The first scene recognizer 42a outputs the classification score with respect to the input frame image 38a, and the first scene at a site with the highest classification score is recognized. Upon recognition of the first scene in the frame image 38a, the first scene recognizer 42a transmits the frame image 38a to any one of the second scene recognizers 43a to 43g corresponding to the recognized first scene. For example, upon recognition of the first scene of a second site inside the stomach from the input frame image 38a, the first scene recognizer 42a transmits the frame image 38a to the second scene recognizer 43b corresponding to the second site. Note that as long as no first scene is recognized in the frame image 38a, the first scene recognizer 42a does not transmit the frame image 38a to the second scene recognizers 43a to 43g.
The one of the second scene recognizers 43a to 43g receives the frame image 38a in which the first scene is recognized by the first scene recognizer 42a, and recognize the second scene. For example, the second scene recognizers 43a to 43g recognize the second scene in the frame image 38a, based on the degree of similarity. Specifically, the second scene recognizers 43a to 43g output the degrees of similarity with respect to the input frame image 38a, and recognize the second scene if the output degree of similarity is greater than or equal to a threshold value, and recognizes no second scene if the output degree of similarity is less than the threshold value.
The second scene recognizers 43a to 43g are provided to correspond to the respective seven sites inside the stomach. Specifically, if the first scene recognizer 42a recognizes the first scene of the first site, the frame image 38a recognized as being of the first site is input to the second scene recognizer 43a. In addition, if the first scene recognizer 42a recognizes the first scene of the second site, the frame image 38a recognized as being of the second site is input to the second scene recognizer 43b. In this manner, in accordance with the site recognized by the first scene recognizer 42a, the frame image 38a is input to the corresponding one of the second scene recognizers 43a to 43g.
In the above-described example, the first scene recognizer 42a recognizes a plurality of first scenes, and each of the second scene recognizers 43a to 43g recognizes the second scene in the corresponding one of the first scenes. Thus, with the trained models obtained through machine learning, the first scene recognizer 42a and the second scene recognizers 43a to 43g can be efficiently configured.
Next, specific examples of a first notification indicating that the first scene is recognized and a second notification indicating that the second scene is recognized will be described.
In addition,
In the case illustrated in
In the case illustrated in
In the case illustrated in
Next, an example of a display manner of the above-described model image 101 on the display 16 will be described.
As illustrated in
Next, a medical image processing method performed by using the medical image processing apparatus 14 will be described.
The medical image acquisition unit 40 receives a medical image (medical image acquisition step: step S101). Subsequently, the first scene recognition unit 42 recognizes a first scene from the received medical image (first scene recognition step: step S102). If the first scene recognition unit 42 recognizes no first scene, the medical image acquisition unit 40 determines whether there is a subsequent image in time series (step S106). If there is a medical image, the medical image acquisition unit 40 receives the medical image (step S101). If there is no medical image, the process ends.
On the other hand, if the first scene recognition unit 42 recognizes the first scene, the first notification unit 44 provides a notification indicating that the first scene is recognized (first notification step: step S103). Subsequently, the second scene recognition unit 43 recognizes a second scene from the medical image (second scene recognition step: step S104). If the second scene recognition unit 43 recognizes the second scene, the second notification unit 45 provides a notification indicating that the second scene is recognized (second notification step: step S105). Subsequently, the medical image acquisition unit 40 determines whether there is a subsequent image (step S106), and if there is a subsequent image, the subsequent medical image is acquired.
As described above, according to this embodiment, the first scene is recognized from the medical image, the user is notified that the first scene is recognized, the second scene is recognized from the medical image, and the user is notified that the second scene is recognized. Accordingly, since the user is notified that the first scene and the second scene are recognized, the user can observe the medical image more efficiently.
Next, a second embodiment will be described. In this embodiment, after the second scene is recognized, the second scene recognition unit 43 does not perform the recognition processing of the second scene in the same first scene. Accordingly, calculation resources can be efficiently used, and it is possible to prevent the observation from being interrupted by repeatedly performing the second notification processing as a result of repeated recognition of the same second scene.
The medical image acquisition unit 40 receives a medical image (step S201). Subsequently, the first scene recognition unit 42 recognizes a first scene from the received medical image (step S202). If the first scene recognition unit 42 recognizes no first scene, the medical image acquisition unit 40 determines whether there is a subsequent image (step S207). If there is a subsequent medical image, the medical image acquisition unit 40 receives the medical image (step S201). If there is no subsequent medical image, the process ends.
On the other hand, if the first scene recognition unit 42 recognizes the first scene, the first notification unit 44 provides a notification indicating that the first scene is recognized (step S203). Subsequently, the second scene recognition unit 43 determines whether a second scene in the recognized first scene has been recognized, based on past recognition records (step S204). Here, if there are a plurality of first scenes (e.g., the examples illustrated in
As described above, according to this embodiment, if the second scene has been recognized in the past, the second scene recognition unit 43 does not recognize the second scene. Accordingly, calculation resources can be efficiently used, and it is possible to prevent the observation from being interrupted by frequently performing the second notification processing as a result of repeated recognition of the same second scene.
Next, a third embodiment will be described. In this embodiment, the first notification unit 44 and the second notification unit 45 alternatively display a notification indication indicating that the first scene is recognized or a notification indication indicating that the second scene is recognized.
The medical image acquisition unit 40 receives a medical image (step S301). Subsequently, the first scene recognition unit 42 recognizes a first scene from the received medical image (step S302). If the first scene recognition unit 42 recognizes no first scene, the medical image acquisition unit 40 determines whether there is a subsequent image in time series (step S306). If there is a medical image, the medical image acquisition unit 40 receives the medical image (step S301). If there is no medical image, the process ends.
On the other hand, if the first scene recognition unit 42 recognizes the first scene, the second scene recognition unit 43 recognizes a second scene (step S303). If the second scene recognition unit 43 recognizes no second scene, the first notification unit 44 provides a notification indicating that the first scene is recognized (step S304).
If the second scene recognition unit 43 recognizes the second scene, the second notification unit 45 provides a notification indicating that the second scene is recognized (step S305).
After the first notification unit 44 provides the notification (step S304), or, after the second notification unit 45 provides the notification (step S305), the medical image acquisition unit 40 determines whether there is a subsequent image (step S306), and if there is a subsequent image, the subsequent medical image is acquired. Note that if the second scene is recognized or an image of the second scene is captured, the first notification unit 44 preferably does not provide a notification even if the corresponding first scene is recognized later. Accordingly, the observation can be prevented from being interrupted by repeated notifications.
As described above, according to this embodiment, the notification indication indicating that the first scene is recognized or the notification indication indicating that the second scene is recognized is alternatively provided, and the user can be explicitly notified. Note that although an example regarding the notification using notification indications has been described in the above example, the notification manner is not limited to this. The first notification unit 44 and the second notification unit 45 may alternatively provide a notification by using audio.
Next, a fourth embodiment will be described. In this embodiment, after the single second scene is recognized, the second scene recognition unit 43 does not perform the recognition processing of the second scene. In addition, in this embodiment, the first notification unit 44 and the second notification unit 45 alternatively display a notification indication indicating that the first scene is recognized or a notification indication indicating that the second scene is recognized.
The medical image acquisition unit 40 receives a medical image (step S401). Subsequently, the first scene recognition unit 42 recognizes a first scene from the received medical image (step S402). If the first scene recognition unit 42 recognizes no first scene, the medical image acquisition unit 40 determines whether there is a subsequent image (step S407). If there is a subsequent medical image, the medical image acquisition unit 40 receives the medical image (step S401). If there is no subsequent medical image, the process ends.
On the other hand, if the first scene recognition unit 42 recognizes the first scene, the second scene recognition unit 43 determines whether a second scene has been recognized (step S403). If the second scene has been recognized, the second scene recognition unit 43 does not recognize the second scene, and the first notification unit 44 provides a notification indicating that the first scene is recognized (step S406). If no second scene has been recognized, the second scene recognition unit 43 recognizes the second scene (step S404). If the second scene is recognized, the second notification unit 45 provides a notification indicating that the second scene is recognized (step S405). If no second scene is recognized, the first notification unit 44 provides a notification indicating that the first scene is recognized (step S406).
As described above, in this embodiment, after the second scene is recognized, the second scene recognition unit 43 does not perform the recognition processing of the second scene. In addition, in this embodiment, a notification indication indicating that the first scene is recognized or a notification indication indicating that the second scene is recognized is alternatively provided. Accordingly, calculation resources can be efficiently used, and the user can be explicitly notified.
Although the endoscope processor apparatus and the medical image processing apparatus are separately provided in the above embodiments, the endoscope processor apparatus and the medical image processing apparatus may be integrated. That is, the endoscope processor apparatus may be provided with the functions of the medical image processing apparatus.
In addition, the measured examination time or treatment time is stored in the memory within the medical image processing apparatus in association with a diagnosis report or the like, but is not limited to this and may also be stored in an external memory (storage unit) connected to the medical image processing apparatus.
Furthermore, the medical images are not limited to endoscopic images captured by an endoscope and may be, for example, time-series images acquired by another modality such as an ultrasound diagnostic apparatus.
In addition, a hardware configuration that performs various controls of the medical image processing apparatus according to the above embodiments is any of the following various processors. Various processors include a central processing unit (CPU), which is a general-purpose processor that executes software (program) and functions as various control units, a programmable logic device (PLD), which is a processor in which the circuit configuration is changeable after manufacture, such as field programmable gate array (FPGA), a dedicated electric circuit, which is a processor having a circuit configuration that is specially designed to execute specific processing, such as an application specific integrated circuit (ASIC), and the like.
One control unit may be constituted by one of these various processors, or may be constituted by two or more processors of the same type or different types (e.g., a combination of a plurality of FPGAs or a combination of a CPU and an FPGA). In addition, a plurality of control units may be constituted by one processor. As examples for constituting a plurality of control units by one processor, firstly, there is a form in which one or more CPUs and software are combined to constitute one processor, and this processor functions as a plurality of control units, as typified by a computer such as a client or a server. Secondly, there is a form of using a processor that implements the functions of the entire system including a plurality of control units by using one integrated circuit (IC) chip, as typified by a system on chip (SoC) or the like. In this manner, various control units are constituted by one or more of the above various processors in terms of hardware configuration.
Furthermore, the present invention includes a medical image processing program to be installed in a computer to cause the computer to function as the medical image processing apparatus according to the present invention, and a non-volatile storage medium on which the medical image processing program is recorded.
Although examples in the present invention have been described above, the present invention is not limited to the above-described embodiments, and it is needless to say that various modifications can be made without departing from the gist of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
2021-034207 | Mar 2021 | JP | national |
The present application is a Continuation of PCT International Application No. PCT/JP2022/008168 filed on Feb. 28, 2022 claiming priority under 35 U.S.C § 119(a) to Japanese Patent Application No. 2021-034207 filed on Mar. 4, 2021. Each of the above applications is hereby expressly incorporated by reference, in its entirety, into the present application.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2022/008168 | Feb 2022 | US |
Child | 18459439 | US |