The present disclosure relates to a processing device, a processing program, a processing method, and a processing system for processing an image of a subject that is captured by a camera.
From the related art, it has been known that a medical doctor observes a change in the state of an oral cavity of a target, and for example, makes a diagnosis such as a viral cold. Here, it is known that a lymphoid follicle that appears in the deepest portion of a pharynx positioned in the oral cavity has a pattern specific to influenza. The lymphoid follicle having such a specific pattern is referred to as an influenza follicle, is a characteristic sign for the influenza, and appears in onset. However, such a pharynx portion is diagnosed by direct inspection of the medical doctor, and is not diagnosed by using an image.
On the other hand, an endoscope for capturing a deep portion in the body such as an abdominal cavity, which includes an imaging optical system on the head, has been known (JP 2009-189496 A). Since such an endoscope is affected by the contamination or the clouding of a capturing window, while cleaning liquid supply means supplies a cleaning liquid, vibration means vibrates at an intensity at which the contamination of a vibrator can be mixed with the cleaning liquid, and then, the vibration at an intensity at which the vibrator can be mixed with the cleaning liquid is switched to a vibration at an intensity weaker than the intensity described above, and the surface of a transparent member is dried, and thus, the influence of the contamination or the clouding of the capturing window is removed.
It is difficult to use a massive device such as an endoscope in the oral cavity such as a pharynx or a tonsil in which a medical opinion is determined routinely, and a diagnosis is made in a short period of time. In addition, the influence of the clouding on the capturing of the oral cavity (in particular, the pharynx portion) where the external air and the air in the body flow in and out at a high frequency due to aspiration has been extremely serious. Further, in the oral cavity, various regions with an active movement, such as teeth or a tongue, are close to each other, and it is difficult to stably capture a sharp image. Therefore, on the basis of the technology as described above, an object of the present disclosure is to provide a processing device, a processing program, a processing method, and a processing system that are suitable for processing an image obtained by capturing an oral cavity to be used in a diagnosis of the oral cavity.
According to one aspect of the present disclosure, a “processing device, including: an input/output interface connected to a camera capturing an image of a subject including at least an oral cavity such that communication is available and configured to receive input of the image captured by the camera; a memory configured to store the image received by the input/output interface, In addition to computer readable instructions; and a processor configured to execute the computer readable instructions stored in the memory to specify a predetermined region in the subject included in the image, on the basis of the image stored in the memory, and to process the image such that the specified predetermined region is identifiable” is provided.
According to one aspect of the present disclosure, a “processing device, including: an input/output interface connected to a camera capturing an image of a subject including at least an oral cavity such that communication is available and configured to receive input of the image captured by the camera; a memory configured to store the image received by the input/output interface, in addition to computer readable instructions; and a processor configured to execute the computer readable instructions stored in the memory to perform processing of removing clouding included in the image from the image stored in the memory” is provided.
According to one aspect of the present disclosure, a “processing device, including: an input/output interface connected to a camera capturing an image of a subject including at least an oral cavity such that communication is available and configured to receive input of the image captured by the camera; a memory configured to store the image received by the input/output interface, in addition to computer readable instructions; and a processor configured to execute the computer readable instructions stored in the memory to perform super-resolution processing with respect to the image for increasing a resolution of the image stored in the memory” is provided.
According to one aspect of the present disclosure, a “processing program allowing a computer, which includes: an input/output interface connected to a camera capturing an image of a subject including at least an oral cavity such that communication is available; and a memory configured to store the image, to function as a processor configured to receive input of the image captured by the camera, to store the image received by the input/output interface, to specify a predetermined region in the subject included in the image, on the basis of the image stored in the memory, and to process the image such that the specified predetermined region is identifiable” is provided.
According to one aspect of the present disclosure, a “processing method performed by executing computer readable instructions with a processor in a computer comprising an input/output interface connected to a camera capturing an image of a subject including at least an oral cavity such that communication is available; and a memory configured to store the image, in addition to the computer readable instructions, the method comprising the steps of: receiving input of the image captured by the camera; a step of storing the image received by the input/output interface; specifying a predetermined region in the subject included in the image, on the basis of the image stored in the memory; and processing the image such that the specified predetermined region is identifiable” is provided.
According to one aspect of the present disclosure, a “processing system, including: a capturing device configured to capture an image of a subject including at least an oral cavity; and the processing device described above” is provided.
According to the present disclosure, it is possible to provide the processing device, the processing program, the processing method, and the processing system that are suitable for processing the image obtained by capturing the oral cavity to be used in the diagnosis of the oral cavity.
Note that, the effect described above is merely an example for explanatory convenience, and is not limited. In addition to the effect described above or instead of the effect described above, any effect that is described in the present disclosure or an effect that is obvious for a person skilled in the art can also be obtained.
Various embodiments of the present disclosure will be described with reference to the attached drawings. Note that, the same reference numerals will be applied to common constituents in the drawings.
1. Outline of Processing System 1
A processing system 1 according to the present disclosure is used for obtaining a subject image by mainly capturing the inside of an oral cavity of a test subject. In particular, the processing system 1 is used for capturing the periphery of the back of a larynx of the oral cavity, specifically, a pharynx. Therefore, hereinafter, a case where the processing system 1 according to the present disclosure is used for capturing the pharynx will be mainly described. Here, the pharynx is an example of a capturing part, and naturally, the processing system 1 according to the present disclosure can also be preferably used in another part such as a tonsil insofar as the part is in the oral cavity.
As an example, the processing system 1 according to the present disclosure is used for specifying a region included in an image obtained by capturing the subject including a pharynx region in the oral cavity of the target, from the image. The image captured by the processing system 1, in which the region is specified, for example, is used by a medical doctor to determine the possibility of incidence of influenza. Here, the possibility of incidence of the influenza is determined by examining the pharynx region or a tonsil region of the target, or by determining the presence or absence of a medical opinion such as a follicle in the pharynx region. Therefore, as an example, the processing system 1 is capable of specifying the pharynx, the tonsil, or a region in which the follicle is expressed, in the obtained image. Note that, the determination of the possibility of incidence of the influenza is an example, and naturally, any determination can also be preferably used insofar as the determination is for a disease in which a difference occurs in the medical opinion of the oral cavity by incidence. Examples of such a disease include a streptococcal infection, an adenovirus infection, an EB virus infection, a mycoplasma infection, arteriosclerosis, and the like.
Note that, in the present disclosure, even though terms such as the “determination” and a “diagnosis” are used for the disease, the terms do not necessarily indicate definitive determination or a definitive diagnosis of the medical doctor. For example, by a capturing target oneself or a user other than the medical doctor using the processing system 1 of the present disclosure, naturally, the determination or the diagnosis can also be performed by a processing device 100 included in the processing system 1.
The captured subject image (the image of the pharynx 715) is transmitted to the processing device 100 connected to the capturing device 200 by wire such that communication is available. A processor of the processing device 100 that receives the subject image processes a program stored in a memory of the processing device 100, and thus, a region is specified on the basis of the subject image, and a result thereof is output to a display or the like.
Note that, as illustrated in
2. Configuration of Processing System 1
At least the head of the capturing device 200 is inserted into the oral cavity of the target, and the oral cavity, in particular, the pharynx is captured. Such specific capturing processing will be described below. The captured subject image is transmitted to the processing device 100 through a wired cable.
The processor 111 functions as a control unit controlling other constituents of the processing system 1, on the basis of a program stored in the memory 112. The processor 111 controls the driving of the camera 211 and the driving of the light source 212, on the basis of the program stored in the memory 112, stores the subject image received from the capturing device 200 in the memory 112, and processes the stored subject image. Specifically, the processor 111 executes processing of receiving instruction input of the user to the manipulation input interface 113 and of turning on the light source 212 to instruct the camera 211 to capture an image, processing for receiving the input of the subject image captured by the capturing device 200, processing for storing the subject image received by the input/output interface 115 in the memory 112, processing for specifying a predetermined region in the subject included in the subject image, on the basis of the subject image stored in the memory 112, processing for processing the subject image such that the specified predetermined region is identifiable, processing for sharpening the subject image, processing for outputting the subject image processed such that the specified specific region is identifiable to the display 114, and the like, on the basis of the program stored in the memory 112. The processor 111 mainly includes one or a plurality of CPUs, and may be suitably combined with a GPU or the like.
The memory 112 includes a RAM, a ROM, a non-volatile memory, an HDD, and the like, and functions as a storage unit. The memory 112 stores a command for various control of the processing system 1 according to this embodiment as a program. Specifically, the memory 112 stores a program for the processor 111 to execute the processing of receiving the instruction input of the user to the manipulation input interface 113 and of turning on the light source 212 to instruct the camera 211 to capture an image, the processing for receiving the input of the subject image captured by the capturing device 200, the processing for storing the subject image received by the input/output interface 115 in the memory 112, the processing for specifying the predetermined region in the subject included in the subject image, on the basis of the subject image stored in the memory 112, the processing for processing the subject image such that the specified predetermined region is identifiable, the processing for sharpening the subject image, the processing for outputting the subject image processed such that specified specific region is identifiable to the display 114, and the like. In addition, the memory 112 stores various information relevant to a target including the target 700 (target identification information and the like), or stores the subject image captured by the camera 211 of the capturing device 200, the subject image subjected to sharpening processing or region specifying processing, and the like in association with the target identification information, in addition to the program. In addition, in some cases, the memory 112 stores each learned model that is used in the sharpening processing or the region specifying processing. Note that, as the memory 112, storage media connected to the outside such that communication is available can be used, or a combination of such storage media can also be used.
The manipulation input interface 113 functions as a manipulation input unit receiving the instruction input of the user to the processing device 100 and the capturing device 200. As illustrated in
The display 114 functions as a display unit for displaying the subject image captured by the capturing device 200 or for outputting a result determined by the processor 111. The display includes a liquid crystal panel, but is not limited to the liquid crystal panel, and may include an organic EL display, a plasma display, or the like.
The input/output interface 115 functions as an input/output unit for receiving various commands relevant to capturing start of the capturing device 200 connected in a wired manner or a wireless manner, or the input/output of image data captured by the capturing device 200. Examples of the input/output interface 115 include various interfaces such as a wired communication connector such as a USB or a SCSI, a wireless communication transceiver such as Bluetooth (Registered Trademark) or an infrared ray, and various connection terminals for a printed circuit board or a flexible printed circuit board. For example, in the example of
The camera 211 functions as a capturing unit that is driven in accordance with an instruction from the processing device 100, and detects the reflected light reflected on the oral cavity that is the subject to generate the subject image. In order to detect the light, as an example, the camera 211 includes a CMOS image sensor, and a lens system and a driving system for attaining a desired function. The image sensor is not limited to the CMOS image sensor, and other sensors such as a CCD image sensor can also be used as the image sensor. Even though it is not particularly illustrated, the camera 211 may have an automatic focus function, and for example, it is preferable that a focal point is set on the front surface of the lens to correspond with the specified region. In addition, the camera 211 may have a zoom function, and it is preferable that the capturing is set at a suitable magnification in accordance with the size of the pharynx or the influenza follicle.
In this embodiment, the camera 211 is inserted into the oral cavity of the target, and is used for capturing the pharynx at the back of the oral cavity, and thus, a distance between the camera 211 and the subject is comparatively short. Therefore, the camera 211 has a field angle (2θ) in which a value calculated by [(Distance from Tip Portion of Camera 211 to Posterior Wall of Pharynx)*tan θ] is 20 mm or more in a vertical way and is 40 mm or more in a horizontal way. By using the camera having such a field angle, it is possible to perform the capturing in a wider range even in a case where the distance between the camera 211 and the subject is short. Therefore, as the camera 211, a general camera can be used, and a camera referred to as a so-called wide-angle camera or super-wide-angle camera can also be used.
In addition, in this embodiment, the influenza follicle captured by the camera 211 is formed in the pharynx in the oral cavity. In general, since the pharynx is formed inside in a depth direction, in a case where the depth of field is shallow, the focal point is misaligned between the anterior portion of the pharynx and the posterior portion of the pharynx, and it is difficult to obtain a suitable subject image used in the determination of the processing device 100 or the medical doctor. Therefore, the camera 211 has a depth of field of at least 20 mm or more, preferably 30 mm or more. By using the camera having such a depth of field, it is possible to obtain an in-focus subject image in any region between the anterior portion of the pharynx and the posterior portion of the pharynx.
The light source 212 functions as a light source unit that is driven in accordance with the instruction from the processing device 100 and is for emitting light to the oral cavity. The light source 212 includes one or a plurality of light sources. In this embodiment, the light source 212 includes one or a plurality of LEDs, and light having a predetermined frequency band is emitted in the direction of the oral cavity from each of the LEDs. In the light source 212, light having a desired band from an ultraviolet light band, a visible light band, and an infrared light band, or a combination thereof is used. In particular, by emitting light having a short wavelength band in the ultraviolet light band to the influenza follicle, a specific component of the influenza follicle reacts, and thus, it is possible to more reliably determine the possibility of a disease.
Note that, in this embodiment, a case has been described in which the processing device 100 and the capturing device 200 are connected by the wired cable such that communication is available. However, naturally, the present disclosure is not limited thereto, the processing device and the capturing device may be connected by wireless communication.
The main body 214 is formed to be columnar in the shape of a hollow cylinder with a perfectly circular sectional surface. A wall portion 224 thereof may contain any material insofar as light can be guided to the inside thereof, and as an example, a thermoplastic resin can be used. As the thermoplastic resin, a polyolefin-based resin such as a chain polyolefin-based resin (a polypropylene-based resin and the like) and a cyclic polyolefin-based resin (a norbornene-based resin and the like), a cellulose ester-based resin such as triacetyl cellulose and diacetyl cellulose, a polyester-based resin, a polycarbonate-based resin, a (meth)acrylic resin, a polystyrene-based resin, or a mixture or a copolymer thereof, and the like can be used. That is, the main body 214 functions as a light guide body for guiding light emitted from the light source in the oral cavity or in the direction of a diffusion plate.
Since the main body 214 is formed into a hollow shape, a housing space 223 is formed on the inner surface thereof by the wall portion 224. The camera 211 is housed in the housing space 223. Note that, it may be sufficient that the main body 214 is formed into the shape of a column including the housing space 223. Therefore, it is not necessary that the housing space 223 is in the shape of a cylinder with a perfectly circular sectional surface, and the sectional surface may be in the shape of an ellipse or a polygon. In addition, it is not necessary that the inside of the main body 214 is formed into a hollow shape.
Here, as an example, the length of the main body 214 is set in a positional relationship with the incisor teeth of the target.
The main body 214 has a distance D1 as a length from the head 221 to the base 220. The distance D1 may be a length of 100% or less of the distance d1 from the incisor teeth 711 to the soft palate 713, preferably a length of 80% or less. In general, in a case where foreign substances are inserted to the back of the larynx, a feeling of vomiting is caused, and in a case where the main body 214 is short, the distance between the camera 211 and the subject excessively increases. With the distance D1 described above, it is possible to prevent the feeling of vomiting, and it is possible to suitably retain a feeling of distance from the subject.
The head of a grip 213 is connected to the base 220 of the main body 214. The user grips the grip 213 and performs a manipulation such as the insertion and removal of the capturing device 200. The grip 213 is relatively narrowed toward the head side connected to the main body 214, in accordance with a shape when gripped with the palm of the user, and has a shape that is bent to swell toward the base side positioned on a side opposite to the main body 214. Note that, in this embodiment, the grip has a shape with a perfectly circular sectional surface, but it is not necessary to have a perfectly circular shape, and the shape may be an ellipse or a polygon.
Here, as an example, the width (a distance D2) of the main body 214 in a direction vertical to a direction connecting the head 221 and the base 220 of the main body 214 is set in a relationship with an opening width in an up-and-down direction of the mouth of the target. According to
The capturing device 200 is inserted at the width of the distance d2 together with the aid 300, and it is necessary for the user to capture the oral cavity while observing the oral cavity from a gap to which the capturing device is inserted. Accordingly, it is favorable not to hinder the visibility of the user in a state where the capturing device 200 is inserted. Therefore, the distance D2 of the main body 214 may be a width of 80% or less, preferably 60% or less of the distance d2 that is the opening width in the up-and-down direction of the mouth, or may be 3.2 cm or less, preferably 2.4 cm or less.
The grip 213 includes an engaging protrusion 217 for positioning the aid 300, and a positioning protrusion 218, in the vicinity of the base 220 of the main body 214. The engaging protrusion 217 engages with an engaging protrusion provided on the aid 300. In addition, the positioning protrusion 218 is inserted to an insert hole provided in the aid 300, and positions the capturing device 200 and the aid 300 to each other. Note that, in this embodiment, as the engaging protrusion 217 of the main body 214, a total of four engaging protrusions are arranged at an equal interval at a position in the vicinity of the base 220 of the main body 214 on the surface of the grip 213. In addition, one positioning protrusion 218 is arranged at a position in the vicinity of the base 220 of the main body 214 between the engaging protrusions 217 on the surface of the grip 213. However, the present disclosure is not limited thereto, and only one of the engaging protrusion 217 and the positioning protrusion 218 may be arranged. In addition, the number of engaging protrusions 217 or positioning protrusions 218 is not limited insofar as there are one or a plurality of engaging protrusions or positioning protrusions.
The grip 213 includes a capture button 215 for starting or ending the capturing of the subject image by receiving the manipulation input from the user. Therefore, the user is capable of instructing the start and the end of the capturing by using the capture button of the manipulation input interface 113 of the processing device 100, and is also capable of instructing the start and the end of the capturing by using the capture button 215.
A diffusion plate 219 is arranged on the head 221 of the main body 214, and diffuses light that is emitted from the light source 212 and passes through the main body 214 toward the inside of the oral cavity. The diffusion plate 219 has a shape corresponding to a sectional shape of a portion of the main body 214 that is configured to be capable of guiding light. In this embodiment, the main body 214 is formed into the shape of a hollow cylinder. Therefore, the sectional surface of the diffusion plate 219 is also formed into a hollow shape corresponding to the shape of the main body.
The camera 211 is used for detecting reflected light that is diffused from the diffusion plate 219, is emitted into the oral cavity, and is reflected on the subject, to generate the subject image. The camera 211 is arranged on the inner surface of the wall portion 224 of the main body 214, that is, in the housing space 223 formed inside the main body 214. Note that, in this embodiment, only one camera 211 is described, but the capturing device 200 may include a plurality of cameras. By generating the subject image using the plurality of cameras, the subject image includes information relevant to a steric shape of the influenza follicle. Accordingly, more accurately, it is possible to specify the region of the influenza follicle. In addition, in this embodiment, the camera 211 is arranged in the housing space 223 of the main body 214, and may be arranged on the head 221 of the main body 214 or in the main body 214 (may be inside the main body 214 or may be on the outer circumference of the main body 214).
Note that, the light sources 212-1 to 212-4 may be configured to be independently controlled. For example, by turning on a part of the light sources 212-1 to 212-4, the shade of the influenza follicle having a steric shape can be included in the subject image. Accordingly, it is possible to include the information relevant to the steric shape of the influenza follicle in the subject image, more accurately, it is possible to specify the region of the influenza follicle.
In addition, in this embodiment, the light sources 212-1 to 212-4 are arranged on the base 220 side of the main body 214, and may be arranged on the head 221 of the main body 214 or in the main body 214 (may be inside the main body 214 or may be on the outer circumference of the main body 214).
The diffusion plate 219 is used for preventing the light emitted from the light source 212 from lighting only a part of the oral cavity to generate homogeneous light. Therefore, as an example, a fine lens array is formed on the surface of the diffusion plate 219, and a lens-shaped diffusion plate having an arbitrary diffusion angle is used. In addition, as a modification, a diffusion plate that is capable of diffusing light by other methods, such as a diffusion plate that attains a light diffusion function by fine irregularities randomly arranged on the surface, may be used. Further, the diffusion plate 219 may be integrated with the main body 214. For example, the integration can be attained by a method of forming fine irregularities in the head portion of the main body 214.
Note that, in the capturing device, at least the head 221 of the main body 214 is inserted into the oral cavity. Therefore, the capturing of the subject is affected by the clouding of the head portion due to the aspiration of the target. Accordingly, in the processing device 100, the sharpening processing for removing the influence of the clouding can be implemented.
3. Processing Flow Executed in Processing Device 100
In a case where the processing flow is started, the processor 111 performs control such that a top screen is displayed on the display 114 (S111). Here,
Returning again to
Here,
The user inserts at least the head portion of the capturing device 200 into the oral cavity of the target, and the subject image of the vicinity of the pharynx that is the subject is displayed on the capturing standby screen in
The processing flow of
Note that, here, the subject image that is captured may be a still image in which one or a plurality of images are continuously captured each time when any capture button is pressed, or may be a moving image that is captured for a predetermined amount of time. In addition, even though it is not particularly illustrated, while a set of processing is executed in the processing device 100, the manipulation input of the user is received in the manipulation input interface 113, and the target to be the test subject is input in advance. Therefore, when the captured subject image is stored in the memory 112, the processor 111 stores the subject image in association with information such as the target identification information or a capturing date of the input target.
Here,
Returning again to
As an example of the clouding removal processing, the processor 111 executes processing such as the correction of a pixel value, and contrast and/or luminance of each pixel, and the application of a clouding removing filter with respect to the obtained subject image. In addition, by applying the obtained subject image to a learned clouding removal image model generated by learning in which a subject image including clouding is set to a clouding learning image, and a subject image not including clouding is set to an excellent learning image, it is also possible to implement the clouding removal processing.
Here, for example, the captured subject image may be enlargedly displayed in order to check a lesion region in detail. However, the subject image may be unsharp due to the degradation or the like of the obtained image, and thus, it may be difficult to sufficiently check the lesion region. Accordingly, the processor 111 performs the super-resolution processing with respect to the subject image subjected to the clouding removal processing (S214). The processor 111 stores the subject image subjected to the super-resolution processing once in the memory 112, in association with the target identification information.
Examples of the super-resolution processing include single-image super-resolution processing using a self correlation of an image, multiple-image super-resolution processing of estimating a fine positional misalignment between a plurality of images to perform interpolation between the pixels, and the like. In addition, by applying the obtained subject image to a learned super-resolution image model generated by learning in which a high-resolution image is set to a high-resolution learning image, and an image obtained by performing degradation processing with respect to the high-resolution learning image is set to a low-resolution learning image, it is also possible to implement the super-resolution processing.
Note that, in this embodiment, the clouding removal processing and the super-resolution processing the processing are performed in this order, but the processing order is not limited only thereto. For example, the order can also be reverse, and other processing can also be implemented by being suitably combined.
Next, the processor 111 performs the region specifying processing with respect to the subject image subjected to the super-resolution processing (S215). Then, the processor 111 stores each of the subject image before the region specifying processing that is stored in the memory 112 and the subject image that is subjected to the region specifying processing and is labeled such that the specified region is identifiable in the memory 112, in association with the target identification information.
Examples of the region specifying processing include labeling processing by the medical doctor or the like through the manipulation input interface 113, processing of retaining a feature amount database stored for each region specified in advance and of labeling on the basis of a feature amount in the obtained subject image, and the like. In addition, by applying the obtained subject image to a learned region specifying image model generated by learning in which the obtained subject image is set as a learning image, and position information of each region that is obtained by performing the region specifying processing with respect to the learning image is set to position information for learning, it is also possible to implement the region specifying processing. Note that, in this embodiment, a case is described in which the region of the pharynx and the follicle is mainly specified, but the same processing is performed even in a case where the tonsil or other regions are specified.
Next, the processor 111 performs processing for outputting each of the subject images subjected to the processing described above to the display 114 (S216). Note that, the display on the display 114 is an example of output processing, and other output methods are also available. As another example of the output, it is also possible to output to an external display device or a printer connected in a wired/wireless manner, or it is also possible to transmit to other terminal devices or a server device connected in a wired/wireless manner.
Here,
Note that, as also illustrated in
According to
As illustrated in
Returning again to
Here,
Note that, as also illustrated in
According to
As also illustrated in
Note that, in a case where the selection manipulation input with respect to the region specifying icon 42 is received, the subject image labeled such that the specified region is identifiable is displayed on the display 114. Since the screen is the same as that in
4. Processing Flow relevant to Generation of Learned Image Model
According to
In a case where each of the clouding learning image and the excellent learning image is obtained, a step of performing machine learning of a clouding removal pattern is executed (S313). As an example of the machine learning, a set of the clouding learning image and the excellent learning image are applied to a neural network configured in combination with neurons, and the learning is repeated while adjusting a parameter of each of the neurons such that the output of the neural network is the same as the excellent learning image. Then, a step of acquiring the learned clouding removal image model (for example, the neural network and the parameter) is executed (S314). The acquired learned clouding removal model may be stored in the memory 112 of the processing device 100 or the other processing device connected to the processing device 100 through wireless communication. Then, by executing the stored learned clouding removal model in the processing device 100 or the other processing device, the clouding removal processing (S213) of the subject image illustrated in
Note that, in order to obtain the subject image after the clouding removal processing in
According to
In a case where each of the high-resolution learning image and the low-resolution learning image is obtained, a step of performing machine learning of a super-resolution pattern by using both the learning images is executed (S414). As an example of the machine learning, a set of the high-resolution learning image and the low-resolution learning image are applied to a neural network configured in combination with neurons, and the learning is repeated while adjusting a parameter of each of the neurons such that the output of the neural network is the same as the high-resolution learning image. Then, a step of acquiring the learned super-resolution image model (for example, the neural network and the parameter) is executed (S415). The acquired learned super-resolution image model may be stored in the memory 112 of the processing device 100 or the other processing device connected to the processing device 100 through wireless communication. Then, by executing the stored learned super-resolution image model in the processing device 100 or the other processing device, the super-resolution processing (S214) of the subject image illustrated in
Note that, in order to obtain the subject image after the super-resolution processing in
According to
In a case where each of the learning image and the position information for learning is obtained, a step of performing machine learning of a region specifying pattern using the learning image and the position information is executed (S514). As an example of the machine learning, a set of the learning image and the position information for learning are applied to a neural network configured in combination with neurons, and the learning is repeated while adjusting a parameter of each of the neurons such that the output of the neural network is the same as the position information for learning. Then, a step of acquiring the learned region specifying image model (for example, the neural network and the parameter) is executed (S515). The acquired learned region specifying image model may be stored in the memory 112 of the processing device 100 or the other processing device connected to the processing device 100 through wireless communication. Then, by executing the stored learned region specifying image model in the processing device 100 or the other processing device, the region specifying processing (S215) of the subject image illustrated in
Note that, in order to obtain the subject image after the region specifying processing in
As described above, in this embodiment, it is possible to provide the processing device, the processing program, the processing method, and the processing system that are suitable for processing the image obtained by capturing the oral cavity to be used in the diagnosis of the oral cavity. In addition, the region such as the pharynx or the tonsil is specified from the image obtained by capturing the oral cavity, and is output to be identifiable, which is preferable for the diagnosis of the medical doctor. In addition, in the capturing of the oral cavity that is easily affected by the expired air, or the movement of the teeth, the tongue, or the like, it is possible to improve the visibility by performing the clouding removal processing or the super-resolution processing.
In the embodiment described above, a case has been described in which the processing device 100 and the capturing device 200 are connected by the wired cable such that communication is available. However, the present disclosure is not limited thereto, and the camera or the light source can also be integrally mounted on the flexible printed circuit board or the printed circuit board in the processing device 100.
Examples of the processing device 100 include a camera-equipped terminal device such as a smart phone, a mobile phone terminal, a tablet terminal, a PDA, a digital camera terminal, a portable game machine, and a laptop personal computer terminal. In this case, the camera 116 is arranged outside the incisor teeth (outside the body) without being inserted to the vicinity of the pharynx in the oral cavity, and captures the oral cavity.
In addition, in one embodiment described above, a case has been described in which the processing device 100 and the capturing device 200 are connected through the wired cable such that communication is available. However, the present disclosure is not limited thereto, and the processing device and the capturing device may be connected by wireless communication. In this case, for example, the subject image captured by the capturing device 200 is transmitted to a server device installed remotely, and the server device is allowed to function as the processing device, and thus, the subject image can be output to the display or the like of the other processing device.
Note that, in such an embodiment, the configuration, the processing, and the procedure are the same as those in one embodiment described in
It is also possible to configure the system by suitably combining or replacing each of the constituents described in each of the embodiments.
The processing and the procedure described in this specification can be attained not only by those explicitly described in the embodiment but also by software, hardware, or a combination thereof. Specifically, the processing and the procedure described in this specification are attained by implementing logic corresponding to the processing in a medium such as an integrated circuit, a volatile memory, a non-volatile memory, a magnetic disk, and an optical storage. In addition, the processing and the procedure described in this specification can be executed in various computers including the processing device and the server device by being implemented as a computer program.
Even in a case where it is described that the processing and the procedure described in this specification are executed by a single device, single software, a single component, or a single module, such processing or procedure can be executed by a plurality of devices, a plurality of software pieces, a plurality of components, and/or a plurality of modules. In addition, even in a case where it is described that various information described in this specification is stored in a single memory or a single storage unit, such information can be dispersedly stored in a plurality of memories provided in a single device or a plurality of memories dispersedly arranged in a plurality of devices. Further, the constituents of the software and the hardware described in this specification can be attained by integrating the constituents to fewer constituents or by disintegrating the constituents to more constituents.
This application is a U.S. National Phase Application under 35 U.S.C. 371 of International Application No. PCT/JP2020/026479, filed on Jul. 6, 2020. The entire disclosure of the above application is expressly incorporated by reference herein.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2020/026479 | 7/6/2020 | WO |