The present invention relates to an image processing device, an endoscope apparatus, an image processing method, an information storage device, and the like.
When observing tissue using an endoscope apparatus, and making a diagnosis, whether or not an early lesion has occurred is generally determined by observing tissue as to the presence or absence of minute concavities and convexities, or determining a difference in color (e.g., reddening or discoloration). When using an industrial endoscope apparatus instead of a medical endoscope apparatus, it is useful to observe the object (i.e., the surface of the object in a narrow sense) as to the presence or absence of a concave-convex structure or the like. In this case, it is possible to detect whether or not a crack has occurred in the inner side of a pipe that is difficult to directly observe with the naked eye, for example. It is also normally useful to detect the presence or absence of a concave-convex structure or the like of the object from the processing target image when using an image processing device other than an endoscope apparatus.
An endoscope apparatus may be designed to perform an image enhancement process that allows the user to easily observe the structure of the object and a difference in color. For example, JP-A-2003-88498 discloses a method that enhances the structure of the object by image processing. JP-A-2005-342234 discloses a method that utilizes a color enhancement process that allows the user to determine a lesion area.
According to one aspect of the invention, there is provided an endoscope apparatus comprising:
an image acquisition section that acquires a captured image that includes an image of an object;
a distance information acquisition section that acquires distance information based on a distance from an imaging section to the object when the imaging section captured the captured image;
a known characteristic information acquisition section that acquires known characteristic information, the known characteristic information being information that represents known characteristics relating to the object; and
an enhancement processing section that performs an enhancement process that corresponds to the distance information based on the known characteristic information,
the enhancement processing section determining an observation state with respect to the object, performing a first enhancement process as the enhancement process on a concave-convex part of the object that agrees with the characteristics represented by the known characteristic information when the enhancement processing section has determined that the observation state is a screening observation state, and performing a second enhancement process as the enhancement process on the concave-convex part of the object that agrees with the characteristics represented by the known characteristic information when the enhancement processing section has determined that the observation state is a zoom observation state.
According to another aspect of the invention, there is provided an image processing method comprising:
acquiring a captured image that includes an image of an object;
acquiring known characteristic information, the known characteristic information being information that represents known characteristics relating to a structure of the object; and
determining an observation state with respect to the object, performing a first enhancement process as an enhancement process on a concave-convex part of the object that agrees with the characteristics represented by the known characteristic information when it has been determined that the observation state is a screening observation state, and performing a second enhancement process as the enhancement process on the concave-convex part of the object that agrees with the characteristics represented by the known characteristic information when it has been determined that the observation state is a zoom observation state.
According to another aspect of the invention, there is provided a non-transitory information storage device storing an image processing program that causes a computer to perform steps of:
acquiring a captured image that includes an image of an object;
acquiring known characteristic information, the known characteristic information being information that represents known characteristics relating to a structure of the object; and
determining an observation state with respect to the object, performing a first enhancement process as an enhancement process on a concave-convex part of the object that agrees with the characteristics represented by the known characteristic information when it has been determined that the observation state is a screening observation state, and performing a second enhancement process as the enhancement process on the concave-convex part of the object that agrees with the characteristics represented by the known characteristic information when it has been determined that the observation state is a zoom observation state.
Exemplary embodiments of the invention are described below. Note that the following exemplary embodiments do not in any way limit the scope of the invention laid out in the claims. Note also that all of the elements described in connection with the following exemplary embodiments should not necessarily be taken as essential elements of the invention.
When examining the digestive tract using an imaging device (e.g., endoscope apparatus), and determining the presence or absence of an early lesion, weight is attached to the structure (e.g., concavities and convexities) of the surface of tissue, and a difference in color (e.g., reddening or discoloration). A method that enhances a color and contrast by spraying a dye onto the digestive tract may be used so that the doctor can more easily observe the structure of the object.
However, since the dye spraying operation is cumbersome for the doctor, and increases the burden imposed on the patient, attempts have been made to enhance a color or a concave-convex structure by image processing so that a lesion can be easily detected (determined) (see JP-A-2003-88498 and JP-A-2005-342234, for example).
When enhancing a color or a concave-convex structure by image processing, an image that is easy to observe (i.e., suitable for observation) can be obtained by changing the enhancement process corresponding to the observation state (observation method). For example, the doctor observes a relatively large structure when sequentially observing the entire digestive tract while relatively quickly moving the imaging device (scope) (e.g., during screening observation). Therefore, it is necessary to present the large structure within the image so that the large structure is not missed. On the other hand, the doctor observes a microscopic structure when the doctor has determined the target by screening, and closely observes the structure of the target by zoom observation (close observation). Therefore, it is necessary to enhance the microscopic structure so that the doctor can determine whether the target is benign or malignant.
However, it is troublesome to manually change (switch) the enhancement process corresponding to the observation state. If the enhancement process is not changed corresponding to the observation state, and a large structure and a microscopic structure are always enhanced, it is difficult to observe the enhanced microscopic structure during screening observation since the image moves quickly, and it is also difficult to observe the enhanced microscopic structure during zoom observation since the large structure is also enhanced.
For example, when using a medical endoscope apparatus, the distance from the scope to the object is relatively long during screening observation in which the user searches the digestive tract for a lesion while the moving the scope (imaging section) relative to the digestive tract. The user brings the scope closer to the object during zoom observation in which the user closely observes the lesion candidate that was found during screening observation. Since the distance to the object differs corresponding to the observation state (i.e., differs between different observation states), it is possible to change the enhancement process corresponding to the observation state by performing the enhancement process that corresponds to the distance information. This makes it possible to selectively enhance the enhancement target in each observation state to provide an image that is suitable for observation.
The enhancement processing section 370 calculates a representative distance that represents the distance to the object based on the distance information, and performs the enhancement process that corresponds to the representative distance on the object (enhancement target) that agrees with the characteristics represented by the known characteristic information (described later). For example, the enhancement processing section 370 may include a target determination section 371 that determines the object that agrees with the characteristic represented by the known characteristic information to be the enhancement target. The enhancement processing section 370 may perform the enhancement process on the determined target. Note that the target determination section 371 is not an indispensable element as described later. The enhancement process may be performed based on the known characteristic information without determining the target.
More specifically, when the enhancement processing section 370 has determined that the representative distance falls under (corresponds to (or is equivalent to) (hereinafter the same)) a distance that corresponds to screening observation, the enhancement processing section 370 performs a first enhancement process on the determined target as the enhancement process. When the enhancement processing section 370 has determined that the representative distance falls under a distance that corresponds to close observation, the enhancement processing section 370 performs a second enhancement process on the determined target as the enhancement process, the second enhancement process differing from the first enhancement process.
This makes it possible to determine the enhancement target based on the known characteristic information, and perform the enhancement process that corresponds to the distance information so that the target is easily observed in each observation state. The target may be a concave-convex part, a reddened part, or a discolored part of tissue (described later). The user pays attention to a specific feature (e.g., a large structure, a small structure, or an area in a given color) in each observation state. The enhancement process may enhance such a feature to which the user pays attention corresponding to the observation state. For example, a relatively large concave-convex structure is enhanced during screening observation in which the scope is moved quickly, and a minute (microscopic) concave-convex structure is enhanced during zoom observation in which a lesion is closely observed.
The term “distance information” used herein refers to information in which each position within the captured image is linked to the distance to the object at each position within the captured image. For example, the distance information is a distance map. The term “distance map” used herein refers to a map in which the distance (depth) to the object in the Z-axis direction (i.e., the direction of the optical axis of the imaging section 200) is specified corresponding to each point (e.g., each pixel) in the XY plane, for example.
Note that the distance information may be arbitrary information acquired based on the distance from the imaging section 200 to the object. For example, when implementing triangulation using a stereo optical system, the distance with respect to an arbitrary point of a plane that connects two lenses that produce a parallax may be used as the distance information. When using a Time-of-Flight method, the distance with respect to each pixel position in the plane of the image sensor may be acquired as the distance information, for example. In such a case, the distance measurement reference point is set to the imaging section 200. Note that the distance measurement reference point may be set to an arbitrary position other than the imaging section 200, such as an arbitrary position within the three-dimensional space that includes the imaging section and the object. The distance information acquired using such a reference point is also intended to be included within the term “distance information”.
The distance from the imaging section 200 to the object may be the distance from the imaging section 200 to the object in the depth direction, for example. For example, the distance from the imaging section 200 to the object in the direction of the optical axis of the imaging section 200 may be used. For example, when a viewpoint is set in the direction orthogonal to the optical axis of the imaging section 200, the distance from the imaging section 200 to the object may be the distance observed at the viewpoint (i.e., the distance from the imaging section 200 to the object along a line that passes through the viewpoint and is parallel to the optical axis).
For example, the distance information acquisition section 380 may transform the coordinates of each corresponding point in a first coordinate system in which a first reference point of the imaging section 200 is the origin, into the coordinates of each corresponding point in a second coordinate system in which a second reference point within the three-dimensional space is the origin, using a known coordinate transformation process, and measure the distance based on the coordinates obtained by transformation. In this case, the distance from the second reference point to each corresponding point in the second coordinate system is identical to the distance from the first reference point to each corresponding point in the first coordinate system (i.e., the distance from the imaging section to each corresponding point).
The distance information acquisition section 380 may set a virtual reference point at a position that can maintain a relationship similar to the relationship between the distance values of the pixels on the distance map acquired when setting the reference point to the imaging section 200 to acquire the distance information based on the distance from the imaging section 200 to the corresponding point. For example, when the actual distances from the imaging section 200 to three corresponding points are respectively “3”, “4”, and “5”, the distance information acquisition section 380 may acquire distance information “1.5”, “2”, and “2.5” respectively obtained by halving the actual distances “3”, “4”, and “5” while maintaining the relationship between the distance values of the pixels.
The term “known characteristic information” used herein refers to information by which a useful structure of the surface of the object can be distinguished from an unuseful structure of the surface of the object. Specifically, the known characteristic information may be information (e.g., the size of a concave-convex part specific to a lesion, hue, or chroma) about a concave-convex part or a color for which the enhancement process is useful (e.g., a concave-convex part or a color that is useful for finding an early lesion). In this case, an object that agrees with the known characteristic information is determined to be the enhancement target. Alternatively, the known characteristic information may be information about a structure for which the enhancement process is not useful. In this case, an object that does not agree with the known characteristic information is determined to be the enhancement target. Alternatively, information about a useful concave-convex part and information about an unuseful structure may be stored, and the range of the useful concave-convex part may be set with high accuracy.
Although an example in which the enhancement process that corresponds to the observation state is performed based on the distance information has been described above, the configuration is not limited thereto. Specifically, the distance information is not indispensable. Whether screening observation or zoom observation is being performed may be determined based on arbitrary information (e.g., operation information input by the user, or a feature quantity within an image), and the enhancement process that corresponds to the determination result may be performed.
A first embodiment of the invention is described in detail below.
The light source section 100 includes a white light source 101, a rotary color filter 102 that includes a plurality of color filters that differ in spectral transmittance, a rotation driver section 103 that drives the rotary color filter 102, and a condenser lens 104 that focuses light (that has passed through the rotary color filter 102 and differs in spectral characteristics) on the incident end face of a light guide fiber 201. The rotary color filter 102 includes three primary-color filters (red color filter, green color filter, and blue color filter), and a rotary motor.
The rotation driver section 103 rotates the rotary color filter 102 at a given rotational speed in synchronization with the imaging period of an image sensor 206 included in the imaging section 200 based on a control signal output from a control section 302 included in the processor section 300. For example, when the rotary color filter 102 is rotated at 20 revolutions per second, each color filter crosses incident white light every 1/60th of a second. In this case, the image sensor 206 captures reflected light from the observation target to which red (R), green (G), or blue (B) light has been applied, and transmits the captured image to an A/D conversion section 209 every 1/60th of a second. Specifically, the endoscope apparatus according to the first embodiment frame-sequentially captures an R image, a G image, and a B image every 1/60th of a second, and the substantial frame rate is 20 fps.
Note that the first embodiment is not limited to the frame sequential method. For example, white light emitted from the white light source 101 may be applied to the object, and captured using an image sensor that includes an RGB Bayer color filter array.
The imaging section 200 is formed to be elongated and flexible so that the imaging section 200 can be inserted into a body cavity (e.g., stomach or large intestine), for example. The imaging section 200 includes the light guide fiber 201 that guides the light focused by the light source section 100, and an illumination lens 203 that diffuses the light guided by the light guide fiber 201 to illuminate the observation target. The imaging section 200 also includes an objective lens 204 that focuses the reflected light from the observation target, the image sensor 206 that detects the focused light, and an A/D conversion section 209 that converts photoelectrically converted analog signals output from the image sensor 206 into digital signals. The imaging section 200 further includes a memory 210 that stores scope ID information and specific information (including production variations) about the imaging section 200, and a connector 212 for removably connecting the imaging section 200 and the processor section 300.
The image sensor 206 is a single-chip monochrome image sensor when implementing the frame sequential method. For example, a CCD image sensor, a CMOS image sensor, or the like may be used as the image sensor 206. The A/D conversion section 209 converts the analog signals output from the image sensor 206 into digital signals, and outputs the digital signals (image) to the image processing section 301. The memory 210 is connected to the control section 302, and transmits the scope ID information and the specific information (including production variations) to the control section 302.
The processor section 300 includes an image processing section 301 that performs image processing on the image transmitted from the A/D conversion section 209, and the control section 302 that controls each section of the endoscope apparatus.
The display section 400 is a display device that can display a movie (moving image), and is implemented by a CRT, a liquid crystal monitor, or the like.
The external I/F section 500 is an interface that allows the user to input information or the like to the endoscope apparatus. For example, the external IN section 500 includes a power switch (power ON/OFF switch), a shutter button (capture start button), a mode (e.g., imaging mode) switch (e.g., a switch for selectively performing an enhancement process on a concave-convex part of the surface of tissue), and the like. The external I/F section 500 outputs the input information to the control section 302.
The image construction section 320 performs given image processing (e.g., OB process, gain process, and gamma process) on the image captured by the imaging section 200 to generate an image that can be output from the display section 400. The image construction section 320 outputs the generated image to the enhancement processing section 370 and the distance information acquisition section 380.
The known characteristic information acquisition section 390 reads (acquires) known characteristic information stored in the storage section 350, and outputs the known characteristic information to the enhancement processing section 370. The known characteristic information is information about the size (i.e., dimensional information (e.g., width, height, and depth)) of the concave-convex part of tissue to be determined to be the enhancement target.
The distance information acquisition section 380 acquires distance information about the distance to the object based on the captured image, and outputs the distance information to the enhancement processing section 370. The distance information is a distance map in which each pixel of the captured image is linked to the distance to the object at each pixel of the captured image, for example. The details of the distance information acquisition section 380 are described later.
The enhancement processing section 370 determines the target based on the known characteristic information and the distance information, performs the enhancement process that corresponds to the distance information on the target, and outputs the resulting image to the display section 400. Specifically, the enhancement processing section 370 calculates a representative distance from the distance information. The representative distance is the distance at the center position of the distance map, or the average distance within a given area of the distance map, for example. The distance variance of the distance map may be used as the representative distance. The distance variance is large during screening observation since the scope is positioned away from the surface of tissue, and is small during zoom observation since the scope almost perpendicularly faces the surface of tissue at a position close to the surface of tissue. The enhancement processing section 370 extracts a concave-convex part that agrees with the desired dimensional characteristics represented by the known characteristic information from the distance information, and determines the extracted concave-convex part to be the target.
When the representative distance is longer than a threshold value (e.g., during screening observation), the enhancement processing section 370 enhances the extracted concave-convex part that has a size (concavity-convexity pattern) larger than a first size threshold value. When the representative distance is shorter than the threshold value (e.g., during zoom observation), the enhancement processing section 370 enhances the extracted concave-convex part that has a size (concavity-convexity pattern) smaller than a second size threshold value. The first size threshold value and the second size threshold value may be set taking account of the size of the enhancement target concave-convex part (that allows the user to easily observe the resulting image) in each observation state. The enhancement process enhances a different color component depending on whether the target is a concave part or a convex part, for example.
Note that the enhancement process need not necessarily be performed on the determined target. The enhancement process may be performed based on the known characteristic information without determining the target. For example, the filter characteristics (e.g., the enhancement target frequency band) of a structural enhancement process (e.g., an enhancement process that enhances a high-frequency component of an image) may be set based on the size of the concave-convex part represented by the known characteristic information so that a concave-convex part having the desired size is enhanced. In this case, the filter characteristics may be changed corresponding to the representative distance.
The luminance signal calculation section 323 calculates a luminance signal Y (luminance value) from the captured image output from the image acquisition section 310 using the following expression (1) under control of the control section 302.
Y=0.299×R+0.587×G+0.114×B (1)
The calculated luminance signal Y is transmitted to the difference calculation section 324, the second derivative calculation section 325, and the storage section 327. The difference calculation section 324 calculates the difference between the luminance signals Y from a plurality of images necessary for calculating a defocus parameter. The second derivative calculation section 325 calculates the second derivative of the luminance signals Y of the image, and calculates the average value of the second derivatives obtained from a plurality of luminance signals Y that differ in the degree of defocus. The defocus parameter calculation section 326 calculates the defocus parameter by dividing the difference between the luminance signals Y of the image calculated by the difference calculation section 324 by the average value of the second derivatives calculated by the second derivative calculation section 325.
The storage section 327 stores the luminance signals Y of the first captured image, and the second derivative results thereof. Therefore, the distance information acquisition section 380 can place the focus lens at different positions through the control section 302, and acquire a plurality of luminance signals Y at different times. The LUT storage section 328 stores the relationship between the defocus parameter and the object distance in the form of a look-up table (LUT).
The object distance calculation method is described below. The control section 302 calculates the optimum focus lens position using a known contrast detection method, a known phase detection method, or the like based on the imaging mode set in advance using the external I/F section 500. The lens driver section 250 drives the focus lens to the calculated focus lens position based on a signal output from the control section 302. The image sensor 206 acquires the first image of the object at the focus lens position to which the focus lens has been driven. The acquired image is stored in the storage section 327 through the image acquisition section 310 and the luminance signal calculation section 323.
The lens driver section 250 then drives the focus lens to a second focus lens position that differs from the focus lens position at which the first image has been acquired, and the image sensor 206 acquires the second image of the object at the focus lens position to which the focus lens has been driven. The second image thus acquired is output to the distance information acquisition section 380 through the image acquisition section 310.
When the second image has been acquired, the defocus parameter is calculated. The difference calculation section 324 included in the distance information acquisition section 380 reads the luminance signals Y of the first image from the storage section 327, and calculates the difference between the luminance signal Y of the first image and the luminance signal Y of the second image output from the luminance signal calculation section 323.
The second derivative calculation section 325 calculates the second derivative of the luminance signals Y of the second image output from the luminance signal calculation section 323. The second derivative calculation section 325 then reads the luminance signals Y of the first image from the storage section 327, and calculates the second derivative of the luminance signals Y. The second derivative calculation section 325 then calculates the average value of the second derivative of the first image and the second derivative of the second image.
The defocus parameter calculation section 326 calculates the defocus parameter by dividing the difference calculated by the difference calculation section 324 by the average value of the second derivatives calculated by the second derivative calculation section 325.
The relationship between the defocus parameter and the focus lens position is stored in the LUT storage section 328 in the form of a table. The defocus parameter calculation section 326 calculates the object distance with respect to the optical system from the defocus parameter by linear interpolation using the defocus parameter and the information included in the table stored in the LUT storage section 328. The calculated object distance is output to the enhancement processing section 370 as the distance information.
Note that the distance information acquisition process may be implemented in various other ways. For example, the distance information may be acquired (calculated) by a Time-of-Flight method that utilizes infrared light or the like. When using the Time-of-Flight method, blue light may be used instead of infrared light, for example.
The distance information may be acquired using a stereo matching process. In this case, the endoscope apparatus may have the configuration illustrated in
The objective lenses 204 and 205 are disposed at given intervals so that a parallax image (hereinafter referred to as “stereo image”) can be captured. The objective lenses 204 and 205 respectively form a left image and a right image (stereo image) on the image sensors 206 and 207. The A/D conversion section 209 performs the A/D conversion process on the left image and the right image respectively output from the image sensors 206 and 207, and outputs the resulting left image and the resulting right image to the image construction section 320 and the distance information acquisition section 380.
The distance information acquisition section 380 performs a matching calculation process on the left image (reference image) and the right image using an epipolar line that passes through an attention pixel of the reference image with respect to a local area that includes the attention pixel and a local area of the right image. The distance information acquisition section 380 calculates a position at which the correlation obtained by the matching calculation process becomes a maximum as a parallax, converts the parallax into the distance in the depth direction (i.e., the Z-axis direction of the distance map), and outputs the distance information to the enhancement processing section 370.
The known characteristic information acquisition section 390 acquires the dimensional information (i.e., information about the size of the extraction target concave-convex part of tissue) from the storage section 350 as the known characteristic information, and determines the frequency characteristics of a low-pass filtering process based on the dimensional information. As illustrated in
As illustrated in
diff(x,y)=dist(x,y)−dist_LPF(x,y) (2)
The target determination section 371 outputs the concavity-convexity map calculated as described above to the enhancement section 372. In this example, a concave-convex part in the surface area of tissue is determined to be the enhancement target by extracting a concave-convex part in the surface area of tissue as the concavity-convexity map.
A process that determines the cut-off frequency (extraction process parameter in a broad sense) from the dimensional information is described in detail below.
The target determination section 371 performs the low-pass filtering process using a given size (e.g., N×N pixels (N is a natural number equal to or larger than 2)) on the input distance information. The target determination section 371 adaptively determines the extraction process parameter based on the resulting distance information (local average distance). Specifically, the target determination section 371 determines the characteristics of the low-pass filtering process that smooth the extraction target concave-convex part of tissue due to a lesion while maintaining the structure of the lumen and the folds specific to the observation target part. Since the characteristics of the extraction target (i.e., concave-convex part) and the exclusion target (i.e., folds and lumen) can be determined from the known characteristic information, the spatial frequency characteristics are known, and the characteristics of the low-pass filter can be determined. Since the apparent size of the structure changes corresponding to the local average distance, the characteristics of the low-pass filter are determined corresponding to the local average distance (see
The low-pass filtering process is implemented by a Gaussian filter represented by the following expression (3), or a bilateral filter represented by the following expression (4), for example. The frequency characteristics of these filters are controlled using σ, σc, and σv. A σ map that corresponds to the pixels of the distance map on a one-to-one basis may be generated as the extraction process parameter. When using the bilateral filter, either or both of a σc map and a σv map may be generated as the extraction process parameter.
For example, σ may be a value that is larger than a value obtained by multiplying a pixel-to-pixel distance D1 of the distance map corresponding to the size of the extraction target concave-convex part by α (>1), and is smaller than a value obtained by multiplying a pixel-to-pixel distance D2 of the distance map corresponding to the size of the lumen and the folds specific to the observation target part by β (<1). For example, σ may be calculated by σ=(α*D1+β*D2)/2*Rσ. Note that Rσ is a function of the local average distance. The value Rσ increases as the local average distance decreases, and decreases as the local average distance increases.
The known characteristic information acquisition section 390 may read the dimensional information corresponding to the observation target part from the storage section 350, and the target determination section 371 may determine the target corresponding to the observation target part based on the dimensional information, for example. The observation target part may be determined using the scope ID stored in the memory 210 (see
Note that the extraction process is not limited to the extraction process that utilizes the low-pass filtering process. For example, extracted concavity-convexity information may be acquired using a morphological process. In this case, an opening process and a closing process using a given kernel size (i.e., the size (sphere diameter) of a structural element) are performed on the distance map (see
The enhancement section 372 determines the observation state (observation method) based on the representative distance, and performs the enhancement process that corresponds to the observation state. Specifically, when the enhancement section 372 has determined that the representative distance is long, and screening observation is being performed, the enhancement section 372 enhances the extracted concave-convex part that has a large size. When the enhancement section 372 has determined that the representative distance is short, and zoom observation is being performed, the enhancement section 372 enhances the extracted concave-convex part that has a small size.
A process that enhances a different color depending on whether the target is a concave part or a convex part is described below as an example of the enhancement process. A pixel for which the concavity-convexity information diff(x, y) is smaller than 0 (diff(x, y)<0) is determined to be a convex part, and a pixel for which the concavity-convexity information diff(x, y) is larger than 0 (diff(x, y)>0) is determined to be a concave part (see the expression (2)). A chroma enhancement process that corresponds to a given hue is performed on each pixel that has been determined to be a convex part, and a chroma enhancement process that corresponds to a hue that differs from the given hue is performed on each pixel that has been determined to be a concave part, for example. The size of a concave-convex part may be determined by comparing the width (i.e., the number of pixels) of a convex area and the width (i.e., the number of pixels) of a concave area with a threshold value, for example. Note that the enhancement process is not limited thereto. Various other enhancement processes may also be used. For example, a given color (e.g., blue) may be enhanced to a larger extent as the concavity-convexity information diff(x, y) increases (i.e., the depth increases) to simulate a state in which a dye (e.g., indigo carmine) has been sprayed. The coloring method may be changed depending on whether the observation state is a screening observation state or a zoom observation state.
According to the first embodiment, the target determination section 371 determines the extraction process parameter based on the known characteristic information, and determines a concave-convex part of the object to be the enhancement target based on the determined extraction process parameter.
This makes it possible to perform the concavity-convexity information extraction (separation) process (target determination process) using the extraction process parameter determined based on the known characteristic information. In order to accurately extract the extracted concavity-convexity information, it is necessary to perform a control process that extracts information about the desired concave-convex part from the information about various structures included in the distance information while excluding other structures (e.g., the structures (e.g., folds) specific to tissue). In the first embodiment, such a control process is implemented by setting the extraction process parameter based on the known characteristic information.
The target determination section 371 may determine the size of the structural element used for the opening process and the closing process as the extraction process parameter based on the known characteristic information, and perform the opening process and the closing process using the structural element having the determined size to extract a concave-convex part of the object as the extracted concavity-convexity information.
This makes it possible to extract the extracted concavity-convexity information based on the opening process and the closing process (morphological process in a broad sense). In this case, the extraction process parameter is the size of the structural element used for the opening process and the closing process. When the structural element is a sphere, the extraction process parameter is a parameter that represents the diameter of the sphere, for example.
The captured image may be an in vivo image that is obtained by capturing the inside of a living body, the object may include a global three-dimensional structure that is a lumen structure inside the living body, and a local concave-convex structure that is formed on the lumen structure, and is more local than the global three-dimensional structure, and the target determination section 371 may extract the concave-convex part of the object that is selected from the global three-dimensional structure and the local concave-convex structure included in the object, and agrees with the characteristics represented by the known characteristic information, as the extracted concavity-convexity information.
This makes it possible to implement a process that extracts the global three-dimensional structure (i.e., a structure having a low spatial frequency as compared with the concave-convex part) and the concave-convex part (that is smaller than the global three-dimensional structure) included in the distance information when applying the method according to the first embodiment to an in vivo image. For example, when the extraction target is a concave-convex part that is useful for finding an early lesion, the three-dimensional structure (e.g., folds and a structure based on the curvature of a wall surface) of tissue can be excluded from the extraction target, and the target determination section 371 extracts only the extraction target concave-convex part. In this case, since the global structure (i.e., a structure having a low spatial frequency) is excluded from the extraction target, and the local structure (i.e., a structure having a high spatial frequency) is determined to be the extraction target, an intermediate spatial frequency or the like is set to be the boundary, for example.
Although an example in which each section included in the processor section 300 is implemented by hardware has been described above, the configuration is not limited thereto. For example, a CPU may perform the process of each section on image signals (acquired in advance using an imaging device) and the distance information. Specifically, the process of each section may be implemented by software by causing the CPU to execute a program. Alternatively, part of the process of each section may be implemented by software.
In this case, a program stored in an information storage device is read, and executed by a processor (e.g., CPU). The information storage device (computer-readable device) stores a program, data, and the like. The function of the information storage device may be implemented by an optical disk (e.g., DVD or CD), a hard disk drive (HDD), a memory (e.g., memory card or ROM), or the like. The processor (e.g., CPU) performs various processes according to the first embodiment based on the program (data) stored in the information storage device. Specifically, a program that causes a computer (i.e., a device that includes an operation section, a processing section, a storage section, and an output section) to function as each section according to the first embodiment (i.e., a program that causes a computer to execute the process of each section according to the first embodiment) is stored in the information storage medium.
The target is determined using the above method (step S3). The representative distance to the object is calculated from the distance information (step S4), and whether or not the representative distance (e.g., the distance at the center of the image, the average distance, or the distance variance) is longer than the threshold value ε is determined (step S5). When the representative distance is longer than the threshold value ε (i.e., when it has been determined that screening observation is being performed), a first enhancement process is performed on the captured image (step S6), and the resulting image is output (step S7). The first enhancement process enhances the determined target that has a size larger than a first size threshold value Tk. When the representative distance is equal to or shorter than the threshold value ε (i.e., when it has been determined that zoom observation is being performed), a second enhancement process is performed on the captured image (step S8), and the resulting image is output (step S9). The second enhancement process enhances the determined target that has a size smaller than a second size threshold value Ts.
According to the first embodiment, the known characteristic information acquisition section 390 acquires the known characteristic information that is information that represents known characteristics relating to the structure of the object. The enhancement processing section 370 performs the enhancement process on a concave-convex part (target) of the object that agrees with the characteristics represented by the known characteristic information. For example, the target determination section 371 extracts a concave-convex part that has the desired characteristics (e.g., size) represented by the known characteristic information from the distance information (distance map in a narrow sense), and determines the extracted concave-convex part to be the target. The enhancement processing section 370 performs the enhancement process on the determined target.
This makes it possible to perform the enhancement process based on the distance information and the known characteristic information, and change the enhancement process corresponding to the observation state. It is possible to determine the observation target that has a specific three-dimensional structure or shape by utilizing the distance information, for example. Specifically, it is possible to extract a concave-convex part that has the desired characteristics (e.g., size), and perform a process that enhances the extracted concave-convex part. The enhancement process can be changed corresponding to the distance information so that the concave-convex part to which the user pays attention in each observation state is enhanced.
More specifically, when the representative distance (e.g., the distance at the center of the image, the average distance, or the distance variance) is longer than the threshold value ε, the enhancement processing section 370 performs the enhancement process on the extracted concave-convex part that has been determined to have a size larger than the first size threshold value Tk. When the representative distance is shorter than the threshold value ε, the enhancement processing section 370 performs the enhancement process on the extracted concave-convex part that has been determined to have a size smaller than the second size threshold value Ts.
According to this configuration, since a small structure is not enhanced during screening observation in which the scope is moved quickly, it is possible to reduce complexity during observation. Since a small structure is enhanced during zoom observation in which a lesion is closely observed, it is possible for the user to clearly observe a minute lesion structure. It is thus possible to improve visibility in each observation state, and suppress a situation in which a lesion is missed while improving the diagnostic accuracy.
A second embodiment illustrates an example in which a reddened part or a discolored part is enhanced. Since a reddened part and a discolored part may not have a specific shape, differing from a concave-convex part, it is desirable to enhance a reddened part and a discolored part using a method that differs from the method employed when enhancing a concave-convex part. Note that the term “reddened part” used herein refers to a part that is observed to have a high degree of redness as compared with its peripheral area, and the term “discolored part” used herein refers to a part that is observed to have a low degree of redness as compared with its peripheral area.
A detailed configuration according to the second embodiment is described below. An endoscope apparatus and an image processing section 301 according to the second embodiment may be configured in the same manner as the endoscope apparatus and the image processing section 301 according to the first embodiment. Note that the same elements as those described above in connection with the first embodiment are respectively indicated by the same reference signs (symbols), and description thereof is appropriately omitted.
The known characteristic information acquisition section 390 acquires color information about the extraction target tissue (e.g., reddened part or discolored part) as the known characteristic information. The target determination section 371 determines an area that agrees with the color represented by the known characteristic information to be the target. For example, the target determination section 371 determines an area that has a high degree of redness as compared with a normal mucous membrane to be the reddened part. For example, the target determination section 371 determines an area for which the ratio “R/G” of the red (R) pixel value to the green (G) pixel value is larger than that of its peripheral area, to be the reddened part. In this case, the known characteristic information represents a condition whereby an area for which the ratio “R/G” is larger to a given extent (e.g., factor) than that of its peripheral area is determined to be the reddened part, for example. Alternatively, the range of the ratio “R/G” or the hue value may be stored as the known characteristic information, and an area that falls under the range of the ratio “R/G” or the hue value may be determined to be the reddened part. An image frequency or structural information (e.g., size or shape) specific to the reddened part may also be used as the known characteristic information.
When the enhancement section 372 has determined that the representative distance is long, and screening observation is being performed, the enhancement section 372 enhances the color of the determined target. For example, the enhancement section 372 performs a process that increases the degree of redness (e.g., the ratio “R/G”, or the chroma within the red hue range) of the reddened part, and decreases the degree of redness of the discolored part. When the enhancement section 372 has determined that the representative distance is short, and zoom observation is being performed, the enhancement section 372 performs a process that enhances the edge component of the image, or a process that enhances a specific frequency region through frequency analysis. The enhancement section 372 may perform a color enhancement process similar to that used during screening observation.
Although an example in which the target is determined has been described above, the enhancement process may be performed without determining the target. For example, a process that increases the degree of redness may be performed on an area having a high degree of redness when the representative distance is long, and the enhancement process that enhances a minute (microscopic) structure with a higher enhancement level may be performed when the representative distance is short. In this case, redness enhancement characteristics are stored as the known characteristic information, for example.
The detection target is not limited to the reddened part and the discolored part. For example, the detection target may be a polyp or the like. When the detection target is a polyp, a shape (or a color) specific to a polyp may be stored as the known characteristic information, and a polyp may be detected by performing a pattern matching process (or a color comparison process), for example. When the enhancement section 372 has determined that the representative distance is long, and screening observation is being performed, the enhancement section 372 may enhance a polyp by performing a contour enhancement process. When the enhancement section 372 has determined that the representative distance is short, and zoom observation is being performed, the enhancement section 372 may perform the color enhancement process in addition to the contour enhancement process.
According to the second embodiment, the known characteristic information acquisition section 390 acquires the known characteristic information that is information that represents known characteristics relating to the color of the object.
This makes it possible to perform the color enhancement process corresponding to the observation state based on the information about the color specific to the observation target. It is possible to perform the enhancement process so that a lesion or the like that differs in color from a normal area can be easily observed, by utilizing the information about the color specific to the observation target, for example.
The enhancement processing section 370 performs the enhancement process that corresponds to the distance information on the object (target) that agrees with the characteristics represented by the known characteristic information. For example, the target determination section 371 determines the object that agrees with the characteristics represented by the known characteristic information to be the enhancement target, and the enhancement processing section 370 performs the enhancement process on the determined target.
More specifically, the enhancement processing section 370 calculates the representative distance (e.g., the distance at the center of the image, the average distance, or the distance variance) based on the distance information. When the representative distance is longer than the threshold value ε, the enhancement processing section 370 performs the enhancement process that enhances the color (e.g., the ratio “R/G”, or the chroma within a given hue range) of the target. When the representative distance is shorter than the threshold value ε, the enhancement processing section 370 performs the enhancement process that enhances at least the structure of the target.
For example, the known characteristic information is information that represents known characteristics relating to the color of the reddened part for which the red component (e.g., ratio “R/G”) is larger than that of a normal mucous membrane. When the representative distance is longer than the threshold value c, the enhancement processing section 370 performs the enhancement process that enhances the red component (e.g., the ratio “R/G”, or the chroma within the red hue range) of the reddened part that has been determined to be the target.
This makes it possible to determine the position of a lesion (e.g., reddened part or discolored part) having a specific color during screening observation in which the scope is moved quickly. Since both a color and a structure can be enhanced during zoom observation in which a lesion is closely observed, it is possible for the user to easily observe a minute lesion structure to which the user pays attention. It is thus possible to improve visibility in each observation state, and suppress a situation in which a lesion is missed while improving the diagnostic accuracy.
The image processing device, the processor section 300 and the like according to the embodiments of the invention may include a processor and a memory. The processor may be a central processing unit (CPU), for example. Note that the processor is not limited to a CPU. Various processors such as a graphics processing unit (GPU) or a digital signal processor (DSP) may also be used. The processor may be a hardware circuit that includes an ASIC. The memory stores a computer-readable instruction. Each section of the image processing device, the processor section 300 and the like according to the embodiments of the invention is implemented by causing the processor to execute the instruction. The memory may be a semiconductor memory (e.g., SRAM or DRAM), a register, a hard disk, or the like. The instruction may be an instruction included in an instruction set of a program, or may be an instruction that causes a hardware circuit of the processor to operate.
The embodiments to which the invention is applied and the modifications thereof have been described above. Note that the invention is not limited to the above embodiments and the modifications thereof. Various modifications and variations may be made of the above embodiments and the modifications thereof without departing from the scope of the invention. A plurality of elements described in connection with the above embodiments and the modifications thereof may be appropriately combined to implement various configurations. For example, some elements may be omitted from the elements described in connection with the above embodiments and the modifications thereof. Some of the elements described in connection with different embodiments or modifications thereof may be appropriately combined. Specifically, various modifications and applications are possible without materially departing from the novel teachings and advantages of the invention. Any term cited with a different term having a broader meaning or the same meaning at least once in the specification and the drawings can be replaced by the different term in any place in the specification and the drawings.
Number | Date | Country | Kind |
---|---|---|---|
2013-035730 | Feb 2013 | JP | national |
This application is a continuation of International Patent Application No. PCT/JP2013/075628, having an international filing date of Sep. 24, 2013, which designated the United States, the entirety of which is incorporated herein by reference. Japanese Patent Application No 2013-035730 filed on Feb. 26, 2013 is also incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2013/075628 | Sep 2013 | US |
Child | 14834905 | US |