This invention relates to an image processing device and an image processing method for generating an image capable of identifying a difference area in an image with another image.
In order to understand damage situation based on disaster such as a flood, a forest fire, a volcanic eruption, an earthquake, a tsunami or a drought, situation of urban development, or movement and retention of cargoes and people, a change detection technology is utilized which detects areas where the ground surface conditions have changed, based on images taken from high locations, for example, images taken by a satellite.
Synthetic aperture radar (SAR) technology is a technology which can obtain an image (hereinafter referred to as a SAR image) equivalent to an image by an antenna having a large aperture, when a flying object such as artificial satellite, aircraft, or the like transmits and receives a radio wave while the flying object moves. The synthetic aperture radar is utilized, for example, for analyzing a ground surface displacement by signal-processing reflected waves from the ground surface, etc.
Hereinafter, an image taken by a satellite, etc. is referred to as an observed image. Unless otherwise specified, both optical and SAR images are acceptable for an observed image.
Generally, in change detection, two images obtained by observing the same area at different times are compared. By comparing two images, a change in one or more bodies (objects) in the area is detected. A change in an object may be, for example, appearance of a new object or disappearance of an object. Hereinafter, each of the two images is referred to as an object presence image or an object map, and the two images are sometimes referred to as an image pair. An image capable of identifying a difference part between two images based on the comparison of the two images is sometimes referred to as a difference map or a synthesized difference map.
Patent literature 1 describes a method for generating a classifier (a trained model) using two types of images (image pair) generated from interferometric SAR images and correct answer data as learning data (training data). In patent literature 1, ground surface change is determined using the trained model.
When the method described in patent literature 1 is used, as described in paragraph 0019 of patent literature 1, the correct answer data is manually generated. Therefore, it takes time to obtain lots of correct answer data. In addition, it is possible that the correct answer data generated by one preparer may differ from the correct answer data generated by another preparer. Therefore, objectivity of correct answer data cannot be guaranteed. In other words, there is a possibility that correct answer data reflecting individual differences may be generated.
When the method described in non-patent literature 1 is used, the first image 331 and the second image 332 that are sources of the synthesized difference map 333 are manually generated. The synthesized difference map 333 may deviate from a difference map obtained from an actual observed image, even if the synthesized difference map 333 that can be used as the correct answer data is automatically generated from the first image 331 and the second image 332. This is because the original first image 331 and the original second image 332 are artificially generated. As a result, when the synthesized difference map 333 is used as the correct difference map, it may deviate from a correct difference map obtained from the actual observed images.
It is an object of the present invention to provide an image processing device and an image processing method that can generate an image capable of identifying a difference part between two input images in a short time without being affected by individual differences, and that can eliminate the deviation of the image from an image obtained from the actual observed image.
An image processing device according to the present invention includes image deformation means for deforming object presence areas in two object presence images, in which one or more objects are present, obtained from each of two observed images to generate two deformed images, based on an observation angle of each of the two observed images and a size of the object appearing in each of the two observed images, and image generation means for generating a synthesized image by synthesizing the two deformed images, determining difference of the object between the two object presence images, and generating an image capable of identifying the determined difference.
An image processing method according to the present invention includes deforming object presence areas in two object presence images, in which one or more objects are present, obtained from each of two observed images to generate two deformed images, based on an observation angle of each of the two observed images and a size of the object appearing in each of the two observed images, and generating a synthesized image by synthesizing the two deformed images, determining difference of the object between the two object presence images, and generating an image capable of identifying the determined difference.
An image processing program according to the present invention causes a computer to execute a process of deforming object presence areas in two object presence images, in which one or more objects are present, obtained from each of two observed images to generate two deformed images, based on an observation angle of each of the two observed images and a size of the object appearing in each of the two observed images, and a process of generating a synthesized image by synthesizing the two deformed images, determining difference of the object between the two object presence images, and generating an image capable of identifying the determined difference.
According to the present invention, it is possible to generate an image capable of identifying a difference part between two input images in a short time without being affected by individual differences, and to eliminate the deviation of the image from an image obtained from the actual observed image.
A set of observed images is input to the object map generating means 10. The object map generation means 10 extracts from each of the observed images an image (object presence image) including an object presence area in which an object that is a target of change detection is present. In other words, the object map generation means 10 generates a set of object maps. The set of object maps correspond to the image pair described above. For example, the object map generation means 10 extracts predetermined areas from the observed images, but it is also possible to manually extract areas from the observed images.
An observation angle (azimuth and incidence angle) and a size (height and width) of the object in each of the observed images are input to the correct difference map generation means 20. The size of the object is predetermined depending on the object that is a target of change detection.
The correct difference map generation means 20 deforms each object map based on the observed angle and the size of the object in each of the observed images. Further, the correct difference map generation means 20 generates an image showing an area where the object has changed between the two object maps, i.e., a difference map, by synthesizing the deformed object maps to generate a synthesized image. The difference map generated by the correct difference map generation means 20 is output as a correct difference map.
Next, examples of an object map generating method and a correct difference map generating method will be explained. Hereinafter, a SAR image is used as an example of the observed image. In addition, an automobile is used as an example of the object.
On the center of the upper row of
In this example, the first object map 111 and the second object map 121 correspond to images of the parking lot 200.
The correct difference map generation means 20 generates a correct difference map 150 using the image A and the image B. In the correct difference map 150, the ellipse surrounded by a solid line indicates an area where the automobile 93 that has not changed from the time t1 to the time t2 exists. In other words, it indicates an area where there is no change. The black ellipse indicates an area where the newly appeared automobile 94 exists. The ellipses surrounded by dashed lines indicate areas where the disappeared automobiles 91, 92 existed. In other words, the black ellipse and the ellipse surrounded by a dashed line indicate a change area.
In the correct difference map 150, the change area and the non-change area can be distinguishable by a different expression than that illustrated in
l
A
=h/tan θA (1)
When the observed image is an optical image, assuming that the incidence angle of sunlight is θA, the collapse amount lA is expressed by the following equation (2).
l
A
=h·tan θA (2)
Since the case of SAR images is used as an example in this example embodiment, hereinafter, the collapse amount with respect to image A is denoted as lA and the collapse amount with respect to image B is denoted as lB (lB=h/tan θB). When an optical image is used, the collapse amount with respect to image B is lB (lB=h/tan θB).
As shown in
In the first object map 112 and the second object map 121 shown in
The correct difference map generation means 20 superimposes the image A after the dilating process, i.e., the first object map 112 on the image B after the dilating process, i.e., the second object map 122.
It is assumed that the observed image that is a source of the image B has been obtained later in time than the observed image that is a source of the image A.
In
The correct difference map generation means 20 generates the difference map 140 based on the concept as illustrated in
In the difference map 140 and the correct difference map 150 shown in
The correct difference map generation means 20 applies a noise elimination process to the difference map 140. The noise elimination process is a process to eliminate areas that are smaller than the object as noises. In the example shown in
Although the difference map in which the noise has been eliminated is used as the correct difference map 150 in this example embodiment, the difference map 140 before the noise elimination process is applied may be used as the correct difference map, in site of the fact that noise remains.
The first collapse parameter calculation means 21 is provided with a range azimuth, an incidence angle and a height of the object regarding the image A (the first object map 111). The first collapse parameter calculation means 21 calculates the collapse amount of the object in the image A using the incidence angle and the height of the object. The first collapse parameter calculation means 21 also determines the collapsing direction of the object in the image A using the range azimuth. The collapsing direction is the same as the direction indicated by the range azimuth αA. The first collapse parameter calculation means 21 outputs the first collapse parameter to the second dilation means 32. The first collapse parameter includes at least data indicating a collapse amount of the object and data indicating a collapsing direction of the object.
The second collapse parameter calculation means 22 is provided with a range azimuth, an incidence angle and a height of the object regarding the image B (the second object map 121). The second collapse parameter calculation means 22 calculates the collapse amount of the object in the image B using the incidence angle and the height of the object. The second collapse parameter calculation means 22 also determines the collapsing direction of the object in image B using the range azimuth. The collapsing direction is the same as the direction indicated by the range azimuth αB. The second collapse parameter calculation means 22 outputs the second collapse parameter to the first dilation means 31. The second collapse parameter includes at least data indicating a collapse amount of the object and data indicating a collapsing direction of the object.
When an optical image is used as the observed image, the first collapse parameter calculation means 21 calculates a direction indicated by the range azimuth αA+180 degrees (or the range azimuth αA−180 degrees) as the collapsing direction in the first collapse parameter. The second collapse parameter calculation means 22 calculates the direction indicated by the range azimuth αB+180 degrees (or the range azimuth αB−180 degrees) as the collapsing direction in the second collapse parameter.
The image A and the second collapse parameter are input to the first dilation means 31. The first dilation means 31 dilates the object in the image A using the second collapse parameter to generate an image A (a first object map 112) in which the object is dilated. The first dilation means 31 outputs the first object map 112 to the difference map generation means 41.
The image B and the first collapse parameter are input to the second dilation means 32. The second dilation means 32 dilates the object in the image B using the first collapse parameter to generate an image B (a second object map 122) in which the object is dilated. The second dilation means 32 outputs the second object map 122 to the difference map generation means 41.
The difference map generation means 41 superimposes the first object map 112 on the second object map 122. In other words, the difference map generation means 41 synthesizes the first object map 112 and the second object map 122. Then, the difference map generation means 41 determines a difference (disappearance or appearance) between the object in the first object map 112 and the corresponding object in the second object map 122 to the object in the first object map 112. The difference map generation means 41 modifies the synthesized image, in which the first object map 112 is superimposed on the second object map 122, to an image capable of distinguishing a change areas from a non-change area, and outputs the image as the difference map 140 to the noise elimination means 51.
The noise elimination means 51 applies an opening process to the difference map 140 and outputs an image in which noises are eliminated as the correct difference map.
Next, the operation of the correct difference map generation means 20 is explained with reference to the flowchart in
As shown in
Meta-information of one observed image is input to the first collapse parameter calculation means 21. Meta-information of the other observed image is input to the second collapsed parameter calculation means 22. In general, an available observed image is accompanied by meta-information (metadata) such as the time of shooting, the shooting location (for example, latitude and longitude of the center of the observed image), and the direction of electromagnetic radiation (observation direction), etc. The first collapse parameter calculation means 21 extracts the range azimuth as and the incidence angle θA from the meta-information of one observed image, and the second collapse parameter calculation means 22 extracts the range azimuth αB and the incidence angle θB from the meta-information of the other observed image (step S12).
It is not essential that the first collapse parameter calculation means 21 and the second collapse parameter calculation means 22 extract a range azimuths and an incidence angle from the meta-information. For example, means other than the first collapse parameter calculation means 21 and the second collapse parameter calculation means 22 may extract a range azimuth and an incidence angle from the meta-information. In such a case, the means provides the extracted range azimuth and the extracted incidence angle to the first collapse parameter calculation means 21 and the second collapse parameter calculation means 22.
Data indicating the height h of the object is input to the first collapse parameter calculation means 21 and the second collapse parameter calculation means 22 (step S13).
The processing order of steps S11-S13 is arbitrary. That is, the processing order of steps S11-S13 does not necessarily have to be the order shown in
The first collapse parameter calculation means 21 and the second collapse parameter calculation means 22 calculate the collapse parameters (step S14). In step S14, the first collapse parameter calculation means 21 calculates the collapse amount lA of the object in the image A by the above equation (1) using the incidence angle θA obtained in the process of step S12 and the height h of the object. The first collapse parameter calculation means 21 regards the range azimuth αA obtained in the process of step S12 as the collapsing direction of the object. The first collapse parameter calculation means 21 regards the obtained collapse amount and the collapsing direction as the first collapse parameter. When there are multiple objects in the image A, the first collapse parameter calculation means 21 determines the collapse amount and collapsing direction of each object, and includes each collapse amount and each collapsing direction in the first collapse parameter.
In step S14, the second collapse parameter calculation means 22 calculates the collapse amount lB of the object in the image B by the above equation (1) using the incidence angle θB obtained in the process of step S12 and the height h of the object. The second collapse parameter calculation means 22 regards the range azimuth αB obtained in the process of step S12 as the collapsing direction of the object. The second collapse parameter calculation means 22 regards the obtained collapse amount and the collapsing direction as the second collapse parameter. When there are multiple objects in the image B, the second collapse parameter calculation means 22 determines the collapse amount and collapsing direction of each object, and includes each collapse amount and each collapsing direction in the second collapse parameter.
When an optical image is used as the observed image, the first collapse parameter calculation means 21 determines a direction which is different from the range azimuth αA by 180 degrees as the collapsing direction in the first collapse parameter. The second collapse parameter calculation means 22 determines a direction which is different from the range azimuth αB by 180 degrees as the collapsing direction in the second collapse parameter.
The first dilation means 31 and the second dilation means 32 dilate the object in the object map (image A or image B) (step S15). In step S15, the first dilation means 31 dilates the object in the image A in the collapsing direction included in the second collapse parameter by the collapse amount lB. The second dilation means 32 dilates the object in image B in the collapsed direction included in the first collapsed parameter by the collapsed amount lA.
The difference map generation means 41 superimposes the image A (the first object map 112: refer to
The difference map generation means 41 determines whether the object has changed or not based on the multiplicity of the object in the synthesized image generated in the process of step S16. For example, the difference map generation means 41 compares the first object map 112 and the second object map 122 pixel by pixel (every pixel) to determine whether the object has changed or not. Then, as illustrated in
The difference map generation means 41 generates the difference map 140 (refer to
Data indicating the width of the object is input to the noise elimination means 51 (step S18). The width of the object is set in advance. For example, when the object is an automobile, the value of the width of an ordinary automobile or a value with a margin to it is input to the noise elimination means 51 as the width of the object. It should be noted that he process of step S18 does not have to be performed at the timing shown in
The noise elimination means 51 applies an opening process to the difference map 140 and outputs an image in which noises are eliminated as the correct difference map (step S19). In the process of step S19, the noise elimination means 51 erodes the object by the number of pixels corresponding to the size (specifically, the width) of the object in the erosion process in the opening process. The number of pixels to be eroded is determined in advance according to the size of the object. Therefore, it is set to the number of pixels that can be eliminated from the collection of pixels that should be determined not to be the object. As an example, when the maximum width of the object is 3 pixels, the noise elimination means 51 performs the erosion process two times so that blocks with a size of less than 3 pixels, i.e., equal to or less than 2 pixels will be eliminated.
As explained above, the image processing device of this example embodiment generates a difference map as a correct answer to be used as training data for machine learning, based on actual observed images. Therefore, the difference map can be generated in a short time without being affected by individual differences, as is the case when the difference map is manually generated. It is also possible to eliminate possibility that the difference map deviates from the image obtained from the actual observed images.
In addition, as described above, it is preferably that the image processing device is configured to dilate the object presence area in the first object map 111 in accordance with the collapsing direction and the collapse amount of the object in the second object map 121, and dilate the object presence area in the second object map 121 in accordance with the collapsing direction and the collapse amount of the object in the first object map 111. In such a configuration, the visibility of the object in one of the two object maps with different observation directions can be brought closer to the visibility of the object in the other object map. Therefore, accuracy of detecting change/non-change in the object presence area using the image obtained by synthesizing the first object map 111 and the second object map 121 is improved.
In addition, as described above, it is preferably that the image processing device is configured to eliminate areas whose sizes are smaller than a predetermined value determined based on the width of the object. In such a configuration, when a small size area is determined to be a change area in the synthesized image, the difference map finally obtained (the difference map used as the correct difference map) becomes to be a map that does not include change areas other than the object. Therefore, the reliability of the correct difference map can be increased.
The image processing device of the above example embodiment can be configured with hardware, but can also be configured with a computer program.
The program memory 1002 is, for example, a non-transitory computer readable medium. The non-transitory computer readable medium is one of various types of tangible storage media. For example, as the program memory 1002, a semiconductor storage medium such as a flash ROM (Read Only Memory) or a magnetic storage medium such as a hard disk can be used. In the program memory 1002, an image processing program for realizing functions of blocks (the object map generation means 10, the correct difference map generation means 20, the first collapse parameter calculation means 21, the second collapse parameter calculation means 22, the first dilation means 31, the second dilation means 32, the difference map generation means 41, the noise elimination means 51) in the image processing device of the above example embodiment is stored.
The processor 1001 realizes the function of the image processing device by executing processing according to the image processing program stored in the program memory 1002. When multiple processors are implemented, they can also work together to realize the function of the image processing device.
For example, a RAM (Random Access Memory) can be used as the memory 1003. In the memory 1003, temporary data that is generated when the image processing device executes processing, etc. are stored. It can be assumed that an image processing program is transferred to the memory 1003 and the processor 1001 executes processing based on the image processing program in the memory 1003. The program memory 1002 and the memory 1003 may be integrated into a single unit.
As shown in
As shown in
A part of or all of the above example embodiments may also be described as, but not limited to, the following supplementary notes.
(Supplementary note 1) An image processing device comprising:
(Supplementary note 2) The image processing device according to Supplementary note 1, wherein
(Supplementary note 3) The image processing device according to Supplementary note 2, wherein
(Supplementary note 4) The image processing device according to Supplementary note 3, further comprising
(Supplementary note 5) The image processing device according to any one of Supplementary notes 1 to 4, further comprising
(Supplementary note 6) An image processing method comprising:
(Supplementary note 7) The image processing method according to Supplementary note 6, wherein
(Supplementary note 8) The image processing method according to Supplementary note 7, wherein
(Supplementary note 9) The image processing method according to any one of Supplementary notes 6 to 8, further comprising
(Supplementary note 10) A computer readable recording medium storing an image processing program, wherein
(Supplementary note 11) The recording medium according to Supplementary note 10, wherein
(Supplementary note 12) The recording medium according to Supplementary note 11, wherein
(Supplementary note 13) The recording medium according to any one of Supplementary notes 10 to 12, wherein
(Supplementary note 14) An image processing program causing a computer to execute:
(Supplementary note 15) The image processing program according to Supplementary note 14, causing the computer to execute
(Supplementary note 16) The image processing program according to Supplementary note 15, causing the computer to execute
(Supplementary note 17) The image processing program according to any one of Supplementary notes 14 to 16, causing the computer to further execute
(Supplementary note 18) An image processing program for realizing the image processing method of any one of Supplementary notes 6 to 9.
Although the invention of the present application has been described above with reference to example embodiments, the present invention is not limited to the above example embodiments. Various changes can be made to the configuration and details of the present invention that can be understood by those skilled in the art within the scope of the present invention.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2020/028065 | 7/20/2020 | WO |