The present application claims priority to and incorporates by reference the entire contents of Japanese Patent Application No. 2014-125656 filed in Japan on Jun. 18, 2014.
1. Field of the Invention
The present invention relates to an information processing device, a computer-readable recording medium, and an information processing method.
2. Description of the Related Art
A technique is known which obtains a three-dimensional model of an entire observation object from a range image of the observation object acquired by a range sensor. Range information included in the range image is range information measured from a single direction. Therefore, when obtaining a three-dimensional model of the entire observation object, images of the observation object are captured from a plurality of different directions and range information corresponding to each image capturing direction is acquired. In general, local feature amounts of respective range images are defined, portions of range images that have a similar feature amount are associated with each other, and relative alignment of the range images acquired from a plurality of directions is performed, so that the three-dimensional model of the entire observation object is generated.
JP 5253066 B1 discloses a position and posture measurement device that stabilizes and streamlines position and posture detection processing by appropriately selecting features used to calculate position and posture from among features extracted from three-dimensional model data of an observation object body.
In the case of this position and posture measurement device, a plurality of geometric features based on geometric information of the observation object body are extracted by drawing the three-dimensional model data which represents a surface shape of the observation object body. Further, a reference image where the position and the posture of an image capturing device with respect to the observation object body have been calculated is searched for image features corresponding to the plurality of geometric features, and geometric features of which corresponding image feature is detected are selected from the plurality of extracted geometric features. Then, the position and the posture of the image capturing device with respect to the observation object body are calculated by associating the selected geometric features with an image of the observation object body in an input image.
Thereby, even when the features extracted from the three-dimensional model data of the observation object body are largely different from features that can be extracted from the observation object body of which image is captured in a captured image, it is possible to stably estimate the position and the posture.
However, there is a problem that a conventional alignment method is difficult to estimate a corresponding point when the amount of overlapping portions between range images is small and when there is no characteristic portion in a three-dimensional structure.
Further, in the case of the position and posture measurement device disclosed in JP 5253066 B1, a plurality of geometric features based on geometric information of the observation object body are extracted by drawing the three-dimensional model data which represents a surface shape of the observation object body. Therefore, there is a problem that it is very difficult to align an unknown object for which no three-dimensional model is prepared in advance.
Therefore, there is a need for an information processing device, a computer-readable recording medium, and an information processing method, which can accurately align a plurality of range images.
It is an object of the present invention to at least partially solve the problems in the conventional technology.
There is provided an information processing device that includes a range image acquisition unit that acquires at least two range images of an observation object, a color image acquisition unit that acquires color images of the observation object, which respectively correspond to the range images, a feature portion detection unit that detects feature portions from the acquired color images, a calibration unit that performs calibration processing that associates each pixel of the color image with each point of the range image, which corresponds to each pixel, and generates calibration information that indicates each point corresponding to each pixel, and an alignment processing unit that performs alignment of the range images so that the detected feature portions overlap with each other by using the detected feature portions and the calibration information.
The above and other objects, features, advantages and technical and industrial significance of this invention will be better understood by reading the following detailed description of presently preferred embodiments of the invention, when considered in connection with the accompanying drawings.
Hereinafter, as an example, an image processing device, which is an embodiment to which an information processing device, an information processing program, and an information processing system are applied, will be described in detail with reference to the attached drawings.
First,
The stereo camera unit 1 is configured by two camera units including a first camera unit for left eye and a second camera unit for right eye, which are incorporated in parallel with each other. Each camera unit includes a lens, an image sensor, and a sensor controller. The image sensor is, for example, a CCD image sensor or a CMOS image sensor. The CCD is an abbreviation of “Charge Coupled Device”. The CMOS is an abbreviation of “Complementary Metal-Oxide Semiconductor”. The sensor controller performs exposure control of the image sensor, image read control, communication with an external circuit, transmission control of image data, and the like.
The HDD 5 (or ROM 3 or RAM 4) stores an image processing program which is an example of the information processing program. The image processing program includes an alignment processing program for performing alignment processing of range images obtained by capturing images of an observation object from different directions by the stereo camera unit 1. Further, the image processing program includes a mesh generation processing program that converts each aligned range image into a mesh and generates a three-dimensional model. Further, the image processing program includes a texture mapping processing program that maps a predetermined texture on the three-dimensional model generated by the mesh generation processing program.
Luminance image data of a captured image that is captured by each camera unit of the stereo camera unit 1 is written to the RAM 4 through the bus line 8. The CPU 2 generates a parallax image (range image) by performing, for example, a gamma correction, a skew correction (paralleling the left and right images with each other), and parallax calculation by block matching, which are processing requiring real-time property, on the luminance image data stored in the RAM 4, and writes the parallax image to the RAM 4 again. Further, the CPU 2 controls an operation of an entire drawing processing device and further performs and controls alignment processing, mesh generation processing, texture mapping processing, and the like of each range image written to the RAM 4 according to a drawing processing program stored in the ROM 3.
Next, to create a three-dimensional model of an observation object, it is necessary to overlap range images captured from different directions. To overlap the range images, it is necessary that relative positional relationship between the range images is known. Processing to derive the relative positional relationship is the alignment processing. For the alignment processing, the image processing device of the embodiment acquires a color image and a range image from the same direction and calculates an image capturing direction of the observation object by using the acquired color image. Then, the image processing device of the embodiment performs alignment of each range image by using a calculated orientation of the object. Thereby, even when a portion where the range images overlap with each other is small, or even when there is no geometric feature, it is possible to perform the alignment by using a relative direction between the range images obtained from posture estimation.
The color image acquisition unit 11 acquires a color image (RGB image) of the observation object. The range image acquisition unit 13 acquires a range image of the observation object. Regarding the color image acquired by the color image acquisition unit 11 and the range image acquired by the range image acquisition unit 13, the color image and the range image which are acquired from the same direction are linked (associated) with each other and stored in the RAM 4. The posture calculation unit 12 calculates an image capturing direction of the observation object by using the color image acquired by the color image acquisition unit 11.
The noise removing unit 14 removes noise such as high-frequency noise from the range image acquired by the range image acquisition unit 13. The alignment processing unit 15 performs alignment of the range image from the posture (the orientation of the observation object) calculated by the posture calculation unit 12. The mesh generation unit 16 converts each aligned range image into a mesh to generate a three-dimensional model. The texture mapping unit 17 maps a predetermined texture on the three-dimensional model generated by the mesh generation processing.
When the CPU 2 functions as the posture calculation unit 12, the CPU 2 functions as a feature point detection unit 21, a feature amount calculation unit 22, and a likelihood calculation unit 23 as illustrated in
In summary, the posture calculation unit 12 detects feature points of an inputted color image, calculates a feature amount at each feature point, and defines a positional relationship between the feature points and the feature amounts as an original property. Then, the posture calculation unit 12 calculates the image capturing direction of the observation object corresponding to the inputted color image by calculating an orientation (posture) where the likelihood is greatest for the original property obtained from the inputted color image by using a learning model.
How to calculate the orientation (posture) of the object is disclosed in detail in S. Savarese and L. Fei-Fei, “3D generic object categorization, localization and pose estimation”, IEEE Intern. Conf. in Computer Vision (ICCV), Brazil, October, 2007.
A method for calculating an orientation of a vehicle at a degree of accuracy of about 90% by creating a learning model by 80 images of the vehicle captured from eight directions is disclosed in Nadia Payet, Sinisa Todorovic, “From contours to 3D object detection and pose estimation”, ICCV 2011: 983-990.
Next, the flowchart in
The range image is an image in which distance information to the observation object is stored in each pixel. It is possible to generate point group data by mapping range image data on an xyz coordinate system. In the ICP, alignment of each range image is performed by minimizing the sum of the distances between the point groups in the processing illustrated in the flowchart of
Specifically, in step S1 in the flowchart of
Next, in step S3, the alignment processing unit 15 overlaps end points of respective point groups of the first range image and the second range image based on the recognized orientations of the range images. For example, it is assumed that when the posture calculation unit 12 calculates the orientation (posture) of the observation object from the color image corresponding to the first range image and the color image corresponding to the second range image, it is calculated that the first range image is a range image obtained by capturing an image of the observation object from the front direction and the second range image is a range image obtained by capturing an image of the observation object from the right direction. The most front right point in the point group of the first range image obtained by capturing an image of the observation object from the front direction and the most front left point in the point group of the second range image obtained by capturing an image of the observation object from the right direction are substantially coincident with each other. In step S3, the alignment processing unit 15 performs processing to overlap end points of point groups of range images which can be overlapped with each other in this manner.
Next, in step S4, the alignment processing unit 15 detects the closest point to each point of the point group A of the first range image from the point group B of the second range image. Further, in step S5, the alignment processing unit 15 calculates the sum of the distances between closest points of the first range image and the second range image. In step S6, the alignment processing unit 15 compares the sum of the distances between closest points of the first range image and the second range image with a predetermined threshold. Then, in step S6, the alignment processing unit 15 determines whether or not the sum of the distances between closest points of the first range image and the second range image is smaller than or equal to the predetermined threshold. The fact that the sum of the distances between closest points is smaller than or equal to the predetermined threshold (step S6: Yes) means that the first range image and the second range image roughly overlap with each other at a position corresponding to each image capturing direction (the alignment is completed). Therefore, the alignment processing unit 15 ends the processing of the flowchart in
On the other hand, the fact that the sum of the distances between closest points is greater than the predetermined threshold (step S6: No) means that the first range image and the second range image do not overlap with each other at a position corresponding to each image capturing direction. Therefore, the alignment processing unit 15 proceeds to step S7 and calculates a conversion parameter R where the sum of the distances between closest points is smallest. The conversion parameter R is a parameter indicating a rotation angle of the first range image or the second range image. For example, when alignment of the first range image obtained by capturing an image of the observation object from the front direction and the second range image obtained by capturing an image of the observation object from the right direction is performed, the conversion parameter R does not represent a rotation of 90 degrees or more. This is because when a rotation of 90 degrees or more is performed, the calculation result of the orientation (posture) of the observation object, which is the front direction and the right direction, in the posture calculation unit 12 is an error.
Therefore, in step S8, the alignment processing unit 15 determines whether or not the calculated conversion parameter R is smaller than or equal to a predetermined threshold (90 degrees). The fact that the calculated conversion parameter R is greater than the predetermined threshold (90 degrees) (step S8: No) means that the calculated conversion parameter R is an error as described above, so that the alignment processing unit 15 ends the processing of the flowchart in
On the other hand, the fact that the calculated conversion parameter R is smaller than or equal to the predetermined threshold (90 degrees) (step S8: Yes) means that the calculated conversion parameter R is a conversion parameter R that can cause the first range image and the second range image to be close to each other so that the first range image and the second range image can be roughly overlapped with each other at a position corresponding to each image capturing direction. Therefore, in step S9, the alignment processing unit 15 rotates the point group A of the first range image by the calculated conversion parameter R and returns to step S4.
Thereafter, the alignment processing unit 15 repeatedly performs the step S4 to step S9 to perform alignment of the first range image with respect to the second range image while gradually rotating the point group A of the first range image by the conversion parameter R calculated each time. Then, in step S6, when the alignment processing unit 15 determines that the sum of the distances between closest points is smaller than or equal to the predetermined threshold (step S6: Yes), it means that the first range image and the second range image roughly overlap with each other at a position corresponding to each image capturing direction (the alignment is completed). Therefore, the alignment processing unit 15 ends the processing of the flowchart in
A specific example of such alignment processing will be described. For example, a case is considered in which the ICP is performed on point groups obtained by measuring curved surfaces. A diagram in
In the same manner,
Next, the mesh generation unit 16 illustrated in FIG. 2 generates a three-dimensional model by converting the range images aligned in this way into a meth as one range image. Then, the texture mapping unit 17 attaches a corresponding texture to the generated three-dimensional model and outputs the three-dimensional model.
As obvious from the above description, when the image processing device of the first embodiment performs alignment of the first range image and the second range image, the image processing device acquires a color image captured from the same image capturing direction as that of the first range image. Further, the image processing device acquires a color image captured from the same image capturing direction as that of the second range image. The posture calculation unit 12 calculates the image capturing direction of each color image (=the image capturing direction of each range image) by referring to a learning model stored in the HDD 5. The alignment processing unit 15 calculates a conversion parameter R that causes the sum of the distances between closest points of the first range image and the second range image to be smaller than or equal to a predetermined threshold. Then, the alignment processing unit 15 performs alignment of the range images by translating or rotating the first range image (or the second range image) by using the calculated conversion parameter R.
Even when the amount of overlapping portions between the range images is small or there is no geometric feature point in the range images, it is possible to more accurately perform the alignment using a relative direction between the range images obtained from the posture estimation by calculating the image capturing direction of the observation object by using the acquired color image.
Further, the alignment processing unit 15 performs the alignment by using the range image from which noise is removed by the noise removing unit 14. When noise is superimposed on the range image, the alignment is performed by using the range image which is partially distorted, so that it is difficult to perform accurate alignment. However, the image processing device of the embodiment performs the alignment after removing noise of each range image by the noise removing unit 14. Therefore, it is possible to perform more accurate alignment.
The image processing device of the embodiment as described above can be applied to, for example, a three-dimensional scanner device and an eye of an industrial robot. A three-dimensional model is required to obtain an output of a three-dimensional printer. The three-dimensional scanner device can be used as a means to generate a three-dimensional model. When performing alignment by acquiring range images from various directions by the three-dimensional scanner device, if a portion where the range images overlap with each other is small, it is difficult to accurately perform the alignment. However, the image processing device of the embodiment can perform accurate alignment even when a portion where the range images overlap with each other is small, so that it is possible to improve convenience of the three-dimensional scanner device.
When the image processing device of the embodiment is applied to an industrial robot, not only a color image, but also range images are used to correctly recognize a component. To correctly recognize the component as a three-dimensional model, it is necessary to acquire range images from a plurality of directions, and it is necessary to acquire the range images from various directions so that there are some overlaps. In a production line or the like, components flowing on the production line are usually determined in advance, so that it is easy to create a learning model and the image processing device of the embodiment will effectively function.
In the above description of the embodiment, a range image generated by performing parallax calculation of each captured image of the stereo camera unit 1 is used. As the range image, it is possible to use image information to which distance information detected by radar ranging or the like for each point of an image is added.
It is all right that a range image and a corresponding color image are stored in a memory, the posture calculation unit 12 reads the color image from the memory and calculates the orientation of the range image, and the alignment processing unit 15 performs alignment of each range image read from the memory by using the calculated orientation of the range image.
Next, an image processing device of a second embodiment will be described. The image processing device of the first embodiment converts an aligned range image into a meth. On the other hand, the image processing device of the second embodiment removes unnecessary points from an aligned range image and then converts the range image into a meth. The second embodiment described below is different from the first embodiment described above only in the point described above. Therefore, in the description below, only the difference between the first embodiment and the second embodiment will be described and redundant description will be omitted.
The alignment processing unit 15 generates one range image by aligning, for example, two range images as described above. Therefore, the generated range image includes unnecessary points when seen as a point image. Therefore, the unnecessary point removing unit 31 forms a range image as an integrated point image by removing points which are unnecessary when two aligned range images are seen as a point image.
Thereby, the mesh generation unit 16 can perform mesh generation processing without being affected by the unnecessary points and it is possible to obtain the same advantageous effect as that of the first embodiment described above.
Next, an image processing device of a third embodiment will be described. In the processing of step S8 in the flowchart illustrated in
In other words, in the case of the image processing device of the third embodiment, the posture calculation unit 12 calculates the orientation of each range image based on the color image in step S1. In this case, the posture calculation unit 12 does not calculate only one piece of information representing the orientation of each image, but calculates a plurality of orientations with a quantitative scale such as, for example, the likelihood of x direction is 90 and the likelihood of y direction is 80.
The alignment processing unit 15 first calculates the orientation of each range image by using the direction of which likelihood is the highest (in this case, the x direction). However, when the value of the conversion parameter R exceeds the threshold in step S8, the alignment processing unit 15 calculates the orientation of each range image by using the direction of which likelihood is the second highest (in this case, the y direction) in step S2.
Thereby, it is possible to recalculate the accurate orientation of each range image and to contribute to accurate alignment of each range image, and further it is possible to obtain the same advantageous effect as that of the first embodiment described above.
Next, an image processing device of a fourth embodiment will be described. In the description below of the image processing device of the fourth embodiment, the sections that operate in the same manner as those in each embodiment described above are denoted by the same reference symbols as those used in the description of each embodiment described above and the detailed description thereof will be omitted.
Each embodiment described above acquires a point group and an image at the same time and estimates the posture of the acquired image. Then, the embodiments obtains a relative angle from the posture estimation result and performs alignment on the premise that end points of point groups are coincident with each other. On the other hand, the image processing device of the fourth embodiment further improves the alignment accuracy by using a feature region detected from a color image (RGB image) acquired by the color image acquisition unit 11 for the alignment.
As an example, a normal vehicle includes door mirrors and tires. The door mirrors and the tires are feature regions of the vehicle. Therefore, in the fourth embodiment, many vehicle images are prepared and learned. Further, in the fourth embodiment, when the color image described above is a vehicle, a feature region (a color image feature region) of the vehicle such as a door mirror and a tire is detected by using a learning result. Further, in the fourth embodiment, a range image feature region corresponding to the color image feature region is detected from a range image that is an image of a large number of point groups acquired by the range image acquisition unit 13. In the fourth embodiment, when the alignment of range images is performed, the alignment is performed so that range image feature regions, such as door mirrors and tires, of respective range images overlap with each other.
In the case of each embodiment described above, the end points of the point groups are used for the alignment. However, in the case of the fourth embodiment, feature portions of an object to be aligned are used for the alignment. Thereby, the amount of information used for the alignment increases, so that it is possible to further improve the accuracy of the alignment.
The color image acquisition unit 11 acquires a color image (RGB image) of the observation object. The range image acquisition unit 13 acquires a range image of the observation object. The noise removing unit 14 removes noise such as high-frequency noise from the range image acquired by the range image acquisition unit 13. The mesh generation unit 16 converts each aligned range image into a mesh to generate a three-dimensional model. The texture mapping unit 17 maps a predetermined texture on the three-dimensional model generated by the mesh generation processing. The unnecessary point removing unit 31 forms a range image as an integrated point image by removing points which are unnecessary when two aligned range images are seen as a point image.
When the color image is, for example, a vehicle, the feature portion detection unit 41 detects feature portions (regions) such as tires and door mirrors which are generally included in a vehicle. Then, the feature portion detection unit 41 generates a feature portion labeling image in which a feature portion (region) is labeling-processed. Specifically, the feature portion detection unit 41 performs labeling processing that adds information, such as the same number, to distinguish pixels in the same region from pixels of another region, to the pixels in the same region (in the case of this example, pixels in a feature portion (region)), and generates the feature portion labeling image.
The calibration unit 42 compares the color image from the color image acquisition unit 11 with the range image which is a point group from the range image acquisition unit 13 and thereby detects points on the point group of the range image to which pixels of the color image correspond respectively (calibration of the color image and the range image (point group)). Then, the calibration unit 42 generates calibration information that indicates a correspondence relationship between each pixel of the color image and each point of the range image. The alignment processing unit 43 performs alignment processing of a plurality of range images (point images) acquired by the range image acquisition unit 13 by using the feature portion labeling image and the calibration information.
In
For example,
In step S12, the feature model generation unit 51 detects feature portions of the vehicle such as a windshield, door mirrors, and headlights as indicated by shaded regions in the diagrams in
Subsequently, in step S13, the feature model generation unit 51 determines whether or not a predetermined number of learning images are obtained. The feature model generation unit 51 repeatedly performs each processing of step S11 to step S13 until the predetermined number of learning images are obtained (step S13: No). When the feature model generation unit 51 determines that the predetermined number of learning images are obtained (step S13: Yes), the feature model generation unit 51 proceeds to step S14.
In step S14, the feature model generation unit 51 detects feature portions between a plurality of learned color images. In step S15, the feature model generation unit 51 integrates the detected feature portions to generate a feature model, stores the feature model in the HDD 5, which is an example of a feature model storage unit, and ends the processing of the flowchart in
Next, the likelihood calculation unit 52 illustrated in
In the flowchart of
Subsequently, in step S23, the likelihood calculation unit 52 generates a feature portion labeling image in which a feature portion of the color image is labeling-processed by using the detected feature model and supplies the feature portion labeling image to the alignment processing unit 43. Thereby, when the observation object of the supplied color image is a vehicle, a feature portion labeling image is generated in which, for example, tires and door mirrors which are feature portions of the vehicle are labeling-processed as illustrated in the diagrams in
Next, in the case of the image processing device of the fourth embodiment, a feature portion is detected from the color image, a portion in the point group of the range image which corresponds to the feature portion of the color image is detected, and alignment is performed so that feature portions of respective point groups overlap with each other. Therefore, the calibration unit 42 performs calibration processing that associates each pixel of the color image with each point of the point group of the range image which corresponds to each pixel and generates calibration information that indicates each point corresponding to each pixel.
There are various method to perform the calibration processing. As an example, images of a fixed rectangular parallelepiped are captured by an RGB camera and a depth sensor, respectively. Thereby, a color image of the rectangular parallelepiped is obtained from the RGB camera and an image of a point group (a point image) of the rectangular parallelepiped is obtained from the depth sensor. Then, projective transformation processing is performed on the color image or the point image so that the rectangular parallelepiped in the color image and the rectangular parallelepiped in the point image have the same shape. Thereby, it is possible to calculate points on the point image which correspond to each pixel of the color image. The calibration unit 42 performs the calibration processing that associates each pixel of the color image with each point of the range image by performing such projective transformation processing as an example. The calibration unit 42 performs the projective transformation processing on the color image or the point image, so that it is possible to calculate a point group of the range image that corresponds to a feature region of the color image.
The number of pixels of the color image is not necessarily the same as the number of points of the point image. In this case, it is difficult to associate a pixel with a point on a one-to-one basis. In this case, the calibration unit 42 performs the calibration processing so that a pixel and a point closest to each other are associated with each other.
Subsequently, the alignment processing unit 43 performs alignment of point groups of respective range images by using information of feature portions of images (the feature portion labeling images described above) and the calibration information between each pixel of the color image and each point of the range image. For example, when an object is a vehicle, it is assumed that the left door mirror is detected by the feature portion detection unit 41. When the left door mirror is detected in a plurality of range images, point groups in respective range images, which correspond to the left door mirror, overlap with each other. In this way, the alignment processing unit 43 performs alignment of point groups in respective range images by using feature portions.
The flowchart in
Subsequently, in step S32, the alignment processing unit 43 calculates a point group corresponding to the feature portion. For example, when the left door mirror is detected from the feature portion labeling images, the alignment processing unit 43 detects point groups on respective range images which correspond to the detected left door mirror.
Subsequently, in step S33, the alignment processing unit 43 aligns the range images so that the detected feature portions of respective range images overlap with each other (the feature portions are overlapped with each other). In the example described above, the alignment processing unit 43 aligns the range images so that the point groups of the left door mirror, which is a feature portion of each range image, overlap with each other. The alignment processing unit 43 detects feature portions as many as possible and aligns the range images so that all the detected feature portions overlap with each other.
Subsequently, in step S34, the alignment processing unit 43 performs alignment so that two point groups overlap with each other as a whole by using, for example, an algorithm of the ICP (Iterative Closest Point) disclosed in Besl, Paul J.; N. D. McKay (1992). “A Method for Registration of 3-D Shapes”. IEEE Trans. on Pattern Analysis and Machine Intelligence (Los Alamitos, Calif., USA: IEEE Computer Society) 14 (2): 239-256 and ends the processing of the flowchart in
In other words, for each point included in a point group of one range image provided as an input, the alignment processing unit 43 detects a closest point in a point group of another range image by using the ICP algorithm and defines these points as temporary corresponding points. The alignment processing unit 43 estimates rigid body conversion that minimizes the distances between the corresponding points. The alignment processing unit 43 estimates a motion that aligns point groups of two range images by repeating the corresponding point detection and the rigid body conversion estimation.
When the alignment is performed on the premise that end points of point groups are coincident with each other, the point groups of respective range images may not overlap with each other as illustrated in
In the image processing device of the fourth embodiment, the noise removing unit 14 illustrated in
The mesh generation unit 16 and the texture mapping unit 17 illustrated in
The unnecessary point removing unit 31 illustrated in
As obvious from the above description, the image processing device of the fourth embodiment acquires a range image of a point group and a color image captured from the same image capturing direction as that of the range image. The calibration unit 42 performs calibration processing on a range image and a color image and associates each point group of the range image with each pixel of the color image (obtains a correspondence relationship). The feature portion detection unit 41 generates in advance a feature model in which feature portions of the observation object are detected in advance and stores the feature model in the HDD. Further, the feature portion detection unit 41 generates a feature portion labeling image, in which a feature portion of the color image is labeling-processed, by using the feature model detected in advance. The alignment processing unit 43 performs alignment of range images of point groups by using the feature portion labeling image. For example, when an object is a vehicle, it is assumed that the left door mirror is detected by the feature portion detection unit 41. When the left door mirror is detected in a plurality of range images, point groups in respective color images, which correspond to the left door mirror, overlap with each other. The alignment processing unit 43 performs alignment of point groups in respective range images by using such feature portions.
In summary, the image processing device of the fourth embodiment captures a color image and a range image of a point group from the same direction, detects feature portions of the observation object (in the case of a vehicle, tires and door mirrors), and uses information of the feature portions for the alignment of the range image of a point group.
Thereby, even when a portion where point groups of respective range images overlap with each other is small, or even when there is no geometric feature (there is not so much geometric feature), it is possible to perform the alignment accurately. Therefore, it is possible to further improve the degree of accuracy of the alignment.
According to the present invention, there is an effect that a plurality of range images can be accurately aligned.
Although the invention has been described with respect to specific embodiments for a complete and clear disclosure, the appended claims are not to be thus limited but are to be construed as embodying all modifications and alternative constructions that may occur to one skilled in the art that fairly fall within the basic teaching herein set forth.
Number | Date | Country | Kind |
---|---|---|---|
2014-125656 | Jun 2014 | JP | national |