The present application claims priority to Japanese Application Number 2019-084349, filed Apr. 25, 2019, the disclosure of which is hereby incorporated by reference herein in its entirety.
The present invention relates to an image processing apparatus, an image processing method, and a robot system.
There is known a robot system configured to recognize an object by a camera and to handle the recognized object by a robot. For example, Japanese Unexamined Patent Publication (Kokai) No. 2016-185573A discloses a robot system including a target object selection unit configured to select a target object; a proximity state determination unit configured to determine whether another object is disposed in proximity to the target object; an avoidance vector determination unit configured to determine an avoidance vector such that no interference with the object occurs; and a picking path correction unit configured to generate a corrected path which is obtained by correcting a picking path, based on the avoidance vector. A system disclosed in Japanese Unexamined Patent Publication (Kokai) No. 2013-186088A performs three-dimensional position/attitude measurement of a target object by using a first sensor configured to acquire two-dimensional (2D) information or three-dimensional (3D) information of the target object and a second sensor configured to acquire 2D information or 3D information of the target object.
When a two-dimensional camera which acquires a two-dimensional (2D) image is used for recognition of an object, a contour of the object is extracted from the 2D image and the contour is used for separation of the object, and there may be a case in which the contour of the object is not properly extracted due to an influence of a pattern of a surface of the object (e.g. a packing tape attached to a surface of a cardboard box that is the object), and erroneous recognition occurs. On the other hand, there is a method of extracting a contour of an object without an influence of a surface of the object, by using a three-dimensional camera for recognition of the object, the 3D camera being capable of acquiring a distance image representative of a distance to the object. However, in such a case that a plurality of objects are arranged close to each other, it may be possible that, with a distance image of a low resolution, the objects are unable to be recognized by separating the objects by a narrow gap between the objects.
According to one aspect of the present disclosure, an image processing apparatus includes a two-dimensional image storage unit configured to store a plurality of two-dimensional image data captured by photographing an identical imaging target object under different exposure conditions; a distance image storage unit configured to store distance image data representative of distance information depending on a spatial position of the imaging target object, the distance image data including a pixel array of a known relationship to a pixel array of the two-dimensional image data; a pixel extraction unit configured to extract, among a plurality of pixels in each of the two-dimensional image data, a first pixel at which a difference in brightness between identical pixels is less than a predetermined value; and a distance image adjusting unit configured to specify a second pixel of the distance image data at a position corresponding to the first pixel in the pixel array, and to set the second pixel as a non-imaging pixel in the distance image data.
According to another aspect of the present disclosure, a robot system includes a robot; a robot controller configured to control the robot; and the above-described image processing apparatus, wherein the robot controller is configured to cause the robot to handle the imaging target object, based on the distance image data acquired as a result of the distance image adjusting unit setting the second pixel as the non-imaging pixel.
According to still another aspect of the present disclosure, an imaging processing method includes storing a plurality of two-dimensional image data captured by photographing an identical imaging target object under different exposure conditions; storing distance image data representative of distance information depending on a spatial position of the imaging target object, the distance image data including a pixel array of a known relationship to a pixel array of the two-dimensional image data; extracting, among a plurality of pixels in each of the two-dimensional image data, a first pixel at which a difference in brightness between identical pixels is less than a predetermined value; and specifying a second pixel of the distance image data at a position corresponding to the first pixel in the pixel array, and setting the second pixel as a non-imaging pixel in the distance image data.
The object, features and advantages of the present invention will be more clearly understood by the description below of an embodiment relating to the accompanying drawings. In the accompanying drawings:
Hereinafter, an embodiment of the present disclosure will be described with reference to the accompanying drawings. Corresponding constituent elements are denoted by the same reference numerals throughout the drawings. These drawings use different scales as appropriate to facilitate understanding. The mode illustrated in each drawing is one example for carrying out the present invention, and the present invention is not limited to the modes illustrated in the drawings.
The robot system 100 further includes an image acquisition apparatus 50 which is connected to the image processing apparatus 30. The image acquisition apparatus 50 includes a function as a three-dimensional (3D) camera, which acquires a distance image representative of distance information depending a spatial position of the imaging target object W, and a function as a two-dimensional (2D) camera, which acquires a two-dimensional (2D) image of the imaging target object W with an identical pixel array to a pixel array of the distance image. For example, the image acquisition apparatus 50 may include a light source, a projector which projects pattern light, and two cameras disposed on both sides of the projector with the projector being interposed, and may be configured to photograph an object, on which the pattern light is projected, by the two cameras disposed at different positions, and to acquire three-dimensional (3D) position information of the object by a stereo method. In this case, the image processing apparatus 50 can acquire a distance image and a 2D image which have an identical pixel array, or a distance image and a 2D image which have pixel arrays of a known correspondence relation. As a method for acquiring 3D position information of the object, other kinds of methods may be used. The image processing apparatus 50 is disposed at a known position in a working space where the robot 10 is disposed, and photographs the imaging target object W from above. Note that the image processing apparatus 50 may be attached to a wrist portion at an arm tip end of the robot 10.
In the present embodiment, as the image acquisition apparatus 50, a configuration in which one camera acquires both the distance image and the 2D image is adopted. However, the embodiment is not limited to this, and the image acquisition apparatus 50 may be configured such that a 3D camera that acquires the distance image of the object and a 2D camera that acquires the 2D image are separately disposed in the robot system 100. In this case, the correspondence relation between the pixel array of the distance image and the pixel array of the 2D image is calibrated in advance and set in a known state.
The robot controller 20 causes the robot 10 to handle the imaging target object W, based on the distance image data which is adjusted by the image processing apparatus 30. For example, as illustrated in
In
An explanation is given of a problem of erroneous recognition, which may possibly occur when an object is recognized by using a 2D image or a distance image. As illustrated in
Consideration is now given to the case in which the cardboard boxes W1 and W2 are individually recognized by pattern matching by using the 2D image 201 acquired as illustrated in
The image processing apparatus 30 according to the present embodiment is configured to solve the above-described problem which may occur when the image recognition of the object is performed by using the distance image. Referring to
In step S3, among the pixels in each of a plurality of 2D image data, a first pixel at which a difference in brightness between identical pixels is less than a predetermined value is extracted. For example, the extraction of the first pixel is performed as follows.
(1) The degree of variation in brightness relative to the variation in exposure time is calculated with respect to all pixels between the 2D image 211 and the 2D image 212.
(2) The degree of variation in brightness relative to the variation in exposure time is calculated with respect to all pixels between the 2D image 212 and the 2D image 213.
(3) The mean value of the above (1) and (2) is calculated with respect to each pixel.
(4) A pixel, at which the “degree of variation in brightness relative to the variation in unit exposure time” calculated in the above (3) is less than a predetermined value, is extracted.
By the process of the above (1) to (4), the first pixel, which constitutes the image of the part of the gap G1, can be extracted. The “predetermined value” may be set by various methods, such as a method of setting a fixed value, for instance, an experimental value or an empirical value, or a method of setting the “predetermined value”, based on the “variation in brightness relative to the variation in exposure time” for each pixel acquired by the above (3). In the case of the latter, there may be a method in which a value calculated by subtracting a certain value from a maximum value in all pixels of the “variation in brightness relative to the variation in exposure time”, or a value calculated by subtracting a certain value from a mean value in all pixels of the “variation in brightness relative to the variation in exposure time”, is set as the “predetermined value”. Note that the extraction of the first pixel can be performed if there are two 2D images captured by photographing an identical imaging target object W under different exposure conditions.
In step S4, the image processing apparatus 30 specifies, in the pixel array in the distance image 302, a second pixel of distance image data existing at a position corresponding to the first pixel extracted in the 2D image in step S3, and sets the second pixel as a non-imaging pixel in the distance image data. To set the second pixel as the non-imaging pixel means that the second pixel is not set as a pixel representative of distance information. For example, the pixel value of the second pixel may be set to zero, or to some other invalid value which is not representative of distance information. Thereby, in the distance image 302, the objects can be separated by the part of the gap G1, and the objects can exactly be recognized. The process in step S4 will be described with reference to
Next, referring to
Next, the image processing apparatus 30 searches, in the 2D images, a pixel at which the degree of variation in brightness relative to the variation in exposure time is less than a predetermined value (step S105). The image processing apparatus 30 recognizes the searched pixel as a pixel (first pixel) of the part corresponding to the gap G1 between the cardboard boxes W1 and W2 (step S106). Next, the image processing apparatus 30 sets, as a non-imaging pixel, a pixel (second pixel) at a position corresponding to the gap G1 in the distance image acquired in step S101 (step S107).
Next, the object recognition unit 21 of the robot controller 20 recognizes the cardboard box W1 or W2 in the distance image by using model data of the cardboard box W1, W2 (step S108). The model data for object recognition is stored in a storage device (not illustrated) of the robot controller 20. Next, the picking operation execution unit 22 of the robot controller 20 calculates the position of the cardboard box W1 or W2 in a robot coordinate system, based on the position of the cardboard box W1 or W2 recognized in the distance image, and the position of the image acquisition apparatus 50 in the robot coordinate system. Based on the calculated position of the cardboard box W1 or W2 in the robot coordinate system, the picking operation execution unit 22 executes the operation of moving the robot 10 and individually grasping and picking the cardboard box W1 or W2 by the grasping device 15 (step S109).
The above-described method, which is a method by extracting, among the pixels in each of the 2D image data captured by photographing an identical imaging target object under different exposure conditions, the first pixel at which the difference in brightness between identical pixels is less than the predetermined value, can be used for, in addition to the extraction of a gap between objects, the extraction of a pixel which holds imaging information of a space from which reflective light of illumination light does not easily return to the camera, such as a hole, a slit or the like formed in an object.
As described above, according to the present embodiment, an object can be recognized with high precision from a distance image.
Although the embodiment of the present disclosure was described above, it is understood by a skilled person that various modifications and changes can be made without departing from the scope of disclosure of Claims which will be stated below.
The configuration of the robot system illustrated in
In the above-described embodiment, the method in which, among the pixels in each of 2D image data, the first pixel at which the difference in brightness between identical pixels is less than the predetermined value is extracted by using a plurality of 2D image data captured by photographing an identical imaging target object under different exposure conditions, and in which the second pixel of distance image data at a position corresponding to the first pixel is set as the non-imaging pixel, can also be expressed as a method in which the information of the distance image is supplemented by using the information of the pixel at which the difference in brightness between identical pixels is less than the predetermined value on the 2D image data.
When step S3 (the extraction of the first pixel at which the difference in brightness between identical pixels is less than the predetermined value) in the image process of
Number | Date | Country | Kind |
---|---|---|---|
JP2019-084349 | Apr 2019 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
5386228 | Okino | Jan 1995 | A |
6453069 | Matsugu | Sep 2002 | B1 |
6988610 | Fromme | Jan 2006 | B2 |
7599533 | Nishimura | Oct 2009 | B2 |
7672517 | Buscema | Mar 2010 | B2 |
7701455 | Tsao | Apr 2010 | B2 |
7756299 | Higaki | Jul 2010 | B2 |
8174611 | Estevez | May 2012 | B2 |
8340464 | Watanabe | Dec 2012 | B2 |
8459073 | Isei | Jun 2013 | B2 |
8654219 | Im | Feb 2014 | B2 |
8699821 | Orr, IV | Apr 2014 | B2 |
8730318 | Nakahira | May 2014 | B2 |
8780113 | Ciurea | Jul 2014 | B1 |
9113142 | Boisson | Aug 2015 | B2 |
9247153 | Umezu | Jan 2016 | B2 |
9270902 | Watanabe | Feb 2016 | B2 |
9275464 | Levy | Mar 2016 | B2 |
9294695 | Kageyama | Mar 2016 | B2 |
9412172 | Sorkine-Hornung | Aug 2016 | B2 |
9582888 | Ramachandra | Feb 2017 | B2 |
9832404 | Ishiga | Nov 2017 | B2 |
10059002 | Miyatani | Aug 2018 | B2 |
10127622 | Nishitani | Nov 2018 | B2 |
10157495 | Bendall | Dec 2018 | B2 |
10198792 | Lai | Feb 2019 | B2 |
10212408 | Pappas-Katsiafas | Feb 2019 | B1 |
10292321 | Neitemeier | May 2019 | B2 |
10313569 | Shiramizu | Jun 2019 | B2 |
10339668 | Takizawa | Jul 2019 | B2 |
10380911 | Hsu | Aug 2019 | B2 |
10434654 | Namiki | Oct 2019 | B2 |
10477175 | Ogasawara | Nov 2019 | B2 |
10498963 | Sorgi | Dec 2019 | B1 |
10502556 | Meng | Dec 2019 | B2 |
10582180 | Nobayashi | Mar 2020 | B2 |
10863105 | Rosengaus | Dec 2020 | B1 |
10869010 | Ogasawara | Dec 2020 | B2 |
10917543 | Ramirez Luna | Feb 2021 | B2 |
10986267 | Aldridge | Apr 2021 | B2 |
20040125206 | Lohmann | Jul 2004 | A1 |
20050162644 | Watanabe | Jul 2005 | A1 |
20090027509 | Giesen | Jan 2009 | A1 |
20100033619 | Kimura | Feb 2010 | A1 |
20140176761 | Takizawa | Jun 2014 | A1 |
20150371398 | Qiao | Dec 2015 | A1 |
20160155235 | Miyatani | Jun 2016 | A1 |
20160175964 | Penoyer | Jun 2016 | A1 |
20160279809 | Nakajima | Sep 2016 | A1 |
20170257540 | Shiramizu | Sep 2017 | A1 |
20180023947 | Meng | Jan 2018 | A1 |
20180084708 | Neitemeier | Mar 2018 | A1 |
20190033067 | Yamagami | Jan 2019 | A1 |
20200021743 | Yoshida | Jan 2020 | A1 |
20200175352 | Cha | Jun 2020 | A1 |
20200226729 | Zhang | Jul 2020 | A1 |
Number | Date | Country |
---|---|---|
2013-186088 | Sep 2013 | JP |
2016-185573 | Oct 2016 | JP |
Entry |
---|
Birchfield et al., “Depth Discontinuities by Pixel-to-Pixel Stereo”, (pp. 269-293) (Year: 1999). |
Number | Date | Country | |
---|---|---|---|
20200342563 A1 | Oct 2020 | US |