This invention relates generally to computer vision, and more particularly to segmenting objects from images, and more particularly to segmenting moving objects using foreground extraction.
Segmenting an image of a scene into foreground and background parts is a fundamental task in many computer vision applications. Typically, the foreground part corresponds to an object in the scene. After the object has been segmented from the background, it is easier to perform subsequent tasks on the object, such as object recognition, composition, pose estimation, and positioning.
Background Subtraction
If the background is stationary or slowly varying, then the background can be estimated by applying a low pass filter, such as a mean or median filter, over an image sequence. A Gaussian mixture model (GMM) can also be used to model intensities of pixels associated with the background. After this is done, background subtraction can be accomplished by detecting the pixels that do not conform to the model of the background. Known color-, shape-, or motion-based prior probabilities for the foreground can improve the background subtraction.
Background subtraction can work well in some environments, such as surveillance applications, in which the majority of the pixels are associated with a static background, while a small number of pixels are associated with foreground objects that are moving. In scenarios where the camera is moving, e.g., a camera mounted on a robotic arm, the background pixels also change in intensity due to camera motion, and therefore conventional background subtraction cannot be used.
Intensity-Based Segmentation
Pixel intensities define appearance properties, such as color and texture, in the image. If the statistical properties of the foreground and background are sufficiently different, then an affinity model and a graph-based method can be used to segment the foreground from the background. Most graph-based methods define a graph in which vertices represent pixels, and the vertices are connected by edges. The affinity, or weight, of each edge depends on the locations and intensities of the two pixels that the edge connects. In general, adjacent pixels have a larger affinity, as do pixels with the same intensity properties. The spatial dependence is typically enforced using a Gaussian function on the Euclidean distance between the corresponding pixel locations. Similarly, color- and texture-based dependence is enforced by using a Gaussian or similar weighting function over the difference in color or texture between the corresponding pixels. These and other costs are multiplied to compute an edge weight for each of the edges in the graph. A cost function can be used to find a “cut” on the graph to separate foreground and background pixels. The cost can be normalized to avoid small segments. This leads to a formulation called normalized cuts. Intensity-based segmentation does not work well on images that have a homogeneous appearance.
Model-Based Segmentation
In model-based segmentation, the appearance properties of the foreground are known a priori. This type of segmentation is frequently used for common objects, such as vehicles and people. Model-based approaches depend on the availability of an accurate model of the object(s) to be segmented from the background. These prior models of the foreground objects can be manually prepared, or determined by machine learning that identifies features that distinguish the objects from the rest of the image.
In general, prior art segmentation methods do not work well with dynamic scenes, homogenous scenes, or images acquired by a moving camera.
The embodiments of the invention provide a method and system for segmenting a set of images into foreground and background parts. Note that “foreground” refers to pixels that are associated with a moving object, while “background” refers to pixels that are associated with the remainder of the scene.
In one embodiment, the foreground corresponds to a moving object in the scene, and the images are acquired by a camera having motion that is substantially identical to the motion of the object. For example, the camera and object are coupled to each other, either physically or logically. The segmentation uses statistical inference on properties of the pixels.
The invention is based on the following idea. Because the camera is moving along with the object, in the frame of reference of the camera, the object appears relatively motionless while the rest of the scene moves due to the camera motion. If the camera acquires multiple images during the motion, then the image features, e.g., color or intensity, of foreground pixels are relatively static, while the features of the background pixels change dynamically over time. If the camera moves during the exposure time of a single image, then in the resulting image, foreground pixels appear sharp while background pixels suffer from motion blur. This is in contrast with prior art background subtraction, in which the foreground generally has motion while the background remains static.
Conventional approaches to segmenting foreground pixels from background pixels typically use background subtraction or variations thereof. To use background subtraction, it is assumed that the background is stationary or static while the foreground is moving or dynamic. Since the foreground is moving in the camera's frame of reference, further processing such as registration is typically necessary even after the background has been subtracted.
In contrast, the invention uses foreground extraction. In our method, the background is assumed to be dynamic or moving with respect to the camera, and the foreground is static or stationary with respect to the camera. As a result, the extracted foreground can be directly used for higher-level computer vision tasks without the need for further low-level processing such as registration.
Because the object is stationary with respect to the camera, the features obtained from the object are not affected by motion of the object and the camera, making the features more reliable, more distinct, and more useful for further computer vision tasks such as pose estimation and object recognition.
It is assumed that the object to be segmented from the background is moving with some velocity (speed and direction), which can vary over time. In order to move the camera with the same velocity as the object, the camera can be physically or logically connected to the object, or the motion of the object can be measured.
The invention can be used in industrial applications where objects are moved by robot arms, by assembly line conveyor systems, or by other mechanical means such as dollies or tracks.
The invention enables the correct positioning of the object as the object is manipulated by automated equipment. For example, if the object is a vehicle part to be painted or drilled, the invention can be used in dynamically orienting the object or the tool for the current processing step. Similarly, the part can be positioned for a next processing step, such as assembly with other parts.
The scene includes an object 125. The system also includes a camera 130 having an output connected to a processor 140. The processor includes input/output interfaces and memories as known in the art. The processor performs steps of a method 200 according to embodiments of the invention.
In an example industrial application, the robot picks parts from a bin 180 or from a moving conveyor belt 190. In one embodiment, the camera is arranged on a robot arm 150. The robot arm includes end effector 160 for holding the object. Foreground extraction is performed on the object that is held by the end effector. We are interested in determining the pose of the object that is held by the end effector, after the object has been picked up, so that the object can be positioned into the correct pose for the current or next stage of the industrial process.
The movement of the object is directly coupled to the motion of the camera. The robot arm is moved such that a relative difference between a motion of the object and a motion of the camera is substantially zero. The motion of the object can be known or unknown. It should be noted that the camera could be mounted on any part of the robot arm, as long as the relative motion between the camera and the object is substantially zero. It should also be noted that the invention can work for a single image when there is camera motion during the exposure time of the image.
In another embodiment, the camera is mounted on a first robotic arm, the object is held by an end effector mounted on a second robotic arm, and the arms move in parallel and in unison. In this case, the motion of the object is typically known or measured, and foreground extraction is performed on the object that is held by the end effector.
Alternatively, foreground extraction can be performed on an object that is arranged on the moving conveyor belt. In this case, we are interested in estimating a pose of the object on the conveyor belt. As defined herein, the 6D-pose is a combination of the translational (x, y, z) location and angular (θ, φ, ψ) orientation of the object.
After the pose of the object is determined, the pose can be used for the current or next stage in the industrial process, which can involve picking up the object, positioning the object, or using a tool on the object, such as a spray gun, arc welder, drill, or grinder. The camera can be mounted on the robot arm, on the conveyor belt, or elsewhere, as long as the relative motion between the camera and the object is substantially zero.
There are numerous ways to ensure that the difference in relative motion is substantially zero. The motion of the camera and object can be directly coupled, the motion of the object can be known, or the motion of the object can be unknown but measured or estimated. It is understood that multiple cameras can also be used. The background 170 in the scene can be cluttered.
As shown in
When the camera and object move during an exposure time of one image, as shown in
In the case that the set of images only includes a single image, the embodiments of the invention assume that there is motion of the background with respect to the camera during the exposure time of the image so that the foreground object can be extracted from blurred background.
When the set of images includes a temporal sequence of images, the motion assumption for a single image can be relaxed. In fact, there does not need to be any apparent motion in a single image. In this case, the assumption is that there is apparent motion in the background across the sequence, and the observed motion across the set of images at different times can be used for our foreground extraction, because the foreground appears static across all the images, while the background changes dynamically from one image to the next.
In this case, instead of using spatial statistical properties of each pixel in an individual image, the method uses statistical properties of each pixel (x, y) 111 across the set of images, as the properties dynamically vary over time.
Let I(x, y, i) represent the intensity of a pixel (x, y) 111 in the ith image of the set of images. If multiple RGB channels are used, then the intensities represent color. We use I(x, y, •) to refer to the intensities of the pixel at location (x, y) 111 in the set of images. Similarly, let g(x, y, i) represent a feature or combination of features of pixel (x, y) in image i, and let g(x, y, •) refer to the feature values of pixel (x, y) 111 across all images in the set 110. Examples of such features can include, but are not limited to, any combination of intensity, spatial or temporal gradient of intensity, and spatial, temporal, or spatio-temporal frequency of intensity.
S is a set of all of the pixels in the set of images. It is desired to partition the set S into a set F 231 of the “foreground” pixels and a set B 232 of the “background” pixels.
These sets are related as follows:
For each pixel associated with the object, one can consider the features at that pixel across the images in the set as being samples drawn from a distribution pf whose statistical properties 225 are known, estimated, or modeled. For pixels associated with the background, it can be the case that no simple model can be derived because background pixels correspond to parts of the scene that move with respect to the camera.
In many cases, where it is difficult or impossible to obtain a specific background model, the potentially arbitrary variation of features across the set of images for pixels associated with the background is much greater than the variation of the same features for the foreground pixels. Therefore, we apply the statistical method to this statistical difference between the distribution of features of pixels across the images in the set.
For example, the set of foreground pixels F 231 can be identified according to the statistical properties by considering the statistics of each pixel and performing the following hypothesis test H independently for each pixel:
Null Hypothesis H0: The features g(x, y, •) are drawn from a foreground distribution pf(x, y).
Alternative Hypothesis H1: The features g(x, y, •) are not drawn from the foreground distribution pf(x, y).
The hypothesis test is solved by using a threshold T on a likelihood that the pixel features are drawn from the foreground distribution pf(x, y):
(x,y) is in the set F if P(g(x,y,•)|F)>T
and
(x,y) is in the set B if P(g(x,y,•)|F)≦T,
where the threshold T can be determined using a desired false alarm rate (FAR). The FAR refers to a fraction of pixels that are erroneously associated with the foreground.
In a convenient special case, the feature values are defined as the intensities of the pixels, and the foreground distribution pf(x, y) is modeled as a normal (Gaussian) distribution whose variance is constant for all foreground pixels. That is, for each pixel (x, y) in the set F, its intensities I(x, y, •) are drawn from the normal distribution N(μf(x, y),σf2), where μf(x, y) is the mean intensity of this distribution, and σf2 is the variance of intensity, which models small variations in intensity due to factors such as sensor noise and quantization noise. In this case, the hypothesis test can be derived as follows.
The standard deviation s estimated from pixel intensities at a particular pixel (x, y) is
where M is the number of images used to compute this statistic, and m(x, y) is the estimated mean:
Then, the partitioning of pixels into sets F and B is performed according to:
(x,y) is in the set F if s(x,y)<T; and
(x,y) is in set the B if s(x,y)≧T.
It is also possible to do a “soft” segmentation probabilistically, using a Bayesian framework, instead of the binary (“hard”) segmentation described above. Another variation of the method partitions the images into smaller parts, and then performs the segmentation on each part independently.
The motion of the object and the camera can be arbitrarily fast and complex, as long as the relative motion between the camera and the object is substantially zero. In addition, the structure and textural properties of the scene and object need not be restricted.
Although the invention has been described with reference to certain preferred embodiments, it is to be understood that various other adaptations and modifications can be made within the spirit and scope of the invention. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the invention.
Number | Name | Date | Kind |
---|---|---|---|
4305130 | Kelley et al. | Dec 1981 | A |
4611292 | Ninomiya et al. | Sep 1986 | A |
4727471 | Driels et al. | Feb 1988 | A |
5608818 | Chini et al. | Mar 1997 | A |
5911767 | Garibotto et al. | Jun 1999 | A |
5915048 | Hill et al. | Jun 1999 | A |
5960125 | Michael et al. | Sep 1999 | A |
6222938 | Melen | Apr 2001 | B1 |
6252974 | Martens et al. | Jun 2001 | B1 |
6320624 | Ayer et al. | Nov 2001 | B1 |
6377269 | Kay et al. | Apr 2002 | B1 |
6728582 | Wallack | Apr 2004 | B1 |
7024054 | Cahill et al. | Apr 2006 | B2 |
7084900 | Watanabe et al. | Aug 2006 | B1 |
7177459 | Watanabe et al. | Feb 2007 | B1 |
7302096 | Kim | Nov 2007 | B2 |
7336814 | Boca et al. | Feb 2008 | B2 |
7454053 | Bryll et al. | Nov 2008 | B2 |
7664315 | Woodfill et al. | Feb 2010 | B2 |
7978925 | Souchard | Jul 2011 | B1 |
8073197 | Xu et al. | Dec 2011 | B2 |
8095237 | Habibi et al. | Jan 2012 | B2 |
8244402 | Wells et al. | Aug 2012 | B2 |
8456527 | Elangovan et al. | Jun 2013 | B2 |
20050244049 | Onishi et al. | Nov 2005 | A1 |
20060245635 | Ishikawa | Nov 2006 | A1 |
20070047834 | Connell | Mar 2007 | A1 |
20080069444 | Wilensky | Mar 2008 | A1 |
20090016636 | Kasashima et al. | Jan 2009 | A1 |
20090167928 | Asukabe et al. | Jul 2009 | A1 |
20100085486 | Chen et al. | Apr 2010 | A1 |
20100097476 | Marks | Apr 2010 | A1 |
20100302365 | Finocchio et al. | Dec 2010 | A1 |
20100303340 | Abraham et al. | Dec 2010 | A1 |
20110087360 | Chen et al. | Apr 2011 | A1 |
20110091065 | Chandrashekar et al. | Apr 2011 | A1 |
20110255775 | McNamer et al. | Oct 2011 | A1 |
20110292205 | Yuguchi et al. | Dec 2011 | A1 |
20120069198 | Steinberg et al. | Mar 2012 | A1 |
Number | Date | Country |
---|---|---|
06-018539 | Jan 1994 | JP |
08-044874 | Feb 1996 | JP |
2008080606 | Jul 2008 | WO |
Entry |
---|
Silvia Anton, Theodor Borangiu, Florin Daniel Anton ,, Recognizing Articulated Objects in Robotic Tasks , Control Engineering and Applied Informatics , CEAI, vol. 9, No. 2, pp. 42-47, 2007. |
Sai K. Vuppala, Sorin M. Grigorescu, Danijela Ristic, and Axel Graser, Robust Color Object Recognition for a Service Robotic Task in the System FRIEND II,Proceedings of the 2007 IEEE 10th International Conference on Rehabilitation Robotics, Jun. 12-15, Noordwijk, The Netherlands. |
Toru Kaneko, “Seeking Applications of Image Processing Technologies.” vol. 33, No. 44. ME2009-178 (Oct. 2009) p. 11-16. |
Number | Date | Country | |
---|---|---|---|
20110141251 A1 | Jun 2011 | US |