Processing information of two (or more) images of the same scene, which originate from cameras slightly separated from one another, can yield depth information for objects in the images. The processing of these images is computationally intensive and requires the handling of large amounts of data. Typically, the processing has substantial latency because it involves a series of calculations where one type of calculation is performed over an entire image before moving on to a next type calculation. In addition, the output arrays from these calculations are usually large, sometimes including for each position in the image, depth information, intensity information, and color information. It would be useful to reduce the output arrays to ease calculation requirements for subsequent processing. It would also be useful to enable real-time image processing by reducing the latency.
Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.
The invention can be implemented in numerous ways, including as a process, an apparatus, a system, a composition of matter, a computer readable medium such as a computer readable storage medium or a computer network wherein program instructions are sent over optical or electronic communication links. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. A component such as a processor or a memory described as being configured to perform a task includes both a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task. In general, the order of the steps of disclosed processes may be altered within the scope of the invention.
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.
An integrated image processor implemented on a substrate is disclosed. An input interface is configured to receive pixel data from two or more images. A pixel handling processor disposed on the substrate is configured to convert the pixel data into depth and intensity pixel data. In some embodiments, a foreground detector disposed on the substrate is configured to classify pixels as foreground or not foreground. In some embodiments, a projection generator disposed on the substrate is configured to generate a projection of the depth and intensity pixel data. In some embodiments, a segmentation processor disposed on the substrate is configured to generate a list of objects and their descriptions including height, extent, number, and/or color. In various embodiments, an application processor disposed on the substrate is configured to use results of the other processors to produced outputs for user applications such as count people, count cars, count objects, detect objects, track objects, identify obstacles, detect navigable paths for vehicles, and or identify tailgaters through a door.
In various embodiments, the source of the images (for example, a camera) does not change location between images and an object of interest does not change locations between images; the source of the images does not change location between images and an object of interest does change locations between images; the source of the images does change location between images and an object of interest does not change locations between images; and the source of the images does change locations between images and an object of interest does change locations between images.
In some embodiments, input interface 102, pixel handling processor 104, and foreground detector processor 106 are arranged in a pipeline architecture where each block is processing in parallel or at the same time. In some embodiments, only a subset of image information is required by each block so that, depending on the latency of each block and the size of the subset of information required by each block, a number of blocks are operating on pixel data from the same frame period at the same time, where a frame period is the time between frames of image data. In various embodiments, input interface 102, pixel handling processor 104, and foreground detector processor 106 are integrated circuits disposed on a common substrate or as part of a common integrated circuit package.
In some embodiments, the projection statistics are resealed. For example, a closer object to the imaging source of a given size will have a greater number of pixels representing it as compared to a farther object of the same given size. A correction to rescale the pixel counts for a given projection cell based on the distance to the image source can be calculated and applied for the 2-dimensional or 3-dimensional projection data set. In various embodiments, a threshold is applied before resealing, after resealing, or before and after resealing.
In some embodiments, input interface 202, pixel handling processor 204, and projection generator 206 are arranged in a pipeline architecture where each block is processing in parallel or at the same time. In some embodiments, only a subset of image information is required by each block so that, depending on the latency of each block and the size of the subset of information required by each block, a number of blocks are operating on pixel data from the same frame period at the same time. In various embodiments, input interface 202, pixel handling processor 204, and projection generator 206 are integrated circuits disposed on a common substrate or as part of a common integrated circuit package.
Foreground detector processor 306 inputs include depth information or disparity information, and/or validity information. Foreground detector processor 306 outputs include an indication of which pixels are part of the foreground and which pixels are not part of the foreground. The indication of which pixels are part of the foreground is indicated by a data array of one bit per pixel. In some embodiments, foreground detection comprises modeling the background. The classification of pixels into foreground or background comes from building a model of the background and updating that model over time. The background pixels are the pixels that remain relatively stable over a relatively long period of time. Projection generator 308 inputs include depth information or disparity information, and/or validity information. Projection generator 308 inputs include depth information or disparity information, an optional input mask, and/or validity information. Projection generator 308 outputs include one or more of the following: a 3-dimensional data set or a 2-dimensional projection of the 3D data.
Segmentation processor 310 inputs include the outputs of projection generator 308. Segmentation processor 310 calculations include smoothing, thresholding, and connecting elements of the 2-dimensional projection data set or 3-dimensional data set to create a list of objects and descriptors of those objects. For example, in a 2-dimensional projection data set parallel to the floor of a room, objects are segmented using a smoothing filter on the counts of neighboring cells, applying a threshold to the smoothed values to help in detecting object boundaries, and identifying locally connected regions remaining as objects of interest. Each object can be described in terms of spatial location 3D physical extent, color, etc. In some embodiments, one or more edge detection filters are used to identify boundaries around objects. In some embodiments, segmentation processor 310 detects which areas in the top or plan view projection represent an object (for example, a person).
Application processor 312 inputs include the outputs of the segmentation processor 310. In various embodiments, application processor 312 can count objects, classify objects, track objects, or detect patterns in the objects. For example, application 312 can count cars by identifying objects with an extent that is large enough to be a vehicle on top of a road surface and counting each object as a car. In another example, application 312 can classify the vehicles to differentiate between cars and trucks using the different descriptors of the objects—cars are not as tall as, are narrower than, and are less long than a truck. As another example, in situations where a card key swipe is required to enter through a door, identifying the pattern of two people moving through the door on one card swipe can indicate that there is breach of security. In some embodiments, application processor 312 tracks people by comparing the location of the people in a current frame to prior locations and assigns person identifiers based on the spatial correspondence of the people in the frames.
In some embodiments, input interface 302, pixel handling processor 304, foreground detector processor 306, projection generator 308, segmentation processor 310, and application processor 312 are arranged in a pipeline architecture where each block is processing in parallel or at the same time. In some embodiments, only a subset of image information is required by each block so that, depending on the latency of each block and the size of the subset of information required by each block, a number of blocks are operating on pixel data from the same frame period at the same time. In various embodiments, input interface 302, pixel handling processor 304, foreground detector processor 306, projection generator 308, segmentation processor 310, and application processor 312 are integrated circuits disposed on a common substrate or as part of a common integrated circuit package.
Rectify processor 404 has as input the outputs from image grabber 402 and source (for example, cameras, color imagers, monochrome imagers, etc.) predetermined calibration parameters (for example, relative imager positions and orientations, lens principal points, focal lengths, and lens distortion parameters). Rectify processor 404 “rectifies”, corrects for geometric misalignment and lens distortion, each input source (in some cases removal of all distortions and misalignments is not possible). Lens and camera distortions and position corrections are removed by rotation, translation, magnification, and distortion removal operations to the images. In some embodiments, rotation, translation, and magnification are achieved using matrix operations on the image information. In some embodiments, distortion removal is achieved using a polynomial distortion removal model. The combinations of the geometric and distortion corrections results in mappings from rectified image pixel coordinates to distorted image pixel coordinates. In some embodiments, mappings from rectified image pixel coordinates to distorted image pixel coordinates can be stored in the form of precalculated lookup tables. In some embodiments, the mappings can be computed on the fly during rectification of the pixel streams. In some embodiments, the mappings do not match any exact distorted image pixel coordinates so that interpolation is required to derive the rectified pixel coordinate and/or pixel value. In various embodiments, interpolation is linear interpolation, spline interpolation, or any other appropriate interpolation. Rectify processor 404 operations require a subset of image information for calculation of each pixel. In some embodiments, a pipelining architecture is used to perform calculations in parallel where only the subset of image information required for calculating a given pixel information is required by rectify processor 404.
Depth/disparity processor 406 inputs include the output information from rectify processor 404. Depth/disparity processor 406 calculates the disparity of pixels in the input images by correlating pixel data between the two or more images and/or the depth, or range from the source, based at least in part on calibration parameters. The depth and/or disparity calculation is performed by correlating one or more pixels along a dimension of the image. In some embodiments, an array of pixel values from one image is correlated with pixel values from another image. In some embodiments, a non-parametric correlation is used. In various embodiments, SLOG (sign bits of Laplacian of Gaussian), SAD (sum of absolute differences), SSD (sum of squared differences), etc., may be used for correlation. In some embodiments, validity information is provided as an output for the depth/disparity processor block. Validity information can be derived from checking the correspondence between calculating the disparity between, for example, image 1 and image 2 and then image 2 and image 1. In some embodiments, if the pixel area in the image is smooth, then the depth information is less likely to be valid. In various embodiments, intensity and/or color information is used to calculate depth and/or disparity.
In some embodiments, image grabbing processor 402, rectify processor 404, and depth/disparity processor 406 are arranged in a pipeline architecture where each block is processing in parallel or at the same time. In some embodiments, only a subset of image information is required by each block so that, depending on the latency of each block and the size of the subset of information required by each block, a number of blocks are operating on pixel data from the same frame period at the same time. In various embodiments, image grabbing processor 402, rectify processor 404, and depth/disparity processor 406 are integrated circuits disposed on a common substrate or as part of a common integrated circuit package.
In some embodiments, pixel handling processor 400 does not include depth/disparity processor 406. In this case, pixel handling processor 400 has as input one or more images of information containing range, or depth, information for each pixel. Foreground detector processor 106 or projection processor 206 can then operate on input range information based at least in part on the one or more input images with range information.
Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.
This application is a continuation of U.S. patent application Ser. No. 11/264,760 now U.S. Pat. No. 7,664,351, entitled INTEGRATED IMAGE PROCESSOR, filed Oct. 31, 2005, which is incorporated herein by reference for all purposes, which claims priority to U.S. Provisional Application No. 60/624,954, entitled DEEP SEAG2 SYSTEM OVERVIEW filed Nov. 3, 2004, which is incorporated herein by reference for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
3901595 | Helava et al. | Aug 1975 | A |
4524455 | Holsztynski et al. | Jun 1985 | A |
4574394 | Holsztynski et al. | Mar 1986 | A |
5054097 | Flinois et al. | Oct 1991 | A |
5531227 | Schneider | Jul 1996 | A |
5559335 | Zeng et al. | Sep 1996 | A |
5592599 | Lindholm | Jan 1997 | A |
5812787 | Astle | Sep 1998 | A |
5870085 | Laksono | Feb 1999 | A |
5982378 | Kato | Nov 1999 | A |
6064760 | Brown | May 2000 | A |
6100905 | Sidwell | Aug 2000 | A |
6215898 | Woodfill et al. | Apr 2001 | B1 |
6351573 | Schneider | Feb 2002 | B1 |
6517201 | Qi | Feb 2003 | B2 |
6556704 | Chen | Apr 2003 | B1 |
6661918 | Gordon et al. | Dec 2003 | B1 |
6707939 | Weinholz et al. | Mar 2004 | B1 |
6891966 | Chen | May 2005 | B2 |
7003136 | Harville | Feb 2006 | B1 |
7227893 | Srinivasa et al. | Jun 2007 | B1 |
7317830 | Gordon et al. | Jan 2008 | B1 |
7386170 | Ronk et al. | Jun 2008 | B2 |
7623674 | Nichani et al. | Nov 2009 | B2 |
7664315 | Woodfill et al. | Feb 2010 | B2 |
20010041003 | Grohs et al. | Nov 2001 | A1 |
20020159628 | Matusik et al. | Oct 2002 | A1 |
20030007703 | Roylance | Jan 2003 | A1 |
20030091225 | Chen | May 2003 | A1 |
20040151366 | Nefian et al. | Aug 2004 | A1 |
20040169724 | Ekpar | Sep 2004 | A1 |
20040177783 | Seymour | Sep 2004 | A1 |
20040196299 | Di Lelle et al. | Oct 2004 | A1 |
20040208364 | Haque et al. | Oct 2004 | A1 |
20050030314 | Dawson | Feb 2005 | A1 |
20050041037 | Dawson | Feb 2005 | A1 |
20050093697 | Nichani et al. | May 2005 | A1 |
20050099795 | Seymour | May 2005 | A1 |
20050128196 | Popescu et al. | Jun 2005 | A1 |
20050162646 | Tedesco et al. | Jul 2005 | A1 |
20050174437 | Iga | Aug 2005 | A1 |
20050270372 | Henninger, III | Dec 2005 | A1 |
20060067592 | Walmsley et al. | Mar 2006 | A1 |
20100085423 | Lange | Apr 2010 | A1 |
Number | Date | Country |
---|---|---|
1445964 | Aug 2004 | EP |
6187434 | Jul 1994 | JP |
9297849 | Nov 1997 | JP |
2003109128 | Apr 2003 | JP |
9967742 | Dec 1999 | WO |
Entry |
---|
Van Der Wal et al. “The Acadia Vision Processor”, Computer Architectures for Machine Perception, 2000. Proceedings Fifth IEEE International Workshop on Padova, Italy Sep. 11-13, 2000. XP010515287. |
Jeong et al. “Trellis-Based Systolic Multi-Layer Stereo Matching”, Signal Processing Systems, 2003. SIPS 2003. IEEE Workshop on Aug. 29, 2003. XP010661025. |
Oda et al. “A Video-Rate Stereo Machine and Its Application to Virtual Reality”, 27th International Symposium on Industrial Robots. Proceedings ISIR, Oct. 1, 1996. XP009110596. |
Guan et al. “Application-Specific Multimedia Processor Architecture”, Multimedia Image and Video Processing, Chapter 3, CRC; 1 Edition, Aug. 23, 2000. ISBN-13: 978-0849334924, XP007907756. |
Number | Date | Country | |
---|---|---|---|
20100104175 A1 | Apr 2010 | US |
Number | Date | Country | |
---|---|---|---|
60624954 | Nov 2004 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11264760 | Oct 2005 | US |
Child | 12655010 | US |