The present disclosure generally relates to the field of object detection, tracking, and counting. In specific, the present disclosure relates to a computer-implemented detection and tracking system and process for detecting and tracking human objects of interest that appear in camera images taken, for example, at an entrance or entrances to a facility, as well as counting the number of human objects of interest entering or exiting the facility for a given time period and distinguishing between human objects and other objects such as shopping carts.
Traditionally, various methods for detecting and counting the passing of an object have been proposed. U.S. Pat. No. 7,161,482 describes an integrated electronic article surveillance (EAS) and people counting system. The EAS component establishes an interrogatory zone by an antenna positioned adjacent to the interrogation zone at an exit point of a protected area. The people counting component includes one people detection device to detect the passage of people through an associated passageway and provide a people detection signal, and another people detection device placed at a predefined distance from the first device and configured to detect another people detection signal. The two signals are then processed into an output representative of a direction of travel in response to the signals.
Basically, there are two classes of systems employing video images for locating and tracking human objects of interest. One class uses monocular video streams or image sequences to extract, recognize, and track objects of interest. The other class makes use of two or more video sensors to derive range or height maps from multiple intensity images and uses the range or height maps as a major data source.
In monocular systems, objects of interest are detected and tracked by applying background differencing, or by adaptive template matching, or by contour tracking. The major problem with approaches using background differencing is the presence of background clutters, which negatively affect robustness and reliability of the system performance. Another problem is that the background updating rate is hard to adjust in real applications. The problems with approaches using adaptive template matching are:
1) object detections tend to drift from true locations of the objects, or get fixed to strong features in the background; and
2) the detections are prone to occlusion. Approaches using the contour tracking suffer from difficulty in overcoming degradation by intensity gradients in the background near contours of the objects. In addition, all the previously mentioned methods are susceptible to changes in lighting conditions, shadows, and sunlight.
In stereo or multi-sensor systems, intensity images taken by sensors are converted to range or height maps, and the conversion is not affected by adverse factors such as lighting condition changes, strong shadow, or sunlight.
Therefore, performances of stereo systems are still very robust and reliable in the presence of adverse factors such as hostile lighting conditions. In addition, it is easier to use range or height information for segmenting, detecting, and tracking objects than to use intensity information.
Most state-of-the-art stereo systems use range background differencing to detect objects of interest. Range background differencing suffers from the same problems such as background clutter, as the monocular background differencing approaches, and presents difficulty in differentiating between multiple closely positioned objects.
U.S. Pat. No. 6,771,818 describes a system and process of identifying and locating people and objects of interest in a scene by selectively clustering blobs to generate “candidate blob clusters” within the scene and comparing the blob clusters to a model representing the people or objects of interest. The comparison of candidate blob clusters to the model identifies the blob clusters that is the closest match or matches to the model. Sequential live depth images may be captured and analyzed in real-time to provide for continuous identification and location of people or objects as a function of time.
U.S. Pat. Nos. 6,952,496 and 7,092,566 are directed to a system and process employing color images, color histograms, techniques for compensating variations, and a sum of match qualities approach to best identify each of a group of people and objects in the image of a scene. An image is segmented to extract regions which likely correspond to people and objects of interest and a histogram is computed for each of the extracted regions. The histogram is compared with pre-computed model histograms and is designated as corresponding to a person or object if the degree of similarity exceeds a prescribed threshold. The designated histogram can also be stored as an additional model histogram.
U.S. Pat. No. 7,176,441 describes a counting system for counting the number of persons passing a monitor line set in the width direction of a path. A laser is installed for irradiating the monitor line with a slit ray and an image capturing device is deployed for photographing an area including the monitor line. The number of passing persons is counted on the basis of one dimensional data generated from an image obtained from the photographing when the slit ray is interrupted on the monitor line when a person passes the monitor line.
Despite all the prior art in this field, no invention has developed a technology that enables unobtrusive detection and tracking of moving human objects, requiring low budget and maintenance while providing precise traffic counting results with the ability to distinguish between incoming and outgoing traffic, moving and static objects, and between objects of different heights. Thus, it is a primary objective of this invention to provide an unobtrusive traffic detection, tracking, and counting system that involves low cost, easy and low maintenance, high-speed processing, and capable of providing time-stamped results that can be further analyzed.
In addition, people counting systems typically create anonymous traffic counts. In retail traffic monitoring, however, this may be insufficient. For example, some situations may require store employees to accompany customers through access points that are being monitored by an object tracking and counting system, such as fitting rooms. In these circumstances, existing systems are unable to separately track and count employees and customers. The present invention would solve this deficiency.
According to one aspect of the present invention, a method for counting and tracking defined objects comprises the step of receiving subset data with a data capturing device, wherein the subset data is associated with defined objects and includes a unique identifier, an entry time, an exit time, and location data for each defined object. The method may further include the steps of receiving subset data at a counting system, counting the defined objects, tracking the defined objects, associating a location of a defined object with a predefined area, and/or generating path data by plotting X and Y coordinates for the defined object within the predefined area at sequential time periods.
In some embodiments, the method may further include the step of receiving location data at the data capturing device, wherein the location data is received from tracking technology that detects cellular signals emitted from one or more mobile handsets or signals emitted from membership cards, employee badges, rail or air tickets, rental car keys, hotel keys, store-sponsored credit or debit cards, or loyalty reward cards with RFID chips.
In some embodiments, the cellular signals include T-IMSI, CDMA, or Wi-Fi signals.
In some embodiments, the method may further include the step of receiving data from another independent system regarding physical characteristics of the object.
In some embodiments, the other independent system may be selected from the group consisting of point of sale systems, loyalty rewards systems, point of sale trigger information, and mechanical turks.
In some embodiments, the method may further include the steps of converting the subset data into sequence records and creating a sequence array of all of the sequence records, wherein a sequence record may include: (a) a unique ID, which is an unsigned integer associated with a mobile handset, a telephone number associated with the mobile handset or any other unique number, character, or combination thereof, associated with object, (b) a start time, which may consist of information indicative of a time when the object was first detected within a coverage area, (c) an end time, which may consist of information indicative of a time when the object was last detected within the coverage area of the data capturing device, and/or (d) an array of references to all tracks that overlap a particular sequence record.
In some embodiments, the method may further include the step of determining a dwell time for an object within a predefined area by subtracting the end time for that predetermined area from the start time for that predetermined area.
In some embodiments, the method may further include the step of using Z coordinates to further track objects within a predefined area defined by multiple floors.
In some embodiments, the method may further include the step of using the subset data to generate reports showing at least one of the number of objects within a predefined area, a number of objects within the predefined area during a specific time period, a number of predefined areas that were visited by an object, or dwell times for one or more predefined areas.
In some embodiments, the method may further include the step of using at least path data and subset data to aggregate the most common paths taken by objects and correlate path data information with dwell times.
In some embodiments, the method may further include the step of generating a report that shows at least one of the following: (a) the most common paths that objects take in a store, including corresponding dwell times, (b) changes in shopping patterns by time period or season, and/or (c) traffic patterns for use by store security or HVAC systems in increasing or decreasing resources at particular times.
In some embodiments, the method may further include the step of generating a conversion rate by (a) loading transaction data related to transactions performed, (b) loading traffic data including a traffic count and sorting by time periods, and/or (c) dividing the transactions by the traffic counts for the time periods.
In some embodiments, the transaction data may include a sales amount, the number of items purchases, the specific items purchased, the data and time of the transaction, the register used for the transaction, and the sales associate that completed the transaction.
In some embodiments, the method may further include the step of generating a report showing comparisons between purchasers and non-purchasers based on at least one of the dwell times, the predefined areas, or the time periods.
In accordance with these and other objectives that will become apparent hereafter, the present invention will be described with particular references to the accompanying drawings.
This detailed description is presented in terms of programs, data structures or procedures executed on a computer or a network of computers. The software programs implemented by the system may be written in languages such as JAVA, C, C++, C #, Assembly language, Python, PHP, or HTML. However, one of skill in the art will appreciate that other languages may be used instead, or in combination with the foregoing.
1. System Components
Referring to
Typically, the image capturing device includes at least one stereo camera with two or more video sensors 46 (
Door zone 84 is an area in the height map 56 marking the start position of an incoming track and end position of an outgoing track. Interior zone 86 is an area marking the end position of the incoming track and the start position of the outgoing track. Dead zone 90 is an area in the field of view 44 that is not processed by the counting system 30.
Video sensors 46 (
Examples of video sensors 46 used in the present invention include CMOS (Complementary Metal-Oxide Semiconductor) sensors and/or CCD (Charge-Coupled Device) sensors. However, the types of video sensors 46 should not be considered limiting, and any video sensor 46 compatible with the present system may be adopted.
The counting system 30 comprises three main components: (I) boot loader 32; (2) system management and communication component 34; and (3) counting component 36.
The boot loader 32 is executed when the system is powered up and loads the main application program into memory 38 for execution.
The system management and communication component 34 includes task schedulers, database interface, recording functions, and TCP/IP or PPP communication protocols. The database interface includes modules for pushing and storing data generated from the counting component 36 to a database at a remote site. The recording functions provide operations such as writing user defined events to a database, sending emails, and video recording.
The counting component 36 is a key component of the system 10 and is described in further detail as follows.
2. The Counting Component.
In an illustrated embodiment of the present invention, at least one image capturing device 20 and the counting system 30 are integrated in a single image capturing and processing device. The single image capturing and processing device can be installed anywhere above the entrance or entrances to the facility 23. Data output from the single image capturing and processing device can be transmitted through the system management and communication component 34 to the database for storage and further analysis.
Referring to
When multiple image capturing devices 20 are used, tracks 88 generated by each image capturing device 20 are merged before proceeding to block 102.
Block 102 uses calibration data of the stereo cameras (not shown) stored in the image capturing device 20 to rectify raw stereo frames 48. The rectification operation corrects lens distortion effects on the raw frames 48. The calibration data include each sensor's optical center, lens distortion information, focal lengths, and the relative pose of one sensor with respect to the other. After the rectification, straight lines in the real world that have been distorted to curved lines in the raw stereo frames 48 are corrected and restored to straight lines. The resulting frames from rectification are called rectified frames 52, 53 (
Block 104 creates a disparity map 50 (
Block 106 converts the disparity map 50 to the height map 56. Each pixel of the height map 56 comprises a height value and x-y coordinates, where the height value is represented by the greatest ground height of all the points in the same location in the field of view 44. The height map 56 is sometimes referred to as a frame in the rest of the description.
2.1 Object Detection
Object detection (block 108) is a process of locating candidate objects 58 in the height map 56. One objective of the present invention is to detect human objects standing or walking in relatively flat areas. Because human objects of interest are much higher than the ground, local maxima of the height map 56 often represent heads of human objects or occasionally raised hands or other objects carried on the shoulders of human objects walking in counting zone 84,86 (
Occasionally, some human objects of interest do not appear as local maxima for reasons such as that the height map 56 is affected by false detection due to snow blindness effect in the process of generating the disparity map 50, or that human objects of interests are standing close to taller objects such as walls or doors. To overcome this problem, the current invention searches in the neighborhood of the most recent local maxima for a suboptimal location as candidate positions for human objects of interest, as will be described later.
A run is a contiguous set of pixels on the same row of the height map 56 with the same non-zero height values. Each run is represented by a four-tuple (row, start-column, end-column, height). In practice, height map 56 is often represented by a set of runs in order to boost processing performance and object detection is also performed on the runs instead of the pixels.
Object detection comprises four stages: 1) background reconstruction; 2) first pass component detection; 3) second pass object detection; and 4) merging of closely located detects.
2.1.1 Component Definition and Properties
Pixel q is an eight-neighbor of pixel p if q and p share an edge or a vertex in the height map 56, and both p and q have non-zero height values. A pixel can have as many as eight-neighbors.
A set of pixels E is an eight-connected component if for every pair of pixels Pi and Pi in E, there exists a sequence of pixels Pi′ . . . , Pi such that all pixels in the sequence belong to the set E, and every pair of two adjacent pixels are eight neighbors to each other. Without further noting, an eight connected component is simply referred to as a connected component hereafter.
The connected component is a data structure representing a set of eight-connected pixels in the height map 56. A connected component may represent one or more human objects of interest. Properties of a connected component include height, position, size, etc. Table 1 provides a list of properties associated with a connected component. Each property has an abbreviated name enclosed in a pair of parentheses and a description. Properties will be referenced by their abbreviated names hereafter.
Several predicate operators are applied to a subset of properties of the connected component to check if the subset of properties satisfies a certain condition. Component predicate operators include:
IsNoisy, which checks whether a connected component is too small to be considered a valid object detect 58. A connected component is considered as “noise” if at least two of the following three conditions hold: 1) its det_minSize is less than two thirds of a specified minimum human body size, which is configurable in the range of [9,36] inches; 2) its det_area is less than four ninths of the area of a circle with its diameter equal to a specified minimum body size; and 3) the product of its det_minSize and det area is less than product of the specified minimum human body size and a specified minimum body area.
IsPointAtBoundaries, which checks whether a square window centered at the current point with its side equal to a specified local maximum search window size is intersecting boundaries of the height map 56, or whether the connected component has more than a specific number of pixels in the dead zone 90. If this operation returns true, the point being checked is considered as within the boundaries of the height map 56.
NotSmallSubComponent, which checks if a subcomponent in the second pass component detection is not small. It returns true if its detrninxize is greater than a specified minimum human head size or its det_area is greater than a specified minimum human head area.
BigSubComponentSeed, which checks if a subcomponent seed in the second pass component detection is big enough to stop the grouping operation. It returns true if its detrninxize is greater than the specified maximum human head size or its det_area is greater than the specified maximum human head area.
SmallSubComponent, which checks if a subcomponent in the second pass component detection is small. It returns true if its detrninxize is less than the specified minimum human head size or its der area is less than the specified minimum human head area.
2.1.2 Background Reconstruction
The background represents static scenery in the field view 44 of the image capturing device 20 and is constructed from the height map 56. The background building process monitors every pixel of every height map 56 and updates a background height map. A pixel may be considered as part of the static scenery if the pixel has the same non-zero height value for a specified percentage of time (e.g., 70%).
2.1.3 First-Pass Component Detection
First pass components are computed by applying a variant of an eight-connected image labeling algorithm on the runs of the height map 56. Properties of first pass components are calculated according to the definitions in Table 1. Predicate operators are also applied to the first pass components. Those first pass components whose “IsNoise” predicate operator returns “true” are ignored without being passed on to the second pass component detection phase of the object detection.
2.1.4 Second Pass Object Detection
In this phase, height map local maxima, to be considered as candidate human detects, are derived from the first pass components in the following steps.
First, for each first pass component, find all eight connected subcomponents whose pixels have the same height. The deigrouped property of all subcomponents is cleared to prepare for subcomponent grouping and the deCID property of each subcomponent is set to the ID of the corresponding first pass component.
Second, try to find the highest ungrouped local maximal subcomponent satisfying the following two conditions: (1) the subcomponent has the highest height among all of the ungrouped subcomponents of the given first pass component, or the largest area among all of the ungrouped subcomponents of the given first pass component if several ungrouped subcomponents with the same highest height exist; and (2) the subcomponent is higher than all of its neighboring subcomponents. If such a subcomponent exists, use it as the current seed and proceed to the next step for further subcomponent grouping. Otherwise, return to step 1 to process the next first pass component in line.
Third, ifBigSubComponentSeed test returns true on the current seed, the subcomponent is then considered as a potential human object detect. Set the det grouped flag of the subcomponent to mark it as grouped and proceed to step 2 to look for a new seed. If the test returns false, proceed to the next step.
Fourth, try to find a subcomponent next to the current seed that has the highest height and meets all of the following three conditions: (I) it is eight-connected to the current seed; (2) its height is smaller than that of the current seed; and (3) it is not connected to a third subcomponent that is higher and it passes the NotSmallSubComponent test. If more than one subcomponent meets all of above conditions, choose the one with the largest area. When no subcomponent meets the criteria, set the deigrouped property of the current seed to “grouped” and go to step 2. Otherwise, proceed to the next step.
Fifth, calculate the distance between centers of the current seed and the subcomponent found in the previous step. If the distance is less than the specified detection search range or the current seed passes the SmallSubComponent test, group the current seed and the subcomponent together and update the properties of the current seed accordingly. Otherwise, set the det_grouped property of the current seed as “grouped”. Return to step 2 to continue the grouping process until no further grouping can be done.
2.1.5 Merging Closely Located Detections
Because the image capturing device 20 is mounted on the ceiling of the facility entrance (
The steps of this phase are as follows.
First, search for the closest pair of local maxima detections. If the distance between the two closest detections is greater than the specified detection merging distance, stop and exit the process. Otherwise, proceed to the next step.
Second, check and process the two detections according to the following conditions in the given order. Once one condition is met, ignore the remaining conditions and proceed to the next step:
a) if either but not all detection is in the background, ignore the one in the background since it is most likely a static object (the local maximum in the foreground has higher priority over the one in the background);
b) if either but not all detection is touching edges of the height map 56 or dead zones, delete the one that is touching edges of the height map 56 or dead zones (a complete local maximum has higher priority over an incomplete one);
c) if the difference between det rnaxlleights of detections is smaller than a specified person height variation threshold, delete the detection with significantly less 3-D volume (e.g., the product of det_maxHeight and det_masArea for one connected component is less than two thirds of the product for the other connected component) (a strong local maximum has higher priority over a weak one);
d) if the difference between maximum heights of detections is more than one foot, delete the detection with smaller det_maxHeight if the detection with greater height among the two is less than the specified maximum person height, or delete the detection with greater det_maxHeight if the maximum height of that detection is greater than the specified maximum person height (a local maxima with a reasonable height has higher priority over a local maximum with an unlikely height);
e) delete the detection whose det area is twice as small as the other (a small local maximum close to a large local maximum is more likely a pepper noise);
f) if the distance between the two detections is smaller than the specified detection search range, merge the two detections into one (both local maxima are equally good and close to each other);
g) keep both detections if the distance between the two detections is larger than or equal to the specified detection search range (both local maxima are equally good and not too close to each other). Update the det., closestDet attribute for each detection with the other detection's ID.
Then, return to step 1 to look for the next closest pair of detections.
The remaining local maxima detections after the above merging process are defined as candidate object detects 58, which are then matched with a set of existing tracks 74 for track extension, or new track initiation if no match is found.
2.2 Object Tracking
Object tracking (block 110 in
To count human objects using object tracks, zones 82 are delineated in the height map 56. Door zones 84 represent door areas around the facility 23 to the entrance. Interior zones 86 represent interior areas of the facility. A track 76 traversing from the door zone 84 to the interior zone 86 has a potential “in” count. A track 76 traversing to the door zone 84 from the interior zone 86 has a potential “out” count. If a track 76 traverses across zones 82 multiple times, there can be only one potential “in” or “out” count depending on the direction of the latest zone crossing.
As illustrated in
An object track 76 can be used to determine whether a human object is entering or leaving the facility, or to derive properties such as moving speed and direction for human objects being tracked.
Object tracks 76 can also be used to eliminate false human object detections, such as static signs around the entrance area. If an object detect 58 has not moved and its associated track 76 has been static for a relatively long time, the object detect 58 will be considered as part of the background and its track 76 will be processed differently than normal tracks (e.g., the counts created by the track will be ignored).
Object tracking 110 also makes use of color or gray level intensity information in the frames 52, 53 to search for best match between tracks 76 and object detects 58. Note that the color or the intensity information is not carried to disparity maps 50 or height maps 56.
The same technique used in the object tracking can also be used to determine how long a person stands in a checkout line.
2.2.1 Properties of Object Track
Each track 76 is a data structure generated from the same object being tracked in both temporal and spatial domains and contains a list of 4-tuples (x, y, t, h) in addition to a set of related properties, where h, x and y present the height and the position of the object in the field of view 44 at time t. (x, y, h) is defined in a world coordinate system with the plane formed by x and y parallel to the ground and the h axis vertical to the ground. Each track can only have one position at any time. In addition to the list of 4-tuples, track 76 also has a set of properties as defined in Table 2 and the properties will be referred to later by their abbreviated names in the parentheses:
2.2.2 Track-Related Predicative Operations
Several predicate operators are defined in order to obtain the current status of the tracks 76. The predicate operators are applied to a subset of properties of a track 76 to check if the subset of properties satisfies a certain condition. The predicate operators include:
IsNoisyNow, which checks if track bouncing back and forth locally at the current time. Specifically, a track 76 is considered noisy if the track points with a fixed number of frames in the past (specified as noisy track duration) satisfies one of the following conditions:
a) the range of track 76 (trkrange) is less than the specified noisy track range, and either the negative distance (trk_negDist) is larger than two thirds of the positive distance (trk_posDist) or the negative steps (trk_negNum) are more than two thirds of the positive steps (trk_posNum);
b) the range of track 76 (trkrange) is less than half of the specified noisy track range, and either the negative distance (trk_negDist) is larger than one third of the positive distance (trk_posDist) or the negative steps (trk_negNum) are more than one third of the positive steps (trk_posNum).
WholeTrackIsNoisy: a track 76 may be noisy at one time and not noisy at another time.
This check is used when the track 76 was created a short time ago, and the whole track 76 is considered noisy if one of the following conditions holds:
a) the range of track 76 (trkrange) is less than the specified noisy track range, and either the negative distance (trk_negDist) is larger than two thirds of the positive distance (trk_posDist) or the negative steps (trk_negNum) are more than two thirds of the positive steps (trk_posNum);
b) the range of track 76 (trkrange) is less than half the specified noisy track range, and either the negative distance trk_negDist) is larger than one third of the positive distance (trk_posDist) or the negative steps (trk_negNum) are more than one third of the positive steps (trk_posNum).
IsSameTrack, which check if two tracks 76, 77 are likely caused by the same human object. All of the following three conditions have to be met for this test to return true: (a) the two tracks 76, 77 overlap in time for a minimum number of frames specified as the maximum track timeout; (b) the ranges of both tracks 76, 77 are above a threshold specified as the valid counting track span; and (c) the distance between the two tracks 76, 77 at any moment must be less than the specified minimum person width.
IsCountIgnored: when the track 76 crosses the counting zones, it may not be created by a human object of interest. The counts of a track are ignored if one of the following conditions is met:
Invalid Tracks: the absolute difference between trk_exits and trk_enters is not equal to one.
Small Tracks: trkrange is less than the specified minimum counting track length.
Unreliable Merged Tracks: trkrange is less than the specified minimum background counting track length as well as one of the following: trk_mergedTracks is equal to trk_smallSearches, or trk_backgroundCount is more than 80% of the life time of the track 76, or the track 76 crosses the zone boundaries more than once.
High Object Test: trk_highPtSteps is larger than half of trk_totaISteps.
Small Child Test: trk_lowPtSteps is greater than ¾ of trk_totaISteps, and trk_maxTrackHt is less than or equal to the specified minimum person height.
Shopping Cart Test: trk_voteFollowing is greater than 3, trk_minFollowingDist is more than or equal to 80% of trk_maxFollowingDist, and trk_maxTrackHt is less than or equal to the specified shopping cart height.
False Track test: trk_voteMirrorTrack is more than 60% of the life time of the track 76, and trk_maxMirrorTrackDist is less than two thirds of the specified maximum person width or trk_totalVoteMirrorTrack is more than 80% of the life time of the track 76.
2.2.3 Track Updating Operation
Referring to
Second, for the current frame, obtain the height by using median filter on the most recent three heights of the track 76 and calculate the new position 56 by averaging on the most recent three positions of the track 76 (block 362).
Third, for the current frame, check the noise status using track predicate operator IsNoisyNow. If true, mark a specified number of frames in the past as noisy. In addition, update noise related properties of the track 76 (block 364).
Fourth, update the span of the track 76 (block 366).
Fifth, if one of the following conditions is met, collect the count carried by track 76 (block 374):
a) the track 76 is not noisy at the beginning, but it has been noisy for longer than the specified stationary track timeout (block 368); or
b) the track 76 is not in the background at the beginning, but it has been in the background for longer than the specified stationary track timeout (block 370).
Finally, update the current zone information (block 372).
2.2.4 Track Prediction Calculation
It helps to use a predicted position of the track 76 when looking for best matching detect 58. The predicted position is calculated by linear extrapolation on positions of the track 76 in the past three seconds.
2.2.5 Analysis and Processing of Old Track
This is the first phase of object tracking. Active tracks 88 are tracks 76 that are either created or extended with human object detects 58 in the previous frame. When there is no best matching human object detect 58 for the track 76, the track 76 is considered as inactive.
This phase mainly deals with tracks 76 that are inactive for a certain period of time or are marked for deletion in previous frame 56. Track analysis is performed on tracks 76 that have been inactive for a long time to decide whether to group them with existing tracks 74 or to mark them for deletion in the next frame 56. Tracks 76 are deleted if the tracks 76 have been marked for deletion in the previous frame 56, or the tracks 76 are inactive and were created a very short period of time before. If the counts of the soon-to-be deleted tracks 76 shall not be ignored according to the IsCountIgnored predicate operator, collect the counts of the tracks 76.
2.2.6 First Pass Matching Between Tracks and Detects
After all tracks 76 are analyzed for grouping or deletion, this phase searches for optimal matches between the human object detects 58 (i.e. the set of local maxima found in the object detection phase) and tracks 76 that have not been deleted.
First, check every possible pair of track 76 and detect 58 and put the pair into a candidate list if all of the following conditions are met:
1) The track 76 is active, or it must be long enough (e.g. with more than three points), or it just became inactive a short period of time ago (e.g. it has less than three frames);
2) The smaller of the distances from center of the detect 58 to the last two points of the track 76 is less than two thirds of the specified detection search range when the track 76 hasn't moved very far (e.g. the span of the track 76 is less than the specified minimum human head size and the track 76 has more than 3 points);
3) If the detect 58 is in the background, the maximum height of the detect 58 must be greater than or equal to the specified minimum person height;
4) If the detect 58 is neither in the background nor close to dead zones or height map boundaries, and the track 76 is neither in the background nor is noisy in the previous frame, and a first distance from the detect 58 to the predicted position of the track 76 is less than a second distance from the detect 58 to the end position of the track 76, use the first distance as the matching distance. Otherwise, use the second distance as the matching distance. The matching distance has to be less than the specified detection search range;
5) The difference between the maximum height of the detect 58 and the height oblast point of the track 76 must be less than the specified maximum height difference; and
6) If either the last point off-track 76 or the detect 58 is in the background, or the detect 58 is close to dead zones or height map boundaries, the distance from the track 76 to the detect 58 must be less than the specified background detection search range, which is generally smaller than the threshold used in condition (4).
Sort the candidate list in terms of the distance from the detect 58 to the track 76 or the height difference between the detect 58 and the track 76 (if the distance is the same) in ascending order.
The sorted list contains pairs of detects 58 and tracks 76 that are not paired. Run through the whole sorted list from the beginning and check each pair. If either the detect 58 or the track 76 of the pair is marked “paired” already, ignore the pair. Otherwise, mark the detect 58 and the track 76 of the pair as “paired”. [0144] 2.2.7 Search of Suboptimal Location For Unpaired Tracks.
Due to sparseness nature of the disparity map 50 and the height map 56, some human objects may not generate local maxima in the height map 56 and therefore may be missed in the object detection process 108. In addition, the desired local maxima might get suppressed by a neighboring higher local maximum from a taller object. Thus, some human object tracks 76 may not always have a corresponding local maximum in the height map 56. This phase tries to resolve this issue by searching for a suboptimal location for a track 76 that has no corresponding local maximum in the height map 56 at the current time. Tracks 76 that have already been paired with a detect 58 in the previous phase might go through this phase too to adjust their locations if the distance between from end of those tracks to their paired detects is much larger than their steps in the past. In the following description, the track 76 currently undergoing this phase is called Track A. The search is performed in the following steps.
First, referring to
Second, if Track A has moved a few steps (block 200) (e.g., three steps) and is paired with a detection (called Detection A) (block 186) that is not in the background and whose current step is much larger than its maximum moving step within a period of time in the past specified by a track time out parameter (block 202,204), proceed to the next step. Otherwise, stop the search process and exit.
Third, search around the end point of Track A in a range defined by its maximum moving steps for a location with the largest height sum in a predefined window and call this location Best Spot A (block 188). If there are some detects 58 deleted in the process of merging of closely located detects in the object detection phase and Track A is long in either the spatial domain or the temporal domain (e.g. the span of Track A is greater than the specified noisy track span threshold, or Track A has more than three frames) (block 190), find the closest one to the end point of Track too. If its distance to the end point of Track A is less than the specified detection search range (block 206), search around the deleted component for the position with the largest height sum and call it Best Spot AI (block 208). If neither Best Spot A nor Best Spot AI exists, stop the search process and exit. If both Best Spot A and Best Spot AI exist, choose the one with larger height sum. The best spot selected is called suboptimal location for Track A. If the maximum height at the suboptimal location is greater than the predefined maximum person height (block 192), stop the search and exit. If there is no current detection around the suboptimal location (block 194), create a new detect 58 (block 214) at the suboptimal location and stop the search. Otherwise, find the closest detect 58 to the suboptimal location and call it Detection B (block 196). If Detection B is the same detection as Detection A in step 2 (block 198), update Detection A's position with the suboptimal location (block 216) and exit the search. Otherwise, proceed to the next step.
Fourth, referring to
1) When the suboptimal location for Track A and Detection B are from the same parent component (e.g. in the support of the same first pass component) and the distance between Track A and Detection B is less than half of the specified maximum person width, create a new detect 58 at the suboptimal location (block 238) if all of the following three conditions are met: (i) the difference between the maximum heights at the suboptimal location and Detection B is less than a specified person height error range; (ii) the difference between the height sums at the two locations is less than half of the greater one; (iii) the distance between them is greater than the specified detection search range and the trk_range values of both Track A and Track B are greater than the specified noisy track offset. Otherwise, ignore the suboptimal location and exit;
2) If the distance between the suboptimal location and Detection B is greater than the specified detection search range, create a new detect 58 at the suboptimal location and exit;
3) If Track A is not sizable in both temporal and spatial domains (block 226), ignore the suboptimal location;
4) If Track B is not sizable in both temporal and spatial domain (block 228), detach Track B from Detection B and update Detection B's position with the suboptimal location (block 246). Mark Detection B as Track A's closest detection;
5) Look for best spot for Track B around its end position (block 230). If the distance between the best spot for Track B and the suboptimal location is less than the specified detection search range (block 232) and the best spot for Track B has a larger height sum, replace the suboptimal location with the best spot for Track B (block 233). If the distance between is larger than the specified detection search range, create a detect 58 at the best spot for Track B (block 250). Update Detection A's location with the suboptimal location if Detection A exists.
Fifth, if the suboptimal location and Detection B are not in the support of the same first pass component, proceed to the next step. Otherwise create a new detection at the suboptimal location if their distance is larger than half of the specified maximum person width, or ignore the suboptimal location and mark Detection B as Track A's closest detection otherwise.
Finally, create a new detect 58 at suboptimal location and mark Detection B as Track A's closest detection (block 252) if their distance is larger than the specified detection search range. Otherwise, update Track A's end position with the suboptimal location (block 254) if the height sum at the suboptimal location is greater than the height sum at Detection B, or mark Detection Bas Track A's closest detection otherwise.
2.2.8 Second Pass Matching Between Tracks and Detects
After the previous phase, a few new detections may be added and some paired detects 72 and tracks 76 become unpaired again. This phase looks for the optimal match between current unpaired detects 72 and tracks 76 as in the following steps.
For every pair of track 76 and detect 58 that remain unpaired, put the pair into a candidate list if all of the following five conditions are met:
1) the track 76 is active (block 262 in
2) the distance from detect 58 to the end point of the track 76 (block 274) is smaller than two thirds of the specified detection search range (block 278) when the track doesn't move too far (e.g. the span of the track 76 is less than the minimal head size and the track 76 has more than three points (block 276));
3) if the detect 58 is in the background (block 280), the maximum height of the detect 58 must be larger than or equal to the specified minimum person height (block 282);
4) the difference between the maximum height and the height of the last point of the track 76 is less than the specified maximum height difference (block 284);
5) the distance from the detect 58 to the track 76 must be smaller than the specified background detection search range, if either the last point of the track 76 or the detect 58 is in background (block 286), or the detect 58 is close to dead zones or height map boundaries (block 288); or if not, the distance from the detect 58 to the track 76 must be smaller than the specified detection search range (block 292).
Sort the candidate list in terms of the distance from the detect 58 to the track 76 or the height difference between the two (if distance is the same) in ascending order (block 264).
The sorted list contains pairs of detects 58 and tracks 76 which are not paired at all at the beginning. Then run through the whole sorted list from the beginning and check each pair. If either the detect 58 or the track 76 of the pair is marked “paired” already, ignore the pair. Otherwise, mark the detect 58 and the track 76 of the pair as “paired” (block 270).
2.2.9 Track Update or Creation
After the second pass of matching, the following steps are performed to update old tracks or to create new tracks:
First, referring to
Second, create a new track 80 for every detect 58 that is not matched to the track 76 if the maximum height of the detect 58 is greater than the specified minimum person height, and the distance between the detect 58 and the closest track 76 of the detect 58 is greater than the specified detection search range (block 306,308). When the distance is less than the specified detection merge range and the detect 58 and the closest track 76 are in the support of the same first pass component (i.e., the detect 58 and the track 76 come from the same first pass component), set the trk_IastCollidingTrack of the closest track 76 to the ID of the newly created track 80 if there is one (block 310,320).
Third, mark each unpaired track 77 as inactive (block 324). If that track 77 has a marked closest detect and the detect 58 has a paired track 76, set the trk_IastCollidingTrack property of the current track 77 to the track ID of the paired track 76 (block 330).
Fourth, for each active track 88, search for the closest track 89 moving in directions that are at most thirty degrees from the direction of the active track 88. If the closest track 89 exists, the track 88 is considered as closely followed by another track, and “Shopping Cart Test” related properties of the track 88 are updated to prepare for “Shopping Cart Test” when the track 88 is going to be deleted later (block 334).
Finally, for each active track 88, search for the closest track 89. If the distance between the two is less than the specified maximum person width and either the track 88 has a marked closest detect or its height is less than the specified minimum person height, the track 88 is considered as a less reliable false track. Update “False Track” related properties to prepare for the “False Track” test later when the track 88 is going to be deleted later (block 338).
As a result, all of the existing tracks 74 are either extended or marked as inactive, and new tracks 80 are created.
2.2.10 Track Analysis
Track analysis is applied whenever the track 76 is going to be deleted. The track 76 will be deleted when it is not paired with any detect for a specified time period. This could happen when a human object moves out of the field view 44, or when the track 76 is disrupted due to poor disparity map reconstruction conditions such as very low contrast between the human object and the background.
The goal of track analysis is to find those tracks that are likely continuations of some soon-to-be deleted tracks, and merge them. Track analysis starts from the oldest track and may be applied recursively on newly merged tracks until no tracks can be further merged. In the following description, the track that is going to be deleted is called a seed track, while other tracks are referred to as current tracks. The steps of track analysis are as follows:
First, if the seed track was noisy when it was active (block 130 in
Second, examine each active track that was created before the specified merging track time period and merge an active track with the seed track if the “Is the Same Track” predicate operation on the active track (block 140) returns true.
Third, if the current track satisfies all of the following three initial testing conditions, proceed to the next step. Otherwise, if there exists a best fit track (definition and search criteria for the best fit track will be described in forthcoming steps), merge the best fit track with the seed track (block 172, 176). If there is no best fit track, keep the seed track if the seed track has been merged with at least one track in this operation (block 178), or delete the seed track (block 182) otherwise. Then, exit the track analysis.
The initial testing conditions used in this step are: (1) the current track is not marked for deletion and is active long enough (e.g. more than three frames) (block 142); (2) the current track is continuous with the seed track (e.g. it is created within a specified maximum track timeout of the end point of the seed track) (block 144); (3) if both tracks are short in space (e.g., the trkrange properties of both tracks are less than the noisy track length threshold), then both tracks should move in the same direction according to the relative offset of the trk_start and trk_end properties of each track (block 146).
Fourth, merge the seed track and the current track (block 152). Return to the last step if the current track has collided with the seed track (i.e., the trk_IastCollidingTrack of the current track is the trk_ID of the seed track). Otherwise, proceed to the next step.
Fifth, proceed to the next step if the following two conditions are met at the same time, otherwise return to step 3: (1) if either track is at the boundaries according to the “is at the boundary” checking (block 148), both tracks should move in the same direction; and (2) at least one track is not noisy at the time of merging (block 150). The noisy condition is determined by the “is noisy” predicate operator.
Sixth, one of two thresholds coming up is used in distance checking. A first threshold (block 162) is specified for normal and clean tracks, and a second threshold is specified for noisy tracks or tracks in the background. The second threshold (block 164) is used if either the seed track or the current track is unreliable (e.g. at the boundaries, or either track is noisy, or trkranges of both tracks are less than the specified noisy track length threshold and at least one track is in the background) (block 160), otherwise the first threshold is used. If the shortest distance between the two tracks during their overlapping time is less than the threshold (block 166), mark the current track as the best fit track for the seed track (block 172) and if the seed track does not have best fit track yet or the current track is closer to the seed track than the existing best fit track (block 170). Go to step 3.
2.2.11 Merging of Tracks
This operation merges two tracks into one track and assigns the merged track with properties derived from the two tracks. Most properties of the merged track are the sum of the corresponding properties of the two tracks but with the following exceptions:
Referring to
If a point in time has multiple positions after the merge, the final position is the average (block 352).
The trk_start property of the merged track has the same trk_start value as the newer track among the two tracks being merged, and the trk_end property of the merged track has the same trk_end value as the older track among the two (block 354).
The buffered raw heights and raw positions of the merged track are the buffered raw heights and raw positions of the older track among the two tracks being merged (block 356).
Another embodiment of the present disclosure may be adapted to a counting system various settings, such as in a retail setting. As shown in
In the retail space 1300, the user may evaluate the counting data, e.g., for use in analyzing statistics such as sales conversion rates and visitor traffic. However, one concern is obtaining accurate counting data from the visitors entering and exiting the retail space 1300. More particularly, distinguishing between a visitor and a visitor with an additional object, such as a shopping cart, is generally of concern in obtaining useful information for customer analysis. For instance, if an individual is pushing a shopping cart, it is possible that counting unit may count the individual and the shopping cart as two distinct individuals, causing false positives in counting mechanisms. For example,
Further, a user might want to evaluate habits of visitors to the retail space 1300 and might not be able to obtain an accurate insight of such habits if shopping carts are included with a count. Further still, a user might want to evaluate habits of visitors that go through the retail space 1300 with a shopping cart compared with visitors that do not. However, because of a relatively significant variability in detected heights for a tracked object, distinguishing between a visitor and a visitor that is next to an object throughout a scene, such as a shopping cart, may be a challenge. Current approaches to doing so may be inaccurate and involve computational complexity.
Embodiments presented herein disclose techniques for distinguishing between objects, such as between an individual and an individual pushing a shopping cart in the retail space 1300 (e.g., during the second pass object detection phase discussed above). More particularly, techniques describe generating height maps from image data captured by the stereo-optical sensor unit 1302 and applying one or more parameters relating to a given object (such as the shopping cart 1304) to the height map. As further described herein, the stereo-optical sensor unit 1302 may identify binary large objects (“blobs”) in the height map that is representative of a desired tracking object. As used herein, a blob is a collection of connected pixels in the image data. The stereo-optical sensor unit 1302 may then evaluate a blob for whether a predefined object is present within the blob. Using the shopping cart example, the stereo-optical sensor unit 1302 may evaluate the blobs for properties of a specified object, such as a minimum shopping cart width and length, a pixel height value indicative of the height of the shopping cart, a threshold minimum number of pixels of the blob used to check for the presence of the shopping cart, and a ratio of a number of pixels of the blob with height values less than the cart height. If a cart is identified based on these parameters within the blob, the stereo-optical sensor unit 1302 may classify the shopping cart separately from the individual. Once classified, the stereo-optical sensor unit 1302 may separately count the shopping cart (e.g., as a shopping cart object associated with the individual). Advantageously, embodiments presented herein disclose techniques for accurately determining a height of a tracked object in a manner that requires less computational complexity.
Referring now to
The shapes 1402 and 1404 are representative of a height map generated by the stereo-optical sensor unit 1302. Each pixel in the height map has a height value. The black area on the images represents height values of zero or a value below a specified threshold. The white areas (the blobs 1402 and 1404) indicate objects of interest (e.g., individuals to be counted). Within each of the blobs 1402 and 1404, each pixel has a non-zero height value above the specified threshold.
As stated, the stereo-optical sensor unit 1302 may apply one or more predefined properties associated with a predefined object, such as a shopping cart. In such a case, the properties include a minimum shopping cart width in pixels (also referred to herein as a “minimum cart width”), a minimum shopping cart length in pixels (also referred to herein as a “minimum cart length”), a pixel height value representing the height of a shopping cart (also referred to herein as a “cart height”), a minimum number of blob pixels used to check for a presence of a shopping cart (also referred to herein as a “cart checking threshold”), and a ratio of a number of blob pixels with height values that are less than the cart height (also referred to herein as a “minimum cart area ratio”). Of course, other properties may be contemplated.
Referring now to
In block 1506, the stereo-optical sensor unit 1302 generates, from the one or more images, a height map. Doing so allows the stereo-optical sensor unit 1302 to distinguish between background (e.g., pixels having zero height values) and objects of interest, such as the individual. Further, in block 1508, the stereo-optical sensor unit 1302 removes, from the height map, pixels that are below a specified threshold. The stereo-optical sensor unit 1302 also identifies local maxima within each height map (e.g., maximum height values on pixels in the scene).
In block 1510, the stereo-optical sensor unit 1302 extracts isolated blobs (e.g., the blob 1402 and 1404) from the height map. The isolated blobs are indicative of first stage detection blobs discussed above. In the previous example, the visitor may be represented as one of the extracted blobs. In addition, the stereo-optical sensor unit 1302 identifies the current point of time used for the height map. For example, the stereo-optical sensor unit 1302 may use timestamp data maintained by a stereo-optical sensor unit 1302 corresponding to the time that the stereo-optical sensor unit 1302 captured the image. The stereo-optical sensor unit 1302 tracks the blob through the counting zone for a specified duration. The specified duration may be a period of time used for a particular counting zone. The stereo-optical sensor unit 1302 may thereafter determine whether the tracking of the blob is complete. For example, the tracking of the blob may be complete once the tracked object is no longer in the scene. As another example, the tracking of the blob may be complete once the specified duration has elapsed.
In block 1512, the stereo-optical sensor unit 1302 applies one or more predefined properties relating to a specified object, such as a shopping cart, to the blobs in the height map. For example, the visitor traversing the retail space 1300 may be pushing a shopping cart, and as a result, the shopping cart may be included with the extracted blob representative of the visitor. As stated, properties may include minimum cart width, minimum cart length, cart height, a cart threshold, minimum cart area ratio, etc.
Further, in block 1514, the stereo-optical sensor unit 1302 calculates properties in each blob relative to the one or more properties. For instance, the stereo-optical sensor unit 1302 determines a total area of the blob (e.g., a total number of pixels associated with the blob), a low blob area (e.g., a number of pixels associated with the blob with height values less than the cart height), a minimal blob size, a maximal blob size, a maximal parallelogram for the blob (if present), and the like. Based on the properties of the specified object and a given blob, the stereo-optical sensor unit 1302 determines whether the specified object is within the blob. Continuing the example, the specified object may correspond to a shopping cart, and the stereo-optical sensor unit 1302 may determine whether a shopping cart is associated with the blob representative of the visitor.
For instance, referring now to
Otherwise, the method 1500 proceeds to block 1606, in which the stereo-optical sensor unit determines whether a maximal parallelogram is present around the blob having a long side and a short side, in which the short side of the parallelogram is longer than the minimum cart width. If not, then the method 1500 proceeds to the steps of
Referring now to
Referring now to
Referring now to
Referring now to
The CPU 2002 retrieves and executes programming instructions stored in memory 2008 as well as stores and retrieves application data residing in the storage 2010. The bus 2012 is used to transmit programming instructions and data between CPU 2002, storage 2010, network interface 2006, and memory 2008. Note, the CPU 2002 is included to be representative of a single CPU, multiple CPUs, a single CPU having multiple processing cores, and the like. The memory 2008 is generally included to be representative of a random access memory. The storage 2010 may be a disk drive storage device. Although shown as a single unit, storage 2010 may be a combination of fixed and/or removable storage devices, such as fixed disc drives, removable memory cards, or optical storage, network attached storage (NAS), or a storage area network (SAN).
The stereo-optical camera 2004 includes multiple sensor lenses, e.g., camera lenses, to capture image data, which the storage 2010 may temporarily maintain. In an embodiment, the stereo-optical camera 2004 provides two camera lenses that are positioned horizontally apart from one another. In using multiple camera lenses, the stereo-optical sensor unit 1302 can measure depth or, in cases in which a distance from the stereo-optical sensor unit 1302 and ground is known, height. The memory 2008 includes program code logic to direct the stereo-optical camera 2004 to capture image data, e.g., of individuals entering and leaving the field-of-view thereof. Logic in the stereo-optical sensor unit 1302 (e.g., maintained in the memory 2008) processes the image data. For instance, the stereo-optical sensor unit 1302 generates a height map of the scene captured by the field-of-view and, based on an evaluation of the generated height map, identifies an individual to be included as counting data. The program code may also identify individuals in the image data (e.g., using the techniques described above) for counting data and generate height map data for each individual and process the height map data (e.g., to distinguish between an individual and an object such as a shopping cart being pushed by the individual). The storage 2010 may include configuration information for the stereo-optical sensor unit 1302, image data, counting data, and the like. The network interface 2006 may be embodied as any communication circuit, device, or collection thereof, capable of enabling communications over a network between the optical sensor unit 1302 and other devices. For instance, the stereo-optical sensor unit 1302 may be connected with a management console application (e.g., executing on a remote computing device) that receives image data and other statistics generated by a given stereo-optical sensor unit 1302 (e.g., height data, counting data, and the like).
During configuration, the heights of the stereo-optical sensor unit 1302 from a surface (e.g., the floor of the retail space 1300) is to be measured to establish an accurate relationship between the stereo-optical sensor unit 1302 and the surface. Doing so allows the stereo-optical sensor unit 1302 to reliably detect and track an individual in a scene (e.g., the image area captured in a field-of-view by the optical sensors). The stereo-optical sensor unit 1302 is also configured to define a counting area. In some embodiments, the counting area corresponds to an area within the captured image data that represents a physical area through which an individual passes to be counted.
This application is a continuation of U.S. patent application Ser. No. 16/123,769, filed Sep. 6, 2018, which is a continuation-in-part of U.S. patent application Ser. No. 15/669,440, filed Aug. 4, 2017, which is a continuation of U.S. patent application Ser. No. 15/057,908, filed Mar. 1, 2016 (issued as U.S. Pat. No. 9,734,388), which is a continuation of U.S. patent application Ser. No. 14/680,123, filed Apr. 7, 2015 (issued as U.S. Pat. No. 9,305,363), which is a continuation of U.S. patent application Ser. No. 13/622,083, filed Sep. 18, 2012 (issued as U.S. Pat. No. 9,177,195), which claims the benefit of priority under 35 U.S.C. § 119(e) to U.S. Provisional Patent Application No. 61/538,554, filed on Sep. 23, 2011, and U.S. Provisional Patent Application No. 61/549,511, filed on Oct. 20, 2011. The disclosures set forth in the referenced applications are incorporated herein by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
4916621 | Bean et al. | Apr 1990 | A |
5973732 | Guthrie | Oct 1999 | A |
6445810 | Darrell et al. | Sep 2002 | B2 |
6674877 | Jojic et al. | Jan 2004 | B1 |
6697104 | Yakobi et al. | Feb 2004 | B1 |
6771818 | Krumm et al. | Aug 2004 | B1 |
6952496 | Krumm | Oct 2005 | B2 |
7003136 | Harville | Feb 2006 | B1 |
7092566 | Krumm | Aug 2006 | B2 |
7161482 | Rider et al. | Jan 2007 | B2 |
7176441 | Sumitomo et al. | Feb 2007 | B2 |
7227893 | Srinivasa et al. | Jun 2007 | B1 |
7400744 | Nichani et al. | Jul 2008 | B2 |
7447337 | Zhang et al. | Nov 2008 | B2 |
7660438 | Camus | Feb 2010 | B2 |
7957652 | Kawanishi et al. | Jun 2011 | B2 |
7965866 | Wang et al. | Jun 2011 | B2 |
9177195 | Marcheselli et al. | Nov 2015 | B2 |
9305363 | Marcheselli et al. | Apr 2016 | B2 |
10402631 | Wang et al. | Sep 2019 | B2 |
20030076417 | Thomas et al. | Apr 2003 | A1 |
20040179736 | Yin | Sep 2004 | A1 |
20050249382 | Schwab et al. | Nov 2005 | A1 |
20060028557 | Watanabe | Feb 2006 | A1 |
20060088191 | Zhang et al. | Apr 2006 | A1 |
20060210117 | Chang et al. | Sep 2006 | A1 |
20070182818 | Buehler | Aug 2007 | A1 |
20070200701 | English et al. | Aug 2007 | A1 |
20070257985 | Estevez et al. | Nov 2007 | A1 |
20080285802 | Bramblet et al. | Nov 2008 | A1 |
20090010490 | Wang et al. | Jan 2009 | A1 |
20090326807 | Ramaswamy et al. | Dec 2009 | A1 |
20100063854 | Purvis et al. | Mar 2010 | A1 |
20100157062 | Baba et al. | Jun 2010 | A1 |
20110169917 | Stephen et al. | Jul 2011 | A1 |
20110175738 | Baumann et al. | Jul 2011 | A1 |
20110286633 | Wang et al. | Nov 2011 | A1 |
20120182390 | Purvis et al. | Jul 2012 | A1 |
20130182114 | Zhang et al. | Jul 2013 | A1 |
20130294646 | Shaw | Nov 2013 | A1 |
20160343146 | Brown et al. | Nov 2016 | A1 |
20190333245 | Zhao et al. | Oct 2019 | A1 |
Number | Date | Country |
---|---|---|
2433856 | Mar 2008 | GB |
2476869 | Jul 2011 | GB |
2008139203 | Nov 2008 | WO |
2009004479 | Jan 2009 | WO |
2009004479 | Jan 2009 | WO |
Entry |
---|
Eveland et al, Background Modeling for Segmentation of Video-Rate Stereo Sequences, Jun. 23-25, 1998. |
Darrell, et al, Integrated Person Tracking Using Stereo, Color, and Pattern Direction, pp. 1-8, Jun. 23-25, 1998. |
Haritaoglu et al, W4: Who? When? Where? What? A Real Time System for Detecting and Tracking People, 3. International Conference on Face and Gesture Recognition, Apr. 14, 16, 1998, Nara, Japan; pp. 1-6. |
Isard et al, Contour Tracking by Stochastic Propagation of Conditional Density, in Prc. European Conf. Computer Vision, 1996, pp. 343-356, Cambridge, UK. |
Paolo Remagnino et al; Correlation Techniques in Adaptive Template Matching With Uncalibrated Cameras, Lifia-Inria Rhones-Alples, Nov. 2, 1994. |
United Kingdom Combined Search and Examination Report dated Feb. 3, 2013, with respect to United Kingdom Patent Application No. GB1216792.0. |
Wren et al., Pfinder: Real-Time Tracking of the Human Body, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, No. 7, Jul. 1997; pp. 780-785. |
Search Report issued in EP Application No. 17207519.4 (dated 2018). |
Number | Date | Country | |
---|---|---|---|
20200074157 A1 | Mar 2020 | US |
Number | Date | Country | |
---|---|---|---|
61538554 | Sep 2011 | US | |
61549511 | Oct 2011 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16123769 | Sep 2018 | US |
Child | 16557166 | US | |
Parent | 15057908 | Mar 2016 | US |
Child | 15669440 | US | |
Parent | 14680123 | Apr 2015 | US |
Child | 15057908 | US | |
Parent | 13622083 | Sep 2012 | US |
Child | 14680123 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15669440 | Aug 2017 | US |
Child | 16123769 | US |