This application claims priority from Korean Patent Application No. 10-2013-0062261 filed on May 31, 2013 in the Korean Intellectual Property Office, the disclosure of which is incorporated herein by reference in its entirety.
1. Field of the Invention
The present invention relates to a people detection apparatus and method and a people counting apparatus and method, and more particularly, to a people counting apparatus and method employed to count the number of people entering and leaving a place using a video captured by a video capture device such as a closed circuit television (CCTV).
2. Description of the Related Art
Counting the number of people entering a shop is one of the important indices that measure a convergence ratio which is one of the important marketing elements. Currently, however, counting the number of people is being carried out manually. Counting the number of people with the human eye requires a lot of time and labor costs and does not guarantee accuracy.
Counting the number of people cannot only be used in shop management or as a marketing element but also be widely used in various fields. Therefore, it is required to develop an automated and highly accurate people counting algorithm.
Aspects of the present invention provide a moving object detection apparatus and method employed to accurately detect moving objects using a video captured by a video capture device.
Aspects of the present invention also provide a people detection apparatus and method employed to accurately detect human moving objects among detected moving objects in a video captured by a video capture device.
Aspects of the present invention also provide a people tracking apparatus and method employed to accurately track an object of interest or a human moving object detected in a video captured by a video capture device.
Aspects of the present invention also provide a people counting apparatus and method employed to accurately count the number of people using a video captured by a video capture device.
However, aspects of the present invention are not restricted to the one set forth herein. The above and other aspects of the present invention will become more apparent to one of ordinary skill in the art to which the present invention pertains by referencing the detailed description of the present invention given below.
According to an aspect of the present invention, there is provided a people counting apparatus including: a reception unit which receives a video of an area including an entrance captured by a video capture device; a line setting unit which sets an inline at the entrance and sets an outline such that a specific region is formed on a side of the inline; a detection unit which detects moving objects in the video using information differences between frames of the received video and detects human moving objects among the detected moving objects; a tracking unit which tracks the movement of each of the detected moving objects; and a counting unit which determines whether each of the moving objects passed the inline and the outline based on the tracked movement of each of the moving objects and counts the number of people based on the determination result, wherein the inline and the outline are virtual lines.
According to another aspect of the present invention, there is provided a people counting method including: receiving a video of an area including an entrance captured by a video capture device; setting an inline at the entrance and setting an outline such that a specific region is formed on a side of the inline; detecting moving objects in the video using information differences between frames of the received video and detecting human moving objects among the detected moving objects; tracking the movement of each of the detected moving objects; and determining whether each of the moving objects passed the inline and the outline based on the tracked movement of each of the moving objects and counting the number of people based on the determination result, wherein the inline and the outline are virtual lines.
According to another aspect of the present invention, there is provided a people detection apparatus including: a reception unit which receives a video frame of a video captured by a video capture device; an accuracy calculation unit which calculates the accuracy of detection for each pixel of the video frame using a Gaussian mixture model (GMM) method and a frame difference method and detects pixels whose calculated detection accuracy values are equal to or greater than a preset accuracy value as moving object regions; and a people detection unit which detects human moving objects among the detected moving objects by using positions of the detected moving object regions, sizes of the detected moving object regions, and a histogram of oriented gradient (HOG) which is a shape feature descriptor.
According to another aspect of the present invention, there is provided a people detection method including: receiving a video frame of a video captured by a video capture device; calculating the accuracy of detection for each pixel of the received video frame using a GMM method and a frame difference method and detecting pixels whose calculated detection accuracy values are equal to or greater than a preset accuracy value as moving object regions; and detecting human moving objects among the detected moving objects by using positions of the detected moving object regions, sizes of the detected moving object regions, and a HOG which is a shape feature descriptor.
The above and other aspects and features of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings, in which:
The present invention will now be described more fully with reference to the accompanying drawings, in which exemplary embodiments of the invention are shown. Advantages and features of the present invention and methods of accomplishing the same may be understood more readily by reference to the following detailed description of exemplary embodiments and the accompanying drawings. The present invention may, however, be embodied in many different forms and should not be construed as being limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete and will fully convey the concept of the invention to those skilled in the art, and the present invention will only be defined by the appended claims. Like reference numerals refer to like elements throughout the specification.
It will be understood that when an element is referred to as being “connected to” or “coupled to” another element, it can be directly connected or coupled to the other element or intervening elements may be present. In contrast, when an element is referred to as being “directly connected to” or “directly coupled to” another element, there are no intervening elements present.
Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated components, steps, operations, and/or elements, but do not preclude the presence or addition of one or more other components, steps, operations, elements, and/or groups thereof.
The present invention can count the number of people entering or leaving a place through an entrance by using a video of the entrance captured by a video capture device such as a closed circuit television (CCTV).
Specifically, the present invention can accurately detect moving objects in a noise-resistant manner in each frame of a video captured by a video capture device using a Gaussian mixture model (GMM) method, which is a long-term background recognition technique, and a frame difference method which is a short-term motion detection technique.
The present invention can also detect human moving objects among detected moving objects using a histogram of oriented gradient (HOG), which is a shape feature descriptor, and position and size information of the detected moving objects.
In addition, the present invention can track a detected human moving object using a Kalman filter, template matching, and scale invariant feature transform (SIFT).
Furthermore, the present invention can count the number of people entering or leaving a place through an entrance without being affected by noise in its performance of detecting and tracking moving objects by calculating a probability using Bayes' Rule.
Hereinafter, the present invention will be described in more detail reference to the attached drawings.
Conventional moving object detection methods include a background subtraction method that uses the difference in brightness between the background and an object and the frame difference method that detects motions from the difference between two successive image frames.
The background subtraction method is used to detect moving objects. If the background is complicated and changes significantly, how accurately the background is learned in real time may determine the accuracy of object detection. The GMM method is used to model the background and uses a probabilistic learning method. The brightness distribution of each pixel in an image is approximated using the GMM, and whether a measured pixel belongs to the background or an object is determined using an approximated model variable value.
The GMM based on long-term learning is robust to noise such as a shadow that instantly appears and then disappears or a change in light. In some cases, however, the GMM wrongly determines a moving object to be noise. On the other hand, the frame difference method sensitively senses a moving object. However, the frame difference method also sensitively senses various noises and fails to sense slow motions.
The moving object detection apparatus 100 according to the current embodiment calculates the accuracy of detection using both the GMM method and the frame difference method. The calculated accuracy of detection and the priority-based region expansion of a region detection unit 130 enable the moving object detection apparatus 100 to detect moving objects more accurately than the conventional methods.
The moving object detection apparatus 100 according to the current embodiment can detect moving objects accurately and in a noise-resistant manner in each frame of a video captured by a video capture device by using the GMM method which is a long-term background recognition technique and the frame difference method which is a short-term motion detection technique. Detecting moving objects is a process that must take precedence in order to count the number of people. Therefore, the accuracy of detecting moving objects is one of the important factors that can guarantee the accuracy of counting the number of people.
Referring to
The reception unit 110 may receive and store a video captured by a video capture device such as a CCTV.
The accuracy calculation unit 120 calculates the accuracy of each pixel in a frame of the video received by the reception unit 110.
Specifically, the accuracy calculation unit 120 may calculate the accuracy of detection using the GMM method and the frame difference method.
More specifically, the accuracy calculation unit 120 may calculate the accuracy of detection using Equation (1):
a(x,y)=ag(x,y)+af(x,y), (1)
where (x,y) represents the position of a pixel in a frame. In Equation (1), a(x,y) represents the accuracy of each pixel located at (x,y) in each frame of a captured video. That is, in the GMM method, the higher the probability calculated by ag(x,y), the higher the accuracy. In the frame difference method, the greater the difference between pixels located at the same positions in different frames, the higher the accuracy. A pixel having a high accuracy value is hardly likely to be noise and highly likely to be a moving object. Conversely, a pixel having a low accuracy value is highly likely to be noise and hardly likely to be a moving object.
In addition, ag(x,y) and af(x,y) may be defined by Equation (2) and Equation (3), respectively:
Equation (2) uses the GMM method and is a probabilistic model that defines brightness changes at a position (x,y) in f frames by using K Gaussian models. Here, f and K may vary according to environment. For example, f may be set to 100, and K may be set to 3. Therefore, ag(x,y) in Equation (2) defines a brightness model of the background. When a new frame is received, the probability that the brightness of a pixel located at (x,y) will be an object may be calculated based on the defined model. In Equation (2), μi is the average of an ith Gaussian model, and σi is the variance of the ith Gaussian model. In addition, I(x,y) is a brightness value of a pixel located at (x,y), and
In Equation (3), It(x,y) represents a brightness value of a pixel located at (x,y) in a tth frame. That is, when a new frame is received by the reception unit 110, the accuracy calculation unit 120 may calculate a difference in brightness between pixels at the same positions in a previous frame and the new frame. Then, the accuracy calculation unit 120 may output a low accuracy value if the brightness difference is small and output a high accuracy value if the brightness difference is large.
Therefore, if Equation (1) is described using Equations (1) and (2), the accuracy calculation unit 120 may detect a pixel as a moving object if the calculated accuracy indicates that the pixel is neither the background nor noise. To detect moving objects, the accuracy calculation unit 120 may set an appropriate reference accuracy value Th1. The reference accuracy value Th1 may vary according to environment such as image resolution, situation, operation processing speed, etc.
The accuracy calculation unit 120 may calculate the accuracy a(x,y) in Equation (1) as a value between zero and one by using Equations (2) and (3).
The region detection unit 130 may normalize the accuracy a(x,y) such that a maximum value of the accuracy a(x,y) is one and may define a pixel having a value of one as a reference pixel. The region detection unit 130 may define a priority π(p) as in Equation (4) below. When a value of the priority π(p) is equal to or greater than a set threshold value Tτ, the region detection unit 130 may expand a region. When priority values of all pixels are less than the set threshold value Tτ, the region detection unit 130 may stop region expansion and detect a moving object region:
Equation (4) is an example of an equation that defines the priority π(p), p represents a pixel located at (x,y), and q represents pixels neighboring the pixel p in four directions (up, down, left and right). In addition, τ(p,q) may be defined by Equation (5):
where I(p) represents a brightness value at a position p, I(q) represents a brightness value at a position q, and Tτ is a threshold value that varies according to situation.
The results of detecting a moving object using the conventional moving object detection methods and the result of detecting a moving object using the moving object detection apparatus 100 of
Specifically,
In
People can be counted only after they are detected among detected moving objects. The people detection apparatus 300 according to the current embodiment can detect human moving objects among moving objects detected in a frame.
Referring to
The people detection unit 310 may detect human moving objects among detected moving objects using a location scale-based HOG (LSHOG) descriptor.
The LSHOG descriptor takes into account position information of moving objects detected by the moving object detection apparatus 100 of
For example, if detected moving objects are two or more people located close to each other or overlapping each other, there may be a problem in detecting the people. To solve this problem, the HOG which is a shape feature descriptor may be used.
The HOG may calculate gradients in m directions in n×n blocks as illustrated in
To overcome this limitation, the people detection apparatus 300 of
Specifically, the people detection unit 310 may establish a database of right and wrong answers using the LSHOG descriptor and detect human moving objects by applying nonlinear support vector machine (SVM) or random forest classification.
The people tracking apparatus 500 according to the current embodiment may track detected human moving objects using the Kalman filter, template matching, and the SIFT.
Specifically, referring to
In order to reduce errors caused by an occluded region and various noises and accurately track objects, the tracking unit 510 may track an object of interest detected in a video using a weight wd based on the accuracy of the object detected by the Kalman filter.
The Kalman filter estimates a predictive value by estimating a measurable variable and a predictive variable and is a method of tracking position and data by calculating a weight according to standard deviation.
Specifically, the tracking unit 510 may track moving people detected by the people detection apparatus 300 of
xt=(1−K)(1−wd)
Referring to Equation (6), the tracking unit 510 may track a detected moving person (object) in a video by using a Kalman filter including a Kalman gain K and the weight wd which is based on the accuracy of the detected object. In Equation (6), xt represents the predicted position information of a moving person in a tth frame. Here, xt can be inferred from
More specifically, the people detection apparatus 300 of
wd=βtemplateαtemplate+βhistogramαhistogram+βSIFTαSIFT, (7)
where β is an accuracy weight for each of αtemplate, αhistogram, and αsift. For example, βtemplate may be set to 0.7, βhistogram may be set to 0.1, and βsift may be set to 0.2. Each weight may vary according to the operating system environment of the present invention. Accuracy not applied due to constraints on the amount of calculation may be set to zero, and αtemplate may be defined by Equation (8):
αtemplate=exp[−SAD(xt,xt-1)] (8).
In Equation (8), SAD is the abbreviation of sum of absolute differences. That is, in Equation (8), SAD(xt,xt-1) is the sum of absolute values of differences between pixel values in templates having xt and xt-1 at their centers. A template may denote a block and may be the same as a block of a block matching algorithm (BMA) used in video compression such as H.26x, MPEG, etc.
The tracking unit 510 may calculate the accuracy αhistogram of Equation (7) according to the amount of calculation allowed in the operating system environment of the present invention. To calculate αhistogram, a probabilistic model is defined by applying the GMM to a color histogram of an object of interest in templates, and then a difference between histograms of the templates is calculated. That is, αhistogram may be the difference between matching template histograms of previous and current frames.
Additionally, the tracking unit 510 may calculate the accuracy αsift of Equation (7) according to the amount of calculation allowed in the operating system environment of the present invention. Here, αsift may be a difference between the same object in previous and current frames calculated using a feature descriptor (such as a brightness value vector in a block) in the SIFT. The same object in the previous and current frames may be a block covering an object in the case of a template.
As mentioned above with reference to Equation (7), accuracy not applied due to constraints on the amount of calculation may be set to zero. That is, whether to apply the accuracy αhistogram and the accuracy αsift may be determined based on the amount of calculation allowed.
Referring to
The counting unit 610 counts the number of people using a probability in order to produce the result of counting people regardless of the confusing detection of moving objects and the performance of the tracking unit 510.
Referring to
The line setter 612 may set an inline at an entrance and set an outline such that a specific region is formed on a side of the inline outside the entrance.
The entrance does not necessarily have a door, and various forms of entrance such as the entrance to a subway station may exist. In addition, the entrance may be a passing region for distinguishing the entry and exit of people. A region for distinguishing the entry and exit of people may be referred to as a passing region. That is, the entrance is not limited to a particular form. The passing region is not necessarily a region having an area but may also be a line. Throughout the Detailed Description of the Invention, a region for distinguishing the entry and exit of people will be referred to as an ‘entrance’ instead of a ‘passing region’ in order to help understand the present invention. That is, the ‘entrance’ is one of the criteria for determining entry or exit, such as social and customary concept, a user's setting, and an automatic door.
The size of the inline set at the entrance may be equal to or a little greater than the size of the entrance. In addition, the inline may be set at a predetermined distance from the entrance. The inline and the outline are virtual lines and may be set to be visible in an image. A region formed by the inline and the outline may vary according to a user's setting, the form of the entrance, the mobility of population, the size of the entrance, etc. The inline and the outline may be set such that the intention to enter or leave a place can be clearly identified.
A specific region formed by the inline and the outline may be quadrilateral. However, the specific region is not limited to the quadrilateral shape and may also have various polygonal shapes. The specific region formed by the inline and the outline may be located outside and/or inside the entrance. The outside and inside of the entrance may be set based on social concept. In an example, if the entrance is an automatic door of a building, the outside of the building may be considered as the outside of the entrance, and the inside of the building may be considered as the inside of the entrance. In another example, if the entrance is an entrance to a subway station, stairs or an escalator descending to the subway station may be considered as the inside of the entrance, and the sidewalk outside the subway station may be considered as the outside of the entrance.
To count the number of people leaving a place through the entrance, the line setter 612 may set the inline and the outline such that the specific region is formed in a region through which people intending to leave the place through the entrance should pass. That is, to count the number of people leaving the place through the entrance, the line setter 612 may set the outline outside the entrance.
Conversely, to count the number of people entering the place through the entrance, the line setter 612 may set the inline and the outline such that the specific region is formed in a region through which people intending to enter the place through the entrance should pass. That is, to count the number of people entering the place through the entrance, the line setter 612 may set the outline inside the entrance.
The counter 614 may count the number of people using the set inline and outline. That is, the counter 614 may count the number of human moving objects who passed the set inline and then the set outline as the number of people entering or leaving the place.
Specifically, if the line setter 612 sets the inline and the outline such that the specific region is set outside the entrance, the counter 614 may increase the number of people (Out count) leaving the place by one when a human moving object detected and tracked in a video passes the inline and then the outline.
Conversely, if the line setter 612 sets the inline and the outline such that the specific region is set inside the entrance, the counter 614 may increase the number of people (In count) entering place by one when a human moving object detected and tracked in a video passes the inline and then the outline.
Specifically, even if a moving object passes the outline after passing the inline several times, the counter 614 may increase the number of people leaving or entering the place by only one. Whether a moving object has passed the inline and/or the outline may be determined using the moving object detection apparatus 100 of
After a human moving object detected and tracked in a video passes the inline, it may move within a space formed by the inline and the outline and then pass the inline again. In this case, the counter 614 may not count the human moving object as a person leaving or entering a place. After a human moving object passes the inline, it may move in the space formed by the inline and the outline and then pass the outline. Only in this case, the counter 614 may count the human moving object as a person leaving the place.
In a conventional method, whether a human moving object detected and tracked in a video enters or leaves a place is determined based on one line or two lines. However, in this conventional method, even if a person moves variously but does not actually leave or enter a place, the person is counted as a person leaving or entering the place. For example, if a person moves around an entrance, the person is highly likely to be wrongly counted as a person entering or leaving a place several times in the conventional method using one line or two lines.
Specifically, the way the counter 614 counts the number of people leaving a place using an inline and an outline more accurately than a conventional method will now be described with reference to
Referring to
However, a moving object {circle around (2)} which moves around the line and moving objects {circle around (3)} and {circle around (4)} which are detected and tracked with low accuracy or the tracking of which used to be stopped cannot be counted accurately using the conventional method.
In the conventional method, the moving object {circle around (2)} may be counted as having entered a place three times and left the place three times. However, although the moving object {circle around (2)} intended to leave the place, it did not actually leave the place. Therefore, the moving object {circle around (2)} should be counted as having entered or left the place zero times.
In addition, the moving object {circle around (3)} may be counted as having left the place once in the conventional method. However, although the moving object {circle around (3)} intended to leave the place, it actually entered the place again instead of leaving the place. The moving object {circle around (4)} may be counted as having left the place twice and entered the place once in the conventional method. However, the moving object {circle around (4)} actually left the place after conflicting whether to leave or enter the place.
Unlike the conventional method, the people counting apparatus 600 of
That is, even if it is difficult to detect and track a moving object because the moving object changes its direction rapidly, the counter 614 according to the present invention can accurately count the number of people leaving or entering a place as compared with the conventional method. Detecting a moving object and tracking the detected moving object may be performed using the people detection apparatus 300 of
Specifically, since the moving object {circle around (1)} was detected and tracked as having passed the inline and then outline, the counter 614 may increase the number of people leaving a place by one. The moving object {circle around (2)} was detected and tracked as having passed the inline but not the outline, the counter 614 may not increase the number of people leaving the place.
In the case of the moving object {circle around (3)}, there was a time (e.g., a dotted portion in the path of the moving object {circle around (3)} in
Like the moving object {circle around (3)}, in the case of the moving object {circle around (4)}, there was a time (a dotted portion in the path of the moving object {circle around (4)} in
Referring to
The counting unit 610 may count moving objects which passed the inline and then the outline as people entering a place among moving objects detected and tracked by the line setter 612. That is, in
Referring back to
If the line setter 612 sets an inline and first and second outlines as illustrated in
Referring to
Conversely, if the line setter 612 sets an outline such that a specific region is formed outside an entrance, an inline can exist inside the entrance. Even if the inline exists inside the entrance, the outline may exist outside the entrance, and a wide specific region may be formed outside the entrance by the inline and the outline.
Various examples of setting an inline and an outline using the line setter 612 have been described above with reference to
Specifically, a method of counting the number of people using the counter 614 based on whether each person has passed an inline and an outline will be described with reference to Equations (9) through (12).
The counter 614 may count the number of people using a probability p(S|X) expressed by Bayes' Rule. The probability p(S|X) is given by Equation (9):
In Equation (9), p(S|X) represents a probability that a state S will belong to one of ‘In count’, ‘Out count’, and ‘Not count’ when a trajectory X is given. In count indicates an entering situation, Out count indicates a leaving situation, and Not count indicates an uncounted situation. In Equation (9), a value of p(S) may be preset. For example, assuming that the number of people entering a place is equal to the number of people leaving the place, p(S=In count) may be set to 0.4, and p(S=Out count) may be set to 0.4. The value of p(S) may vary according to the operating system environment of the present invention, a place captured by a CCTV, etc.
As described above, when a moving object passes an outline and then an inline, In count is increased by one. In addition, when a moving object passes the inline and then the outline, Out count is increased by one. Not count may indicate all situations excluding the entering and leaving situations.
The trajectory X denotes {x0, x1, . . . , xt-1, xt}, where xt is position information of a moving person predicted and tracked in a tth frame. Therefore, the trajectory X is a set of xt in frames. p(S|X=IN) may be given by Equation (10):
p(X|S=IN)=p(xt|S=IN)·p(x0|S=IN) (10).
Referring to Equation (10), the counter 614 may count the number of people using a position x0 detected first in the trajectory X and a position xt detected last in the trajectory X. p(xt|S=IN) and p(x0|S=IN) in Equation (10) will be described in more detail using Equation (11) and Equation (12), respectively.
Specifically, when the detected position xt is given, a probability that a person entered a place may be defined by Equation (11):
In Equation (11), αin is a probability value in a case where xt exists on an outline and may be set experimentally. For example, αin may be set to a value equal to or greater than 0.7. p(x0|S=IN) may be defined by Equation (12):
In Equation (12), |ΣitΔyi| represents the sum of y-axis gradients of a trajectory X, and L represents a y-axis length of a counting region. In addition, a value of βin may be determined experimentally. For example, βin may be set experimentally to a value equal to or greater than 0.9. A value of γ may be zero if the sum of the y-axis gradients is a negative number and may be one if the sum of the y-axis gradients is a positive number.
If p(S=IN|X)>0 and p(S=IN|X)/p(S=NOT|X)>T, the counter 614 increases In Count. Here, p(S=NOT|X) is given by 1−p(S=IN|X)−p(S=OUT|X). Assuming that p(X) is a value greater than zero, since the value exists in both a denominator and a numerator, it may be a variable that can be offset. T may be set according to the operating system environment of the present invention and may be set to, e.g., approximately 0.3.
The moving object detection apparatus 100 of
The people detection apparatus 300 of
The people tracking apparatus 500 of
The people counting apparatus 600 of
The people counting apparatus 600 of
Referring to
The accuracy calculation unit 120 may calculate the accuracy of detection for each pixel of the video frame received by the reception unit 110 by using a GMM method and a frame difference method and detect pixels whose calculated detection accuracy values are equal to or greater than a preset accuracy value as moving object regions (operation S1320).
The region expansion unit 130 may expand the moving object regions detected by the accuracy detection unit 120 based on a priority π(p) that uses a difference in brightness between the pixels (operation S1330). A detailed description of the priority π(p) can be found in the description of
The people detection unit 310 may detect human moving objects among the detected moving objects by using positions of the moving object regions expanded by the region expansion unit 130, sizes of the moving object regions expanded by the region expansion unit 130, and a HOG which is a shape feature descriptor (operation S1340).
Referring to
The line setter 612 may set an inline at the entrance and set an outline such that a specific region is formed on a side of the inline, wherein the inline and the outline are virtual lines (operation S1420).
The accuracy calculation unit 120 and the region detection unit 130 may detect moving objects in the video frame received by the reception unit 110 (operation S1430), and the people detection unit 310 may detect human moving objects among the detected moving objects (operation S1440).
The tracking unit 510 may track the movement of each detected human moving object (operation S1450). Information detected and tracked in operations S1430, S1440 and S1450 may be stored in a database before or after operation S1420.
The counting unit 610 may determine whether each moving object has passed the inline and the outline based on information obtained by tracking the movement of each human moving object and count the number of people based on the determination result (operation S1460).
Specifically, if the line setter 612 sets the outline outside the entrance, the counting unit 610 may increase the number of people leaving a place only when a moving object passes the inline and then the outline. On the other hand, if the line setter 612 sets the outline inside the entrance, the counting unit 610 may increase the number of people entering the place only when a moving object passes the inline and then the outline (operation S1460).
Each component described above with reference to
The present invention can accurately detect moving objects in a frame of a video captured by a video capture device.
The present invention can also accurately detect people among the moving objects detected in the frame of the video captured by the video capture device.
The present invention can also accurately track moving people detected in the frame of the video captured by the video capture device.
The present invention can also accurately count the number of people using the video captured by the video capture device.
The present invention counts the number of people using an inline and an outline. Therefore, the present invention can count the number of people with high accuracy as compared with a conventional method of counting the number of people using only one line.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the present invention as defined by the following claims. The exemplary embodiments should be considered in a descriptive sense only and not for purposes of limitation.
Number | Date | Country | Kind |
---|---|---|---|
10-2013-0062261 | May 2013 | KR | national |
Number | Name | Date | Kind |
---|---|---|---|
20050201612 | Park et al. | Sep 2005 | A1 |
20100092030 | Golan | Apr 2010 | A1 |
20110274315 | Fan | Nov 2011 | A1 |
20120057640 | Shi | Mar 2012 | A1 |
20120057748 | Katano | Mar 2012 | A1 |
20130155229 | Thornton | Jun 2013 | A1 |
20140010456 | Merler | Jan 2014 | A1 |
20140037147 | Yoshio | Feb 2014 | A1 |
20140139660 | Zhu | May 2014 | A1 |
20150009332 | Fuhrmann | Jan 2015 | A1 |
Number | Date | Country |
---|---|---|
10-049718 | Feb 1998 | JP |
10-2004-0079550 | Sep 2004 | KR |
10-0519782 | Oct 2005 | KR |
10-2010-0121817 | Nov 2010 | KR |
10-1064927 | Sep 2011 | KR |
Entry |
---|
Communication dated Sep. 17, 2014 issued by Korean Intellectual Property Office in counterpart Korean Application No. 10-2013-0062261. |
Lefloch, “Real-Time People Counting system using Video Camera”, Master of Computer Science, Image and Artificial Intelligence, 2007, 60 pages total. |
Communication dated Dec. 22, 2014, issued by the Korean Intellectual Property Office in counterpart Korean Application No. 10-2013-0062261. |
Chuanying Gao, “Research of people counting by video sequence based on DSP”, China Master's Theses Full-text Database, 2011, No. 4. |
Communication dated Jul. 14, 2016, issued by the State Intellectual Property Office of P.R. China in counterpart Chinese application No. 201410234167.4. |
Number | Date | Country | |
---|---|---|---|
20140355829 A1 | Dec 2014 | US |