The present application is a U.S. national stage application under 35 U.S.C. § 371 of PCT Application No. PCT/CN2017/079520, filed Apr. 6, 2017, which claims priority to Chinese Patent Application No. 2016104785058, filed Jun. 27, 2016. The disclosures of the aforementioned priority applications are incorporated herein by reference in their entireties.
The invention relates to the field of digital image processing, in particular to a surveillance video based video foreground extraction method, which can be applied to preprocessing of a surveillance video so as to further analyze the surveillance video.
Foreground extraction of a video refers to separating a foreground object in the video from a background. As an increasingly common video type, surveillance video has the characteristics of long time, complex foreground and increasing resolution. In real life, surveillance videos are necessary for analysis of some cases and accidents since the surveillance videos can provide more evidence of the cases.
Existing foreground extraction technologies mainly use a Gaussian mixture model, an expectation maximization (EM) algorithm, saliency detection, a Trimap algorithm, a GrabCut algorithm and a superpixel method to extract foreground objects. However, the video foreground extraction methods have the following disadvantages:
(1) the above methods can partly extract a salient region in a video, but the extraction process does not take into account the difference between a static foreground and a dynamic foreground, which leads to a globally uniform background, while a foreground object in the stationary state should be seen as a new background;
(2) for surveillance videos with an increasingly large size, the processing speed of the above methods is slower and slower, and the processing efficiency for large-size videos is low; and
(3) most of the above methods have high algorithm complexity and poor processing performance.
In order to overcome the shortcomings of the prior art described above, the present invention provides a video foreground extraction method for a surveillance video, which adjusts a size of a block to adapt to different video resolutions based on an image block processing method; and then extracts a foreground object in a moving state by establishing a background block model. The method can eliminate the influence on video analysis due to increasing size of the surveillance video.
For convenience of explanation, the following terms and parameters are agreed in the specification: (1) defining variables the dimension of which is the same as that of an original image: original image I; temporary background β; background BB; temporary foreground γ; foreground F; and
(2) defining variables the dimension of which is the same as that of an image after dimension reduction: block image I′; temporary block background α; block background A; background block weight Ψ; foreground block weight Φ; foreground block end ω.
Thresholds ε1, ε2 and ε3 are set to determine the weight.
The principle of the invention is that a surveillance video contains multiple frames of images, while a background of the surveillance video is static, and the invention processes each frame of image in the surveillance video based on a block method and adjusts a size of a block to adapt to different video resolutions. A foreground object in a moving state is extracted by establishing a background block model. The dimension of each frame of image in the surveillance video is reduced, and the influence of the illumination on the block method is eliminated by means of the pixel-based comparison. Meanwhile, the influence on video analysis due to increasing size of the surveillance video is eliminated. Therefore, the method of the invention can extract a dynamic foreground object, adapt to a larger size of a surveillance video, and has a lower algorithm complexity.
The technical scheme provided by the invention is as follows:
Disclosed is a video foreground extraction method for a surveillance video, which adjusts a size of a block to adapt to different video resolutions based on an image block processing method; and then extracts a foreground object in a moving state by establishing a background block model. The method specifically comprises the following steps:
The first step is to gray each frame of image I and calculate an integral image, through which the sum of pixels in the block of different sizes can be conveniently calculated, and the sum of pixels in the block represents all pixels in the block, thus obtaining an image matrix block image I′ with a smaller dimension instead of the original image. There are m*n pixel values in a block. The present invention uses the sum of the m*n pixel values in the block to represent the block, that is, the original m*n dimension has changed to the current 1*1 dimension.
The second step is initializing. For the first frame of image, the background B is initialized with the first frame of image in the video, wherein the background refers to the rest of each frame of image except the foreground object. The foreground F is initialized with a zero matrix, wherein the foreground refers to the foreground object in each frame of image. Intermediate results during processing are preserved with the temporary background β and the temporary foreground γ. The block background A and the temporary block background α are initialized with the block image I′. The background block weight ψ and the foreground block weight Φ are initialized with the zero matrix, wherein the weight refers to the duration that a certain block in the image lasts in a video stream, the initializing value of all the weights is 0, and the dimension is consistent with the dimension of the block background. The foreground block end ω is initialized with the zero matrix, wherein the end is used for judging the continuity of a block flow, and the dimension is consistent with that of the block image I′.
The third step is to compare each element in the current frame of block image I′i with a corresponding element in the previous frame of block image I′i-1, and if the absolute value of the difference between the two exceeds the threshold ε1, it is indicated that the difference between the corresponding blocks in the two frames is large and the current block may be the foreground. The background block weight ψ is auto-increased by 1, and the temporary block background α is obtained by weighting and averaging the elements in the previous and current frames of block images. Otherwise, the background block weight ψ is reset to be 0, and the temporary block background α is reset to be the current element, wherein the temporary block background refers to that the current block may be part of the background, and whether it is the background depends on the fourth step.
In the fourth step, if the elements in the background block weight ψ exceed the threshold ε2, it is indicated that the duration of the current block is long enough, and the block background A is updated with the elements in the temporary block background α.
In the fifth step, the first case is to compare an element in the block image I′i with a corresponding element in the previous frame of block image I′i-1, and the difference between the two exceeds the threshold ε1; the second case is to compare an element in the block image I′ with a corresponding element in the block background A, and the difference between the two exceeds the threshold ε1; and the above two cases both indicate that the current block may be a potential foreground block. If one of the above conditions is met, the sixth step is executed; otherwise, the third step is executed.
The sixth step is to compare the pixels in the original image I corresponding to the elements in the current block image I′ with the pixels corresponding to the background B, if the pixel difference exceeds the number of pixels of the threshold ε3, that is, half of the number of pixels in the block, the seventh step is executed; otherwise, the third step is executed.
The seventh step is to save the difference between the pixel value in the original image I and the pixel value corresponding to the background B as the foreground F. For the second case of the fifth step, whether the foreground block is continuous or not is judged according to the foreground block end ω, and if yes, the background block weight ψ is not 0, which indicates that the current foreground is likely to be a new background, and the foreground block weight Φ is auto-increased by 1. The temporary foreground γ is obtained by weighting and averaging the pixels in the corresponding blocks in the current and previous frames. The temporary foreground herein refers to that the current block may be part of the background, and whether it is the background depends on the eighth step.
In the eighth step, if the foreground block weight Φ exceeds the threshold ε2, the temporary foreground γ is set as the background B, the temporary block background is calculated, the background block weight is set to be ε2, and the temporary foreground γ is set to be 0.
The ninth step is to binarize the foreground F obtained in the seventh step, and the current frame number is saved as the foreground block end ω.
The tenth step is to repeat steps 3-9 till all frames in the video are processed.
Compared with the prior art, the invention has the beneficial effects that:
Disclosed in the present invention is a video foreground extraction method for a surveillance video, which adjusts a size of a block to adapt to different video resolutions based on an image block processing method; and then extracts a foreground object in a moving state by establishing a background block model. The dimension of each frame of image in the surveillance video is reduced, and the influence of the illumination on the block method is eliminated by means of the pixel-based comparison. Meanwhile, the influence on video analysis due to increasing size of the surveillance video is eliminated. Therefore, the method of the invention can extract a dynamic foreground object, adapt to a larger size of a surveillance video, has a lower algorithm complexity, can be applied to the preprocessing of surveillance videos, provides support for further analysis of surveillance videos, and has important significance in current days when more and more surveillance cameras are used. Compared with the prior art, the invention has the following beneficial effects:
(1) based on the block processing method, the invention changes the comparison of a single pixel in the prior art to the comparison of the pixel sum, which relieves most of the calculation pressure in practical application, greatly reduces the calculation complexity, and has high execution efficiency;
(2) the invention realizes stable extraction of dynamic objects (i.e., foreground extraction) in the surveillance video by establishing the background block model and updating the background; and
(3) by introducing the comparison of pixels in the block, the invention eliminates the influence of abrupt changes in illumination on the block method, and the method has high precision.
wherein the resolution of the image is 1920*1080 and the block size is 60*60, so the original image I can be represented by a block image′ of 32*18.
The present invention will be further described by way of embodiments with reference to the accompanying drawings, but the scope of the invention will not be limited in any way.
The invention provides a video foreground extraction method for a surveillance video, which adjusts a size of a block to adapt to different video resolutions based on an image block processing method; and then extracts a foreground object in a moving state by establishing a background block model. The method can eliminate the influence on video analysis due to increasing size of the surveillance video.
Variables, defined in the following method or equations, the dimension of which is the same as that of an original image include: temporary background β; background B; temporary foreground γ; foreground F; Variables defined after dimension reduction of the original image include: temporary block background α; block background A; background block weight Ψ; foreground block weight Φ; foreground block end ω. Thresholds 1, 2 and 3 are used to determine the weight.
Step 1: as shown in
bs,t=Σi=1pΣj=1qpi,j(s∈[0,m],t∈[0,n]) (equation 1)
wherein s and t represent rows and columns of a corresponding new matrix respectively; and pi, j represents pixel values of the grayscale images.
Thus, the original image can be represented by a matrix I′ as shown in equation 2:
Step 2: initializing the temporary background β and the background B with I, wherein the temporary foreground is represented with a zero matrix γ with the size of P*Q, the foreground is represented with F, the block background A and the temporary block background α are initialized with a matrix I′, the background block weight is represented with a zero matrix ψ with the size of m*n, the foreground block weight is represented with Φ, and the foreground block end is represented with ω, as shown in equation 3 and equation 4:
ω/ψ/φ=[0]m*n (equation 3)
/γ=[0]P*Q (equation 4)
Step 3: comparing the element bs,tI′
wherein s and t represent all pixels in the image block corresponding to the current block;
in this embodiment, ε1 represents the threshold of the difference between pixel sums (i.e., blocks herein),
wherein the threshold ε3 is used to compare the pixels to obtain a difference; when ε3 is 15, ε1=60, and the reason why the denominator of the equation is 2 is to offset the influence of the block method on the accuracy.
Step 4: representing the duration that a block stream remains unchanged with the threshold ε2; If the value of a block background weight image exceeds the threshold ε2, the block background is updated with the temporary block background, and the background is updated with a temporary background, as shown in equation 6:
s,t=αs,t, s,t=βs,t(ψs,t>ε2) (equation 6)
In this embodiment, ε2=15, indicating that the size of the block stream is at least 15 frames, in this case, it is believed that the image conforms to the assumption that the background of the image in the surveillance video is static.
Step 5: when the difference between the current block and the background block exceeds the threshold ε1 or the background block weight is 0, executing step 6; otherwise, executing step 3.
Specifically, the first case is to compare an element in the block image I′ with a corresponding element in the block background A, and if the difference between the two exceeds the threshold ε1, as shown in equation 7, step 6 is continued:
|bs,tI′−|>ε1 (equation 7)
in the second case, if the background block weight is ψs, t is 0, as shown in equation 8, step 6 is continued:
ψs,t=0 (equation 8)
otherwise, step 3 is executed.
Step 6: comparing the pixels in the current block with the pixels in a block corresponding to a background frame, if the pixel difference exceeds the number of pixels of the threshold ε3, that is, half of the number of pixels in the block, as shown in equation 9:
wherein f( ) represents the count of pixels within the block meeting the conditions; Ib
Step 7: saving the difference between the pixel value in the original image I and the pixel value corresponding to the background B as the foreground F. For the first case (the difference between the current block and the background block exceeds the threshold ε1) of step 5, whether the foreground block is continuous or not is judged, if yes and the difference between the current and previous foreground blocks is less than the threshold ε1, the foreground block weight is auto-increased by 1 and the foreground blocks are weighted and averaged, i.e. as shown in equation 10:
Step 8: updating the background B with the temporary foreground if the value of the weight of the block foreground image exceeds the threshold ε2, and updating the block background and the temporary block background by summing pixel values in each block of the updated background according to equation 11, wherein the block background and the temporary block background have the same value, and resetting the block background weight to be the threshold ε2, namely equation 11:
s,t=γs,ts,t=αs,t=Σp(p∈γs,t), ψs,t=φs,tφs,t>ε2 (equation 11)
wherein B is the background; s and t represent rows and columns of a corresponding new matrix respectively; γ represents the temporary foreground; A represents the block background; α represents the temporary block background; ψ represents the background block weight; Φ represents the foreground block weight.
This step enables the background to be updated in real time. For example, a car is a foreground when moving, but is updated to be a background when stopped. What is obtained in step 7 is a foreground. The step controls the real-time update of the background by setting a threshold. Judge according to the duration that the foreground stays static, if it does not move for a long time, the foreground is taken as a background.
Step 9: binarizing the foreground F obtained in step 7 to obtain the final foreground result, and saving the current frame number as the foreground block end ωs,t;
The order of step 8 and step 9 can be exchanged.
Step 10: repeating steps 3-9 till all frames in the video are processed so as to obtain the video foreground through extraction.
It should be noted that the purpose of the disclosed embodiments is to help further understand the invention, but those skilled in the art will understand that various substitutions and modifications are possible without departing from the spirit and scope of the invention and the appended claims. Therefore, the invention should not be limited to what is disclosed in the embodiments, and the scope of protection of the invention is subject to the scope defined in the claims.
Number | Date | Country | Kind |
---|---|---|---|
2016 1 0478505 | Jun 2016 | CN | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CN2017/079520 | 4/6/2017 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2018/000882 | 1/4/2018 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
4896446 | Gregory | Jan 1990 | A |
5392553 | Carey | Feb 1995 | A |
5410833 | Paterson | May 1995 | A |
5970642 | Martin | Oct 1999 | A |
7428794 | Oz | Sep 2008 | B2 |
7640688 | Oz | Jan 2010 | B2 |
7647719 | Fitzpatrick et al. | Jan 2010 | B2 |
7930849 | Abraham et al. | Apr 2011 | B2 |
8061072 | Crose | Nov 2011 | B1 |
8141288 | Dodd et al. | Mar 2012 | B2 |
8186090 | Chiarolanza et al. | May 2012 | B1 |
8381427 | Nill | Feb 2013 | B2 |
8402683 | Cabahug et al. | Mar 2013 | B2 |
8453365 | Ballard | Jun 2013 | B1 |
D704294 | Buttstock | May 2014 | S |
8720099 | Sisk | May 2014 | B1 |
8763296 | Chvala | Jul 2014 | B1 |
8844185 | Jarboe | Sep 2014 | B2 |
8950099 | Rogers et al. | Feb 2015 | B2 |
8955245 | Chvala | Feb 2015 | B2 |
D736336 | Mayberry et al. | Aug 2015 | S |
9360272 | Hopkins | Jun 2016 | B2 |
9404708 | Chow et al. | Aug 2016 | B1 |
9410764 | Jarboe | Aug 2016 | B2 |
9464863 | Mather et al. | Oct 2016 | B2 |
9546845 | Mather | Jan 2017 | B2 |
9664476 | Robinson et al. | May 2017 | B1 |
9664479 | Robinson et al. | May 2017 | B1 |
20060254111 | Giauque et al. | Nov 2006 | A1 |
20070289190 | Oz | Dec 2007 | A1 |
20120036757 | Larue | Feb 2012 | A1 |
20140028794 | Wu et al. | Jan 2014 | A1 |
20140360074 | Jarboe | Dec 2014 | A1 |
20150338186 | Hopkins | Nov 2015 | A1 |
20150345881 | Irvin | Dec 2015 | A1 |
20160084612 | Robinson et al. | Mar 2016 | A1 |
20170059271 | Bowen | Mar 2017 | A1 |
20170358093 | Baltsen | Dec 2017 | A1 |
20180181814 | Dong | Jun 2018 | A1 |
20180197294 | Pan | Jul 2018 | A1 |
20180293457 | Li | Oct 2018 | A1 |
Number | Date | Country |
---|---|---|
101216888 | Jul 2008 | CN |
102103751 | Jun 2011 | CN |
103077520 | May 2013 | CN |
106204636 | Dec 2016 | CN |
2014022170 | Jun 2014 | WO |
Entry |
---|
International Search Report and Written Opinion for Corresponding Application No. PCT/CN2017/079520, dated Jun. 26, 2017. WO. |
Magpul PRS GEN3 Precision Adjustable Stock (Options)—MSR Arms, LLC, https://www.msrarms.com, 2017, pp. 1-2. |
Reaser, Rob, The AR-15 You Can't Afford NOT to Buy, Shooting Daily, 2017, pp. 1-4. |
International Search Report and Written Opiniion, Patent Cooperation Treaty, dated Jan. 7, 2019, pp. 1-16. |
Number | Date | Country | |
---|---|---|---|
20190108642 A1 | Apr 2019 | US |