This application claims the benefit of non-provisional patent application number 1753/CHE/2010 titled “Capturing Events Of Interest By Spatio-temporal Video Analysis”, filed on Jun. 23, 2010 in the Indian Patent Office.
The specification of the above referenced patent application is incorporated herein by reference in its entirety.
Video content analysis (VCA) refers to processing of video for determining events of interest and activity in a video, such as, a count of people or vehicles passing through a zone, the direction and speed of their movement, breach of boundary, speeding vehicles, etc. using computer vision techniques. VCA finds numerous applications in video surveillance, customer behavior in super markets, vehicular traffic analysis, and other areas. VCA is becoming a necessity given the extent of breach of security, requirement for surveillance, and threats to or potential compromise of human life and property in cities, defense establishments and at industrial and commercial premises.
Existing VCA algorithms process an entire image or a patch of the image using pure spatial image processing of the video. Hence, these algorithms are inherently suboptimal with respect to their computational efficiency and memory utilization. For example, processing visual graphics array (VGA) image frames of 640×480 pixels each in pure spatial domain requires segmentation of the entire 307200 pixels of each image frame, which is highly memory intensive.
Hence, there is a long felt but unresolved need for a computer implemented method and system for capturing events of interest by performing spatio-temporal analysis of a video using user-defined lines of analysis.
This summary is provided to introduce a selection of concepts in a simplified form that are further described in the detailed description of the invention. This summary is not intended to identify key or essential inventive concepts of the claimed subject matter, nor is it intended for determining the scope of the claimed subject matter.
The computer implemented method and system disclosed herein addresses the above stated need for capturing events of interest by optimally performing spatio-temporal analysis of a video using one or more user-defined lines of analysis. A continuous video stream containing a series of image frames is acquired over time. Each of the image frames is represented by horizontal spatial coordinates and vertical spatial coordinates of a two dimensional (2D) plane. A temporal dimension is assigned across the image frames of the video stream. The temporal dimension represents the order of the image frames using predefined temporal coordinates. A spatio-temporal analysis image is constructed based on a user-defined line of analysis on each of one or more of the image frames. The spatio-temporal analysis image is constructed by concatenating a series of temporally-successive linear pixel arrays along the temporal dimension. Each of the linear pixel arrays comprises an array of pixels selected along the line of analysis defined on each of one or more of the image frames. For example, the spatio-temporal analysis image is constructed by concatenating linear pixel arrays from every Nth successive image frame in the series of image frames. The width of each of the linear pixel arrays is substantially identical to the width of the line of analysis. The line of analysis is, for example, at least a single pixel wide. The constructed spatio-temporal analysis image is segmented for capturing the events of interest. The constructed spatio-temporal analysis image provides a summary of events occurring over the image frames at the line of analysis for the duration of the video or the duration of the concatenation of the series of temporally-successive linear pixel arrays. A user can define one or more lines of analysis for parallely constructing multiple spatio-temporal analysis images that are related in time.
The capture of the events of interest at the line of analysis comprises, for example, detecting presence of an object, detecting traversal of an object, determining speed of the traversal of the object, determining an object count based on the traversal of one or more objects, and determining duration of the presence of the object.
A user defines the line of analysis, having any orientation in the 2D plane, on one or more of the image frames. The user-inputted line of analysis on one of the image frames is used to automatically accumulate the linear pixel arrays from the remaining successive image frames by replicating the same coordinates of the line of analysis over these image frames. Accordingly, the line of analysis is defined by the horizontal spatial coordinates, or the vertical spatial coordinates, or by a combination of the horizontal spatial coordinates and the vertical spatial coordinates in each of the image frames. The constructed spatio-temporal analysis image is represented by a fixed spatial coordinate, a range of variable spatial coordinates, and variable temporal coordinates. For example, the constructed spatio-temporal analysis image is represented by a combination of a fixed horizontal spatial coordinate, a range of the vertical spatial coordinates and variable temporal coordinates, or a combination of a fixed vertical spatial coordinate, a range of the horizontal spatial coordinates and variable temporal coordinates.
The segmentation of the constructed spatio-temporal analysis image comprises foreground segmentation for detecting objects and the events of interest. A background of the constructed spatio-temporal analysis image is modeled, for example, by determining a moving average of the linear pixel arrays along the temporal dimension. The modeled background is subtracted from the constructed spatio-temporal analysis image for obtaining the foreground of the constructed spatio-temporal analysis image. Segmentation of the obtained foreground is then performed for detecting the objects and the events of interest.
Different events of interest are captured using the spatio-temporal analysis image based on the actual events occurring in the actual scenario being captured. One of the captured events is the speed of traversal of the object, for example, a moving vehicle. The speed of traversal of the object is determined by receiving one or more lines of analysis spaced apart from each other by a separation distance on the image frames from a user. The separation distance is based on an actual distance in the actual scene being captured. One or more spatio-temporal analysis images are constructed based on the lines of analysis. The presence and the times of presence of the object on the constructed spatio-temporal analysis images are determined using foreground segmentation. The speed of traversal of the object is determined based on the separation distance, frame rate of the video, and difference between times of presence of the object on the constructed spatio-temporal analysis images.
The foregoing summary, as well as the following detailed description of the invention, is better understood when read in conjunction with the appended drawings. For the purpose of illustrating the invention, exemplary constructions of the invention are shown in the drawings. However, the invention is not limited to the specific methods and instrumentalities disclosed herein.
The capture of events of interest at the line of analysis comprises, for example, detecting presence of an object, detecting traversal of the object, determining speed of the traversal of the object, determining an object count based on the traversal of one or more objects, determining duration of presence of the object, etc.
The user defines the line of analysis, having any in-plane orientation with respect to the two dimensional (2D) plane, on one or more of the image frames. The user-inputted line of analysis on one of the image frames is used to automatically accumulate the linear pixel arrays from the remaining successive image frames by replicating the same coordinates of the line of analysis over these image frames. Accordingly, the line of analysis is defined by the horizontal spatial coordinates or the vertical spatial coordinates, or by a combination of the horizontal spatial coordinates and the vertical spatial coordinates in one or more of the image frames. The constructed spatio-temporal analysis image is represented by a fixed spatial coordinate, a definite range of variable spatial coordinates and variable temporal coordinates. For example, the constructed spatio-temporal analysis image is represented by a combination of a fixed horizontal spatial coordinate, a range of vertical spatial coordinates and variable temporal coordinates, or a combination of a fixed vertical spatial coordinate, a range of horizontal spatial coordinates and variable temporal coordinates.
A normal static 2D image frame can be expressed as lines arranged horizontally or vertically such that each line is obtained from pixels Pij at j=n for all values of i or i=m for all values of j to a maximum number of rows or columns. The static 2D image frame comprises pure spatial data in XY coordinates. Similarly, a video stream comprising 2D static image frames stacked in time (T) is visualized as a cube 201 of image frames, wherein the third dimension or the transverse dimension is time (T).
Mathematically, each individual frame of the moving images is represented by I(y, x) and the video stream is represented by I(y, x, t), where x, y and t represent columns, rows and a time-axis, respectively, with the assumption that “x” varies from 0 to N, “y” varies from 0 to M, and “t” varies from 0 to R. A user selects the line of analysis 202 by providing coordinates of the end points of the line 202 using a user interface, for example, a graphical user interface or a textual user interface. The user is presented with one of the image frames, for example, I(y, x, 3) which is the image frame at t=3, as exemplarily illustrated in
In an example, if the line of analysis 202a is selected along row number 10 and entire “x”, the resulting spatio-temporal analysis image 203 is given by I(10, x, t), where “x” varies from 0 to N, and “t” varies from 0 to R. The values of “x” can be specified by a range “x1” to “x2”, providing the image I(10, β, t), where “β” varies from x1 to x2, and “t” varies from 0 to R. In another example, if the line of analysis 202b is selected along column number 23 and entire “y”, the resulting spatio-temporal analysis image 203 is mathematically given by I(y, 23, t), where “y” varies from 0 to M, and “t” varies from 0 to R. The values of “y” can be specified within a range “y1” to “y2”, providing the image I(α, 23, t), where “α” varies from y1 to y2, and “t” varies from 0 to R. Similarly, for a line of analysis 202c selected between any two points, at any inclination and length within the image resolution results in an spatio-temporal analysis image I(α, β, t), where (α, β) specifies the points on the line of analysis 202c varying between y and x values of the two end-points and “t” varies from 0 to R. In an embodiment, the inclined linear arrays of pixels along the line of analysis 202c from the image frames are re-arranged to form vertical lines on the spatio-temporal analysis image 203, yielding a vertical, that is column, spatial axis and a horizontal, that is row, time axis. Alternatively, the inclined linear arrays of pixels along the line of analysis 202c from the image frames are re-arranged to form horizontal lines on the spatio-temporal analysis image 203, thereby yielding a horizontal spatial axis and a vertical time axis.
Consider an example where the user selects a vertical line of analysis 202b, at a point on the x-axis, which varies over the entirety of the y-axis, that is, from 0 to M, as exemplarily illustrated in
A moving image capture device, for example, a closed-circuit television (CCTV) camera used for surveillance is mounted on rooftops or in tunneling toll booths and may be adapted for panning over a limited sweep angle or subjected to undue jitter. The linear array of pixels selected from each image frame based on the line of analysis 202 reproduces a specific section of the actual image captured at a single instance of time (t). In an embodiment, where the camera is panning over a limited sweep angle, the coordinates of the line of analysis 202 may be dynamically shifted to continue to focus the analysis at a specific part of the actual scene. The coordinates of the line of analysis 202 may be dynamically shifted by obtaining the pan direction and pan magnitude during every transition from one image frame to another. Also, the coordinates of the line of analysis 202 may be dynamically shifted by obtaining the average pan direction and pan magnitude, or the velocity over a single sweep of the camera. This ensures that the linear array of pixels extracted from each successive image frame represents identical section or part of the actual scene, when the moving image capture device is panning over a limited sweep angle. This however requires that the focused part(s) of the scene always remains within the field of view of the camera. Where the camera may be subjected to a momentary jitter, a wider line of analysis 202 is used to compensate for the jitter. Additionally or alternatively, the spatio-temporal analysis image 203 may be constructed by concatenating linear pixel arrays from every Nth successive image frame, for example, every 3rd successive frame in the series of image frames, drowning out momentary fluctuations in the constructed spatio-temporal analysis image 203.
The video content analysis of the spatio-temporal analysis image 203 using foreground segmentation may be performed using different techniques for background modeling and segmentation of the constructed spatio-temporal analysis image 203. The background modeling and segmentation according to the computer implemented method disclosed herein employs, for example, the moving average technique to determine the moving average of each vertical line of the spatio-temporal analysis image 203 of
I(y,α,t);
where I(y, α, t) is taken at x=α, where a is a value in the range of “x” between 0 to N, “y” varies from 0 to M and “t” varies from 0 to R.
The background of the spatio-temporal analysis image 203 is determined by applying a weighted moving average of the linear pixel arrays over time “t”. To begin with, the background of the image frame at “t0” corresponding to the linear pixel array in the spatio-temporal analysis image 203 at “t0” is modeled as follows:
B(y,α,t0)=I(y,α,t0);
Similarly, B(y, α, t1)=I(y, α, t1)*δ+B(y, α, t0)*(1−δ), where “δ” is a background adapting factor and is always 0≦δ≦1; and “t0” and “t1” are the either absolute or relative time indices, representing the image frames from which the linear arrays of pixels are taken. “(y, α, t1)” corresponds to the linear array taken at “α” column (LOA), for selected range of rows “y” on the image frame at time “t1”. B(y, α, t0) is the background modeled at image frame “t0” and can be a line array with all zeros. B(y, α, t1) is the background modeled at image frame “t1”. The background image, exemplarily illustrated in
The foreground is segmented by obtaining the absolute difference between the spatio-temporal analysis image 203 of
D(y,α,t1)=1 if “absolute(B(y,α,t1)−I(y,α,t1))≧threshold;
D(y,α,t1)=0 if “absolute(B(y,α,t1)−I(y,α,t1))<threshold;
The segmented image SI(y, t) is obtained by accumulating D(y, α, t0), D(y, α, t1), D(y, α, t2) . . . so on till D(y, α, tR), since “t” varies from 0 to R. The foreground segmented image SI(y, t) is illustrated in
The spatio-temporal or XYT analysis provides a background model using a single line of pixels as opposed to the conventional background models that process the entire image. Given a video graphics array (VGA) image of 640×480 pixels, the background modeling and segmentation according to the computer implemented method disclosed herein processes 480 pixels for analysis instead of the entire 307200 pixels. The line of analysis 202 is, for example, one pixel wide and hence precludes any duplication of objects, unless the objects cross the line of analysis 202 more than once. The spatio-temporal analysis for video content is, for example, used for people or vehicle count with better accuracy. The XYT analysis summarizes the video of a few gigabytes (GB) for monitoring a highway, a tollgate, an entrance to an office, or an industry or commercial premises, etc., in a few megabytes (MB) of memory. The spatio-temporal analysis image 203 is used for visualization, with each line along the time axis (t) of the spatio-temporal analysis image 203 specifying the unit of time that is used for time indexing for navigating through the video.
The spatio-temporal analysis image 203 constructed by the spatio-temporal analysis of the video exhibits a few notable unique characteristics. For example, if an object remains stationary on the line of analysis 202, the reproduction of the object in the spatio-temporal analysis image appears stretched or elongated along the temporal dimension (t). On the other hand, if an object traverses the line of analysis 202 at a high speed, the object appears compressed in the spatio-temporal analysis image. The spatio-temporal analysis according to the computer implemented method disclosed herein is a valuable technique to summarize an entire day's traffic at a toll gate and the shopper traffic at the entrance of a mall, to determine traffic line violations, the speed of vehicles, perimeter breach in a mining area, industrial or commercial campuses, etc.
The video content analyzer 501a provided on the computing device 501 assigns a temporal dimension across the image frames of the video stream. The temporal dimension represents the order of the image frames using predefined temporal coordinates. The video content analyzer 501a constructs a spatio-temporal analysis image 203 based on a user-defined line of analysis 202 on each of the image frames. The video content analyzer 501a constructs the spatio-temporal analysis image 203 by concatenating a series of temporally-successive linear pixel arrays along the temporal dimension. Each of the linear pixel arrays comprises an array of pixels underlying the line of analysis 202 defined on each of the image frames. The image segmentation module 501b on the computing device 501 segments the constructed spatio-temporal analysis image 203 for capturing the events of interest.
The user interface 501f, for example, a graphical user interface or a textual user interface on the computing device 501 renders one or more image frames of a stored video or a live video to the user. The user interface 501f enables the user to define the line of analysis 202, having any in-plane orientation, length and width, on the image frames. In an embodiment, the user interface 501f enables the user to define one or more lines of analysis 202 for parallely constructing multiple spatio-temporal analysis images. The constructed spatio-temporal analysis image 203 provides a summary of events occurring over the image frames at the line of analysis 202 for the duration of the video or the duration of the concatenation of the series of temporally-successive linear pixel arrays. The spatio-temporal analysis image 203 is represented by, for example, a combination of a fixed horizontal spatial coordinate, a range of the vertical spatial coordinates, and variable temporal coordinates, or a combination of a fixed vertical spatial coordinate, a range of the horizontal spatial coordinates, and variable temporal coordinates.
The image segmentation module 501b performs foreground segmentation of the constructed spatio-temporal analysis image 203 for detecting objects and events of interest. The image segmentation module 501b comprises a background modeler 501c, a background subtractor 501d, and a foreground segmentation module 501e. The background modeler 501c models the background of the constructed spatio-temporal analysis image 203, for example, by determining a moving average of the linear pixel arrays along the temporal dimension. The background subtractor 501d subtracts the modeled background from the constructed spatio-temporal analysis image 203 by obtaining the absolute difference between the constructed spatio-temporal analysis image 203 and the modeled background for obtaining foreground of the spatio-temporal analysis image 203. The foreground segmentation module 501e performs segmentation of the obtained foreground for detecting objects and events of interest, for example, presence of an object, traversal of the object, speed of traversal of the object, an object count based on the traversal of one or more objects, duration of the presence of an object, etc.
The processor 601 is an electronic circuit that can execute computer programs. The memory unit 602 is used for storing programs, applications, and data. For example, the video content analyzer 501a and the image segmentation module 501b are stored on the memory unit 602 of the computer system 600. The memory unit 602 is, for example, a random access memory (RAM) or another type of dynamic storage device that stores information and instructions for execution by processor 601. The memory unit 602 also stores temporary variables and other intermediate information used during execution of the instructions by the processor 601. The computer system 600 further comprises a read only memory (ROM) or another type of static storage device that stores static information and instructions for the processor 601. The data bus 605 permits communication between the modules, for example, 501a, 501b, 501c, 501d, 501e, and 501f of the computer implemented system 500 disclosed herein.
Computer applications and programs are used for operating the computer system 600. The programs are loaded onto the fixed media drive 608 and into the memory unit 602 of the computer system 600 via the removable media drive 609. In an embodiment, the computer applications and programs may be loaded directly through the network 503. Computer applications and programs are executed by double clicking a related icon displayed on the display unit 606 using one of the input devices 607. The user interacts with the computer system 600 using a user interface 501f of the display unit 606. The user selects the line of analysis 202 of any orientation, size, and thickness on one or more image frames on the user interface 501f of the display unit 606 using one of the input devices 607, for example, a computer mouse.
The computer system 600 employs an operating system for performing multiple tasks. The operating system manages execution of, for example, the video content analyzer 501a and the image segmentation module 501b provided on the computer system 600. The operating system further manages security of the computer system 600, peripheral devices connected to the computer system 600, and network connections. The operating system employed on the computer system 600 recognizes keyboard inputs of a user, output display, files and directories stored locally on the fixed media drive 608, for example, a hard drive. Different programs, for example, a web browser, an e-mail application, etc., initiated by the user are executed by the operating system with the help of the processor 601, for example, a central processing unit (CPU). The operating system monitors the use of the processor 601.
The video content analyzer 501a and the image segmentation module 501b are installed in the computer system 600 and the instructions are stored in the memory unit 602. The captured moving images are transferred from the moving image capture device 502 to the video content analyzer 501a installed in the computer system 600 of the computing device 501 via the interface 604 or a network 503. A user initiates the execution of the video content analyzer 501a by double clicking the icon for the video content analyzer 501a on the display unit 606 or the execution of the video content analyzer 501a is automatically initiated on installing the video content analyzer 501a on the computing device 501. Instructions for capturing events of interest by performing spatio-temporal analysis of the video are retrieved by the processor 601 from the program memory in the form of signals. The locations of the instructions from the modules, for example, 501a, 501b, 501c, 501d, and 501e, are determined by a program counter (PC). The program counter stores a number that identifies the current position in the program of the video content analyzer 501a and the image segmentation module 501b.
The instructions fetched by the processor 601 from the program memory after being processed are decoded. The instructions are placed in an instruction register (IR) in the processor 601. After processing and decoding, the processor 601 executes the instructions. For example, the video content analyzer 501a defines instructions for assigning a temporal dimension across the image frames of the video stream. The video content analyzer 501a further defines instructions for constructing a spatio-temporal analysis image 203 based on a user-defined line of analysis 202 on each of the image frames. The image segmentation module 501b defines instructions for segmenting the constructed spatio-temporal analysis image 203 for capturing the events of interest. The background modeler 501c defines instructions for modeling the background of the constructed spatio-temporal analysis image 203. The background subtractor 501d defines instructions for subtracting the modeled background from the constructed spatio-temporal analysis image 203 for obtaining the foreground of the constructed spatio-temporal analysis image 203. The foreground segmentation module 501e defines instructions for performing segmentation of the obtained foreground for detecting objects and events of interest. The instructions are stored in the program memory or received from a remote server.
The processor 601 retrieves the instructions defined by the video content analyzer 501a, the image segmentation module 501b, the background modeler 501c, the background subtractor 501d, and the foreground segmentation sub-module 501e, and executes the instructions.
At the time of execution, the instructions stored in the instruction register are examined to determine the operations to be performed. The specified operation is then performed by the processor 601. The operations include arithmetic and logic operations. The operating system performs multiple routines for performing a number of tasks required to assign input devices 607, output devices 610, and memory for execution of the video content analyzer 501a and the image segmentation module 501b. The tasks performed by the operating system comprise assigning memory to the video content analyzer 501a, the image segmentation module 501b and data, moving data between the memory 602 and disk units and handling input/output operations. The operating system performs the tasks on request by the operations and after performing the tasks, the operating system transfers the execution control back to the processor 601. The processor 601 continues the execution to obtain one or more outputs. The outputs of the execution of the video content analyzer 501a and the image segmentation module 501b are displayed to the user on the display unit 606.
The spatio-temporal analysis of a video also referred to as “video content analysis” according to the computer implemented method and system 500 disclosed herein for constructing and segmenting the spatio-temporal analysis image provides different video content analysis applications for monitoring and surveillance in different scenarios. A few applications in different scenarios are disclosed herein. The spatio-temporal analysis is used to determine a count of people over a predetermined duration. The people count is determined by the count of foreground objects in the spatio-temporal analysis image. The spatio-temporal analysis image is constructed by accumulating the linear pixels arrays from the image frames of the video, based on the line of analysis 202, over the predetermined duration. In this case, the line of analysis 202 is defined across a doorway, a passageway, or a corridor dominated by incoming and outgoing people.
In another scenario, where the line of analysis 202 is selected to depict a perimeter line, the spatio-temporal analysis at the line of analysis 202 is used to detect a perimeter breach or simulate a tripwire. When an object crosses the line of analysis 202, the object is reproduced in the spatio-temporal analysis image 203, as illustrated in
In another scenario, the speed of traversal of the object, for example, a moving vehicle is determined using one or more spatio-temporal analysis images.
Other applications based on similar principles include, for example, determining the winner in a track race. In this case, the line of analysis 202 is defined along the finish line of a track. A spatio-temporal analysis image is constructed based on the line of analysis 202 to determine the race winner based on the time of appearance of the participants or athletes on the spatio-temporal analysis image. The spatio-temporal analysis disclosed herein also enables automated parking lot management. The length of the stretch or elongation of an object, for example, a stationary vehicle on the spatio-temporal analysis image provides the parking duration of the stationary vehicle. The line of analysis 202 is defined within the parking slot with a suitable orientation and length. In the same way, the vacant slots in a parking lot are determined by considering the binary difference image of the spatio-temporal (XYT) image of the parking lot. For example, white patches on the binary difference image correspond to occupied slots, while black patches correspond to the unoccupied slots. In road traffic management, where video monitoring is installed in the area around traffic lights and intersections, the line of analysis 202 is suitably defined along the road markings to enable video content analysis. At an unmanned railway crossing installed with video surveillance, the line of analysis 202 is defined along the rail track for video content analysis. In a museum with video surveillance for monitoring exhibits such as paintings or statues, video content analysis is performed at specific parts of the scene by defining the line of analysis 202 at those parts.
The events of interest in a scene may be specific to the application scenario and requirement, and the general arrangement or landscape of the captured scene. A person of ordinary skill in the art or any user in the monitoring and surveillance domain can easily recognize parts of a scene that require content analysis, and accordingly define one or more lines of analysis 202 on one or more image frames, including the orientation, length and width of the lines of analysis 202, for performing spatio-temporal analysis.
It will be readily apparent that the various methods and algorithms described herein may be implemented in a computer readable medium appropriately programmed for general purpose computers and computing devices. Typically a processor, for example, one or more microprocessors will receive instructions from a memory or like device, and execute those instructions, thereby performing one or more processes defined by those instructions. Further, programs that implement such methods and algorithms may be stored and transmitted using a variety of media, for example, computer readable media in a number of manners. In one embodiment, hard-wired circuitry or custom hardware may be used in place of, or in combination with, software instructions for implementation of the processes of various embodiments. Thus, embodiments are not limited to any specific combination of hardware and software. A “processor” means any one or more microprocessors, central processing unit (CPU) devices, computing devices, microcontrollers, digital signal processors or like devices. The term “computer readable medium” refers to any medium that participates in providing data, for example instructions that may be read by a computer, a processor or a like device. Such a medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media include, for example, optical or magnetic disks and other persistent memory volatile media include dynamic random access memory (DRAM), which typically constitutes the main memory. Transmission media include coaxial cables, copper wire and fiber optics, including the wires that comprise a system bus coupled to the processor. Common forms of computer readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a compact disc-read only memory (CD-ROM), digital versatile disc (DVD), any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a random access memory (RAM), a programmable read only memory (PROM), an erasable programmable read only memory (EPROM), an electrically erasable programmable read only memory (EEPROM), a flash memory, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read. In general, the computer readable programs may be implemented in any programming language. Some examples of languages that can be used include C, C++, C#, Perl, Python, or JAVA. The software programs may be stored on or in one or more mediums as an object code. A computer program product comprising computer executable instructions embodied in a computer readable medium comprises computer parsable codes for the implementation of the processes of various embodiments.
The present invention can be configured to work in a network environment including a computer that is in communication, via a communications network, with one or more devices. The computer may communicate with the devices directly or indirectly, via a wired or wireless medium such as the Internet, a local area network (LAN), a wide area network (WAN) or the Ethernet, token ring, or via any appropriate communications means or combination of communications means. Each of the devices may comprise computers, such as those based on the Intel® processors, AMD® processors, UltraSPARC® processors, Sun® processors, IBM° processors, etc. that are adapted to communicate with the computer. Any number and type of machines may be in communication with the computer.
The foregoing examples have been provided merely for the purpose of explanation and are in no way to be construed as limiting of the present invention disclosed herein. While the invention has been described with reference to various embodiments, it is understood that the words, which have been used herein, are words of description and illustration, rather than words of limitation. Further, although the invention has been described herein with reference to particular means, materials and embodiments, the invention is not intended to be limited to the particulars disclosed herein; rather, the invention extends to all functionally equivalent structures, methods and uses, such as are within the scope of the appended claims. Those skilled in the art, having the benefit of the teachings of this specification, may effect numerous modifications thereto and changes may be made without departing from the scope and spirit of the invention in its aspects.
Number | Date | Country | Kind |
---|---|---|---|
1753/CHE/2010 | Jun 2010 | IN | national |
Number | Name | Date | Kind |
---|---|---|---|
5784023 | Bluege | Jul 1998 | A |
6335985 | Sambonsugi et al. | Jan 2002 | B1 |
6529132 | Boucourt | Mar 2003 | B2 |
6856696 | Ajioka | Feb 2005 | B1 |
7180529 | Covannon et al. | Feb 2007 | B2 |
7203693 | Carlbom et al. | Apr 2007 | B2 |
7304649 | Hori et al. | Dec 2007 | B2 |
7347555 | Grover | Mar 2008 | B2 |
7356164 | Aliaga et al. | Apr 2008 | B2 |
7429997 | Givon | Sep 2008 | B2 |
7450165 | Ahiska | Nov 2008 | B2 |
7519907 | Cohen et al. | Apr 2009 | B2 |
7528881 | Ahiska | May 2009 | B2 |
7545516 | Jia et al. | Jun 2009 | B2 |
7856137 | Yonezawa et al. | Dec 2010 | B2 |
7872665 | Grover | Jan 2011 | B2 |
7876325 | Sharma | Jan 2011 | B1 |
7949295 | Kumar et al. | May 2011 | B2 |
7949474 | Callahan et al. | May 2011 | B2 |
7957565 | Sharma et al. | Jun 2011 | B1 |
7962283 | Zhang et al. | Jun 2011 | B2 |
7991837 | Tahan | Aug 2011 | B1 |
7995078 | Baar | Aug 2011 | B2 |
8009863 | Sharma et al. | Aug 2011 | B1 |
8107699 | Hong et al. | Jan 2012 | B2 |
8145007 | Davey et al. | Mar 2012 | B2 |
8254679 | Marchesotti | Aug 2012 | B2 |
8274552 | Dahi et al. | Sep 2012 | B2 |
8275175 | Baltatu et al. | Sep 2012 | B2 |
8280136 | Gotardo et al. | Oct 2012 | B2 |
8306265 | Fry et al. | Nov 2012 | B2 |
8311915 | Baar et al. | Nov 2012 | B2 |
20010012379 | Amemiya et al. | Aug 2001 | A1 |
20020008785 | Yamaguchi et al. | Jan 2002 | A1 |
20020009719 | Walt et al. | Jan 2002 | A1 |
20020027616 | Jun et al. | Mar 2002 | A1 |
20020028003 | Krebs et al. | Mar 2002 | A1 |
20020037103 | Hong et al. | Mar 2002 | A1 |
20020039387 | Auvray et al. | Apr 2002 | A1 |
20020094135 | Caspi et al. | Jul 2002 | A1 |
20020110195 | Auvray et al. | Aug 2002 | A1 |
20020142477 | Lewis et al. | Oct 2002 | A1 |
20020159536 | Gelblum et al. | Oct 2002 | A1 |
20020164084 | Baggs | Nov 2002 | A1 |
20030028005 | Bazan | Feb 2003 | A1 |
20030088532 | Hampshire, II | May 2003 | A1 |
20030108873 | Dahlberg et al. | Jun 2003 | A1 |
20030185420 | Sefcik et al. | Oct 2003 | A1 |
20040014143 | Haskins et al. | Jan 2004 | A1 |
20040093166 | Kil | May 2004 | A1 |
20040115683 | Medford et al. | Jun 2004 | A1 |
20040190092 | Silverbrook et al. | Sep 2004 | A1 |
20040194129 | Carlbom et al. | Sep 2004 | A1 |
20040233987 | Porikli et al. | Nov 2004 | A1 |
20040249848 | Carlbom et al. | Dec 2004 | A1 |
20050002572 | Saptharishi et al. | Jan 2005 | A1 |
20050042230 | Anderson et al. | Feb 2005 | A1 |
20050048527 | Allawi et al. | Mar 2005 | A1 |
20050104958 | Egnal et al. | May 2005 | A1 |
20050134685 | Egnal et al. | Jun 2005 | A1 |
20050165789 | Minton et al. | Jul 2005 | A1 |
20050166163 | Chang et al. | Jul 2005 | A1 |
20050203927 | Sull et al. | Sep 2005 | A1 |
20050226502 | Cohen et al. | Oct 2005 | A1 |
20060045346 | Zhou | Mar 2006 | A1 |
20060050958 | Okada et al. | Mar 2006 | A1 |
20060056056 | Ahiska et al. | Mar 2006 | A1 |
20060101060 | Li et al. | May 2006 | A1 |
20060176209 | Shu et al. | Aug 2006 | A1 |
20060221181 | Garoutte | Oct 2006 | A1 |
20060222206 | Garoutte | Oct 2006 | A1 |
20060262184 | Peleg et al. | Nov 2006 | A1 |
20060282236 | Wistmuller | Dec 2006 | A1 |
20060291695 | Lipton et al. | Dec 2006 | A1 |
20070052803 | Chosak et al. | Mar 2007 | A1 |
20070058717 | Chosak et al. | Mar 2007 | A1 |
20070116347 | Hong | May 2007 | A1 |
20070160274 | Mashiach | Jul 2007 | A1 |
20070165936 | Yonezawa et al. | Jul 2007 | A1 |
20070230781 | Yamamoto | Oct 2007 | A1 |
20070231791 | Olsen et al. | Oct 2007 | A1 |
20070263915 | Mashiach | Nov 2007 | A1 |
20070269078 | Lee | Nov 2007 | A1 |
20080031521 | Can et al. | Feb 2008 | A1 |
20080131920 | Levin | Jun 2008 | A1 |
20080152192 | Zhu et al. | Jun 2008 | A1 |
20080195284 | Hammadou | Aug 2008 | A1 |
20080201101 | Hebert et al. | Aug 2008 | A1 |
20080233595 | Nie et al. | Sep 2008 | A1 |
20080298674 | Baker et al. | Dec 2008 | A1 |
20080320403 | Glaberson | Dec 2008 | A1 |
20090024547 | Lu et al. | Jan 2009 | A1 |
20090040301 | Sandler et al. | Feb 2009 | A1 |
20090041328 | Hong et al. | Feb 2009 | A1 |
20090073265 | Greenhill et al. | Mar 2009 | A1 |
20090185078 | Van Beek et al. | Jul 2009 | A1 |
20090209858 | Oelze | Aug 2009 | A1 |
20090232353 | Sundaresan et al. | Sep 2009 | A1 |
20090285469 | Callahan et al. | Nov 2009 | A1 |
20090310444 | Hiroe | Dec 2009 | A1 |
20090310883 | Moriya | Dec 2009 | A1 |
20090316990 | Nakamura et al. | Dec 2009 | A1 |
20100021009 | Yao | Jan 2010 | A1 |
20100021378 | Rousso et al. | Jan 2010 | A1 |
20100033574 | Ran et al. | Feb 2010 | A1 |
20100046799 | Saptharishi et al. | Feb 2010 | A1 |
20100070483 | Delgo et al. | Mar 2010 | A1 |
20100070523 | Delgo et al. | Mar 2010 | A1 |
20100092085 | Marchesotti | Apr 2010 | A1 |
20100097526 | Jacob | Apr 2010 | A1 |
20100098289 | Tognoli et al. | Apr 2010 | A1 |
20100100001 | Aguilar et al. | Apr 2010 | A1 |
20100104135 | Nakajima | Apr 2010 | A1 |
20100111374 | Stoica | May 2010 | A1 |
20100202662 | Chung | Aug 2010 | A1 |
20100245157 | Wicks et al. | Sep 2010 | A1 |
20100245582 | Harel | Sep 2010 | A1 |
20100245583 | Harel | Sep 2010 | A1 |
20100245670 | Takeda et al. | Sep 2010 | A1 |
20100262374 | Hwang et al. | Oct 2010 | A1 |
20110002517 | Mollus et al. | Jan 2011 | A1 |
20110065072 | Duffy | Mar 2011 | A1 |
20110142335 | Ghanem et al. | Jun 2011 | A1 |
20110157355 | Ivanov et al. | Jun 2011 | A1 |
20110317009 | Kumaraswamy et al. | Dec 2011 | A1 |
20120002047 | An et al. | Jan 2012 | A1 |
20120038776 | Ahiska et al. | Feb 2012 | A1 |
20120092494 | Garoutte et al. | Apr 2012 | A1 |
20120114245 | Lakshmanan et al. | May 2012 | A1 |
20120128254 | Nakajima | May 2012 | A1 |
20120206606 | Marchese | Aug 2012 | A1 |
Number | Date | Country |
---|---|---|
1845477 | Oct 2007 | EP |
Entry |
---|
K. Toyama, J. Krumm, B. Brumitt, and B. Meyers, “Wallflower: Principles and Practice of Background Maintenance,” in International Conference on Computer Vision, Sep. 1999. |
C. Wren, A. Azarbayejani, T. Darrell, and A. Pentland, “Pfinder: Real-Time Tracking of the Human Body,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, No. 7, pp. 780-785, Jul. 1997. |
Niyogi, S.A., Adelson, E.H., “Analyzing and Recognizing Walking Figures in XYT”, CVPR94(469-474), CVPR 1994 IEEE Computer Society Conference Date: Jun. 21-23, 1994. |
Aissa Saoudi, Hassane Essafi ,“Spatio-Temporal Video Slice Edges Analysis for Shot Transition Detection and Classification”, Waset, vol. 4 No. 3 Summer 2008. |
Number | Date | Country | |
---|---|---|---|
20110317009 A1 | Dec 2011 | US |