Generating and reviewing motion metadata

Information

  • Patent Grant
  • 10515117
  • Patent Number
    10,515,117
  • Date Filed
    Tuesday, February 14, 2017
    7 years ago
  • Date Issued
    Tuesday, December 24, 2019
    5 years ago
Abstract
Aspects of the instant disclosure relate to methods for generating motion metadata for a newly captured video feed. In some aspects, methods of the subject technology can include steps for recording a video feed using the video capture system, partitioning the image frames into a plurality of pixel blocks, and processing the image frames to detect one or more motion events. In some aspects, the method may further include steps for generating motion metadata describing each of the one or more motion events. Systems and computer-readable media are also provided.
Description
TECHNICAL FIELD

The disclosed technology relates to systems and methods for facilitating the retroactive review of video data and in particular, for performing real-time processing on a recorded video feed and storing motion metadata for the subsequent search of motion events.


BACKGROUND

The proliferation of video cameras, such as those deployed in security surveillance setups, has drastically increased the amount of collected video data. Because human review of camera footage is tiresome and expensive, some processing techniques have been developed that enable the pre-selection of areas-of-interest within the feed for which processing is to be performed. However, such processing methods require that the user identify regions of interest before recording, a constraint that is often impractical for security deployments in which a location of events-of-interest within the frame is often hard to predict.





BRIEF DESCRIPTION OF THE DRAWINGS

The embodiments herein may be better understood by referring to the following description in conjunction with the accompanying drawings in which like reference numerals indicate similar elements. It is understood that these drawings depict only example embodiments of the disclosure and are not to be considered limiting in scope. The principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings in which:



FIG. 1 illustrates an example environment in which some aspects of the technology can be implemented.



FIG. 2A illustrates an example of an image frame that is parsed into coarse-cells for facilitating motion detection, according to some aspects of the technology.



FIG. 2B illustrates an example image frame depicted in FIG. 2A, that is parsed into macro-blocks for facilitating motion detection, according to some aspects of the technology.



FIG. 2C illustrates an example use of macro-blocks to indicate areas of motion activity, as well as a corresponding motion vector, according to some aspects of the technology.



FIG. 3A illustrates steps of an example process for associating motion metadata with recorded video information.



FIG. 3B illustrates of an example process for receiving a motion event search query, and performing a search through motion metadata to identify one or more motion events.



FIG. 4 illustrates an example processor-based device on which aspects of the technology can be implemented.





OVERVIEW

Aspects of the instant disclosure relate to methods for generating motion metadata for a newly captured video feed. In some aspects, a process of the technology can include steps for recording a video feed (e.g., storing video data) using a video capture system, partitioning image frames of the video feed into a multiple pixel blocks, and processing the image frames to detect one or more motion events. In some aspects, the process can further include steps for generating motion metadata describing each of the motion events.


DESCRIPTION

Aspects of the disclosed technology address some of the limitations of searching for motion events in legacy video content by providing systems/methods to facilitate the real-time processing of motion events within a recorded video feed. In some aspects, motion events are associated with frame and/or frame-location before storage is completed. As such, the disclosed technology facilitates the quick and easy searching of motion events in stored video data, including the selection and searching of motion events within a frame-subset that defines and area within the video feed that can be selected for the search of motion events.



FIG. 1 illustrates an example environment 100 in which some aspects of the technology can be implemented. Environment 100 includes an image capture device 102 configured for recording video information (e.g., a video feed) as a series of image frames 104. Environment 100 further includes an image processing module 106, metadata/video database 108, and a video server 110. Additionally, environment 100 includes a user 114A, for example, that is associated with user device 114B, and communicatively coupled to video server 110, via network 105.


Environment 100 conceptually illustrates various hardware and/or software modules that can be used to implement aspects of the technology; however, it is understood that additional hardware and/or software modules can be used without departing from the scope of the technology. For example, network 105 can represent a private network such as a local area network (LAN), a wide area network (WAN), or a network of public and/or private networks, such as the Internet. Additionally, it is understood that the various software modules (e.g., image processing module 106, metadata/video database 108, and/or metadata search module 112) can be implemented on one or more hardware devices, such as an image capture device 102, or one or more computers/servers, such as server 110.


In operation, image capture device 102 is configured to record and store video data in a series of image (or video) frames, e.g. frames 104. Although image capture device 102 can include any video recording device capable of capturing or recording and storing digital video information, by way of example, image capture device 102 can be a video surveillance camera, or the like. Additionally, storage frames 104 can be performed locally (e.g., on a memory device visually coupled with image capture device 102), or made from remotely, such as via storage on one or more remote servers communicatively coupled with image capture device 102 via network 105.


Image frames 104 can be processed and stored as video capture is performed. For example, image capture device 102 can be configured to provide a real-time feed of captured image frames 104 to an image processing module configured to detect motion events. As discussed in further detail below, motion events can be detected by calculating a sum of absolute differences between frames on a frame-by-frame basis. Detected motion events can then be indexed based on various indicators, such as, a time stamp, duration, and/or an intensity associated with the frame. Subsequently, data describing the detected motion events (i.e. “motion metadata”) is stored in a database, such as metadata/video database 108.


It is understood that image processing module 106 and/or metadata/video database 108 can be implemented on a common hardware platform, such image capture device 102, and/or a computer/server such as video server 110. Similarly, any of the described software modules environment 100 may be operated using one or more servers that are communicatively coupled via network 105, for example, as various software processes instantiated in a computing cluster (not illustrated).


In practice, metadata/video database 108 is accessible by a metadata search module 112 that can be used to facilitate queries relating to motion metadata for a particular set of image frames 104.


Once motion metadata has been stored, a user (e.g., user 114A) can provide a motion search query (e.g., via user device 114B) to video server 110 over a computer network, such as network 105. The search query can then be received by metadata search module 112 and used to search metadata/video database 108. As discussed in further detail below with respect to FIG. 3B, the motion search query can include boundary conditions indicating a particular area or region within the video frames in which motion events are to be searched. By permitting users to specify a particular “region of interest” for their motion search query, legacy video data can be efficiently searched only for events occurring at indicated locations of interest, increasing the speed and efficiency of the search, as well as reducing processing overhead.


By way of example, a security surveillance system can capture video footage that includes the theft of an item of interest located within a particular region within the image frame. Thus, a user may wish to only search for motion events that occur within the particular region of interest (e.g., at a location where the item was located). In such instances, the user may provide a search query in the form of a selection of the image frame area for which motion events are to be searched, for example, by indicating the area with a cursor or other input device. The search query can then be used to search motion metadata stored in a time-series database, such as that hosted by video server 110.


As discussed in further detail below, the time-series database of motion metadata can be comprised of a series of motion vectors that described various attributes of frames within the video feed. Although the motion vectors can describe essentially any attribute of an associated image frame, in some approaches, motion vectors contain information regarding the location of motion events as placed in a coarse-cell and/or macro-block (e.g., “fine cell”) grid. As discussed in further detail with respect to FIG. 2A, coarse/fine-cell grids can be superimposed over an image area, effectively providing a bit mask that permits irrelevant information to be efficiently disregarded.


Specifically, FIG. 2A illustrates an example of an image frame 201 that is parsed into coarse-cells 203. In the example of FIG. 2A, coarse-cells 203 are arranged in a 9×10 grid superimposed over image frame 201. However, it is understood that other coarse-cell overlay dimensions can be used, without departing from the scope of the technology. As discussed in further detail below, motion detection can be performed using subdivided units of coarse-cells 203 (i.e., macro-blocks), which represent the smallest collection of adjacent pixels for which motion detection events are registered. As discussed in greater detail below, motion events associated with a particular macro-block can be identified by calculating absolute differences (e.g., in pixel color or intensity) as between common macro-blocks in adjacent frames, i.e., on a frame-to-frame basis.



FIG. 2B illustrates an example of image frame 201 depicted in FIG. 2A, wherein each of coarse-cells 203 are further parsed into macro-blocks 205. Macro-blocks 205 can be used to distinguish regions within a particular coarse-cell in which motion is detected. In the example of FIG. 2B, each coarse-cell 203 contains twenty-four total macro-blocks 205; however, it is understood that any other number of macro-blocks may be used to subdivide a coarse-cell, without departing from the technology.


Calculations of motion events can be determined based on changes in pixel values for a particular macro-block across multiple image frames. In the example of FIG. 2B, macro-blocks 205 cover an area of image frame 201 (e.g., coarse-cell A1) that does not contain any motion events. In contrast, FIG. 2C illustrates an example of a coarse-cell (D7) in which a motion event is detected.


Specifically, FIG. 2C illustrates an example of macro-blocks 205 used to indicate locations of motion activity, for example, in coarse-cell D7, as well as a corresponding motion vector 207. In some aspects, motion vectors, such as motion vector 207, can be used to provide a compressed data format for storing motion event data.


As indicated in the example of FIG. 2C, motion events within coarse-cell D7 are identified at macro-blocks 13, 10, 14, 18, 7, 11, 15, and 19. Motion events for each macro-block are described by motion vector 207 that provides an indication of the row/column of the associated coarse-cell, as well as the associated macro-blocks in which motion has been detected. Motion vectors, such as motion vector 207, can be associated with a respective image frame that is sorted as a times-series, for example, in a database such as metadata/video database 108, discussed above.


In some aspects, the motion vector depicted by motion vector 207 is advantageous in that each macro-block can be represented as a single 32-bit integer, wherein the first four bits indicate the macro-block row, and the next four bits indicate the macro-block column. In this example, the remaining 24 bits act as a bit-mask used indicate the presence of motion within each of the macro-block cells. Therefore, in the example of FIG. 2C, the entire image can be stored as up to 90 of these 32-bit integers (e.g., 360 bytes).


In some aspects, the motion vector format also permits sparse data to be vastly compressed. For example, macro-blocks that contain no motion events are not saved, and a motion event that activates only a single macro-block cell consumes only 4 bytes of data.



FIG. 3A illustrates steps of an example process 300 for associating motion metadata with recorded video information. Process 300 begins with step 302, in which a video feed is captured using a video capture system, such as image capture device 102 discussed above with respect to FIG. 1. The captured video feed includes multiple image frames each of a predefined area (e.g., a predefined dimension).


In step 304, the video feed is partitioned into a plurality of pixel blocks, such as coarse-cells, and macro-blocks (fine cells), as illustrated above with respect to FIG. 2A-C. It is understood that video feed partitioning can be performed at the image capture device (e.g. a security camera), and/or at one or more computing devices, such as one or more computers/servers and/or a computing cluster.


At step 306, the image frames are processed to detect one or more motion events. Each detected motion event is associated with at least one pixel block (e.g., a macro-block). As discussed above, determinations of motion events can be based on a calculation of a sum of absolute differences (SAD) between individual macro-blocks (also “frame subtraction”). In other approaches, different types of background subtraction calculations can be implemented; however, it is understood that other motion detection methods may be used, without departing from the scope of the technology.


In step 308, motion metadata is generated based on the one or more motion events detected at step 306. As discussed above, motion metadata may be generated as a motion vector, (e.g., motion vector 207), that is used to describe various attributes associated with the motion event. Such attributes can include indications of time, coarse-block location, macro-block location, and/or intensity, etc.


Next, in step 310, the motion metadata is associated with one or more image frames and/or one or more specific pixel blocks within a respectively associated frame. As discussed in further detail below, the association of motion metadata with image frames can facilitate the convenient searching of metadata events in legacy content, for example, within a region of the image frame specified after the video feed has already been recorded.



FIG. 3B illustrates of an example process 301 for performing a search through motion metadata to identify one or more motion events. Further to the example of step 310, with respect to process 300, the searching of motion metadata in process 301 begins with step 312 in which a motion event search query is received, for example, from a user or system administrator. The motion event search query can be received by a metadata search module (e.g., metadata search module 112) configured to access the metadata/video database where motion metadata is stored, such as metadata/video database 108.


Although the format of the motion event search query can depend on the desired implementation, in some aspects the motion event search query can contain information that defines at least a portion of the predefined pixel area for which the desired search is to be performed. For example, the motion event search query can take the form of a bounded area defining at least a portion of the image frame across which all of the video content is to be searched.


In some aspects, the motion event search query can contain indications of boundaries within the video feed. For example, the motion event search query may specify an area of the image frame, in conjunction with start and stop times (corresponding with different image frames or frame numbers in the feed), for which the search is to be performed. Other search parameters can include, but are not limited to, searches based on event length and/or intensity (e.g., to filter out sudden increases/decreases in ambient lighting). It is understood that other user-specified criteria may be included in the motion event search query, without departing from the scope of the technology.


Subsequently, at step 314, the motion metadata is searched to identify one or more image frames that include motion events of interest. In some aspects, the search query may return more specific information, such as, identifications of a specific pixel block (e.g., macro-block and/or micro-block), that include motion events of interest.


Further to the example provided above with respect to FIGS. 2A-2C, if a user specifies an area of the image frame that includes macro-block D7, the search could be performed across all macro-blocks corresponding with the user-specified area, across all image frames. In this example, because motion events have been identified to occur in those blocks, for example, at least in block D7 (as specified by motion vector 207), the associated frames and/or blocks would be returned to the user for further review.


EXAMPLE DEVICE


FIG. 4 illustrates an example processor-based device 400 that can be used to implement various aspects of the technology. For example, processor-based device 400 may be used to implement an image capture device (e.g., image capture device 102), or a video server, such as video server 110 discussed above with respect to FIG. 1. It is further understood that processor-based device 400 may be used in conjunction with one or more other processor-based devices, for example, as part of a computer network or computing cluster.


Processor-based device 400 includes a master central processing unit (CPU) 462, interfaces 468, and a bus 415 (e.g., a PCI bus). When acting under the control of appropriate software and/or firmware, the CPU 462 can be configured for executing operations necessary to process video content to determine motion events within one or more image frames, as described above.


CPU 462 preferably accomplishes all these functions under the control of software including an operating system and any appropriate applications software. CPU 462 can include one or more processors 463 such as a processor from the Motorola family of microprocessors or the MIPS family of microprocessors. In an alternative embodiment, processor 463 is specially designed hardware for controlling the operations of processor-based device 400. In a specific embodiment, a memory 461 (such as non-volatile RAM and/or ROM) also forms part of CPU 462. However, there are many different ways in which memory could be coupled to the system.


Interfaces 468 can be provided as interface cards (sometimes referred to as “line cards”). Generally, they control the sending and receiving of data packets over the network and sometimes support other peripherals used with the router 410. Among the interfaces that may be provided are Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like. In addition, various very high-speed interfaces may be provided such as fast token ring interfaces, wireless interfaces, Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces and the like. Generally, these interfaces may include ports appropriate for communication with the appropriate media. In some cases, they may also include an independent processor and, in some instances, volatile RAM. The independent processors may control such communications intensive tasks as packet switching, media control and management. By providing separate processors for the communications intensive tasks, these interfaces allow the master microprocessor 462 to efficiently perform routing computations, network diagnostics, security functions, etc.


Although the system shown in FIG. 4 is one specific network device of the present invention, it is by no means the only device architecture on which the present invention can be implemented. For example, an architecture having a single processor that handles communications as well as routing computations, etc. is often used. Further, other types of interfaces and media could also be used with the router.


Regardless of the network device's configuration, it may employ one or more memories or memory modules (including memory 461) configured to store program instructions for the general-purpose network operations and mechanisms for roaming, route optimization and routing functions described herein. The program instructions may control the operation of an operating system and/or one or more applications, for example. The memory or memories may also be configured to store tables such as mobility binding, registration, and association tables, etc.


For clarity of explanation, in some instances the present technology may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.


In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.


Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.


Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rackmount devices, standalone devices, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.


The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.


Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims. Moreover, claim language reciting “at least one of” a set indicates that one member of the set or multiple members of the set satisfy the claim.

Claims
  • 1. A video capture system, comprising: one or more processors; anda non-transitory memory coupled to the processors, the memory storing instructions configured to cause the processors to perform operations comprising: recording a video feed using the video capture system, the video feed comprising a plurality of image frames of a predefined pixel area;partitioning the image frames into a plurality of pixel blocks, wherein each of the pixel blocks comprises a subset of the predefined pixel area of the corresponding image frame;processing each of the plurality blocks of the image frames to detect one or more motion events that occurred across one or more image fames in the corresponding pixel block; andgenerating motion metadata describing each of the one or more motion events;wherein the processing detects motion events within the image frames without being previously directed to specific areas of the image frame.
  • 2. The video capture system of claim 1, further comprising: associating the motion metadata with one or more pixel blocks, wherein the motion metadata associated with each of the pixel blocks corresponds with motion events that occurred across one or more image fames in the corresponding pixel block.
  • 3. The video capture system of claim 1, further comprising: receiving a motion event search query, wherein the search query specifies at least a portion of the predefined pixel area for which a search is to be performed for motion events in the recorded video feed; andsearching the motion metadata to identify one or more image frame and pixel block pairs that include motion events corresponding with the search query.
  • 4. The video capture system of claim 3, wherein the search query specifies a start time and an end time for which motion metadata of the recorded video feed is to be searched.
  • 5. The video capture system of claim 1, wherein processing the image frames to detect one or more motion events further comprises: quantifying motion events for one or more of the pixel blocks; anddetermining if the quantity of motion in each of the one or more pixel blocks exceeds a predetermined threshold.
  • 6. The video capture system of claim 5, wherein the motion events are quantified based on an absolute difference of pixels between two or more image frames of the video feed.
  • 7. The video capture system of claim 1, wherein the motion metadata comprises one or more of: duration information, time stamp information, or spatial data.
  • 8. A method of generating motion metadata, the method comprising: recording a video feed using the video capture system, the video feed comprising a plurality of image frames of a predefined pixel area;partitioning the image frames into a plurality of pixel blocks, wherein each of the pixel blocks comprises a subset of the predefined pixel area of the corresponding image frame;processing each of the plurality blocks of the image frames to detect one or more motion events that occurred across one or more image fames in the corresponding pixel block; andgenerating motion metadata describing each of the one or more motion events;wherein the processing detects motion events within the image frames without being previously directed to specific areas of the image frame.
  • 9. The method of claim 8, further comprising: associating the motion metadata with one or more pixel blocks, wherein the motion metadata associated with each of the pixel blocks corresponds with motion events that occurred across one or more image fames in the corresponding pixel block.
  • 10. The method of claim 8, further comprising: receiving a motion event search query, wherein the search query specifies at least a portion of the predefined pixel area for which a search is to be performed for motion events in the recorded video feed; andsearching the motion metadata to identify one or more image frame and pixel block pairs that include motion events corresponding with the search query.
  • 11. The method of claim 10, wherein the search query specifies a start time and an end time for which motion metadata of the recorded video feed is to be searched.
  • 12. The method of claim 8, wherein processing the image frames to detect one or more motion events further comprises: quantifying motion events for one or more of the pixel blocks; anddetermining if the quantity of motion in each of the one or more pixel blocks exceeds a predetermined threshold.
  • 13. The method of claim 12, wherein the motion events are quantified based on an absolute difference of pixels between two or more image frames of the video feed.
  • 14. The method of claim 8, wherein the motion metadata comprises one or more of: duration information, time stamp information, or spatial data.
  • 15. A non-transitory computer-readable storage medium comprising instructions stored therein, which when executed by one or more processors, cause the processors to perform operations comprising: recording a video feed using the video capture system, the video feed comprising a plurality of image frames of a predefined pixel area;partitioning the image frames into a plurality of pixel blocks, wherein each of the pixel blocks comprises a subset of the predefined pixel area of the corresponding image frame;processing each of the plurality blocks of the image frames to detect one or more motion events that occurred across one or more image fames in the corresponding pixel block; andgenerating motion metadata describing each of the one or more motion events;wherein the processing detects motion events within the image frames without being previously directed to specific areas of the image frame.
  • 16. The non-transitory computer-readable storage medium of claim 15, wherein the instructions are further configured to cause the processors to perform operations comprising: associating the motion metadata with one or more pixel blocks, wherein the motion metadata associated with each of the pixel blocks corresponds with motion events that occurred across one or more image fames in the corresponding pixel block.
  • 17. The non-transitory computer-readable storage medium of claim 15, wherein the instructions are further configured to cause the processors to perform operations comprising: receiving a motion event search query, wherein the search query specifies at least a portion of the predefined pixel area for which a search is to be performed for motion events in the recorded video feed; andsearching the motion metadata to identify one or more image frame and pixel block pairs that include motion events corresponding with the search query.
  • 18. The non-transitory computer-readable storage medium of claim 17, wherein the search query specifies a start time and an end time for which motion metadata of the recorded video feed is to be searched.
  • 19. The non-transitory computer-readable storage medium of claim 15, wherein the instructions are further configured to cause the processors to perform operations comprising: quantifying motion events for one or more of the pixel blocks; anddetermining if the quantity of motion in each of the one or more pixel blocks exceeds a predetermined threshold.
  • 20. The non-transitory computer-readable storage medium of claim 15, wherein the motion events are quantified based on an absolute difference of pixels between two or more image frames of the video feed.
US Referenced Citations (248)
Number Name Date Kind
4460807 Kerr et al. Jul 1984 A
4890257 Anthias et al. Dec 1989 A
5694563 Belfiore et al. Dec 1997 A
5699082 Marks et al. Dec 1997 A
5745711 Kitahara et al. Apr 1998 A
5767897 Howell Jun 1998 A
5825858 Shaffer et al. Oct 1998 A
5874962 de Judicibus et al. Feb 1999 A
5995096 Kitahara et al. Nov 1999 A
6040817 Sumikawa Mar 2000 A
6075531 DeStefano Jun 2000 A
6085166 Beckhardt et al. Jul 2000 A
6191807 Hamada et al. Feb 2001 B1
6300951 Filetto et al. Oct 2001 B1
6392674 Hiraki et al. May 2002 B1
6424370 Courtney Jul 2002 B1
6463473 Gubbi Oct 2002 B1
6554433 Holler Apr 2003 B1
6573913 Butler et al. Jun 2003 B1
6646997 Baxley et al. Nov 2003 B1
6665396 Khouri et al. Dec 2003 B1
6700934 Lin Mar 2004 B2
6711419 Mori Mar 2004 B1
6754321 Innes et al. Jun 2004 B1
6754335 Shaffer et al. Jun 2004 B1
RE38609 Chen et al. Oct 2004 E
7058164 Chan et al. Jun 2006 B1
7058710 McCall et al. Jun 2006 B2
7062532 Sweat et al. Jun 2006 B1
7085367 Lang Aug 2006 B1
7124164 Chemtob Oct 2006 B1
7149499 Oran et al. Dec 2006 B1
7180993 Hamilton Feb 2007 B2
7209475 Shaffer et al. Apr 2007 B1
7418664 Ben-Shachar et al. Aug 2008 B2
7441198 Dempski et al. Oct 2008 B2
7478339 Pettiross et al. Jan 2009 B2
7530022 Ben-Shachar et al. May 2009 B2
7620902 Manion et al. Nov 2009 B2
7634533 Rudolph et al. Dec 2009 B2
7792277 Shaffer et al. Sep 2010 B2
7881450 Gentle et al. Feb 2011 B1
7920160 Tamaru et al. Apr 2011 B2
7956869 Gilra Jun 2011 B1
8059557 Sigg et al. Nov 2011 B1
8081205 Baird et al. Dec 2011 B2
8243801 Yamauchi Aug 2012 B2
8274893 Bansal et al. Sep 2012 B2
8290998 Stienhans et al. Oct 2012 B2
8340268 Knaz Dec 2012 B2
8358327 Duddy Jan 2013 B2
8428234 Knaz Apr 2013 B2
8434019 Nelson Apr 2013 B2
8456507 Mallappa et al. Jun 2013 B1
8462103 Moscovitch et al. Jun 2013 B1
8478848 Minert Jul 2013 B2
8630208 Kjeldaas Jan 2014 B1
8638354 Leow et al. Jan 2014 B2
8675847 Shaffer et al. Mar 2014 B2
8694587 Chaturvedi et al. Apr 2014 B2
8694593 Wren et al. Apr 2014 B1
8706539 Mohler Apr 2014 B1
8831505 Seshadri Sep 2014 B1
8890924 Wu Nov 2014 B2
8892646 Chaturvedi et al. Nov 2014 B2
8924862 Luo Dec 2014 B1
8930840 Riskó et al. Jan 2015 B1
9031839 Thorsen et al. May 2015 B2
9032028 Davidson et al. May 2015 B2
9118612 Fish et al. Aug 2015 B2
9131017 Kurupacheril et al. Sep 2015 B2
9137376 Basart et al. Sep 2015 B1
9141860 Vunic Sep 2015 B2
9143729 Anand et al. Sep 2015 B2
9197701 Petrov et al. Nov 2015 B1
9197848 Felkai et al. Nov 2015 B2
9201527 Kripalani et al. Dec 2015 B2
9204099 Brown Dec 2015 B2
9268398 Tipirneni Feb 2016 B2
9298342 Zhang et al. Mar 2016 B2
9323417 Sun et al. Apr 2016 B2
9335892 Ubillos May 2016 B2
9349119 Desai et al. May 2016 B2
9367224 Ananthakrishnan et al. Jun 2016 B2
9369673 Ma et al. Jun 2016 B2
9407621 Vakil et al. Aug 2016 B2
9432512 You Aug 2016 B2
9449303 Underhill et al. Sep 2016 B2
9495664 Cole et al. Nov 2016 B2
9516022 Borzycki et al. Dec 2016 B2
9525711 Ackerman et al. Dec 2016 B2
9553799 Tarricone et al. Jan 2017 B2
9563480 Messerli et al. Feb 2017 B2
9602700 Ahiska Mar 2017 B2
9609030 Sun et al. Mar 2017 B2
9614756 Joshi Apr 2017 B2
9667799 Olivier et al. May 2017 B2
9762709 Snyder et al. Sep 2017 B1
9892518 Lazar Feb 2018 B2
20020018051 Singh Feb 2002 A1
20020076003 Zellner et al. Jun 2002 A1
20020078153 Chung et al. Jun 2002 A1
20020188522 McCall et al. Dec 2002 A1
20030028647 Grosu Feb 2003 A1
20030046421 Horvitz et al. Mar 2003 A1
20030174826 Hesse Sep 2003 A1
20030197739 Bauer Oct 2003 A1
20030227423 Arai et al. Dec 2003 A1
20040054885 Bartram et al. Mar 2004 A1
20040098456 Krzyzanowski et al. May 2004 A1
20040210637 Loveland Oct 2004 A1
20040267938 Shoroff et al. Dec 2004 A1
20050014490 Desai et al. Jan 2005 A1
20050055405 Kaminsky et al. Mar 2005 A1
20050055412 Kaminsky et al. Mar 2005 A1
20050085243 Boyer et al. Apr 2005 A1
20050099492 Orr May 2005 A1
20050108328 Berkeland et al. May 2005 A1
20050215229 Cheng Sep 2005 A1
20060004911 Becker et al. Jan 2006 A1
20060026255 Malamud et al. Feb 2006 A1
20060083305 Dougherty et al. Apr 2006 A1
20060164552 Cutler Jul 2006 A1
20060224430 Butt Oct 2006 A1
20060250987 White et al. Nov 2006 A1
20070005752 Chawla et al. Jan 2007 A1
20070041366 Vugenfirer et al. Feb 2007 A1
20070058842 Vallone et al. Mar 2007 A1
20070091831 Croy et al. Apr 2007 A1
20070100986 Bagley et al. May 2007 A1
20070116225 Zhao et al. May 2007 A1
20070139626 Saleh et al. Jun 2007 A1
20070198637 Deboy et al. Aug 2007 A1
20070248244 Sato et al. Oct 2007 A1
20070250567 Graham et al. Oct 2007 A1
20080068447 Mattila et al. Mar 2008 A1
20080080532 O'Sullivan et al. Apr 2008 A1
20080209452 Ebert et al. Aug 2008 A1
20090012963 Johnson et al. Jan 2009 A1
20090100142 Stern et al. Apr 2009 A1
20090193327 Roychoudhuri et al. Jul 2009 A1
20090254619 Kho et al. Oct 2009 A1
20090256901 Mauchly et al. Oct 2009 A1
20090292999 LaBine et al. Nov 2009 A1
20090309846 Trachtenberg et al. Dec 2009 A1
20100005142 Xiao et al. Jan 2010 A1
20100061538 Coleman et al. Mar 2010 A1
20100070640 Allen, Jr. et al. Mar 2010 A1
20100073454 Lovhaugen et al. Mar 2010 A1
20100094867 Badros et al. Apr 2010 A1
20100121959 Lin et al. May 2010 A1
20100157978 Robbins et al. Jun 2010 A1
20100211872 Rolston et al. Aug 2010 A1
20100245535 Mauchly Sep 2010 A1
20100262925 Liu et al. Oct 2010 A1
20100303227 Gupta Dec 2010 A1
20100316207 Brunson Dec 2010 A1
20100318399 Li et al. Dec 2010 A1
20110075830 Dreher et al. Mar 2011 A1
20110131498 Chao et al. Jun 2011 A1
20110264928 Hinckley Oct 2011 A1
20110270609 Jones et al. Nov 2011 A1
20110271211 Jones et al. Nov 2011 A1
20110283226 Basson et al. Nov 2011 A1
20110314139 Song et al. Dec 2011 A1
20120013704 Sawayanagi et al. Jan 2012 A1
20120084714 Sirpal et al. Apr 2012 A1
20120179502 Farooq et al. Jul 2012 A1
20120190386 Anderson Jul 2012 A1
20120246229 Carr et al. Sep 2012 A1
20120246596 Ording et al. Sep 2012 A1
20120284635 Sitrick et al. Nov 2012 A1
20120296957 Stinson et al. Nov 2012 A1
20120306757 Keist et al. Dec 2012 A1
20120306993 Sellers-Blais Dec 2012 A1
20130038675 Malik Feb 2013 A1
20130047093 Reuschel et al. Feb 2013 A1
20130050398 Krans et al. Feb 2013 A1
20130063542 Bhat et al. Mar 2013 A1
20130090065 Fisunenko et al. Apr 2013 A1
20130091205 Kotler et al. Apr 2013 A1
20130091440 Kotler et al. Apr 2013 A1
20130135837 Kemppinen May 2013 A1
20130141371 Hallford et al. Jun 2013 A1
20130148789 Hillier et al. Jun 2013 A1
20130185672 McCormick et al. Jul 2013 A1
20130198629 Tandon et al. Aug 2013 A1
20130215215 Gage et al. Aug 2013 A1
20130219278 Rosenberg Aug 2013 A1
20130222246 Booms et al. Aug 2013 A1
20130225080 Doss et al. Aug 2013 A1
20130227433 Doray et al. Aug 2013 A1
20130290421 Benson et al. Oct 2013 A1
20130297704 Alberth, Jr. et al. Nov 2013 A1
20130300637 Smits et al. Nov 2013 A1
20130329865 Ristock et al. Dec 2013 A1
20130335507 Aarrestad et al. Dec 2013 A1
20140012990 Ko Jan 2014 A1
20140028781 MacDonald Jan 2014 A1
20140040819 Duffy Feb 2014 A1
20140068452 Joseph et al. Mar 2014 A1
20140078182 Utsunomiya Mar 2014 A1
20140108486 Borzycki et al. Apr 2014 A1
20140157338 Pearce Jun 2014 A1
20140161243 Contreras et al. Jun 2014 A1
20140185686 Wu Jul 2014 A1
20140198175 Shaffer et al. Jul 2014 A1
20140237371 Klemm et al. Aug 2014 A1
20140280595 Mani et al. Sep 2014 A1
20140282213 Musa et al. Sep 2014 A1
20140317561 Robinson et al. Oct 2014 A1
20140351327 Huang et al. Nov 2014 A1
20140372908 Kashi et al. Dec 2014 A1
20150004571 Ironside et al. Jan 2015 A1
20150009278 Modai et al. Jan 2015 A1
20150029301 Nakatomi et al. Jan 2015 A1
20150067552 Leorin et al. Mar 2015 A1
20150074189 Cox et al. Mar 2015 A1
20150081885 Thomas et al. Mar 2015 A1
20150085060 Fish et al. Mar 2015 A1
20150089393 Zhang et al. Mar 2015 A1
20150089394 Chen et al. Mar 2015 A1
20150113050 Stahl Apr 2015 A1
20150172120 Dwarampudi et al. Jun 2015 A1
20150304120 Xiao et al. Oct 2015 A1
20150304366 Bader-Natal et al. Oct 2015 A1
20150319113 Gunderson et al. Nov 2015 A1
20150373063 Vashishtha et al. Dec 2015 A1
20160050079 Martin De Nicolas et al. Feb 2016 A1
20160050160 Li et al. Feb 2016 A1
20160050175 Chaudhry et al. Feb 2016 A1
20160072862 Bader-Natal et al. Mar 2016 A1
20160094593 Priya Mar 2016 A1
20160105345 Kim et al. Apr 2016 A1
20160110056 Hong et al. Apr 2016 A1
20160173537 Kumar et al. Jun 2016 A1
20160182580 Nayak Jun 2016 A1
20160188980 Martin Jun 2016 A1
20160277461 Sun et al. Sep 2016 A1
20160283909 Adiga Sep 2016 A1
20160307165 Grodum et al. Oct 2016 A1
20160309037 Rosenberg et al. Oct 2016 A1
20160379373 Givon et al. Dec 2016 A1
20170006162 Bargetzi et al. Jan 2017 A1
20170006446 Harris et al. Jan 2017 A1
20170070706 Ursin et al. Mar 2017 A1
20170104961 Pan et al. Apr 2017 A1
20170324850 Snyder et al. Nov 2017 A1
Foreign Referenced Citations (11)
Number Date Country
101055561 Oct 2007 CN
102572370 Jul 2012 CN
102655583 Sep 2012 CN
102938834 Feb 2013 CN
103141086 Jun 2013 CN
204331453 May 2015 CN
959585 Nov 1999 EP
1488630 Dec 2004 EP
2341686 Aug 2016 EP
WO 2012167262 Dec 2012 WO
WO 2014118736 Aug 2014 WO
Non-Patent Literature Citations (29)
Entry
International Search Report and Written Opinion from the International Searching Authority, dated Apr. 20, 2018, 12 pages, for the corresponding International Application PCT/US2018/017947.
Author Unknown, “A Primer on the H.323 Series Standard,” Version 2.0, available at http://www.packetizer.com/voip/h323/papers/primer/, retrieved on Dec. 20, 2006, 17 pages.
Author Unknown, ““I can see the future” 10 predictions concerning cell-phones,” Surveillance Camera Players, http://www.notbored.org/cell-phones.html, Jun. 21, 2003, 2 pages.
Author Unknown, “Active screen follows mouse and dual monitors,” KDE Community Forums, Apr. 13, 2010, 3 pages.
Author Unknown, “Implementing Media Gateway Control Protocols” A RADVision White Paper, Jan. 27, 2002, 16 pages.
Author Unknown, “Manage Meeting Rooms in Real Time,” Jan. 23, 2017, door-tablet.com, 7 pages.
Averusa, “Interactive Video Conferencing K-12 applications,” “Interactive Video Conferencing K-12 applications” copyright 2012. http://www.averusa.com/education/downloads/hvc brochure goved.pdf (last accessed Oct. 11, 2013).
Cisco Systems, Inc., “Cisco WebEx Meetings Server System Requirements release 1.5.” 30 pages, Aug. 14, 2013.
Cisco White Paper, “Web Conferencing: Unleash the Power of Secure, Real-Time Collaboration,” pp. 1-8, 2014.
Clarke, Brant, “Polycom Announces RealPresence Group Series,” “Polycom Announces RealPresenoe Group Series” dated Oct. 8, 2012 available at http://www.323.tv/news/polycom-realpresence-group-series (last accessed Oct. 11, 2013).
Clauser, Grant, et al., “Is the Google Home the voice-controlled speaker for you?,” The Wire Cutter, Nov. 22, 2016, pp. 1-15.
Cole, Camille, et al., “Videoconferencing for K-12 Classrooms,” Second Edition (excerpt), http://www.iste.org/docs/excerpts/VIDCO2-excerpt.pdf (last accessed 1011-2013), 2009.
Epson, “BrightLink Pro Projector,” BrightLink Pro Projector. http://www.epson.com/cgi-bin/Store/jsp/Landing/brightlink-pro-interactive-projectors.do?ref=van brightlink-pro—dated 2013 (last accessed 10-112013).
Infocus, “Mondopad,” Mondopad. http://www.infocus.com/sites/default/files/InFocus-Mondopad-INF5520a-INF7021-Datasheet-EN.pdf (last accessed Oct. 11, 2013), 2013.
MacCormick, John, “Video Chat with Multiple Cameras,” CSCW '13, Proceedings of the 2013 conference on Computer supported cooperative work companion, pp. 195-198, ACM, New York, NY, USA, 2013.
Microsoft, “Positioning Objects on Multiple Display Monitors,” Aug. 12, 2012, 2 pages.
Mullins, Robert, “Polycom Adds Tablet Videoconferencing,” Mullins, R. “Polycom Adds Tablet Videoconferencing” available at http://www.informationweek.com/telecom/unified-communications/polycom-adds-tablet-videoconferencing/231900680 dated Oct. 12, 2011 (last accessed Oct. 11, 2013).
Nu-Star Technologies, “Interactive Whiteboard Conferencing,” Interactive Whiteboard Conferencing. http://www.nu-star.com/interactive-conf.php dated 2013 (last accessed Oct. 11, 2013).
Polycom, “Polycom RealPresence Mobile: Mobile Telepresence & Video Conferencing,” http://www.polycom.com/products-services/hd-telepresence-video-conferencing/realpresence-mobile.html#stab1 (last accessed Oct. 11, 2013), 2013.
Polycom, “Polycom Turns Video Display Screens into Virtual Whiteboards with First Integrated Whiteboard Solution for Video Collaboration,” Polycom Turns Video Display Screens into Virtual Whiteboards with First Integrated Whiteboard Solution for Video Collaboration—http://www.polycom.com/company/news/press-releases/2011/20111027 2.html—dated Oct. 27, 2011.
Polycom, “Polycom UC Board, Transforming ordinary surfaces into virtual whiteboards” 2012, Polycom, Inc., San Jose, CA, http://www.uatg.com/pdf/polycom/polycom-uc-board- datasheet.pdf, (last accessed Oct. 11, 2013).
Stevenson, Nancy, “Webex Web Meetings for Dummies” 2005, Wiley Publishing Inc., Indianapolis, Indiana, USA, 339 pages.
Stodle. Daniel, et al., “Gesture-Based, Touch-Free Multi-User Gaming on Wall-Sized, High-Resolution Tiled Displays,” 2008, 13 pages.
Thompson, Phil, et al., “Agent Based Ontology Driven Virtual Meeting Assistant,” Future Generation Information Technology, Springer Berlin Heidelberg, 2010, 4 pages.
TNO, “Multi-Touch Interaction Overview,” Dec. 1, 2009, 12 pages.
Toga, James, et al., “Demystifying Multimedia Conferencing Over the Internet Using the H.323 Set of Standards,” Intel Technology Journal Q2, 1998, 11 pages.
Ubuntu, “Force Unity to open new window on the screen where the cursor is?” Sep. 16, 2013, 1 page.
VB Forums, “Pointapi,” Aug. 8, 2001, 3 pages.
Vidyo, “VidyoPanorama,” VidyoPanorama—http://www.vidvo.com/products/vidvobanorama/ dated 2013 (last accessed Oct. 11, 2013).
Related Publications (1)
Number Date Country
20180232889 A1 Aug 2018 US