The present invention relates generally to image processing, and in particular embodiments, to a method and system for image processing to classify an object in an image.
Systems and methods have been developed for defining an object in video and for tracking that object through the frames of the video. In various applications, a person may be the “object” to be tracked. For example, sports images are interested in following the actions of a person such as the players and/or the referees.
Players and referees are displayed in sports videos. Localization and labeling of them can be done in IPTV systems so that a regular TV broadcast (MPEG-2/-4) is augmented with additional information (MPEG-7 encoded) that defines those objects in the video, along with additional content to be displayed when they are selected. Specification of objects with additional content (metadata) is usually implemented by an authoring tool that includes such functions as extraction of shots and key frames, specification of the interactive regions, and tracking of the specified regions to get the region locations in all frames.
Team classification-based interactive services by clicking the player in hypervideo or iTV has been discussed. Team information search and retrieval and team data (statistics results, articles and other media) can be linked assuming the player can be localized by the interaction service system. Various methods for locating the players/referees can be split in two groups. The first group makes use of fixed cameras (usually they are calibrated in advance) in a controlled environment while the second group uses only regular broadcasting videos. While the former can provide better performance, the latter are more flexible. In the second group, some approaches tried to overcome difficulties by finding the playfield first, using color segmentation and post-processing with morphological operations, such as connected component analysis, in order to limit the search area.
In accordance with a first embodiment of the present invention, an image processing method is performed, e.g., on a processor. An object is located within an image, such as a video or still image. An area around the object is determined and divided into at least first and second portions based upon image information within the area. The object can then be classified based upon both image information in the first portion of the area and image information in the second portion of the area.
In another embodiment, an interactive television system includes an authoring tool configured to receive a video image, locate an object within the image, divide an area around the object into first and second portions; and generate metadata based upon first image information within the first portion and upon second image information within the second portion. An aggregator is configured to receive the video image and metadata and generate a video stream that is enhanced with the metadata and a delivery system is configured to transmit the video stream that is enhanced with the metadata.
For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, in which:
The making and using of the presently preferred embodiments are discussed in detail below. It should be appreciated, however, that the present invention provides many applicable inventive concepts that can be embodied in a wide variety of specific contexts. The specific embodiments discussed are merely illustrative of specific ways to make and use the invention, and do not limit the scope of the invention.
Team classification of player/referee is used to distinguish their identities as Team A or Team B or Referee. Issues for this task include the selection of features and the clustering method for matching. Template and histogram methods have been used. Feature selection is based on discrimination of different classes, robustness and computational cost. Histograms are a good trade-off between these requisites. Clustering methods can be supervised or unsupervised. The present invention improves the efficiency of feature extraction and discrimination in histogram matching simultaneously.
In a first embodiment, the present invention discloses a sports team based interactive service for IPTV systems, including real time and on demand video delivery. For example, a sporting event video is processed and the team/referee visual objects are analyzed and categorized in real time. A multi-histogram matching scheme and a method for separating the player/referee blob (obtained by playfield model-based segmentation) into two parts (upper and lower) via a discriminate function are proposed. The proposed scheme achieves good classification accuracy with low computational complexity. The team classification-based interaction functions include team information search and retrieval and team data (statistics result, articles and other media) linking when the player is clicked. The proposed team classification method still has a promising potential usage in sporting events and tactics analysis as well as an interactive service for sports programs in IPTV systems.
In other embodiments, a proposed team classification-based interactive service for sports programs in an IPTV system is provided. In another embodiment, a method for team classification of player/referee in soccer game videos based on a multi-histogram matching scheme, which offers better classification accuracy with low computational complexity, is provided. In yet another embodiment, a method can be used to separate the player/referee blob (obtained by playfield model-based segmentation) into two parts (upper and lower) by a proposed discriminate function.
In hyperlinked video, objects are selectable resulting in an associated action, akin to linked rich media content about the objects of interest. Possible venues for hyperlinked video include broadcast TV, streaming video and published media such as DVD. Hyperlinked video offers new interaction possibilities with streaming media.
Interactive TV (iTV) is a popular application area of hyperlinked video with the convergence between broadcast and network communications. For example, the European GMF4iTV (Generic Media Framework for Interactive Television) project has developed such a system where active video objects are associated to metadata information, embedded in the program stream at production time and can be selected by the user at run time to trigger the presentation of their associated metadata. Another European PorTiVity (Portable Interactivity) project is developing and experimenting with a complete end-to-end platform providing Rich Media Interactive TV services for portable and mobile devices, realizing direct interactivity with moving objects on handheld receivers connected to DVB-H (broadcast channel) and UMTS (unicast channel).
IPTV (Internet Protocol Television) is a system where a digital television service is delivered using Internet Protocol over a network infrastructure, which may include delivery by a broadband connection. An IP-based platform also allows significant opportunities to make the TV viewing experience more interactive and personalized. Interactive TV services will be a key differentiator for the multitude of IPTV offerings that are emerging. Interactivity via a fast two-way connection will lift IPTV ahead of today's television.
Aspects of the present disclosure relate to a scenario related to a rich media interactive TV application, for example, an IPTV application. The focus is interaction with moving objects in sports programs. Based on direct interaction with certain objects, the TV viewer can retrieve and link rich media content about objects of interest. The term “television” or “TV” is used to denote any video image displayed to a user. For example, this image can be displayed on a computer screen, mobile device or an actual television and still be within the scope of the term television.
Players and referees are two examples of moving objects in sports videos. Localization and labeling of them is useful for interactive services in IPTV systems, so that a regular TV broadcast (MPEG-2/-4) is augmented with additional information (MPEG-7 encoded) which defines those objects in the video, along with additional content to be displayed when they are selected. Specification of objects with additional content (metadata) can be implemented by an authoring tool that includes such functions as extraction of shots and key frames, specification of the interactive regions, and tracking of the specified regions to get the region locations in all frames.
In embodiments of the present invention, a player team classification-based interactive service for soccer game programs in an IPTV system is proposed.
The interaction is based on the combination of information prepared on the IPTV server side and real time team classification on the IPTV client side and/or in a network middle box (such as the Content and Metadata Sources block 105 shown in
The system can be applied to a number of sports. For example, sports such as soccer, football, basketball, baseball, hockey, cricket and others can utilize the concepts described herein.
In the embodiment, the user is registered with the IMS infrastructure. The TV content is to be enhanced with metadata information for the playfield description and team target models represented as the multi-color histograms. The IPTV client is enhanced with such a service, which implies an environment to run additional services and respectively execute advanced program code on the IPTV client for on-line player localization (segmentation or tracking).
The IPTV Service Control Function 150 manages all user-to-content and content-to-user relationships and controls the Content Delivery and Storage 140 and the Content Aggregator 110. The IPTV Application Function 145 supports various service functions and provides an interface to the user 160 to notice the IPTV service information and accept the service request of the user (such as registration or authentication). The IPTV Application Function 145, in conjunction with the Service Control Function 150 provides users with the value added services they request.
The Content Preparation 130 sends a content distribution request to the Content Delivery Control 135. The Content Delivery Control 135 produces a distribution task between Content Preparation 130 and the Content Delivery and Storage 140 according to the defined distribution policy when it receives the request of content distribution. The Content Delivery and Storage 140 delivers aggregated and metadata-enhanced content to the user 160, and may perform player localization and team classification in implementations where these task are not performed at the IPTV Client 155.
The Content Aggregator 110 links the content 120 to the metadata 125 via the Authoring Tool 115 and aggregates content that is enhanced with metadata information for interactive service purposes. The Authoring Tool 115 runs play field learning and team model acquisition and generates the MPEG-7 metadata.
Although the present invention targets interactive services in IPTV systems, the invention is not so limited. The proposed scheme can be used in other video delivery systems with improved accuracy and low computational complexity.
The player/referee localization can be performed in a number of ways. For example, various methods for locating the players/referees can be split in two groups. The first group makes use of fixed cameras (usually calibrated in advance) in a controlled environment. Such a technique is taught by M. Xu, J. Orwell, G. Jones, “Tracking football players with multiple cameras”. ICIP 2004, pp. 2909-2912, the content of which is incorporated herein by reference. The second group uses only regular broadcasting videos. While the former can provide better performance, the latter are more flexible. In the second group, difficulties in localization can be overcome by finding the playfield first, using color segmentation and post-processing with morphological operations, such as connected component analysis, in order to limit the search area.
The issues for the localization task are selection of features and the clustering method for matching. In one aspect, the present invention improves the efficiency of feature extraction and discrimination in histogram matching simultaneously. Embodiments adopt a multi (e.g., two)-histogram based matching method to classify the players and referees in game videos.
The playfield extraction 205 includes playfield pixel detection 210, connected component analysis 215, morphological filtering (e.g., dilation, erosion) 220 and size filtering 225. Further details on playfield extraction will now be discussed.
The playfield can be used for analyzing several types of sports videos, such as soccer, football, baseball and tennis. For example, the playfield is grass for a soccer, baseball or football game. While the color of grass is generally green, this color can vary depending on the individual playfield, the presence of shadows or the viewing angle. In another example, the playfield is ice for a hockey game, but similar issues will occur.
Embodiments of the invention will now be described with to respect soccer. It is understood that the same concepts apply to other sports as well.
Due to the nature of the soccer game, there are many wide shots where the majority of the image is the playfield. Based on this observation, an unsupervised segmentation technique can obtain the playfield model. However, the playfield area in any frame is not always big enough to make the assumption of dominant color valid. Therefore, supervised methods for learning the playfield model can be used. A drawback of the supervised methods is the requirement of many labeled data, where hand-labeling is tedious and expensive.
In one embodiment, two options are defined. The first option is a small set of labeled data, i.e., the pixels in a given playfield area, is used to generate a rough playfield model with a single Gaussian or a mixture of Gaussian distributions (for the latter one, more labeled data is compulsory). Then, this model can be modified by collecting more playfield pixels based on an unsupervised method using dominant color detection.
In a second option, one frame, where the dominant color assumption is satisfied, is selected. Then its dominant mode is extracted to generate the initial playfield model. Like the first option, this model can be modified by collecting more playfield pixels based on dominant color detection.
The determination of the playfield model is discussed in greater detail in Provisional Patent Application Ser. No. 61/144,386, which is incorporated herein by reference. Further information can be derived from that application.
Players and referees are foreground objects in the soccer playfield. Since soccer is a spectator sport, the play fields, the lines, the ball and the uniforms of the players and referees are designed to be visually distinctive in color. Therefore, the framework in
The object detection 230 includes interior filtering 235. A comparison is made between the pre-filtered and the filtered image data as illustrated by the XOR gate. The result can be thought of as the image without the background. This result can then go through connected component analysis 240 and shape filtering 245. The shape filtering can deal with size, roundness and/or eccentricity, as examples.
Given the segmented blobs for players and referees, each will be labeled with an identity of Team A, Team B or Referee. Sometimes two team goalies are classified as well. In order to do that, each team player or referee's appearance model is used to acquire by learning the labeled data.
Since the player's jersey is mostly discriminate from the shorts and the former occupies more area in the player blob, two color histograms can be used to represent the player's appearance model, with one being given a bigger weight in histogram matching. In this context, the terms jersey and shorts are used to denote the upper portion of the player and the lower portion of the player regardless of whether the player is actually wearing a jersey or shorts. For example, the term “shorts” includes shorts worn by a basketball and also pants worn by a baseball player. Similarly, the term “jersey” can be applied to both teams in a “shirts v. skins” game.
The separation of the jersey 315 and shorts 320 in each player/referee blob 300 into upper and lower parts (here it is assumed the player stands approximately upright) is based on a discriminate function. Given a rectangle 325 of size w×h (width w and height h), the cutting line 310 is searched around the middle position which maximizes the objective function
ρ[p,q]=1.0−Σu=1m√{square root over (puqu)}, (1)
where the pu is the color histogram of the upper part and the qu is the color histogram of the lower part. m is the number of bins in the histogram. In one embodiment, a histogram with 8×8×8 bins of color RGB color space can be used. The measure in equation (1) is called the Bhattacharyya distance.
To accelerate the above process, the color histogram for each row in the search gap of the target window (rectangle 325) is calculated and so only the row histogram is added or subtracted from histograms of the upper or lower part respectively when the cutting line 310 scrolls up and down one row per iteration. Since the position range of the cutting line 310 is narrow, an exhaustive search is feasible to find the best cutting line that discriminates most of the upper part from the lower part. For example, the search gap my include a range of less than 25% of the rows, or preferably, less than 10% of the rows.
Eventually two color histograms are utilized for player/referee labeling. In appearance model learning, two histograms are generated and saved as p1i and p2i for each team or referee, i.e., i=Team A, Team B and Referee. If multiple samples for each type (either team player or referee), obtained from either manually segmented or playfield-model-based segmentation, are applied, all the pixels in the upper part (jersey mainly) and the lower part (shorts mainly) of the segmented blob are collected to build the upper and lower color histogram respectively. Eventually both histograms are normalized.
In testing or running the classification, for an unknown blob (extracted from playfield-model-based segmentation process) in the soccer game videos two normalized color histograms for jersey and short regions are built as well, i.e., q1 and q2, then its label i is estimated by minimizing a weighted Bhattacharyya distance as
where w is the weight (0<w<1.0, suggested as 0.7).
The discussion above provides details on the determination of which team a player is associated with. With additional processing, the identity of the player may also be determined. For example, number recognition techniques could be used to locate and identify the player's number. In a typical game, the roster of the players is known so that a comparison could be made between the known player information and the derived player information.
A specific example of an interactive television system will now be described with respect to
This scenario describes a rich media interactive television application. It focuses on new concepts for interaction with moving objects in the sport programs. Based on direct interaction with certain objects, the viewer can retrieve rich media content about objects of his choice.
The interaction is based on the combination of information prepared on the IPTV server side and real time team classification on the IPTV client side. The information on the server side is stored as metadata in the MPEG-7 format and describes the play field, team templates and related media information about the teams. The client side does the real time object processing and, based on the MPEG-7 metadata to do the team classification, presents the related media information on a screen for user interaction.
The TV content is enhanced with metadata information for the description of the field and team templates represented as the color histogram. The user has to be registered with the IMS infrastructure. The IPTV client has to be enhanced with such a service, which implies an environment to run additional services and respectively execute advanced program code on the IPTV client for content processing and object highlighting. Charging can be used for transaction and accounting.
Referring now to
The IPTV client 420, for example a set top box (STB), is responsible to provide the viewer 430 with the functionality to make use of the interaction, in terms of real time object processing, to spot high lighting of objects containing additional content, to select objects and to view additional content. The IMS based IPTV client 420 is enabled with techniques such as real time object processing for providing the interactive service. In another example, if the video content is not enhanced with the metadata information, the IPTV client 420 can provide a user interface to the user 430 for collecting the team templates.
The user 430 makes use of the service by selecting objects, and consuming additional content. The delivery system 440, typically owned by the service provider 410, delivers aggregated and metadata-enhanced content to the user 430, provides trick functions and highly efficient video and audio coding technologies.
The content aggregator 450 links the content 460 to the metadata 470 via the authoring tool 480. This aggregator 450 aggregates content which is enhanced with metadata information for interactive service purposes. The content aggregator 450 provides the delivery system 440 with aggregated content and attaches them with enhanced content. Therefore MPEG7 as standard for multimedia metadata descriptions should be considered.
The authoring tool 480 disposes algorithms for field learning and team template acquisition in video streams and an MPEG-7 metadata generator.
In operation of the system 400, the user 430 registers with the service provider 410 and requests the desired service. For this example, the user 430 is able to click on a player to start tracking the player and get the team information about his/her team information and related video by clicking on the appropriate colored button on remote control.
In response to the request from the user 430, the service provider 410 causes the aggregator 450 to prepare the enhanced content. In doing so, the aggregator 450 communicates with the authoring tool 480, which processes the content image and enhances the content 460 with the metadata 470. The aggregator 450 can then provide the aggregated content to the delivery system 440.
The delivery system 440 forwards the enhanced content to the IPTV client 420, which interacts with the user 430. The user 430 also provides stream control to the delivery system 440, either via the IPTV client 420 or otherwise.
Features of each of the functional units shown in
Features of the service provider 410 include:
Features of the IPTV client 420 include
Features of the user 430 include:
Features of the delivery system 440 include:
Features of the aggregator 450 include:
Features of the authoring tool 480 include:
In
Aspects of the invention have been described in the context of specific examples. It is understood, however, that the invention is not limited to just these examples. For example, the invention has been discussed with respect to video images. The concepts described herein could equally apply to still images. The concepts have also been applied to sports images. Any other images, whether photographic, drawn, computer generated or other, can also utilize concepts described herein.
While this invention has been described with reference to illustrative embodiments, this description is not intended to be construed in a limiting sense. Various modifications and combinations of the illustrative embodiments, as well as other embodiments of the invention, will be apparent to persons skilled in the art upon reference to the description. It is therefore intended that the appended claims encompass any such modifications or embodiments.
This is a continuation application of U.S. patent application Ser. No. 12/686,902, entitled “Player Team Classification-Based Interactive Service for Soccer Game Programs in an IPTV System” which was filed on Jan. 13, 2010, which application claims the benefit of U.S. Provisional Application No. 61/144,380, filed on Jan. 13, 2009, entitled “Player Team Classification-Based Interactive Service for Sports Game Programs in an IPTV System” and also claims the benefit of U.S. Provisional Application No. 61/144,386, filed on Jan. 13, 2009, entitled “A Semi-Supervised Method For Learning and On-Line Updating Playfield Model in Sports Videos.” All three of these applications are incorporated herein by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
5638465 | Sano et al. | Jun 1997 | A |
5961571 | Gorr et al. | Oct 1999 | A |
6360008 | Suzuki et al. | Mar 2002 | B1 |
6363160 | Bradski et al. | Mar 2002 | B1 |
6587576 | Wesolkowski | Jul 2003 | B1 |
6587593 | Matsuoka et al. | Jul 2003 | B1 |
6798897 | Rosenberg | Sep 2004 | B1 |
6870945 | Schoepflin | Mar 2005 | B2 |
7251364 | Tomita, Jr. et al. | Jul 2007 | B2 |
7457440 | Scott et al. | Nov 2008 | B2 |
7542608 | Yeh | Jun 2009 | B2 |
7921036 | Sharma et al. | Apr 2011 | B1 |
8064716 | Shaw | Nov 2011 | B2 |
8081846 | Ignatchenko et al. | Dec 2011 | B2 |
8218077 | Kondo et al. | Jul 2012 | B2 |
8249343 | Perronnin et al. | Aug 2012 | B2 |
8363884 | Amsterdam et al. | Jan 2013 | B2 |
8391547 | Huang et al. | Mar 2013 | B2 |
8456528 | Brown et al. | Jun 2013 | B2 |
8532336 | Bobbitt et al. | Sep 2013 | B2 |
8542872 | Gornick et al. | Sep 2013 | B2 |
8587666 | Albu et al. | Nov 2013 | B2 |
8588465 | Geiss | Nov 2013 | B2 |
8687918 | Tamura | Apr 2014 | B2 |
8705894 | Albu et al. | Apr 2014 | B2 |
8718363 | Desai et al. | May 2014 | B2 |
8731304 | Baba | May 2014 | B2 |
9269154 | Huang | Feb 2016 | B2 |
9390506 | Asvatha Narayanan | Jul 2016 | B1 |
9681125 | Wang | Jun 2017 | B2 |
9881380 | Sorkine Hornung | Jan 2018 | B2 |
9898677 | An jelkovi | Feb 2018 | B1 |
20020028021 | Foote et al. | Mar 2002 | A1 |
20020031256 | Hiramatsu et al. | Mar 2002 | A1 |
20020085096 | Mertens | Jul 2002 | A1 |
20020140813 | Trajkovic et al. | Oct 2002 | A1 |
20020168091 | Trajkovic | Nov 2002 | A1 |
20020172417 | Nicolas | Nov 2002 | A1 |
20030056799 | Young et al. | Mar 2003 | A1 |
20030095709 | Zhou et al. | May 2003 | A1 |
20030107649 | Flickner et al. | Jun 2003 | A1 |
20030169921 | Yamamoto | Sep 2003 | A1 |
20040001612 | Gutta | Jan 2004 | A1 |
20040040041 | Crawford | Feb 2004 | A1 |
20040126038 | Aublant et al. | Jul 2004 | A1 |
20050013486 | Wiedemann et al. | Jan 2005 | A1 |
20050069225 | Schneider et al. | Mar 2005 | A1 |
20050100204 | Afzal et al. | May 2005 | A1 |
20050145791 | Fujita et al. | Jul 2005 | A1 |
20050220348 | Chiu et al. | Oct 2005 | A1 |
20050271269 | Errico et al. | Dec 2005 | A1 |
20050286759 | Zitnick et al. | Dec 2005 | A1 |
20060018516 | Masoud et al. | Jan 2006 | A1 |
20060114480 | Chang et al. | Jun 2006 | A1 |
20060265232 | Katariya et al. | Nov 2006 | A1 |
20060280335 | Tomita, Jr. et al. | Dec 2006 | A1 |
20060280363 | Umeda | Dec 2006 | A1 |
20060288291 | Lee et al. | Dec 2006 | A1 |
20060290780 | Porikli | Dec 2006 | A1 |
20060291696 | Shao et al. | Dec 2006 | A1 |
20070003164 | Takata | Jan 2007 | A1 |
20070016999 | Harber | Jan 2007 | A1 |
20070064974 | Ayachitula et al. | Mar 2007 | A1 |
20070098257 | Shah et al. | May 2007 | A1 |
20070104384 | Shaw | May 2007 | A1 |
20070154066 | Lin et al. | Jul 2007 | A1 |
20070183661 | El-Maleh et al. | Aug 2007 | A1 |
20070183662 | Wang | Aug 2007 | A1 |
20070250901 | McIntire et al. | Oct 2007 | A1 |
20080032797 | Harris et al. | Feb 2008 | A1 |
20080050015 | Lu et al. | Feb 2008 | A1 |
20080056568 | Porikli | Mar 2008 | A1 |
20080089581 | Pitie et al. | Apr 2008 | A1 |
20080101652 | Zhao et al. | May 2008 | A1 |
20080144931 | Yan et al. | Jun 2008 | A1 |
20080201116 | Ozdemir et al. | Aug 2008 | A1 |
20080298571 | Kurtz et al. | Dec 2008 | A1 |
20080304735 | Yang et al. | Dec 2008 | A1 |
20090016643 | Johnson | Jan 2009 | A1 |
20090034932 | Oisel et al. | Feb 2009 | A1 |
20090060352 | Distante et al. | Mar 2009 | A1 |
20090110236 | Huang | Apr 2009 | A1 |
20090141949 | Lee et al. | Jun 2009 | A1 |
20090153744 | Morita et al. | Jun 2009 | A1 |
20090220070 | Picard et al. | Sep 2009 | A1 |
20090225183 | Tamura | Sep 2009 | A1 |
20090238406 | Huang et al. | Sep 2009 | A1 |
20090238549 | Atsushi | Sep 2009 | A1 |
20090257649 | Yamauchi et al. | Oct 2009 | A1 |
20090262217 | Mack et al. | Dec 2009 | A1 |
20090317061 | Jung et al. | Dec 2009 | A1 |
20100026809 | Curry | Feb 2010 | A1 |
20100027845 | Kim | Feb 2010 | A1 |
20100027846 | Xu | Feb 2010 | A1 |
20100066667 | MacDougall et al. | Mar 2010 | A1 |
20100098307 | Huang et al. | Apr 2010 | A1 |
20100134614 | Aman | Jun 2010 | A1 |
20100177194 | Huang et al. | Jul 2010 | A1 |
20100177969 | Huang et al. | Jul 2010 | A1 |
20100214472 | Tomonaga | Aug 2010 | A1 |
20110013836 | Gefen et al. | Jan 2011 | A1 |
20110044538 | Vanderhoff | Feb 2011 | A1 |
20110107370 | Cassanova et al. | May 2011 | A1 |
20110142289 | Barenbrug | Jun 2011 | A1 |
20110216167 | Katz | Sep 2011 | A1 |
20110291925 | Israel et al. | Dec 2011 | A1 |
20110310100 | Adimatyam et al. | Dec 2011 | A1 |
20120062700 | Antonellis et al. | Mar 2012 | A1 |
20120106837 | Partis | May 2012 | A1 |
20120121174 | Bhagavathy et al. | May 2012 | A1 |
20120170803 | Millar | Jul 2012 | A1 |
20120213438 | Quan | Aug 2012 | A1 |
20130170557 | Wang | Jul 2013 | A1 |
20130336535 | Bobbin et al. | Dec 2013 | A1 |
20140056473 | Tojo | Feb 2014 | A1 |
20140074263 | Balakrishnan et al. | Mar 2014 | A1 |
20140254881 | Jin | Sep 2014 | A1 |
20150221069 | Shaburova | Aug 2015 | A1 |
20150310370 | Burry | Oct 2015 | A1 |
20150310624 | Bulan | Oct 2015 | A1 |
20160078279 | Pitre | Mar 2016 | A1 |
20160217338 | Li | Jul 2016 | A1 |
20160275375 | Kant | Sep 2016 | A1 |
20170039432 | Bulan | Feb 2017 | A1 |
20170075993 | Matsushita | Mar 2017 | A1 |
20170236290 | Sorkine Hornung | Aug 2017 | A1 |
20170278289 | Marino | Sep 2017 | A1 |
Number | Date | Country |
---|---|---|
1991864 | Jul 2007 | CN |
101079109 | Nov 2007 | CN |
101243448 | Aug 2008 | CN |
101287056 | Oct 2008 | CN |
2008118147 | Oct 2008 | WO |
2008139399 | Nov 2008 | WO |
Entry |
---|
First Office Action of Chinese Application No. 201080003841.8, dated Jan. 17, 2013, 14 pages. (Partial Translation). |
Avidan, S., “Ensemble Tracking,” Mitsubishi Electric Research Laboratories, TR2005-065, CVPR, Jun. 2005, 9 pages. |
Babu, R. V., et al., “Robust Tracking with Motion Estimation and Kernel-Based Color Modelling,” IEEE, ICOP'05, 2005, 4 pages. |
Balfanz, D., et al., “Interaction with Content-Augmented Video Via Off-Screen Hyperlinks for Direct Information Retrieval,” The 10-th Int. Conf. in Central Europe on Computer Graphics, Visualization and Computer Vision, Czech Republic, Feb. 4-8, 2002, 8 pages. |
Barnard, M., et al., “Robust Playfield Segmentation using MAP Adaptation,” ICPR'04, 2004, 4 pages. |
Beetz, M., et al., “Camera-based Observation of Football Games for Analyzing Multi-agent Activities,” Proc. of Int. Joint Conf. on Autonomous Agents and Multiagent Systems, Japan, 2006, 8 pages. |
Beetz, et al., “Visually Tracking Football Games Based on TV Broadcasts,” (2007) Proceedings of the Twentieth International Joint Conference on Artificial Intelligence (IJCAI), pp. 2066-2017. |
Black, M. J., et al., “EigenTracking: Robust Matching and Tracking of Articulated Objects Using a View-Based Representation,” EECV'96, 1996, 14 pages. |
Bradski, G. R., “Computer Vision Face Tracking for Use in a Perceptual User Interface,” IEEE Workshop on Applications of Computer Vision, 1998, 15 pages. |
Choi, S., et al., “Where are the ball and players?: Soccer Game Analysis with Color-based Tracking and Image Mosaick,” Int. Conf. on Image Analysis and Processing, Sep. 1997, 15 pages. |
Issue document received in CN 102576412, dated Nov. 5, 2014, 18 pgs. |
Chinese Office Action received in Application No. 201080003841.8 dated Aug. 16, 2013, 5 pages. |
Chinese Search Report received in Application No. 2010800038418 dated Aug. 8, 2013, 3 pages. |
Collins, R., “Mean-Shift Blob Tracking through Scale Space,” CVPR'03, 2003, 7 pages. |
Collins, R. T., et al., “On-Line Selection of Discriminative Tracking Features,” CVPR'03, 2003, 14 pages. |
Comaniciu, D., et al., “Kernel-Based Object Tracking,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, Issue 5, May 2003, 30 pages. |
Gillet, A., et al., “Color image segmentation by analysis of 3D histogram with fuzzy morphological filters,” Chapter in the book: “Fuzzy filters for image processing—Studies in fuzziness and soft computing,” Springer-Verlag Editor, 25 pages, 2002. |
Gomez, et al., “Detection and Classification of Football Players with Automatic Generation of Models,” Optical Engineering 49(1), Jan. 2010, pp. 1-17. |
Hager, G. D., et al., “Multiple Kernel Tracking with SSD,” CVPR'04, 2004, 8 pages. |
Huang, Y., et al., “A Method of Small Object Detection and Tracking Based on Particle Filters,” ICPR'08, 2008, Florida, Dec. 2008, 4 pages. |
Huang, Y., “Player Highlighting and Team Classification in Broadcast Soccer Videos for the Next Generation TV,” WOCC'09, 2009, 27 pages. |
Huang, Y., et al., “Players and Ball Detection in Soccer Videos Based on Color Segmentation and Shape Analysis,,” Int. Workshop on Multimedia Content Analysis and Mining (MCAM'07), Jun. 2007, 10 pages. |
International Search Report and Written Opinion, International Application No. PCT/US2010/020927, dated May 4, 2010, 12 pages. |
Jepson, A. D., et al., “Robust Online Appearance Models for Visual Tracking,” IEEE Conference on Computer Vision and Pattern Recognition, Kauai, 2001, vol. 1, pp. 415-422. |
Kaneko, T., et al., “Template Update Criterion for Template Matching of Image Sequences,” ICPR'02, IEEE, 2002, 5 pages. |
Lim, J., et al., “Incremental Learning for Visual Tracking,” MIT Press, 2005, 8 pages. |
Liu, J., et al., “Automatic Player Detection, Labeling and Tracking in Broadcast Soccer Video,” European Conference on Computer Vision, 2007, 10 pages. |
Matthews, L, et al., “The Template Update Problem,” The Robotics Institute, Carnegie Mellon University, IEEE T-PAMI, 26(6), 2004, 16 pages. |
Muselet, D., et al., “Color person image indexing and retrieval,” Machine Graphics and Vision, vol. 11, No. 2/3, 21 pages, 2002. |
Nguyen, H. T., et al., “Fast Occluded Object Tracking by a Robust Appearance Filter,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, No. 8, Aug. 2004, 6 pages. |
Porikli, F., et al., “Object Tracking in Low-Frame-Rate Video,” SPIE Image and Video Communications and Processing, vol. 5685, Mar. 2005, 10 pages. |
Rakotomalala, V., et al., “Identification of retinal vessels by color image analysis,” Machine Graphics and Vision, vol. 7, No. 4, 23 pages, 1998. |
Rees, D., et al., “CLICK-IT: Interactive Television Highlighter for Sports Action Replay,” Fourteenth International Conference on Pattern Recognition, 1998, 4 pages. |
Rittscher, J., et al., “A Probabilistic Background Model for Tracking,” ECCV'00, 2000, 15 pages. |
Spagnolo, P., et al., “Unsupervised Algorithms for Segmentation and Clustering Applied to Soccer Players Classification,” Int. Conf. on Signal Processing and Multimedia Applications, Barcelona, Jul. 2007, pp. 129-134. |
Tong, X-F., et al., “An Effective and Fast Soccer Ball Detection and Tracking Method,” Proceedings of the 17th International Conference on Pattern Recognition (ICPR'04), 4 pages. |
Ultré, V., et al., “Determination of compatibility coefficients for color edge detection by relaxation,” Proc. of the IEEE Int. Conf. on Image Processing (ICIP 96), vol. 3 Lausanne (Suisse), 5 pages, 1996. |
Utsumi, O., et al., “An Object Detection Method for Describing Soccer Games from Video,” IEEE ICME'02, 2002, 4 pages. |
Vandenbroucke, N., et al., “Color image segmentation by pixel classification in an adapted hybrid color space. Application to soccer image analysis,” Computer Vision and Image Understanding, vol. 90, No. 2, 27 pages, 2003. |
Vanderbroucke, et al., “Color Pixels Classification in Hybrid Space,” IEEE (1998), pp. 1-5. |
Xu, M., et al., “Tracking Football Players with Multiple Cameras,” ICIP 2004, 4 pages. |
Yilmaz, A., “Object Tracking by Asymmetric Kernel Mean Shift with Automatic Scale and Orientation Selection,” CVPR'07, IEEE, 2007, 6 pages. |
Yin, Z., et al., “Spatial Divide and Conquer with Motion Cues for Tracking through Clutter,” IEEE CVPR'06, Jun. 2006, 8 pages. |
Yu, X., et al., “Trajectory-Based Ball Detection and Tracking with Applications to Semantic Analysis of Broadcast Soccer Video,” ACM Multimedia, 2003, pp. 11-20. |
Zhang, H., et al., “Affine Object Tracking with Kernel-based Spatial-Color Representation,” CVPR'05, 2005, 8 pages. |
Zivkovic, Z., et al., “An EM-like algorithm for color-histogram-based object tracking,” CVPR'04, 2004, pp. 798-803. |
Number | Date | Country | |
---|---|---|---|
20160117835 A1 | Apr 2016 | US |
Number | Date | Country | |
---|---|---|---|
61144380 | Jan 2009 | US | |
61144386 | Jan 2009 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12686902 | Jan 2010 | US |
Child | 14990441 | US |