The invention relates generally to audio/visual content and more particularly to an apparatus and method for improved chroma-keying using instrumentation data.
Typical television sport event coverage includes many video cameras covering different parts of the event. Some televised football games have as many as 20 video cameras covering the football field and are capable of providing a viewpoint from many different directions.
In many televised events, it is desirable to include overlays in portions of the televised broadcast. These overlays may include graphical or captured images such as scoreboards, games statistics, advertisements, logos, and play-by-play graphics. To produce a television program of a live event with overlays such as a football game, a large amount of manual input is typically required to create a television program displaying scenes of the football game in conjunction with the overlays placed in an appropriate position. For the overlays to be effective, they should not block an important portion of the live programming. However, positioning the overlays is typically a manually intensive operation.
The invention illustrates a system and method of displaying a base image and an overlay image comprising: capturing a base image of a real event; receiving an instrumentation data based on the real event; identifying a visual segment within the base image based on the instrumentation data; and rendering an overlay image within the visual segment.
Other aspects and advantages of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrated by way of example of the principles of the invention.
Specific reference is made in detail to the embodiments of the invention, examples of which are illustrated in the accompanying drawings. While the invention is described in conjunction with the embodiments, it will be understood that the embodiments are not intended to limit the scope of the invention. The various embodiments are intended to illustrate the invention in different applications. Further, specific details are set forth in the embodiments for exemplary purposes and are not intended to limit the scope of the invention. In other instances, well-known methods, procedures, and components have not been described in detail as not to unnecessarily obscure aspects of the invention.
The invention includes a system and method for employing an improved chroma-keying system that utilizes instrumentation data. The invention utilizes techniques for seamlessly displaying an overlay image within a base image in response to the instrumentation data gathered by video cameras and/or sensors. For the sake of simplicity and clarity, the invention is described with MPEG-2 being chosen as the delivery mechanism. However, any delivery mechanism suitable for use with the invention may be utilized.
A plurality of cameras 120 is utilized to capture visual and audio signals of the event at the event site 110. In addition, the plurality of cameras 120 also captures camera instrumentation data concurrently with the visual and audio signals. Camera instrumentation data may include, for each video frame, the camera location, tilt, zoom, pan, field of view, focus setting, iris setting, and other information related to the optics of each of the plurality of cameras 120.
A plurality of cameras 120 is utilized to capture visual and audio signals of the event at the event site 110. In addition, the plurality of cameras 120 also captures camera instrumentation data concurrently with the visual and audio signals. Camera instrumentation data may include, for each video frame, the camera location, tilt, zoom, pan, field of view, focus setting, iris setting, and other information related to the optics of each of the plurality of cameras 120.
A plurality of sensors 140 are utilized within the event site 110 to capture performance instrumentation data. The performance instrumentation data describes the real event at the event site 110. The plurality of sensors 140 may capture the performance instrumentation data concurrently with the data camera instrumentation data captured by the plurality of cameras 120. In this example of a televised football game, each football player may utilize a global positioning satellite unit in their helmet as one of the plurality of sensors 140 to provide the performance instrumentation data in the form of the position related to the football player. In another embodiment, one of the plurality of sensors 140 may include force sensor within each helmet to provide the performance instrumentation data in the form of the force exerted on the football player. These specific examples of the plurality of sensors 140 are shown for exemplary purposes only. Any type of sensor used to measure a physical aspect of the event at the event site 110 may be utilized.
An audio/visual equipment module 130 is configured to process the audio visual signals. In one embodiment, the audio/visual equipment module 130 is configured to receive the audio/visual signals from the plurality of cameras 120.
A data acquisition module 150 is configured to process instrumentation data. In one embodiment, the data acquisition module 150 is configured to receive the camera instrumentation data from the plurality of cameras 120 and the performance instrumentation data from the plurality of sensors 140. Thus, the performance data collected in the data acquisition module 150 includes both the camera instrumentation data which relates to particular parameters associated with the plurality of cameras 120 while recording the event and the performance instrumentation data which relates to data captured by the plurality of sensors 140 which describes aspects of the event.
The multiplex and modulate module 160 is configured to receive the audio visual signals from the audio visual equipment module 130 and the instrumentation data from the data acquisition module 150. In one embodiment, the module 160 is configured to multiplex and modulate the audio visual signals with the instrumentation data into a unified signal relative to time. A transmitter module 170 is configured to receive the unified signal from the multiplex and modulate module 160 and to transmit this unified signal. A television 180 a shown as an exemplary device to receive the unified signal via the transmitter module 170.
With reference to
The instrumentation data corresponds with the audio and video signals in real time; the instrumentation data and the audio and video signals are temporally correlated. In one embodiment, they are temporally correlated by the use of timestamps. In another embodiment, they may be temporally correlated by relative signal timing.
In one embodiment, the system 200 includes an audio/visual (A/V) source 210, an MPEG-2 encoder 212, a data injector 214, a real-time data streamer 216, a carousel streamer 218, a trigger generator 220, an A/V and data transport stream 222, a modulator 224, a transmitter 226, a tuner 228, a demultiplexer 230, an MPEG-2 decoder 232, a presentation engine 234, a broadcast data handler 236, and an application module 238. Additional specific elements common in computer system such as processors, memory, user interfaces, system busses, storage devices, and the like are not shown to prevent unnecessarily obscuring the aspects of the invention.
The components 210-238 are merely illustrated in
The A/V source 210 is connected to the MPEG-2 encoder 212 and provides the MPEG-2 encoder with A/V content. In one embodiment, the A/V source 210 includes a video camera. However, in another embodiment, the A/V source 210 may also include a video cassette recorder, a digital recorder, or other means for providing A/V content. The MPEG-2 encoder 212 receives the A/V content and encodes this content to form an encoded A/V data stream according the MPEG-2 standard which is well known in the art. In other embodiments, other A/V encoders such as MPEG-1 or MPEG-4 may be utilized.
The MPEG-2 encoder 212, the real-time data streamer 216, the carousel streamer 218 and the trigger generator 220 are connected to the data injector 214. The real-time data streamer 216 provides the data injector 214 with instrumentation data which describes and corresponds in real-time with the A/V content from the A/V source 110. Instrumentation data describes in real-time physical aspects or conditions that correspond with the AN content.
The carousel streamer 218 provides the data injector 214 with assets (e.g., images, audio clips, text files) related to the user interface. The trigger generator 220 provides the data injector 214 with data used to activated predefined actions on the receiver (e.g., authored questions for a trivia game or poll, advertisement names for pop-up ad inserts).
The data injector 214 receives incoming data from the MPEG-2 encoder 212, the real-time data streamer 216, the carousel streamer 218, and the trigger generator 220. The data injector 214 synchronizes the incoming data such that the data from the real-time data streamer 216, carousel streamer 218, and trigger generator 220 are timed with the corresponding encoded A/V data stream. The data injector 214 is connected to the A/V and data transport stream 222 and feeds the synchronized data through the A/V and data transport stream 222 to the modulator 224.
The modulator 224 receives the synchronized data. The synchronized data includes the encoded A/V data stream and associated instrumentation data from the real-time data streamer 216, carousel streamer 218, and trigger generator 220. The modulator 224 broadcasts this synchronized data through the transmitter 226. The transmitter 226 may broadcast through air, cable, phone lines, and the like.
The tuner 228 receives the synchronized data which is broadcast through the transmitter 226. The demultiplexer 230 is connected to the tuner 228 and receives the synchronized data from the tuner 228. The demultiplexer 230 separates the encoded A/V data stream from other data originally from the real-time data streamer 216, carousel streamer 218, and trigger generator 220. The MPEG-2 decoder 232 is connected to the demultiplexer 230 and receives the encoded A/V data stream from the demultiplexer 230. The broadcast data handler 236 is connected to the demultiplexer. The data from the real-time data streamer 216, carousel streamer 218, and trigger generator 220, is received by the broadcast data handler 236 from the demultiplexer 230.
The MPEG-2 decoder processes the encoded A/V data stream and returns a decoded A/V data stream which is either identical or nearly identical to the original A/V data stream from the A/V source 210. Similar to the MPEG-2 encoder 212, the MPEG-2 decoder 232 may be substituted with other A/V encoders such as MPEG-1 or MPEG-4. The MPEG-2 decoder 232 is connected with the presentation engine 234. The presentation engine 234 receives the decoded A/V data stream from the MPEG-2 decoder 232.
The broadcast data handler 236 is connected to the application module 138. The broadcast data handler 236 reformats the data from the transport stream into data that the application module 238 can utilize. The data from the real-time data streamer 216, carousel streamer 218, and trigger generator 220 is received by the application module 238. The application module 238 utilizes the data from the real-time data streamer 216, carousel streamer 218, and trigger generator 220. The application module 238 also interacts with the presentation engine 234.
With reference to
The components 310-340 are merely illustrated in
In one embodiment, the sensor 310 and the camera 320 are configured on the broadcast side and the rendering module 330 and the compositing module 340 are configured to be placed on the receiver side. However, in other embodiments, the rendering module 330 and the compositing module 340 are configured to be placed on the broadcast side.
In one embodiment, the camera 320 is configured to capture both image data 360 and camera instrumentation data 365. The image data 360 is sent the compositing module 340. The camera instrumentation data 365 is sent to the rendering module 330. The camera instrumentation data 365 may include field-of-view data, camera position data, zoom data, and pan data of the event being captured by the camera 320. There may also be multiple cameras within the system 300 wherein each camera is uniquely identified.
The sensor 310 is configured to capture performance instrumentation data 370 for use by the rendering module 330. In one embodiment, an auto racing event is utilized to demonstrate various performance instrumentation data 370 within the system 300. In other embodiments, the system 300 may be applied to other events. For example, the performance instrumentation data 370 may include car speed, car engine performance parameters, forces exerted onto the car, car position, and the like. Multiple sensors may be utilized within the system 300.
The segment identifier module 315 receives the camera instrumentation data 365 and the performance instrumentation data 370. The segment identifier module 315 is configured to identify a visual segment through the use of camera instrumentation data and performance instrumentation data. The visual segment refers to a part of the base image which is captured by the camera 320. In one embodiment, the visual segment is the portion of the base image which gets superimposed or keyed in. In other words, the visual segment identifies the pixels which will be affected by the graphics insertion. Specific examples of graphics insertions are shown in the following figures.
The visual segment can be identified by coordinates in the base image which can be obtained from the camera instrumentation data and the performance instrumentation data. In one embodiment, a three-dimensional model of the base image is utilized in conjunction with the dimensions and coordinates of a surface used to display to correlate the camera instrumentation data with the size and coordinates to insert the graphic.
In another embodiment, the camera instrumentation data is utilized to identify the corner points or contours of the area bounding the visual segment in two dimensional space. In doing so, additional signals may be identified within the area bounding the visual segment. In another embodiment, the insertion surface may be more complex than a simple two dimensional surface. In this case, a more complete three dimensional model may be utilized to accurately define the corresponding visual segment.
The camera instrumentation data allows the definition of the boundaries of the visual segment where the projections of graphics occurs. In the case of a visual segment defined by a moving surface, such as a helmet of a football player, performance instrumentation data tracking the moving helmet and camera instrumentation data tracking the moving helmet relative to the viewpoint perspective are utilized. In this specific example, the visual segment moves with the corresponding football player.
Further, even in a specific example of a stationary object being the visual segment, such as the football field, the change in focus or zoom of the camera may change the coordinates of the visual segment. The coordinates of the visual segment may be constantly updated and recalculated based on the instrumentation data.
The color histogram module 325 is configured to identify the range of colors within the visual segment. The color histogram module 325 is also configured to calculate the amount and frequency of the colors contained within the visual segment. In one embodiment, the color histogram module 325 utilizes a specialized graph or plot to represent the number of colors and the amount and frequency of these colors within the visual segment.
In one embodiment, the color histogram module 325 also identifies a new key color value 375 and transmits this value 375 to the segment identifier module 315. In one embodiment, the new key color value 375 is chosen based on the nearest peak or nearest summit on the graph representing the various colors within the visual segment. In another embodiment, the new key color value 375 is related to color distribution where parameters such as sensitivity determine how closely the colors must match the key color to be considered part of the color key.
In another embodiment, various alternate systems besides the color histogram module 325 may be utilized to identify the key color value. For example, a variety of statistical algorithms and system may be utilized to identify the key color value.
In one embodiment, the rendering module 330 receives the instrumentation data, the key color value, and the visual segment information. In one embodiment, the rendering module 330 generates an overlay image based on the visual segment information. In another embodiment, the rendering module 330 generates an overlay image based on the key color value. In yet another embodiment, the rendering module 330 generates an overlay image based on the instrumentation data.
In one embodiment, the rendering module 330 is configured to generate an overlay image that incorporates the insertion of graphics to be displayed in conjunction with the base image. The overlay image is rendered by the rendering module 330 in response to the visual segment, the key color value, and/or the instrumentation data.
The compositing module 340 receives the overlay image from the rendering module 330 and the image data 360 from the camera 320. In one embodiment, the compositing module 340 integrates the image data 360 within the overlay image. In other words, the compositing module 340 blends the image data 360 within the overlay image to create a single, combined overlay image wherein the combined overlay image includes the overlay image from the rendering module 330 combined with the image data 360 which depicts a real event captured by the camera 320.
For the sake of clarity, the embodiment shown in the system 300 is illustrated utilizing the overlay image created by the rendering module 330 and image data representing a single base image captured by the camera 320. In another embodiment, multiple overlay images and image data representing multiple base images may be utilized to create a stream of images representing a video stream. Further, this stream of images both overlay and base may be combined by the compositing module 340.
A televised football game has been utilized as an example within various embodiments of the invention. However, any type of live event is suitable as application for use with the invention. In a televised football game, the static portions of the world model include the football field and surrounding stadium. The dynamic objects include the football players and the ball. If the instrumentation data includes tracking the position of the football players, then the football player positions may be tracked using a technique such as inverse kinematics in one embodiment. If the instrumentation data includes tracking the particular motions of the football players, then the football player motions may be tracked using a technique such as joint position and/or orientation in one embodiment.
The flow diagrams as depicted in
The blocks within the flow diagram may be performed in a different sequence without departing from the spirit of the invention. Further, blocks may be deleted, added or combined without departing from the spirit of the invention.
In Block 400, a real event is captured by a camera and/or sensor. A series of base images are captured by the camera and a corresponding series of instrumentation data are captured by the camera and/or sensor.
In Block 410, the chroma keying parameters are selected by a user. Various chroma keying parameters may be selected such as a key color value, spill removal, softness, garbage mate, key shrink, glossing, shadows, flare suppression, and the like.
In Block 420, the key color value is selected. In one embodiment, the selection of the color key value may be manually performed by the user. In another embodiment, the selection of the color key value may be automated. The color key value may be defined over a finite range of color values.
For example, a weatherman on television typically appears in front of a monochromatic background. During a televised production, a background scene is overlayed onto the monochromatic background while still showing the weatherman in front of the background scene if the key color value is properly matched and has an appropriate range relative to the monochromatic background. The background scene is often a video stream showing various map regions. However, if the key color value has a range that is too broad, portions of the weatherman are erroneously covered with the background scene. Additionally, if the key color value has a range that is too narrow, there will be holes in the background scene which will display the monochromatic background. Further, if the key color value is centered poorly, then both of these effects will occur.
In Block 430, the visual segment is identified according to the segment identifier module 315 (
In Block 440, a color histogram is calculated for the visual segement according to the color histogram module 325 (
In Block 450, a new key color value is calculated from the color histogram. The new key color value updates the key color value according to the visual segment. The visual segment may be dynamic. For example, as the visual segment changes in color due to movement, shadows, varying environment, and the like. The new key color value is transmitted to the Block 430 thereby updating the key color value from the Block 420.
In Block 460, the rendering module 330 generates an overlay image. Multiple overlay images represents a video data stream. The overlay image is configured to overlay the visual segment within the base image. In one embodiment, the overlay image may include a graphics image, a captured image, and the like. The overlay image may take the form of an advertisement, a scoreboard, a logo, and the like.
In Block 470, the compositing module 340 (
The Blocks 400-470 are performed within the context of the segment identifier module 315, the color histogram module 325, the rendering module 330 and the compositing module 340 for exemplary purposes only. In other embodiments, the Blocks 400-470 may be performed in any generalized processor or any graphics specific processor.
In Block 510, the depth map information may be utilized to pre-segment a portion of the visual segment. For example, the depth map information is utilized to identify an object that is located in front of an area in the visual segment where the overlay image is inserted. The portion of the visual segment that is blocked by the object may be excluded from the visual segment and from the color histogram calculations.
In Block 520, the pre-segment information, the instrumentation data, and the key color value are utilized to identify boundaries of the visual segment.
For the sake of clarity,
The screen shot 650 incorporates the base scene of the screen shot 600 and an overlay image. The screen shot 650 includes a football field 660, a plurality of football players 670, and a plurality of graphics 680. The plurality of graphics 680 represent a team logo and are part of the overlay image which is integrated with the base scene image.
In one embodiment, the system 300 (
The screen shot 750 incorporates the base scene of the screen shot 700 and an overlay image. The screen shot 750 includes a football field 760, a plurality of football players 770, and a plurality of graphics 780. The plurality of graphics 780 represent a team logo and are part of the overlay image which is integrated with the base scene image. A group of football players 775 are positioned in front of the plurality of graphics 780. Because of the position of the group of football players 775 relative to the plurality of graphics 780, the visual segment excludes the area occupied by the group of football players 775. Accordingly, the plurality of graphics 780 are rendered behind the group of football players 775 to prevent obstructing their view by a user.
In one embodiment, the system 300 (
The foregoing descriptions of specific embodiments of the invention have been presented for purposes of illustration and description. For example, the invention is described within the context of auto racing and football as merely embodiments of the invention. The invention may be applied to a variety of other theatrical, musical, game show, reality show, and sports productions.
They are not intended to be exhaustive or to limit the invention to the precise embodiments disclosed, and naturally many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to explain the principles of the invention and its practical application, to thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the claims appended hereto and their equivalents.
This present application is a Divisional of application Ser. No. 11/439,304, filed May 22, 2006 entitled “System and Method for Data Assisted Chroma-Keying”, which is a continuation of U.S. patent application Ser. No. 10/215,520, filed Aug. 8, 2002 entitled “A System and Method for Data Assisted Chroma-Keying”, which claims priority from Provisional Application entitled “Method and Apparatus for Mixed Reality Broadcast” filed on Aug. 10, 2011, with Ser. No. 60/311,477.
Number | Name | Date | Kind |
---|---|---|---|
4202008 | King | May 1980 | A |
4716458 | Heitzman et al. | Dec 1987 | A |
4970666 | Welsh et al. | Nov 1990 | A |
4987552 | Nakamura | Jan 1991 | A |
5398075 | Freytag et al. | Mar 1995 | A |
5457370 | Edwards | Oct 1995 | A |
5577188 | Zhu | Nov 1996 | A |
5600368 | Matthews et al. | Feb 1997 | A |
5610653 | Abecassis | Mar 1997 | A |
5673401 | Volk et al. | Sep 1997 | A |
5689442 | Swanson et al. | Nov 1997 | A |
5714997 | Anderson | Feb 1998 | A |
5729471 | Jain et al. | Mar 1998 | A |
5742521 | Ellenby et al. | Apr 1998 | A |
5745126 | Jain et al. | Apr 1998 | A |
5796991 | Shimizu | Aug 1998 | A |
5823786 | Easterbrook | Oct 1998 | A |
5835667 | Wactlar et al. | Nov 1998 | A |
5838310 | Uya | Nov 1998 | A |
5850232 | Engstrom et al. | Dec 1998 | A |
5860862 | Junkin | Jan 1999 | A |
5865624 | Hayashigawa | Feb 1999 | A |
5878174 | Stewart et al. | Mar 1999 | A |
5900868 | Duhault et al. | May 1999 | A |
5953076 | Astle et al. | Sep 1999 | A |
5966132 | Kakizawa et al. | Oct 1999 | A |
5986718 | Barwacz et al. | Nov 1999 | A |
5995941 | Maquire et al. | Nov 1999 | A |
6002982 | Fry | Dec 1999 | A |
6031545 | Ellenby et al. | Feb 2000 | A |
6044397 | Eleftheriadis et al. | Mar 2000 | A |
6061056 | Menard et al. | May 2000 | A |
6067653 | Tsukagoshi | May 2000 | A |
6072504 | Segen | Jun 2000 | A |
6080063 | Khosla | Jun 2000 | A |
6133962 | Proehl et al. | Oct 2000 | A |
6144375 | Jain et al. | Nov 2000 | A |
6147716 | Ohki | Nov 2000 | A |
6148280 | Kramer | Nov 2000 | A |
6151009 | Kanade et al. | Nov 2000 | A |
6159016 | Lubell et al. | Dec 2000 | A |
6178007 | Harrington | Jan 2001 | B1 |
6193610 | Junkin | Feb 2001 | B1 |
6219011 | Aloni et al. | Apr 2001 | B1 |
6229550 | Gloudemans et al. | May 2001 | B1 |
6236395 | Sezan et al. | May 2001 | B1 |
6266100 | Gloudemans et al. | Jul 2001 | B1 |
6282317 | Luo et al. | Aug 2001 | B1 |
6295115 | Zhang et al. | Sep 2001 | B1 |
6330486 | Padula | Dec 2001 | B1 |
6353461 | Shore et al. | Mar 2002 | B1 |
6359585 | Bechman et al. | Mar 2002 | B1 |
6366316 | Parulski et al. | Apr 2002 | B1 |
6378132 | Grandin et al. | Apr 2002 | B1 |
6380933 | Sharir et al. | Apr 2002 | B1 |
6384871 | Wilf et al. | May 2002 | B1 |
6408257 | Harrington et al. | Jun 2002 | B1 |
6414696 | Ellenby et al. | Jul 2002 | B1 |
6445815 | Sato | Sep 2002 | B1 |
6449540 | Rayner | Sep 2002 | B1 |
6466250 | Hein et al. | Oct 2002 | B1 |
6466275 | Honey et al. | Oct 2002 | B1 |
6483511 | Snyder | Nov 2002 | B1 |
6483523 | Feng | Nov 2002 | B1 |
6525780 | Bruno et al. | Feb 2003 | B1 |
6535114 | Suzuki et al. | Mar 2003 | B1 |
6535226 | Sorokin et al. | Mar 2003 | B1 |
6537076 | McNitt et al. | Mar 2003 | B2 |
6545705 | Sigel et al. | Apr 2003 | B1 |
6571012 | Pettigrew | May 2003 | B1 |
6571054 | Tonomura et al. | May 2003 | B1 |
6571193 | Unuma et al. | May 2003 | B1 |
6593936 | Huang et al. | Jul 2003 | B1 |
6597406 | Gloudemans et al. | Jul 2003 | B2 |
6642939 | Vallone et al. | Nov 2003 | B1 |
6674461 | Klapman | Jan 2004 | B1 |
6681395 | Nishi | Jan 2004 | B1 |
6697072 | Russell et al. | Feb 2004 | B2 |
6707456 | Marrin et al. | Mar 2004 | B1 |
6710822 | Walker et al. | Mar 2004 | B1 |
6720990 | Walker et al. | Apr 2004 | B1 |
6741241 | Jaubert et al. | May 2004 | B1 |
6750919 | Rosser | Jun 2004 | B1 |
6760916 | Holtz et al. | Jul 2004 | B2 |
6771272 | Deering | Aug 2004 | B2 |
6778085 | Faulkner et al. | Aug 2004 | B2 |
6791574 | Hoch et al. | Sep 2004 | B2 |
6792321 | Sepe, Jr. | Sep 2004 | B2 |
6799180 | McGrath et al. | Sep 2004 | B1 |
6810397 | Qian et al. | Oct 2004 | B1 |
6825875 | Strub et al. | Nov 2004 | B1 |
6833865 | Fuller et al. | Dec 2004 | B1 |
6860806 | Kojima et al. | Mar 2005 | B2 |
6868440 | Gupta et al. | Mar 2005 | B1 |
6882793 | Fu et al. | Apr 2005 | B1 |
6961954 | Maybury et al. | Nov 2005 | B1 |
6987535 | Matsugu et al. | Jan 2006 | B1 |
7000245 | Pierre et al. | Feb 2006 | B1 |
7016540 | Gong et al. | Mar 2006 | B1 |
7065250 | Lennon | Jun 2006 | B1 |
7120586 | Loui et al. | Oct 2006 | B2 |
7184959 | Gibbon et al. | Feb 2007 | B2 |
7313808 | Gupta et al. | Dec 2007 | B1 |
7369130 | Kawabe et al. | May 2008 | B2 |
7673321 | Yurt et al. | Mar 2010 | B2 |
20010003715 | Jutzi et al. | Jun 2001 | A1 |
20010005218 | Gloudemans et al. | Jun 2001 | A1 |
20010012324 | Normile | Aug 2001 | A1 |
20010014175 | Tavor | Aug 2001 | A1 |
20010023202 | Okubo | Sep 2001 | A1 |
20020010928 | Sahota | Jan 2002 | A1 |
20020016971 | Berezowski et al. | Feb 2002 | A1 |
20020024450 | Townsend et al. | Feb 2002 | A1 |
20020027617 | Jeffers et al. | Mar 2002 | A1 |
20020064764 | Fishman et al. | May 2002 | A1 |
20020069265 | Bountour et al. | Jun 2002 | A1 |
20020115047 | McNitt et al. | Aug 2002 | A1 |
20020152462 | Hoch et al. | Oct 2002 | A1 |
20020170068 | Rafey et al. | Nov 2002 | A1 |
20020178450 | Morita et al. | Nov 2002 | A1 |
20020188943 | Freeman et al. | Dec 2002 | A1 |
20030032484 | Ohshima et al. | Feb 2003 | A1 |
20030033318 | Carlbom et al. | Feb 2003 | A1 |
20030038892 | Wang et al. | Feb 2003 | A1 |
20030169335 | Monroe | Sep 2003 | A1 |
20040107439 | Hassell et al. | Jun 2004 | A1 |
20040170321 | Gong et al. | Sep 2004 | A1 |
20050028194 | Elenbaas et al. | Feb 2005 | A1 |
Number | Date | Country |
---|---|---|
WO 0067867 | Nov 2000 | WO |
Entry |
---|
Rick Cavallaro, “The Fox Trax Hockey Puck Tracking System”, Mar.-Apr. 1997, pp. 6-12, IEEE. |
Audio-Visual: Definition from Answers.com, http:www/answers.com/topic/audio-visuals, Nov. 25, 2008, three pages. |
Number | Date | Country | |
---|---|---|---|
20120057750 A1 | Mar 2012 | US |
Number | Date | Country | |
---|---|---|---|
60311477 | Aug 2001 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11439304 | May 2006 | US |
Child | 13225183 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10215520 | Aug 2002 | US |
Child | 11439304 | US |