System and method for the creation, synchronization and delivery of alternate content

Abstract
According to one embodiment, A nonlinear editing system comprises a video encoder, an audio compressor and an engine unit. The engine unit is adapted to synchronize the video with the first alternate video and to synchronize the audio with the first alternate audio so that, where the scene of the program includes objectionable content, either (i) the video is substituted with the first alternate video, (ii) the audio is substituted with the first alternate audio, or (iii) both the video and the audio are substituted with the first alternate video and the first alternate audio.
Description
BACKGROUND

1. Field


Embodiments of the invention relate to the field of digital content transmissions and display. More specifically, one embodiment of the invention relates to a system and method for creation, synchronization and delivery of alternate content for dynamic substitution of unwanted video or audio on a scene-by-scene basis.


2. General Background


Television is used to deliver content, such as entertainment and educational programs, to viewers. A growing number of parents are now watching and listening to content perceived by their children, in an effort to better shield their young children from exposure to objectionable content, namely violence, sexual situations, indecent or suggestive language and the like.


In television programming, for example, the removal/restoration of potentially objectionable content is implemented during content authoring/editing using a nonlinear editing system.


As illustrated in FIG. 1, a conventional nonlinear editing system 100 is used to select desired portions of audio and video sequences (scenes) taken from one or more libraries 110, which contain all of the raw content footage (e.g., video, audio, etc.) under consideration, and then to establish their temporal relationships. The temporal relationships involve the arrangement of adjacent sequences of the same content type.


For instance, as shown in FIG. 2, video sequences 202, 204, 206 and 208, namely uncompressed digital video 200, are arranged in a temporal relationship to each other. Audio sequences 212, 214, 216 and 218, namely audio 210, are similarly arranged with respect to each other. Thereafter, each video sequence 202, 204, 206 and 208 is synchronized with one or more audio sequences which may represent different forms of audio such as dialog tracks 212 and 214, music track 216, sound effects 218, or the like. Normally, the audio is in the form of Pulse Code Modulation (PCM) audio samples. Even though the end product consists of one continuous video sequence with a single synchronized audio track (containing a composition of multiple forms of audio accomplished by a digital mixdown unit 120 of FIG. 1), all digital components 200 and 210 comprising the content remain distinct and separate while in nonlinear editing system 100.


Referring back to FIG. 1, after the final edited version of the content is completed, it is then assembled into a compilation including the selected video and audio sequences. This involves the selected video and audio sequences being compressed, perhaps using MPEG-2 video encoder 130 and AC-3 audio compressor 140 for video and audio respectively, and synchronized with each other to produce a continuous transport stream 150 of a reduced size. This size reduction is what makes the transmission, broadcast and/or storage of digital video content practical, allowing it to fit on inexpensive media like DVDs or to be carried in standard six megahertz (6 MHz) Radio Frequency (RF) spectral channel concurrent with eight or more other A/V services.


After compression and placement into transport stream 150, the video and audio cannot be edited. Transport stream 150 features video running at a constant frame per second rate and compressed audio synchronized with the video. The audio is continuous as well, where silence is represented by specific data values. Conventional nonlinear editing system 100 offers no mechanism to produce any transport streams with alternate content that can be substituted for selected video and audio.


At the destination, with respect to the rendering of video programs, televisions have been implemented with technologies that allow parents to preclude the playback of programs that exceed a selected age-based rating. A black screen is shown for any program that exceeds the selected age-based rating in order to prevent such viewing. This program blocking can be disabled with a parental password.


However, there are a number of major shortcomings of these technologies. For instance, the filtering only occurs at the program level, namely a viewer gets to view the entire program or nothing at all. In other words, there is no selective blocking capability, and clearly no ability to substitute alternate content provided with the transport stream.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the invention are illustrated by way of example and not by way of limitation in the figures of the accompanying drawings, in which like references indicate similar elements and in which:



FIG. 1 is a block diagram of the operations of a conventional nonlinear editing system.



FIG. 2 is a block diagram of a display of content processed by the conventional nonlinear editing system.



FIG. 3 is a block diagram of a display of content processed by a dynamic nonlinear editing system.



FIG. 4 is a block diagram of the operations of the dynamic nonlinear editing system developed in accordance with one embodiment of the invention.



FIG. 5 is an exemplary block diagram of a content delivery network consistent with certain embodiments of the invention.



FIG. 6 is an exemplary block diagram of a content delivery back-end consistent with certain embodiments of the invention.



FIG. 7 is an exemplary diagram of a content filter unit implemented within a digital device of the content delivery system of FIG. 5.



FIGS. 8A and 8B are exemplary diagrams of an Adaptation field of an I-frame, being a portion of a MPEG packet of the digital data stream.



FIG. 9 is an exemplary diagram of a first embodiment of a content replacement process consistent with certain embodiments of the invention.



FIG. 10 is an exemplary diagram of a second embodiment of a content replacement process consistent with certain embodiments of the invention.



FIG. 11 is an illustrative flowchart of a process for determining whether a segment of content associated with a program should be prevented from being displayed.



FIG. 12 is an illustrative flowchart of a process for substituting replacement segments of content based on parental control parameters.



FIG. 13 is an exemplary embodiment of a Age-based Rating scheme for modification or substitution of scenes including objectionable content.



FIG. 14 is an exemplary embodiment of a Content-based Rating scheme for modification or substitution of scenes including objectionable content.



FIGS. 15A and 15B are exemplary embodiments of a Tag Rating scheme for modification or substitution of scenes including objectionable content.





DETAILED DESCRIPTION

Various embodiments of the invention relate to a system and method for creation, synchronization and delivery of alternate content for dynamic substitution of unwanted video or audio on a scene-by-scene basis. According to one embodiment of the invention, a nonlinear editing system supporting dynamic content replacement (hereinafter referred to as “dynamic nonlinear editing system”) is implemented within the content delivery network. At a digital device, a content filter unit is implemented within the digital device and is configured to provide parental control by performing scene-by-scene evaluation of content associated with the incoming transport stream (e.g., video, images, audio or any combination thereof) and subsequent modification or elimination of objectionable content. Of course, the content filter unit may be implemented at the front-end in lieu of at the back-end that is described below in detail.


In short, the combination of the dynamic nonlinear editing system and the content filter unit provides scene-by-scene filtration, and even to the granularity of frame-by-frame, to block or replace individual scenes or words. Such blocking may be accomplished by the content filter unit residing with the customer through screen blocking or obscuring pixels for a particular image or muting audio. Alternatively, content within the scene can be replaced in order to reduce the rating of the program (referred to as “down-rating”), or may even be replaced to reinsert content that was previously obscured/removed in order for the programming to achieve the higher rating (referred to as “up-rating”).


According to one embodiment of the invention, the replacement content is transmitted in the same digital transport stream as the primary content, and is synchronized with the primary content that it is replacing for swapping on-the-fly. Thus, the rendering of higher or lower rated content is accomplished, where programs can be created and broadcast with multiple versions having different content and ratings. Also, all content, including advertisements, news, sporting events and the like may be filtered as well in lieu of just movies and television shows.


For purposes of the present description, the term “digital device” may refer to consumer electronics that is adapted to tune, receive, decrypt, descramble and/or decode transmissions from any content provider. The “consumer electronics” may include, but is not limited or restricted to a television, a set-top box, a personal digital assistant (PDA), a computer, a cellular telephone, a video game console, a portable music player, a stereo system, a personal video recorder, or the like. Examples of “content providers” may include a terrestrial broadcaster, a cable or satellite television distribution system, or a company providing content for download over the Internet or other Internet Protocol (IP) based networks like an Internet service provider.


In the following description, certain terminology is used to describe features of the invention. For example, in certain situations, the terms “component,” “unit” and “logic” are representative of hardware and/or software configured to perform one or more functions. For instance, examples of “hardware” include, but are not limited or restricted to an integrated circuit such as a processor (e.g., a digital signal processor, microprocessor, application specific integrated circuit, a micro-controller, etc.). Of course, the hardware may be alternatively implemented as a finite state machine or even combinatorial logic.


An example of “software” includes executable code in the form of an application, an applet, a routine or even a series of instructions. The software may be stored in any type of machine readable medium such as a programmable electronic circuit, a semiconductor memory device such as volatile memory (e.g., random access memory, etc.) and/or non-volatile memory (e.g., any type of read-only memory “ROM”, flash memory, etc.), a floppy diskette, an optical disk (e.g., compact disk or digital video disc “DVD”), a hard drive disk, a tape, or the like.


In addition, the term “program” generally represents a stream of digital content that is configured for transmission to one or more digital devices for viewing and/or listening. For instance, the program may be configured as a transport stream. According to one embodiment, the program may contain multiple identifiers such as Packet Identifiers (PIDs) when the program is MPEG (Moving Pictures Expert Group) compliant compressed video. These multiple PIDs provide for replacement of one scene of content for another. The “scene” is defined as one or more frames of content, namely video, images, audio or any combination thereof. For instance, a scene may be one or more frames of video and/or audio, or even a packet of music downloaded into a portable music player.


More specifically, multiple PIDs may be associated with a single program in order to provide content replacement for this program. Content replacement provides an enhanced level of viewing/listening customization. For example, content replacement can be used to provide targeted advertising to an audience by the substitution of one advertisement determined to be “objectionable” for another. Content replacement can also be used to provide multiple scenes, endings or other segments for a program, or perhaps to provide multiple views in a sportscast or other program. Other applications for the content replacement of the invention fall within the spirit and scope of the invention.


Certain types of content, such as movies, have an associated content rating established by a supervisory body to inform potential viewers of the type of material that may form a part of the content. General ratings of G, PG, PG-13, R and NC-17 are commonly established for most movies produced in the United States by the Classification and Ratings Administration (CARA) of the Motion Picture Association (MPA) or the Motion Picture Association of America (MPAA). Television programming is similarly rated by the TV Parental Guidelines Monitoring Board, which provides ratings of TV-Y, TV-Y7, TV-Y7-FV, TV-G, TV-14 and TV-MA.


For purposes of this document, content having a rating provided under such rating systems will be referred to as having a “higher” rating if it is intended for more mature audiences, and a “lower” rating if it is intended for a more general or less mature audience. Hence, the previously described rating systems are listed from lowest to highest. Therefore, within this terminology, an R rating would be considered a “higher” rating than a G rating. Such nomenclature is defined herein for simplicity of description of the above rating system as well as equivalent systems, without intent for the nomenclature to pass any judgment on the quality of such content.


Additionally, the term “rating” is intended to embrace any suitable content rating system, domestic or foreign, public, private or even governmental that is formed to serve a similar function. Such rating information may be provided as data embedded within program specific information such as data inserted into an Adaptation field of a MPEG packet for example when the content is video.


While this invention is susceptible of embodiment in many different forms, there is shown in the drawings and will herein be described in detail specific embodiments, with the understanding that the present disclosure is to be considered as an example of the principles of the invention and not intended to limit the invention to the specific embodiments shown and described.


I. General Architecture of Content Delivery Front-End


In order to support dynamic content substitution, the authoring process associated with the dynamic nonlinear editing system is modified to allow concurrent editing of one or more alternate video tracks and additional audio tracks. Each of the alternate video tracks is subordinate to the primary video track and is temporally synchronized with it. After post-processing, these video tracks will be carried in the same transport stream, namely as part of the same program. The dynamic nonlinear editing system is extended to accommodate the additional tracks for alternate video and audio and is complementary to the editing paradigm established for this tool.


As described earlier, the “final cut” is assembled, mixed (audio) and streamed to compression equipment (encoders). A video encoder can only accept a single, continuous video stream. The primary video meets the criterion. The alternate video, however, is normally a non-continuous sequence of video that is transmitted concurrently with the primary video and, in some situations, used to replace the primary video or portions thereof.


Referring to FIG. 3, a block diagram of a display of content processed by a dynamic nonlinear editing system being part of a content delivery front-end is shown. The dynamic nonlinear editing system 300 is modified to support synchronization and delivery of replacement content, such as alternate video 315 and/or audio tracks 360 for example.


As shown, alternate video 320 is provided during a first time period 312 (from T2 and T3) of a master timeline 310, which corresponds to the display duration of video scene 330. Also, alternate video 340 is provided during a second time period 314 (from T6 and T7) of master timeline 310, corresponding to the duration of video scene 350. The duration of first and second time periods 312 and 314 may constitute the display time for one or more frames of video (e.g. 15 frames of video constituting one-half second of display time, or perhaps more frames).


As further shown, alternate audio 360 includes multiple audio segments 362-366. A first alternate audio segment 362 occurs at the start of time T0 and ends at time T0A while a second alternate audio segment 363 occurs at time T2A and ends at time T3. Between time periods T4 and T5, a third and fourth alternate audio segments 364-365 are placed, namely between T4A-T4B and T4C-T4D. Similarly, a fifth alternate audio segment 366 occurs at time T6A and ends at time T6B. For this embodiment, first, fourth and fifth alternate audio segments 362, 365 and 366 constitute alternate dialog. Second audio segment 363 constitutes an alternate music track and third audio segment 364 constitutes an alternate special effects track.


Referring to FIG. 4, a block diagram of the operations of dynamic nonlinear editing system 300 developed in accordance with one embodiment of the invention is shown. Dynamic nonlinear editing system 300 receives information from content storage units 410 and 412, which are configured to store video and audio content. Dynamic nonlinear editing system 300 is adapted to sequence video and audio retrieved from content storage units 410 and 412, and then to establish their temporal relationships. The temporal relationships involve the arrangement of adjacent sequences of the same content type.


In order to remain compatible with video encoder 420, dynamic nonlinear editing system 300 fills the time periods before, after and between alternate video 320 and 340 with synchronized black screen in order to create a continuous video stream for processing by video encoder 420. It is contemplated that the encoding of the primary and alternate video will occur on the same encoder, and therefore, the two processes will occur serially. As a result, dynamic nonlinear editing system 300 will need to communicate with video encoder 420 so that synchronization information may be inserted by video encoder 420 into both resultant compressed video streams 430 and 440 for post-encoding reprocessing where video streams 430 and 440 are combined with proper synchronization.


Similarly, dynamic nonlinear editing system 300 fills the time periods before, after and between alternate audio segments 362-366 with muted audio in order to create a continuous audio stream, for processing by audio compressor 450. It is contemplated that the encoding of the primary and alternate audio may occur serially on the same audio compressor 450, or on multiple audio compressors. When using the same audio compressor 450 as shown, dynamic nonlinear editing system 300 will need to communicate with audio compressor 450 so that synchronization information may be inserted by audio compressor 450 into both resultant compressed audio streams 460 and 470 for post-encoding reprocessing.


A transport processing engine unit 480 receives video streams 430 and 440 and audio streams 460 and 470. Transport processing engine unit 480 conducts a trimming operation of alternate content as well as performs content synchronization, content re-multiplexing and meta data insertion. These operations are complementary with the functionality of device 510 of FIG. 5, and perhaps substantially similar where content replacement is conducted in the front end.


First, the alternate content contains blank video (black screen) and muted audio between alternate content segments. This is a byproduct of preparing the content for compression. Transport processing engine unit 480 trims all black screen content and muted audio purposefully added to create the continuous audio. This allows the alternate content to be multiplexed into the primary transport stream.


As an illustrative example, referring to FIG. 3, video scene 330 and alternate video 320 need to be processed to ensure that these video segments are temporally equivalent within the transport stream. In other words, these video segments need to be virtually identical in duration and aligned temporally where such editing occurs after being encoded/compressed and during placement within the transport stream. The same temporal alignment is required for video scene 350 and its alternate video 340.


Secondly, transport processing engine unit 480 uses synchronization marks inserted by video encoder 420 and audio compressor 450 to ensure that the digital device will display either the primary or alternate video at the same time. As an example, for MPEG transmissions, three synchronization elements are used for video: PCR (primary clock reference)—clocking at approximately 27 megahertz (MHz); DTS (decoding time stamp)—placed on video itself to start conversion from MPEG to analog video format; and PTS (presentation time stamp)—when picture should be displayed. Both primary video 330 and alternate video 320 should have the same synchronization marks.


Similar synchronization marks may be used for audio, and thus, the primary and alternative audio will have the same synchronization marks. This allows alternate content to be correctly located temporally within the transport stream so that primary/alternate video as well as the primary/secondary audio can be contextually located adjacent to each other.


Third, transport processing engine unit 480 merges the alternate content into the transport stream and inserts meta data as needed to enable suitably equipped playback devices, such as digital device as described below, to dynamically select any combination or primary/alternate content during broadcast or playback of the resultant composite content. As part of the merging process, the re-multiplexer corrects PCR synchronization marks and other tasks normally encountered and associated with the re-multiplexing process.


According to one embodiment, the meta data is inserted into an Adaptation field of one or more MPEG-based packets forming the transport stream. The meta data may provide Age-based rating information, content tag information or the like as described below.


The composite, homogeneous output 490 from transport processing engine unit 480 is then returned to the normal content process flow, where it is stored or forwarded to a distribution phase by a content provider as describe below.


II. General Architecture of Content Delivery Back-End


Referring to FIG. 5, an exemplary block diagram of a content delivery back-end 500 consistent with certain embodiments of the invention is shown. Content delivery back-end 500 comprises a digital device 510 that receives digital content such as a program from one or more content providers 520. The program may be propagated as a digital data stream for example in compliance with any data compression scheme. Examples of a data compression scheme include, but are not limited or restricted MPEG standards.


According to one embodiment of the invention, digital device 510 decodes an incoming program, and evaluates the content of the incoming program on a scene-by-scene basis to determine whether such scenes include any objectionable content. Such evaluation may involve a review of the rating assigned to each scene or perhaps access data within tags assigned to each scene to determine whether the scene includes any objectionable content. These evaluation schemes are described below.


Since digital device 510 may be adapted to increase or decrease the rating of the program, the term “objectionable content” has two meanings. One meaning is that it is content featuring certain subject matter that may be inappropriate for the viewing or listening audience at this time. Another meaning is that it is low-rated content that can be substituted for higher rated content due to the mature nature of the viewing or listening audience and their request for up-rating of content.


Content provider 520 provides the digital content to digital device 510 through transmission medium 530, which operates as a communication pathway for the program within the content delivery network. The transmission medium 530 may include, but is not limited to electrical wires, optical fiber, cable, a wireless link established by wireless signaling circuitry, or the like.


Content provider 520 may be adapted with a satellite antenna 600 to receive a multiplexed stream of content from a satellite transmission as shown in FIG. 6. The stream of content is received, demodulated and decrypted at a satellite receiver 605 before being routed to PID Mapping logic 610. The stream of content may include replacement content associated with secondary PIDs. Alternatively, the replacement content may be retrieved from a local content database 615 or other sources of content.


PID mapping logic 610 maps the incoming content from whatever source to a set of primary PIDs and a set of secondary PIDs. The incoming content includes (i) content originally intended for rendering (referred to as the “primary content”) and (ii) substitution content for various scenes (referred to as “replacement content”). The primary content is associated with the primary PIDs and the replacement content is associated with the secondary PIDs. According to this embodiment, PID mapping logic 610 may also be adapted to insert flags (or markers) into the content in order to identify the location where replacement content is to be inserted.


The mapped content then is routed to PSI/PMT insertion logic 620 that inserts Program Specific Information (PSI) and Program Map Tables (PMT) into the content for use by the decoding side in decoding the programming. If the content is to be encrypted, it may optionally pass through encryption logic 625 prior to modulation at a modulator (e.g., a QAM modulator) 630. The modulated stream of content is then transmitted via transmission medium 530 to content delivery back-end 500, which features one or more digital device(s). For instance, the modulated stream of content may be routed to via a cable plant 635 to content delivery back-end 500. The operation performed by content provider 520, including but not limited to the PID mapping for content replacement, is carried out under control of a control computer 640.


In general, a system can be used to form a content substitution encoder consistent with certain embodiments of the invention in which input data inclusive of primary content and replacement content. PID mapping logic 610 assigns a primary PID to the primary content and assigns a secondary PID to the replacement content. Private data to identify the primary content by the primary PID and replacement content by the secondary PID, along with the primary content mapped to the primary PID and the replacement content mapped to the secondary PID, are assembled into a data stream.


Of course, it is contemplated that PID mapping logic 610 may be instructed to remap the PIDs to select only the desired content based on parental control parameters routed to content provider 520 from various digital devices. Hence, if utilized, secondary PIDs may be associated with previously substituted segments of primary content.


Referring now to FIG. 7, an exemplary diagram of digital device 110 of the content delivery system 100 is shown. Digital device 110 comprises a power supply unit 700, a tuner/demodulation unit 710 and data processing unit 720. Power supply unit 700 is adapted to provide powered to tuner/demodulation unit 710 and data processing unit 720. Power supply unit 700 may provide regulated power based on input power received from a wall socket (e.g., 110 volts alternating current “VAC”, 220 VAC, etc.). Tuner/demodulation unit 710 tunes to a channel to receive the program and demodulates an incoming program to recover the content. Then, the content is routed as a digital bit stream to data processing unit 720.


According to one embodiment of the invention, data processing unit 720 comprises a processor 730 (e.g., central processing unit) and/or a decoder 740. These components may be implemented as separate integrated circuits or as a single integrated circuit. As described herein, data processing unit 720 features logic operating to decode and process incoming information as well as a content filter unit 750 that is adapted to alternatively select replacement content during playback of video and/or audio.


In general, according to one embodiment of the invention, content filter unit 750 is designed to select alternate PID(s) from the digital bit stream based on findings by data processing unit 720, which accesses meta data in Adaptation fields within MPEG packets to obtain information for use in evaluating whether a scene has objectionable content.


According to one embodiment, data processing unit 720 is adapted to process the closed captioning and emergency alert system (EAS) data as well as guide data. In addition, data processing unit 720 may be adapted to access the meta data in Adaptation fields of MPEG packets in order to provide signaling to content filter unit 750 regarding the subject matter within each scene that is scheduled to be played back in the near future. This allows content filter unit 750 to determine whether the subject matter constitutes objectionable content.


Data processing unit 720 (or perhaps a separate graphics component) is adapted for rendering on-screen menus or guide data. For instance, data processing unit 720 may be adapted to collect guide data and for storage of such guide data in local memory (not shown). When such data is requested, it is displayed and may even overlay video concurrently being displayed. Also, data processing unit 720 may be adapted to distort or blank out displayed images.


Referring to FIGS. 8A and 8B, exemplary diagrams of an Adaptation field 810 of an I-frame 800, being a portion of a MPEG packet 805 of the digital data stream is shown. Adaptation field 810 features control information 815 (also referred to as “private data”) to initiate a filtering operation is shown. According to one embodiment of the invention, Adaptation field 810 includes control information 815 used by data processing unit 720 of FIG. 7 to (i) determine if content replacement is enabled or disabled and (ii) evaluate whether replacement content should be used. Such evaluation is conducted ahead of any scheduled rendering (playback for visual or audible perception) of scenes of digital content.


Adaptation field 810 includes control information 815 that features a plurality of sub-fields, including but not limited or restricted to one or more of the following: (1) PID sub-field 820, (2) Conversion type sub-field 825; (3) Content type sub-field 830; (4) Replacement type sub-field 835; (5) Length sub-field 840; (6) Rating sub-field 845; (7) Tag sub-field 850; (8) Intensity level sub-field 855; (9) Filter blocking summary sub-field 860 and (10) Flag deletion sub-field 865.


As illustrated in FIG. 5B, the PID sub-field 820 is adapted to identify what PIDs are provided for a scene started by I-frame 800. These PIDs may include a single (Primary) PID where no replacement content is available, and/or one or more secondary PIDs (e.g., PID-2, PID-3, etc.).


Conversion type sub-field 825 identifies whether the replacement content associated with a secondary PID is an up-rating (e.g., increasing the rating level and making the program for more mature audiences) or a down-rating (e.g., decreasing the rating level and making the displayed program for more general audiences).


Content type sub-field 830 identifies the type of content associated with I-frame 800 and the scene associated with I-frame 800. For instance, content type sub-field 830 may identify the content as a movie, a television series, a news clip, a live sporting event, a live entertainment event such as a concert, etc. This information is used to allow content filter unit 750 of FIG. 7 to filter a variety of types of content, especially those types of content that have not been subject to a ratings scheme.


Replacement type sub-field 835 identifies the type of replacement scheme available. For instance, the replacement scheme may involve full scene replacement or inter-scene data replacement (e.g., pixel obfuscation, audio muting, etc.) to reduce transmission bandwidth requirements. Alternatively, the replacement scheme may simply involve the insertion of content markers that are used by personal video recorders (PVRs) or other types of digital devices to adjust content playback during post-recording solutions.


Length sub-field 840 identifies the number of frames (I, B, P) associated with the length of the digital data stream forming the content.


Rating sub-field 845 provides an age or content rating associated with the scene that are used to determine if any modification, replacement or preclusion involving the rendering of a scene is required.


Tag sub-field 850 provided expanded filtering capabilities that enable filtering to be conducted based on specific types of content found in each scene and not the overall rating of the scene.


Intensity level sub-field 855 provides an alphanumeric value of the intensity of certain subject matter, such as levels of violence, sex or language for example, to provide greater granularity for the filtering operation. This granularity may avoid false-positives where some content may feature content that is in an objectionable category, but such inclusion is minimal or not graphic.


Filter blocking summary sub-field 860 provides for a contextual (tag) overrides described below.


Flag deletion sub-field 865 is a flag (or marker) used by PVR and perhaps other recording digital devices to remove blocking content.


III. Content Replacement Techniques


Referring to FIG. 9, an exemplary diagram of a first embodiment of a content replacement process consistent with certain embodiments of the invention is shown. For this embodiment, no replacement content is provided in the digital data stream, but control information is placed within the MPEG packets such as within the Adaptation field of the I-frame as described above. In response to detecting objectionable content, accomplished by comparing preset parental control parameters with information contained in rating, tag and/or intensity sub-fields 845-855 of FIG. 8B, the content filter unit causes a black frame 900 to be displayed in lieu of the objectionable content. Alternatively, in lieu of a black screen, a pre-stored banner 910, namely selected advertisements paid by various companies having a rating no greater than the limits (e.g., ratings, absence of certain subject matter, etc.) established for the parental control parameter(s) (hereinafter referred to as “parental control limits”), may be displayed if the blocking is for a longer duration.


Referring now to FIG. 10, an exemplary diagram of a second embodiment of a content replacement process consistent with certain embodiments of the invention is shown. For this embodiment, primary content 1000 formed by a plurality of primary content segments 10101-1010N (where N≧1), which are associated with corresponding primary identifiers (ID-1), such as a Packet Identifier (PID), and loaded into content filter unit 750 of FIG. 7. Replacement content 1020, namely replacement content segments 10302, 10305-10307 & 10352 associated with certain secondary identifiers (e.g., ID-2, ID-3), is also provided to the content filter unit. Replacement content segments 10302, 10305-10307 & 10352 may be lower rated or higher rated content than corresponding primary content segments 10102 & 10105-10107.


In response to detecting objectionable content in primary content segments 10102 and 10107, replacement content segments 10352 and 10307 are substituted for content segment 10102 and 10107, respectively. The collection of primary content segments and replacement content segments are output for viewing and/or listening by the end user. Upon evaluation of the primary content segment 10103, it is determined that this also includes objectionable content, but no replacement content segment is provided. As a result, screen blanking or audio muting is conducted as described in FIG. 7 as denoted by “X”.


Referring now to FIG. 11, an illustrative flowchart of a process for determining whether a segment of content associated with a program should be prevented from being displayed is shown. As illustrated, initially programmed parental control parameters are examined to determine what constitutes “objectionable content” for this content filter unit (block 1100). The parental control parameters are preset as shown in FIGS. 13, 14 and 15A.


Next, a determination is made whether the content filter unit has been enabled (block 1110). If not, the V-chip is used as an alternative source for regulating the content to be viewed (arrow 1115). If the content filter unit has been enabled, a first evaluation is whether the program being transmitted can be altered and maintained within the parental control limits set for the content filter unit (block 1120). This can be accomplished by analysis of a meta data regarding the content that summarizes the ratings (age or content based), content labels or content tags on a per scene or per program basis. As a result, replacement content having a lower rating (referred to as “down-rating”) or having a higher rating may be substituted for the primary (chosen) content, provided such content still falls within the parental control limits (also referred to as “up-rating”).


For each scene, a portion of the incoming content segment, such as the Adaptation field for example, is accessed to retrieve information therefrom (block 1130). A determination is made whether the received content segment includes control information to enable filtering of the content (block 1140). If not, the received content segment is rendered without filtering (block 1150). If the received content segment includes control information, a determination is made whether the primary content should be substituted for replacement content (blocks 1160 and 1170). The replacement scheme is set forth in FIG. 12.



FIG. 12 provides an illustrative flowchart of a process for substituting replacement content based on parental control parameters. Initially, as shown in block 1200, parental control parameters associated with the scene are examined. Next, a determination is made whether the display of the scene is controlled based on an Age Rating (block 1210). If so, Age-based rating processing is conducted where the Age Rating of the scene is compared to Aged-based parental control limits programmed by the end user and monitored by the content filter unit (block 1215). In the event that the Age Rating of the scene exceeds the Aged-based parental control limits, the scene is not rendered, but instead may be down-rated through substitution of replacement content or blocked. This substitution may be regulated through use of the synchronization marks described above.


As an option, the content filter unit can be programmed to substitute replacement content having a higher rating than the primary content. The higher rated content would be displayed if the following conditions are met: (1) replacement content has a higher rating than the primary content, and (3) the Age Rating of the replacement content does not exceed the Aged-based parental control limits.


For instance, as shown in FIG. 13, an exemplary ratings table 1300 for objectionable content is shown. Herein, the content filter unit is programmed by setting an Age-based parental control limit within a first parental control parameter 1310. As an example, the Age Rating is established to cover age-based rating levels G, PG, PG-13, R and NC-17 ratings used in movies as well as rating levels TV-Y, TV-Y7, TV-Y7-FV, TV-G, TV-14 and TV-MA used in television broadcasts. For instance, TV-14 may have the same Age rating level as PG-13 (V,S), where “V” denotes violence and “S” denotes sexual content.


As an illustrative example, a first scene 1320 of a program may feature PG-13 content as identified by a primary identifier (ID-1), and thus, no substitution is conducted if the Aged-based parental control limit is set to “PG-13”. However, if the Aged-based parental control limit is set to allow only “G” (or lesser rated) movies or related television broadcasting (e.g. TV-G), replacement content associated with a first alternative identifier (ID-2) would be rendered for the first scene. However, if no age threshold was set, the “R” version of the scene would be rendered by substitution of the “PG-13” content associated with ID-1 with “R” scene content associated with a second alternative identifier (ID-3). The same operations would occur for each scene.


As a result, for “PG-13” Aged-based parental control limits set within first parental control parameter 1310, content segments for scenes 1-8 associated with ID-1 would be rendered. For “G” Aged-based limits set within first parental control parameter 1310, content segments associated with ID-1 (scenes 4 and 8) and ID-2 (scenes 1-3 & 5-7) would be shown. Where no Aged-based limits are established and the content filter unit is programmed to increase the rating of the content where possible, content associated with ID-1 (scene 3), ID-2 (scene 8) and ID-3 (scenes 1-2 & 4-7) would be shown.


Referring back to FIG. 12, another determination is whether the display of the scene is controlled based on content label scheme (blocks 1220 and 1225). With respect to content label evaluation scheme, a determination is made if any particular category of content is present. This is similar to Content Rating scheme describe below where the “Intensity” parental control limits are set to “0”, which also indicates that if there is any content with this subject matter, replacement content without such subject matter (and within the limits set for other types of subject matter) is selected. If no replacement content is available, the content is blocked or muted.


Yet another determination is whether the display of the scene is controlled based on content rating scheme (block 1230). If so, an Intensity rating is applied to a variety of subject matter categories that parents may find objectionable for their children (block 1235). For instance, these categories may be broadly defined such as violence, sex, language, and the like. For those content segments that include this subject matter, “Intensity” values are assigned for each scene based on the maturity level of the subject matter.


As an illustrative example, as shown in FIG. 14, if a first scene 1400 for primary content 1410 of a program includes a fist fight along with a curse word, the violence rating may be set to a first prescribed value, which will likely be fairly low (e.g., 2 out of 15). Depending on the curse word spoken and the audible nature of the curse word, the language rating may be set to a second prescribed value (e.g., 6 out of 15). Replacement content 1412 is provided and associated with ID-2, it may include the elimination of the fist flight and merely a shouting match without any curse words. This replacement content may be substituted if the violence parental control parameters for violence and language are less than “2” and “6”, respectively. Otherwise, the primary content is shown since these Intensity values are less than or equal to the parental control limits of “6”, “7” and “3” for violence, language and sex parental control parameters 1420, 1422 and 1424.


Of course, as an alternative, there may be replacement content 1414 where the fist fight is substituted with a knife fight as originally intended by the director for this program. This replacement content would have a higher violence rating (e.g., 6 out of 15). Since the parental control limits for violence were set at “6”, replacement content associated with ID-3 would be shown if up-rating of the content is desired.


For these parent control limits, if no up-rating is desired, scenes 1, 5-6 and 8 would render content associated with ID-1 (primary ID) and scenes 2-4 and 7 would render content associated with ID-2. With respect to scenes 5 and 7, since no up-rating is desired, the primary content is normally retained as a default, providing scenes 5 and 7 fall within the parental control limits selected. However, it is contemplated that the default setting can be configured to select the content segment having a rating closest to the parental control limits.


In the event that up-rating of content has been selected, scenes 2 and 8 would render content associated with ID-1 (primary ID), scenes 3-6 would render content associated with ID-2 (first alternative ID) and scenes 1 and 7 would render content associated with PID-3 (second alternative ID).


Referring back to FIG. 12, another determination is whether the display of the scene is controlled based on a tag rating scheme (blocks 1240 and 1245). As shown in FIGS. 15A and 15B, the “tag rating” scheme provides a substantial granularity as to what specific content may be excluded. According to one embodiment, each category 1500-1540 and sub-category is uniquely assigned a unique bit pattern so that a multi-byte tag value associated with each scene may be used to identify multiple categories where potential objectionable content may be found. Hence, specific renderings of violence, language, sex, etc. may be prevented.


As an optional configuration, each category may be assigned a particular bit and sub-categories are identified by bit combinations inclusive of the bit assigned to the category. As an example, as shown in FIG. 15B, a tag table may be created where a fourth least significant bit (bbb00001xxx) of a double word tag (4-bytes, where “b” denotes byte) is set when the scene involves violence. For the violence categories, a number of subcategories may be used to denote blood and gore (bbb00001001); sexual assault (bbb00001010), disturbing images (bbb00001011), etc.


Hence, for each scene, the tag is extracted and processed, such as conducting Exclusive OR'ed with the parent control parameter tags set by the user, in order to determine if any objectionable content is found in the scene. If not, the original content may be displayed. Otherwise, if objectionable content is found, a determination is made whether the replacement content features the objectionable content. If not, the replacement content is displayed. Otherwise, if down-rating is needed, no override is indicated, and no replacement content is available, the primary content is blocked or muted.


Upon determining that the scene includes objectionable content, as an optional feature, a determination may be conducted to understand the context that the objectionable content is featured.


For instance, as an illustrative example, certain nudity may be prevalent for a medical procedure featured in the program. Therefore, if the program is a documentary on medical procedures, various bits reserved in the parental control parameters or separate context identifiers coupled be selected as illustrated in FIG. 15B. Hence, certain sub-categories of sex tags may be ignored. Similarly, a gun fight may be prevalent in a documentary regarding the Wild West so any violence tags below a predetermined rating (e.g., rating “6”) would be ignored. Hence, the context parameter can be used to override the content segment replacement operations described above.


In the foregoing description, the invention is described with reference to specific exemplary embodiments thereof. It will, however, be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the present invention as set forth in the appended claims. The specification and drawings are accordingly to be regarded in an illustrative rather than in a restrictive sense.

Claims
  • 1. A nonlinear editing system comprising: a video encoder to receive a first video stream including video originally intended as video for a scene of a program and a second video stream including a first alternate video for replacing the video, the video encoder inserting video synchronization marks into the video and the first alternate video;an audio compressor to receive a first audio stream including audio originally intended as audio for the scene of the program and a second audio stream including a first alternate audio for replacing the audio, the audio compressor inserting audio synchronization marks into the audio and the first alternate audio, wherein the video and audio synchronization marks, respectively, include a primary clock reference, decoding time stamp, and presentation time stamp; andan engine unit to synchronize the video with the first alternate video using the video synchronization marks and to synchronize the audio with the first alternate audio using the audio synchronization marks so that, where the scene of the program includes objectionable content, (i) the video is substituted with the first alternate video, (ii) the audio is substituted with the first alternate audio, or (iii) both the video and the audio are substituted with the first alternate video and the first alternate audio.
  • 2. The nonlinear editing system of claim 1, wherein the video encoder includes a first video encoder to process the first video stream and a second video encoder to process the second video stream.
  • 3. The nonlinear editing system of claim 1, wherein the second video stream includes a blank screen video between the first alternate video and a second alternate video.
  • 4. The nonlinear editing system of claim 3, wherein the second audio stream includes muted audio between the first alternate audio and a second alternate audio.
  • 5. The nonlinear editing system of claim 4, wherein the engine unit trims the blank screen video from the second video stream and the muted audio from the second audio stream.
  • 6. The nonlinear editing system of claim 5, wherein the engine unit insert a reminder of the second video stream after trimming of the blank screen video to be multiplexed into a primary transport stream including the video and the audio.
  • 7. The nonlinear editing system of claim 6, wherein the engine unit insert a reminder of the second audio stream after trimming of the muted audio to be multiplexed into the primary transport stream.
  • 8. A nonlinear editing system comprising: a video encoder to receive a first video stream including video originally intended as video for a scene of a program and a second video stream including a first alternate video for replacing at least a portion of the video, the video encoder inserting video synchronization marks into the video and the first alternate video;an audio compressor to receive a first audio stream including audio originally intended as audio for the scene of the program and a second audio stream including a first alternate audio for replacing at least a portion of the audio, the audio compressor inserting audio synchronization marks into the audio and the first alternate audio, wherein the video and audio synchronization marks, respectively, include a primary clock reference, decoding time stamp, and presentation time stamp; andan engine unit to determine if the scene of the program includes objectionable content and to substitute main content, being at least one of the video and the audio, with replacement content using at least one of the audio and video synchronization marks if the scene of the program includes objectionable content and the replacement does not feature the objectionable content.
  • 9. The nonlinear editing system of claim 8, wherein the replacement content includes the alternate video.
  • 10. The nonlinear editing system of claim 8, wherein the replacement content includes the alternate audio.
  • 11. The nonlinear editing system of claim 8, wherein the replacement content is one of an ad banner and distorted display images.
  • 12. The nonlinear editing system of claim 8, wherein the engine unit determining if the scene of the program features objectionable content by accessing information provided with main content including the video and the audio and comparing the information with prior established parental control limits.
  • 13. The nonlinear editing system of claim 12, wherein the engine unit accessing information provided with the main content and comparing the information with prior established parental control limits by determining whether an Age-based rating assigned to the scene of the program associated with the main content exceeds an Age-based parental control limit being the prior established parental control limit and to substitute the main content with the replacement content if (i) the Age-based rating of the scene associated with the main content exceeds the Age-based parental control limit and (ii) an Age-based rating of the replacement content is less than or equal to the Age-based parental control limit.
  • 14. The nonlinear editing system of claim 12, wherein the engine unit comparing of the information by analyzing tag information provided with the main content for the scene, determining by analysis of the tag information if the main content includes content identified as objectionable, and substituting the main content with the replacement content if the main content includes objectionable content as identified by the tag information and the replacement content does not include the objectionable content.
  • 15. A machine readable non-transitory medium to tangibly store software that, if executed by an engine unit, will cause a non-linear editing system to perform the following operations: receiving a first video stream including video originally intended for a scene of a program and a second video stream including a first alternate video for replacing the video, the video and the first alternate video including video synchronization marks;receiving a first audio stream including audio originally intended for the scene of the program and a second audio stream including a first alternate audio for replacing the audio, the audio and the first alternate audio including audio synchronization marks, wherein the video and audio synchronization marks, respectively, include a primary clock reference, decoding time stamp, and presentation time stamp; andsynchronizing the video with the first alternate video using the video synchronization marks and synchronizing the audio with the first alternate audio using the audio synchronization marks so that, where the scene of the program includes objectionable content, (i) the video is substituted with the first alternate video, (ii) the audio is substituted with the first alternate audio, or (iii) both the video and the audio are substituted with the first alternate video and the first alternate audio.
  • 16. The machine readable non-transitory medium of claim 15, wherein the synchronizing performed by the software includes trimming blank screen video from the second video stream and muted audio from the second audio stream.
  • 17. The machine readable non-transitory medium of claim 16, wherein the synchronizing performed by the software further includes inserting a reminder of the second video stream after trimming of the blank screen video to be multiplexed into a primary transport stream including the video and the audio.
  • 18. The machine readable non-transitory medium of claim 17, wherein the synchronizing performed by the software further includes inserting a reminder of the second audio stream after trimming of the muted audio to be multiplexed into the primary transport stream.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit of priority on U.S. Provisional Patent Application No. 60/636,504 filed Dec. 15, 2004 and U.S. Provisional Patent Application No. 60/637,305 filed Dec. 16, 2004, the contents of which are incorporated by reference.

US Referenced Citations (552)
Number Name Date Kind
3852519 Court Dec 1974 A
4374399 Ensinger et al. Feb 1983 A
4381519 Wilkinson et al. Apr 1983 A
4419693 Wilkinson Dec 1983 A
4521853 Guttag Jun 1985 A
4634808 Moerder Jan 1987 A
4700387 Hirata Oct 1987 A
4703351 Kondo Oct 1987 A
4703352 Kondo Oct 1987 A
4710811 Kondo Dec 1987 A
4712238 Gilhousen et al. Dec 1987 A
4722003 Kondo Jan 1988 A
4739510 Jeffers et al. Apr 1988 A
4772947 Kono Sep 1988 A
4785361 Brotby Nov 1988 A
4788589 Kondo Nov 1988 A
4815078 Shimura Mar 1989 A
4845560 Kondo et al. Jul 1989 A
4881263 Herbison et al. Nov 1989 A
4887296 Horne Dec 1989 A
4890161 Kondo Dec 1989 A
4914515 Van Luyt Apr 1990 A
4924310 von Brandt May 1990 A
4944006 Citta et al. Jul 1990 A
4953023 Kondo Aug 1990 A
4964126 Musicus et al. Oct 1990 A
4989245 Bennett Jan 1991 A
4995080 Bestler et al. Feb 1991 A
5018197 Jones et al. May 1991 A
5023710 Kondo et al. Jun 1991 A
5091936 Katznelson et al. Feb 1992 A
5122873 Golin Jun 1992 A
5138659 Kelkar et al. Aug 1992 A
5142537 Kutner et al. Aug 1992 A
5144662 Welmer Sep 1992 A
5144664 Esserman et al. Sep 1992 A
5151782 Ferraro Sep 1992 A
5159452 Kinoshita et al. Oct 1992 A
5159633 Nakamura Oct 1992 A
5195135 Palmer Mar 1993 A
5196931 Kondo Mar 1993 A
5208816 Seshardi et al. May 1993 A
5237424 Nishino et al. Aug 1993 A
5237610 Gammie et al. Aug 1993 A
5241381 Kondo Aug 1993 A
5247575 Sprague et al. Sep 1993 A
5258835 Kato Nov 1993 A
5319707 Wasilewski et al. Jun 1994 A
5319712 Finkelstein et al. Jun 1994 A
5325432 Gardeck et al. Jun 1994 A
5327502 Katata Jul 1994 A
5341425 Wasilewski et al. Aug 1994 A
5359694 Concordel Oct 1994 A
5379072 Kondo Jan 1995 A
5381481 Gammie et al. Jan 1995 A
5398078 Masuda et al. Mar 1995 A
5400401 Wasilewski et al. Mar 1995 A
5414852 Kramer et al. May 1995 A
5416651 Uetake et al. May 1995 A
5416847 Boze May 1995 A
5420866 Wasilewski et al. May 1995 A
5428403 Andrew et al. Jun 1995 A
5434716 Sugiyama et al. Jul 1995 A
5438369 Citta et al. Aug 1995 A
5444491 Lim Aug 1995 A
5444782 Adams et al. Aug 1995 A
5455862 Hoskinson Oct 1995 A
5463565 Cookson et al. Oct 1995 A
5469216 Takahashi et al. Nov 1995 A
5471501 Parr et al. Nov 1995 A
5473692 Davis Dec 1995 A
5477263 O'Callaghan et al. Dec 1995 A
5481296 Cragun et al. Jan 1996 A
5481554 Kondo Jan 1996 A
5481627 Kim Jan 1996 A
5485577 Eyer et al. Jan 1996 A
5491748 Auld, Jr. et al. Feb 1996 A
5515107 Chiang et al. May 1996 A
5526427 Thomas et al. Jun 1996 A
5528608 Shimizume Jun 1996 A
5535276 Ganesan Jul 1996 A
5539823 Martin et al. Jul 1996 A
5539828 Davis Jul 1996 A
5553141 Lowry et al. Sep 1996 A
5555305 Robinson et al. Sep 1996 A
5561457 Cragun et al. Oct 1996 A
5561713 Suh Oct 1996 A
5568552 Davis Oct 1996 A
5574787 Ryan Nov 1996 A
5582470 Yu Dec 1996 A
5583576 Perlman et al. Dec 1996 A
5583863 Darr, Jr. et al. Dec 1996 A
5590202 Bestler et al. Dec 1996 A
5594507 Hoarty Jan 1997 A
5598214 Kondo et al. Jan 1997 A
5600378 Wasilewski Feb 1997 A
5600721 Kitazato Feb 1997 A
5606359 Youden et al. Feb 1997 A
5608448 Smoral et al. Mar 1997 A
5615265 Coutrot Mar 1997 A
5617333 Oyamada et al. Apr 1997 A
5625715 Trew et al. Apr 1997 A
5629866 Carrubba et al. May 1997 A
5629981 Nerlikar May 1997 A
5644507 Ostrover et al. Jul 1997 A
5652615 Bryant et al. Jul 1997 A
5652795 Eillon et al. Jul 1997 A
5663764 Kondo et al. Sep 1997 A
5666293 Metz et al. Sep 1997 A
5696906 Peters et al. Dec 1997 A
5699429 Tamer et al. Dec 1997 A
5703655 Corey et al. Dec 1997 A
5703889 Shimoda et al. Dec 1997 A
5717814 Abecassis Feb 1998 A
5726702 Hamaguchi et al. Mar 1998 A
5726711 Boyce Mar 1998 A
5732185 Hirayama Mar 1998 A
5732346 Lazaridis et al. Mar 1998 A
5742680 Wilson Apr 1998 A
5742681 Giachetti et al. Apr 1998 A
5751280 Abbott et al. May 1998 A
5751743 Takizawa May 1998 A
5751813 Dorenbos May 1998 A
5754650 Katznelson May 1998 A
5754658 Aucsmith et al. May 1998 A
5757417 Aras et al. May 1998 A
5757909 Park May 1998 A
5761180 Murabayashi et al. Jun 1998 A
5768539 Metz et al. Jun 1998 A
5796786 Lee Aug 1998 A
5796829 Newby et al. Aug 1998 A
5796840 Davis Aug 1998 A
5802176 Audebert Sep 1998 A
5805700 Nardone et al. Sep 1998 A
5805712 Davis Sep 1998 A
5805762 Boyce et al. Sep 1998 A
5809147 De Lange et al. Sep 1998 A
5815146 Youden et al. Sep 1998 A
5818934 Cuccia Oct 1998 A
5825879 Davis Oct 1998 A
5835668 Yanagihara Nov 1998 A
5838873 Blatter et al. Nov 1998 A
5850218 LaJoie et al. Dec 1998 A
5852290 Chaney Dec 1998 A
5852470 Kondo et al. Dec 1998 A
5859662 Cragun et al. Jan 1999 A
5870474 Wasilewski et al. Feb 1999 A
5892900 Ginter et al. Apr 1999 A
5894320 Vancelette Apr 1999 A
5894516 Brandenburg Apr 1999 A
5905732 Fimoff et al. May 1999 A
5914719 Herz et al. Jun 1999 A
5915018 Aucsmith Jun 1999 A
5917830 Chen et al. Jun 1999 A
5920625 Davies Jul 1999 A
5920626 Durden et al. Jul 1999 A
5922048 Eumura Jul 1999 A
5923755 Birch Jul 1999 A
5930361 Hayashi et al. Jul 1999 A
5933498 Schneck et al. Aug 1999 A
5933500 Blatter et al. Aug 1999 A
5940738 Rao Aug 1999 A
5943605 Koepele, Jr. Aug 1999 A
5949877 Traw et al. Sep 1999 A
5949881 Davis et al. Sep 1999 A
5963909 Warren et al. Oct 1999 A
5973679 Abbott et al. Oct 1999 A
5973722 Wakai et al. Oct 1999 A
5973726 Iijima et al. Oct 1999 A
5999622 Yasukawa et al. Dec 1999 A
5999698 Nakai et al. Dec 1999 A
6005561 Hawkins et al. Dec 1999 A
6005940 Kulinets Dec 1999 A
6011849 Orrin Jan 2000 A
6012144 Pickett Jan 2000 A
6016348 Blatter et al. Jan 2000 A
6021199 Isibashi Feb 2000 A
6021201 Bakhle et al. Feb 2000 A
6026164 Sakamoto et al. Feb 2000 A
6028932 Park Feb 2000 A
6049613 Jakobsson Apr 2000 A
6055314 Spies et al. Apr 2000 A
6055315 Doyle et al. Apr 2000 A
6057832 Lev et al. May 2000 A
6057872 Candelore May 2000 A
6058186 Enari May 2000 A
6058192 Guralnick et al. May 2000 A
6061451 Muratani et al. May 2000 A
6061471 Coleman, Jr. May 2000 A
6064676 Slattery et al. May 2000 A
6064748 Hogan May 2000 A
6065050 DeMoney May 2000 A
6069647 Sullivan et al. May 2000 A
6070245 Murphy, Jr. et al. May 2000 A
6072872 Chang et al. Jun 2000 A
6072873 Bewick Jun 2000 A
6073122 Wool Jun 2000 A
6088450 Davis et al. Jul 2000 A
6105134 Pinder et al. Aug 2000 A
6108422 Newby et al. Aug 2000 A
6115057 Kwoh Sep 2000 A
6115821 Newby et al. Sep 2000 A
6118873 Lotspiech et al. Sep 2000 A
6134237 Brailean et al. Oct 2000 A
6134551 Aucsmith Oct 2000 A
6138237 Ruben et al. Oct 2000 A
6148082 Slattery et al. Nov 2000 A
6154206 Ludtke Nov 2000 A
6157719 Wasilewski et al. Dec 2000 A
6160548 Lea et al. Dec 2000 A
6166780 Bray Dec 2000 A
6170075 Schuster et al. Jan 2001 B1
6181334 Freeman et al. Jan 2001 B1
6181364 Ford et al. Jan 2001 B1
6185369 Ko et al. Feb 2001 B1
6185546 Davis Feb 2001 B1
6189096 Haverty Feb 2001 B1
6192131 Geer, Jr. et al. Feb 2001 B1
6199053 Herbert et al. Mar 2001 B1
6201927 Comer et al. Mar 2001 B1
6204843 Freeman et al. Mar 2001 B1
6209098 Davis Mar 2001 B1
6215484 Freeman et al. Apr 2001 B1
6216263 Elam Apr 2001 B1
6219358 Pinder et al. Apr 2001 B1
6222924 Salomaki Apr 2001 B1
6223290 Larsen et al. Apr 2001 B1
6226385 Taguchi et al. May 2001 B1
6226618 Downs et al. May 2001 B1
6229576 Tarr et al. May 2001 B1
6229895 Son et al. May 2001 B1
6230194 Frailong et al. May 2001 B1
6230266 Perlman et al. May 2001 B1
6236727 Ciacelli et al. May 2001 B1
6240553 Son et al. May 2001 B1
6246720 Kutner et al. Jun 2001 B1
6256747 Inohara et al. Jul 2001 B1
6263506 Ezaki et al. Jul 2001 B1
6266416 Sigbjornsen et al. Jul 2001 B1
6266480 Ezaki et al. Jul 2001 B1
6272538 Holden et al. Aug 2001 B1
6278783 Kocher et al. Aug 2001 B1
6289455 Kocher et al. Sep 2001 B1
6292568 Akins et al. Sep 2001 B1
6292892 Davis Sep 2001 B1
6307939 Vigarie Oct 2001 B1
6311012 Cho et al. Oct 2001 B1
6314111 Nandikonda et al. Nov 2001 B1
6314409 Schneck et al. Nov 2001 B2
6323914 Linzer Nov 2001 B1
6324288 Hoffman Nov 2001 B1
6327421 Tiwari et al. Dec 2001 B1
6337947 Porter et al. Jan 2002 B1
6351538 Uz Feb 2002 B1
6351813 Mooney et al. Feb 2002 B1
6377589 Knight et al. Apr 2002 B1
6378130 Adams Apr 2002 B1
6389533 Davis et al. May 2002 B1
6389537 Davis et al. May 2002 B1
6415031 Colligan et al. Jul 2002 B1
6415101 deCarmo et al. Jul 2002 B1
6418169 Datari Jul 2002 B1
6424717 Pinder et al. Jul 2002 B1
6430361 Lee Aug 2002 B2
6445738 Zdepski et al. Sep 2002 B1
6449718 Rucklidge et al. Sep 2002 B1
6452923 Gerszberg et al. Sep 2002 B1
6453115 Boyle Sep 2002 B1
6453116 Ando et al. Sep 2002 B1
6456985 Ohtsuka Sep 2002 B1
6459427 Mao et al. Oct 2002 B1
6460018 Kasai et al. Oct 2002 B1
6463445 Suzuki et al. Oct 2002 B1
6466671 Maillard et al. Oct 2002 B1
6469749 Dimitrova Oct 2002 B1
6473459 Sugano et al. Oct 2002 B1
6480551 Ohishi et al. Nov 2002 B1
6490728 Kitazato Dec 2002 B1
6505032 McCorkle et al. Jan 2003 B1
6505299 Zeng et al. Jan 2003 B1
6510554 Gordon et al. Jan 2003 B1
6519693 Debey Feb 2003 B1
6526144 Markandey et al. Feb 2003 B2
6529526 Schneidewend Mar 2003 B1
6543053 Li et al. Apr 2003 B1
6549229 Kirby et al. Apr 2003 B1
6550008 Zhang et al. Apr 2003 B1
6557031 Mimura et al. Apr 2003 B1
6587561 Sered et al. Jul 2003 B1
6590979 Ryan Jul 2003 B1
6621866 Florencio et al. Sep 2003 B1
6621979 Eerenberg et al. Sep 2003 B1
6640145 Hoffberg et al. Oct 2003 B2
6640305 Kocher et al. Oct 2003 B2
6643298 Brunheroto et al. Nov 2003 B1
6650754 Akiyama et al. Nov 2003 B2
6654389 Brunheroto et al. Nov 2003 B1
6678740 Rakib et al. Jan 2004 B1
6681326 Son et al. Jan 2004 B2
6684250 Anderson et al. Jan 2004 B2
6697489 Candelore Feb 2004 B1
6697944 Jones et al. Feb 2004 B1
6701258 Kramb et al. Mar 2004 B2
6704733 Clark et al. Mar 2004 B2
6707696 Turner et al. Mar 2004 B1
6714650 Maillard et al. Mar 2004 B1
6741795 Takehiko et al. May 2004 B1
6754276 Harumoto et al. Jun 2004 B1
6771285 McGrath et al. Aug 2004 B1
6772340 Peinado et al. Aug 2004 B1
6788690 Harri Sep 2004 B2
6788882 Geer et al. Sep 2004 B1
6826185 Montanaro et al. Nov 2004 B1
6853728 Kahn et al. Feb 2005 B1
6867818 Bray Mar 2005 B2
6883050 Safadi Apr 2005 B1
6891565 Dieterich May 2005 B1
6895128 Bohnenkamp May 2005 B2
6898799 Jarman May 2005 B1
6899383 Hwang May 2005 B2
6904520 Rosset et al. Jun 2005 B1
6917684 Tatebayashi et al. Jul 2005 B1
6925180 Iwamura Aug 2005 B2
6938162 Nagai et al. Aug 2005 B1
6972802 Bray Dec 2005 B2
6973187 Gligor et al. Dec 2005 B2
6976166 Herley et al. Dec 2005 B2
6976265 Yang et al. Dec 2005 B1
6988238 Kovacevic et al. Jan 2006 B1
7023924 Keller et al. Apr 2006 B1
7039802 Eskicioglu et al. May 2006 B1
7039938 Candelore May 2006 B2
7054547 Abecassis May 2006 B1
7055166 Logan et al. May 2006 B1
7065213 Pinder Jun 2006 B2
7079752 Leyendecker Jul 2006 B1
7089579 Mao et al. Aug 2006 B1
7096481 Forecast et al. Aug 2006 B1
7096487 Gordon et al. Aug 2006 B1
7110659 Fujie et al. Sep 2006 B2
7120250 Candelore Oct 2006 B2
7124303 Candelore et al. Oct 2006 B2
7127619 Unger et al. Oct 2006 B2
7139031 Bray Nov 2006 B1
7139398 Candelore et al. Nov 2006 B2
7146007 Maruo et al. Dec 2006 B1
7151831 Candelore et al. Dec 2006 B2
7151833 Candelore et al. Dec 2006 B2
7155012 Candelore et al. Dec 2006 B2
7158185 Gastaldi Jan 2007 B2
7194758 Waki et al. Mar 2007 B1
7221706 Zhao et al. May 2007 B2
7292692 Bonan et al. Nov 2007 B2
7298959 Hallberg et al. Nov 2007 B1
7336785 Lu et al. Feb 2008 B1
7360234 Robson et al. Apr 2008 B2
7391866 Fukami et al. Jun 2008 B2
7421729 Zenoni Sep 2008 B2
7490236 Wasilewski Feb 2009 B2
7490344 Haberman et al. Feb 2009 B2
7496198 Pinder Feb 2009 B2
7508454 Vantalon et al. Mar 2009 B1
7555123 Pinder et al. Jun 2009 B2
20010013123 Freeman et al. Aug 2001 A1
20010017885 Asai et al. Aug 2001 A1
20010024471 Bordes et al. Sep 2001 A1
20010030959 Ozawa et al. Oct 2001 A1
20010036271 Javed Nov 2001 A1
20010051007 Teshima Dec 2001 A1
20020003881 Reitmeier et al. Jan 2002 A1
20020021805 Schumann et al. Feb 2002 A1
20020023013 Hughes et al. Feb 2002 A1
20020026478 Rodgers et al. Feb 2002 A1
20020026587 Talstra et al. Feb 2002 A1
20020036717 Abiko et al. Mar 2002 A1
20020044558 Gobbi et al. Apr 2002 A1
20020046406 Chelehmal et al. Apr 2002 A1
20020047915 Misu Apr 2002 A1
20020056093 Kunkel et al. May 2002 A1
20020059425 Belfiore et al. May 2002 A1
20020065678 Peliotis et al. May 2002 A1
20020066101 Gordon et al. May 2002 A1
20020067436 Shirahama et al. Jun 2002 A1
20020083317 Ohta et al. Jun 2002 A1
20020083438 So et al. Jun 2002 A1
20020083439 Eldering Jun 2002 A1
20020097322 Monroe et al. Jul 2002 A1
20020100054 Feinberg et al. Jul 2002 A1
20020108035 Herley et al. Aug 2002 A1
20020109707 Lao et al. Aug 2002 A1
20020116705 Perlman et al. Aug 2002 A1
20020122136 Safadi et al. Sep 2002 A1
20020126890 Katayama et al. Sep 2002 A1
20020129243 Nanjundiah Sep 2002 A1
20020144116 Giobbi Oct 2002 A1
20020144260 Devara Oct 2002 A1
20020147782 Dimitrova et al. Oct 2002 A1
20020150239 Carny et al. Oct 2002 A1
20020157115 Lu Oct 2002 A1
20020157116 Jasinschi Oct 2002 A1
20020164022 Strasser et al. Nov 2002 A1
20020170053 Peterka et al. Nov 2002 A1
20020184506 Perlman Dec 2002 A1
20020188567 Candelore Dec 2002 A1
20020194589 Cristofalo et al. Dec 2002 A1
20020194613 Unger Dec 2002 A1
20020194619 Chang et al. Dec 2002 A1
20020196939 Unger et al. Dec 2002 A1
20020199208 Chang et al. Dec 2002 A1
20030002854 Belknap et al. Jan 2003 A1
20030009669 White et al. Jan 2003 A1
20030012286 Ishtiaq et al. Jan 2003 A1
20030021412 Candelore et al. Jan 2003 A1
20030026423 Unger et al. Feb 2003 A1
20030026523 Chua et al. Feb 2003 A1
20030028879 Gordon et al. Feb 2003 A1
20030034997 McKain et al. Feb 2003 A1
20030035482 Klompenhouwer et al. Feb 2003 A1
20030035540 Freeman et al. Feb 2003 A1
20030035543 Gillon Feb 2003 A1
20030046686 Candelore et al. Mar 2003 A1
20030046687 Hodges et al. Mar 2003 A1
20030059047 Iwamura Mar 2003 A1
20030063615 Iuoma et al. Apr 2003 A1
20030072555 Yap et al. Apr 2003 A1
20030077071 Lin et al. Apr 2003 A1
20030081630 Mowery et al. May 2003 A1
20030081776 Candelore May 2003 A1
20030084284 Ando et al. May 2003 A1
20030097662 Russ et al. May 2003 A1
20030105679 Krishnan et al. Jun 2003 A1
20030108199 Pinder et al. Jun 2003 A1
20030112333 Chen et al. Jun 2003 A1
20030118059 Sugahara Jun 2003 A1
20030118243 Sezer et al. Jun 2003 A1
20030123664 Pedlow, Jr. et al. Jul 2003 A1
20030123849 Nallur et al. Jul 2003 A1
20030126086 Safadi Jul 2003 A1
20030133570 Candelore et al. Jul 2003 A1
20030140257 Peterka et al. Jul 2003 A1
20030145329 Candelore Jul 2003 A1
20030152224 Candelore et al. Aug 2003 A1
20030152232 Pirila et al. Aug 2003 A1
20030156718 Candelore et al. Aug 2003 A1
20030159139 Candelore et al. Aug 2003 A1
20030159140 Candelore Aug 2003 A1
20030159152 Lin et al. Aug 2003 A1
20030163832 Tsuria et al. Aug 2003 A1
20030174837 Candelore et al. Sep 2003 A1
20030174844 Candelore Sep 2003 A1
20030188154 Dallard Oct 2003 A1
20030188164 Okimoto Oct 2003 A1
20030190054 Troyansky et al. Oct 2003 A1
20030193973 Takashimizu et al. Oct 2003 A1
20030194211 Abecassis Oct 2003 A1
20030198223 Mack et al. Oct 2003 A1
20030204717 Kuehnel Oct 2003 A1
20030206631 Candelore Nov 2003 A1
20030222994 Dawson Dec 2003 A1
20030226149 Chun et al. Dec 2003 A1
20030228018 Vince Dec 2003 A1
20040003008 Wasilewski et al. Jan 2004 A1
20040010717 Simec et al. Jan 2004 A1
20040021764 Driscoll, Jr. et al. Feb 2004 A1
20040028227 Yu et al. Feb 2004 A1
20040037421 Truman Feb 2004 A1
20040039648 Candelore et al. Feb 2004 A1
20040047470 Candelore Mar 2004 A1
20040049688 Candelore et al. Mar 2004 A1
20040049690 Candelore et al. Mar 2004 A1
20040049691 Candelore et al. Mar 2004 A1
20040049694 Candelore Mar 2004 A1
20040049708 Thomas et al. Mar 2004 A1
20040049780 Gee Mar 2004 A1
20040064688 Jacobs Apr 2004 A1
20040068659 Diehl Apr 2004 A1
20040073917 Pedlow, Jr. et al. Apr 2004 A1
20040078575 Morten et al. Apr 2004 A1
20040081333 Grab et al. Apr 2004 A1
20040083177 Chen et al. Apr 2004 A1
20040083484 Ryal Apr 2004 A1
20040086127 Candelore May 2004 A1
20040088552 Candelore May 2004 A1
20040088558 Candelore et al. May 2004 A1
20040091109 Son et al. May 2004 A1
20040091239 Poslinski et al. May 2004 A1
20040100510 Milic-Frayling et al. May 2004 A1
20040123094 Sprunk Jun 2004 A1
20040136532 Pinder et al. Jul 2004 A1
20040139337 Pinder et al. Jul 2004 A1
20040151314 Candelore Aug 2004 A1
20040158721 Candelore Aug 2004 A1
20040165586 Read et al. Aug 2004 A1
20040168121 Matz Aug 2004 A1
20040172650 Hawkins et al. Sep 2004 A1
20040181666 Candelore Sep 2004 A1
20040187161 Cao Sep 2004 A1
20040189690 Poslinski et al. Sep 2004 A1
20040193550 Siegel Sep 2004 A1
20040205812 Candelore Oct 2004 A1
20040216169 Fujiwara Oct 2004 A1
20040240668 Bonan et al. Dec 2004 A1
20040247122 Hobrock et al. Dec 2004 A1
20040255321 Matz Dec 2004 A1
20040261099 Durden et al. Dec 2004 A1
20040264924 Campisano et al. Dec 2004 A1
20040267602 Gaydos et al. Dec 2004 A1
20050002385 Candelore Jan 2005 A1
20050004875 Kontio et al. Jan 2005 A1
20050015816 Christofalo et al. Jan 2005 A1
20050028193 Candelore et al. Feb 2005 A1
20050036067 Ryal et al. Feb 2005 A1
20050058291 Candelore Mar 2005 A1
20050063541 Candelore Mar 2005 A1
20050066357 Ryal Mar 2005 A1
20050071669 Medvinsky et al. Mar 2005 A1
20050071888 Girouard et al. Mar 2005 A1
20050094808 Pedlow, Jr. et al. May 2005 A1
20050094809 Pedlow, Jr. et al. May 2005 A1
20050097596 Pedlow, Jr. May 2005 A1
20050097597 Pedlow, Jr. et al. May 2005 A1
20050097598 Pedlow, Jr. et al. May 2005 A1
20050097614 Pedlow, Jr. et al. May 2005 A1
20050102702 Candelore et al. May 2005 A1
20050129233 Pedlow, Jr. Jun 2005 A1
20050141713 Genevois Jun 2005 A1
20050169473 Candelore Aug 2005 A1
20050192904 Candelore Sep 2005 A1
20050198586 Sekiguchi et al. Sep 2005 A1
20050259813 Wasilewski et al. Nov 2005 A1
20050265547 Strasser et al. Dec 2005 A1
20050283797 Eldering et al. Dec 2005 A1
20060064715 Poslinski Mar 2006 A1
20060115083 Candelore et al. Jun 2006 A1
20060130119 Candelore Jun 2006 A1
20060130121 Candelore et al. Jun 2006 A1
20060136976 Coupe et al. Jun 2006 A1
20060153379 Candelore et al. Jul 2006 A1
20060168616 Candelore Jul 2006 A1
20060174264 Candelore Aug 2006 A1
20060236220 Jarman Oct 2006 A1
20060262219 Molaro Nov 2006 A1
20060262926 Candelore et al. Nov 2006 A1
20060269060 Candelore et al. Nov 2006 A1
20070006253 Pinder et al. Jan 2007 A1
20070091886 Davis et al. Apr 2007 A1
20070118801 Harshbarger et al. May 2007 A1
20070154176 Elcock et al. Jul 2007 A1
20070186236 Jarman Aug 2007 A1
20070288516 Jarman Dec 2007 A1
20080130119 Candelore et al. Jun 2008 A1
20090210698 Candelore Aug 2009 A1
Foreign Referenced Citations (47)
Number Date Country
2389247 May 2001 CA
0471373 Feb 1992 EP
0527611 Jul 1992 EP
0558016 Feb 1993 EP
0596826 Apr 1993 EP
0610587 Dec 1993 EP
0680209 Apr 1995 EP
0674440 Sep 1995 EP
0674441 Sep 1995 EP
0696141 Feb 1996 EP
0720374 Jul 1996 EP
0382764 Apr 1997 EP
0833517 Apr 1998 EP
0866615 Sep 1998 EP
0926894 Jun 1999 EP
1187483 Mar 2002 EP
61-264371 Nov 1986 JP
2000125260 Nov 1986 JP
07046575 Feb 1995 JP
7067028 Mar 1995 JP
10336624 Dec 1998 JP
2001069480 Mar 2001 JP
2001117809 Apr 2001 JP
2001242786 Sep 2001 JP
11243534 Oct 2002 JP
2003122710 Apr 2003 JP
100299634 Aug 2008 KR
WO-8607224 Dec 1986 WO
WO-9309525 May 1993 WO
WO-9410775 May 1994 WO
WO-9413081 Jun 1994 WO
WO-9413107 Jun 1994 WO
WO-9528058 Oct 1995 WO
WO-9738530 Oct 1997 WO
WO-9746009 Dec 1997 WO
WO-9808341 Feb 1998 WO
WO-0031964 Jun 2000 WO
WO-0059203 Oct 2000 WO
WO-0060846 Oct 2000 WO
WO-0064164 Oct 2000 WO
WO-0070817 Nov 2000 WO
WO-0126372 Apr 2001 WO
WO-0135669 May 2001 WO
WO-0167667 Sep 2001 WO
WO-0165762 Sep 2001 WO
WO-0178386 Oct 2001 WO
WO-0251096 Jun 2002 WO
Related Publications (1)
Number Date Country
20060130121 A1 Jun 2006 US
Provisional Applications (2)
Number Date Country
60636504 Dec 2004 US
60637305 Dec 2004 US