1. Field of the Invention
The present principles of the embodiments generally relate to a method and an apparatus for selecting or removing one or more audio component types in an audio source associated with a video source at an electronic device. For example, the present invention allows a user of an electronic device to selectively choose or remove, e.g., speech, music, and/or some other component type of an audio source for a selected video program at a receiver. In another embodiment, the selected audio source may be from another program.
2. Background Information
Typically, a program such as a television program comprises both video content as well as audio content. In a typical sports program such as, e.g., figure skating competition, the audio content may further comprise different audio component types such as, e.g., speech of a sports announcer, sound of the accompanying music, and other sounds (e.g., ambience noise, background crowd noise, noise of the skating action on the ice), etc. The present inventors recognize that in certain instances, it would be advantageous for a user of an electronic device to be able to select a particular audio component type such as e.g., the accompanying music, or remove, e.g., boring comments from a TV announcer while watching the skating event. In addition, it would be advantageous to be able to, e.g., replace the boring TV announcer with another more entertaining announcer or a home-team announcer from another video program from another television channel or some other sources, if the same event is being simulcast on the other channel or streamed from the Internet. Additionally, the user may just want to watch the skating event with some other background music from another totally different program with no announcer voice at all.
The present inventors recognize the needs for improvement to the existing electronic devices as described above. Therefore, an object of the present invention is to provide solutions to those needs in accordance with the principles of the present invention.
In accordance with an aspect of the present invention, an apparatus is presented, comprising:
a controller for controlling processing of a first video signal and a first audio signal;
an audio decoder for providing a plurality of audio component types in the first audio signal; and
wherein the audio decoder provides a first modified audio signal for the first video signal in response to selection of one or more of the plurality of audio component types.
In another embodiment, a method is presented comprising steps of:
receiving a first video signal and a first audio signal;
indicating to a user a plurality of audio component types available for selection in the first audio signal;
receiving a user command to select one or more of the available audio component types from the first audio signal to form a first modified audio signal; and
playing back the modified audio signal along with the first video signal.
In accordance with another aspect of the present invention, an apparatus is presented, comprising:
a controller for controlling processing of a first program having a first video signal and a first audio signal, and another program having another audio signal;
an audio decoder for providing a plurality of different audio component types in the received audio signals;
a user interface for receiving a user command to select one or more of the audio component types; and
wherein the audio decoder provides the modified audio signal for the first video signal from the selected one or more of the audio component types.
In another embodiment, a method is presented comprising steps of:
receiving a program having a first video signal and a first audio signal;
receiving another program having another audio signal;
indicating to a user a plurality of different audio component types available for selection in the received audio signals;
receiving a user command to select one or more of the available audio component types to form a modified audio signal; and
playing back the modified audio signal along with the first video signal.
In accordance with another aspect of the present invention, an apparatus is presented, comprising:
a controller for controlling processing of a plurality of programs;
an audio decoder for providing a plurality of audio component types from a plurality of audio sources of the pluralities of programs;
a video decoder for providing a display signal for displaying selected audio component types and indications from which respective programs the selected audio component types are to be found; and
the audio decoder provides a modified audio signal for a first program from one or more of the selected audio component types.
In another embodiment, a method is presented comprising steps of: receiving a first program;
displaying a plurality of audio component types associated with the program;
for a selected audio component type, receiving a user command for choosing which program the selected audio component type is to be found; and
forming a modified audio signal for the first program from one or more of the selected audio component types.
The aforementioned brief summary of exemplary embodiments of the present invention is merely illustrative of the inventive concepts presented herein, and is not intended to limit the scope of the present invention in any manner.
The above-mentioned and other features and advantages of this invention, and the manner of attaining them, will become more apparent and the invention will be better understood by reference to the following description of embodiments of the invention taken in conjunction with the accompanying drawings, wherein:
The examples set out herein illustrate exemplary embodiments of the invention. Such examples are not to be construed as limiting the scope of the invention in any manner.
Referring now to the drawings, and more particularly to
Other exemplary devices may include mobile devices such as cellular telephones, tablets, PCs, or devices combining computer and television functions such as the so-called “PC/TVs”. The term “program” as used herein represents any form of content data such as digital video and audio information, including streaming and stored data content received via cable, satellite, broadcast and other telecommunications networks, or via local networks or connections, such as WiFi, USB or Firewire connections. The content can be from a remote source (e.g., a server) or from a local source, such as from a local storage media (e.g., hard drives, memory cards or USB memory sticks, etc.)
As an overview, in the video receiver system of
Video and audio decoders 85 and 80 respectively, decode the compressed data from system 25 to provide outputs for display 89 and speakers 88. Data port 75 provides an interface for communication of the compressed data from system 25 to/from other devices such as a computer or a High Definition Television (HDTV) receiver, for example. Storage device 90 stores compressed data from system 25 on storage medium 105. Device 90, in a playback mode also supports retrieval of the compressed or uncompressed video and audio data from storage medium 105 for processing by system 25 for decoding, communication to other devices or storage on a different storage medium (not shown to simplify drawing).
Considering
Multiplexer (mux) 37 of service detector 33 is provided, via selector 35, with either the output from decoder 30, or the decoder 30 output is further processed by a descrambling unit 40. Descrambling unit 40 may be, for example, a removable unit such as a smart card in accordance with ISO 7816 and NRSS (National Renewable Security Standards) Committee standards (the NRSS removable conditional access system is defined in EIA Draft Document IS-679, Project PN-3639), or a CableCARD used in U.S. cable systems. Selector 35 detects the presence of an insertable, compatible, descrambling card and provides the output of unit 40 to mux 37 only if the card is currently inserted in the video receiver unit. Otherwise selector 35 provides the output from decoder 30 to mux 37. The presence of the insertable card permits unit 40 to descramble additional premium program channels, for example, and provide additional program services to a viewer. It should be noted that in the preferred embodiment NRSS unit 40 and smart card unit 130 (smart card unit 130 is discussed later) share the same system 25 interface such that only either an NRSS card or a smart card may be inserted at any one time. However, the interfaces may also be separate to allow parallel operation.
The data provided to mux 37 from selector 35 is in the form of an MPEG compliant packetized transport data stream as defined e.g., in MPEG2 Systems Standards ISO_IEC 13818-1 and may include program guide information and the data content of one or more program channels. The individual packets that comprise particular program channels are identified by Packet Identifiers (PIDs). The transport stream contains Program Specific Information (PSI) for use in identifying the PIDs and assembling individual data packets to recover the content of all the program channels that comprise the packetized data stream. Transport system 25, under the control of the system controller 115, acquires and collates program guide information from the input transport stream, storage device 90 or an Internet service provider via the communication interface unit 116. The individual packets that comprise either particular program channel content or program guide information, are identified by their Packet Identifiers (PIDs) contained within header information. Program guide information may contain descriptions for a program which may comprise different program descriptive fields such as title, star, rating, detailed event description, and etc., relating to a program.
The user interface incorporated in the video receiver shown in
Using remote control unit 125 (or other selection means such as a mouse, voice or other gesture activated devices, etc.) a user can select from the OSD menu items such as, for example, an icon to be selected, a plurality of audio component types to be selected, a program to be selected or viewed, a program to be stored or viewed from storage, the type of storage media and manner of storage, and scrolling of the EPG, etc.
A user is able to select or remove one or more of the audio component types of an audio source for a selected video program via remote control unit 125. System controller 115 uses the selection information, provided via remote unit interface 120, to configure the system shown in
In addition, when a user selects programs for viewing or storage, system controller 115 generates PSI suitable for the selected storage device and media. Controller 115 also configures system 25 elements 45, 47, 50, 55, 65 and 95 by setting control register values within these elements via a data bus and by selecting signal paths via muxes 37 and 110 with control signal C.
In response to control signal C, mux 37 selects either, the transport stream from unit 35, or in a playback mode, a data stream retrieved from storage device 90 via store interface 95. In normal, non-playback operation, the data packets comprising the program that the user selected to view are identified by their PIDs by selection unit 45. If an encryption indicator in the header data of the selected program packets indicates the packets are encrypted, unit 45 provides the packets to decryption unit 50. Otherwise unit 45 provides non-encrypted packets to transport decoder 55. Similarly, the data packets comprising the programs that the user selected for storage are identified by their PIDs by selection unit 47. Unit 47 provides encrypted packets to decryption unit 50 or non-encrypted packets to mux 110 based on the packet header encryption indicator information.
The functions of decryptors 40 and 50 may be implemented in a single removable smart card which is compatible with the NRSS standard. This approach places all security related functions in one removable unit that easily can be replaced if a service provider decides to change encryption technique or to permit easily changing the security system, e.g., to descramble a different service.
Units 45 and 47 employ PID detection filters that match the PIDs of incoming packets provided by mux 37 with PID values pre-loaded in control registers within units 45 and 47 by controller 115. The pre-loaded PIDs are used in units 47 and 45 to identify the data packets that are to be stored and the data packets that are to be decoded for use in providing a video image. The pre-loaded PIDs are stored in look-up tables in units 45 and 47. The PID look-up tables are memory mapped to encryption key tables in units 45 and 47 that associate encryption keys with each pre-loaded PID. The memory mapped PID and encryption key look-up tables permit units 45 and 47 to match encrypted packets containing a pre-loaded PID with associated encryption keys that permit their decryption. Non-encrypted packets do not have associated encryption keys. Units 45 and 47 provide both identified packets and their associated encryption keys to decryptor 50. The PID look-up table in unit 45 is also memory mapped to a destination table that matches packets containing pre-loaded PIDs with corresponding destination buffer locations in packet buffer 60. The encryption keys and destination buffer location addresses associated with the programs selected by a user for viewing or storage are pre-loaded into units 45 and 47 along with the assigned PIDs by controller 115. The encryption keys are generated by ISO 7816-3 compliant smart card system 130 from encryption codes extracted from the input datastream. The generation of the encryption keys is subject to customer entitlement determined from coded information in the input datastream and/or pre-stored on the insertable smart card itself (International Standards Organization document ISO 7816-3 of 1989 defines the interface and signal structures for a smart card system).
The packets provided by units 45 and 47 to unit 50 are encrypted using an encryption techniques such as the Data Encryption Standard (DES) defined in Federal Information Standards (FIPS) Publications 46, 74 and 81 provided by the National Technical Information Service, Department of Commerce. Unit 50 decrypts the encrypted packets using corresponding encryption keys provided by units 45 and 47 by applying decryption techniques appropriate for the selected encryption algorithm. The decrypted packets from unit 50 and the non-encrypted packets from unit 45 that comprise the program for display are provided to decoder 55. The decrypted packets from unit 50 and the non-encrypted packets from unit 47 that comprise the program for storage are provided to mux 110.
Unit 60 contains four packet buffers accessible by controller 115. One of the buffers is assigned to hold data destined for use by controller 115 and the other three buffers are assigned to hold packets that are destined for use by application devices 75, 80 and 85. Access to the packets stored in the four buffers within unit 60 by both controller 115 and by application interface 70 is controlled by buffer control unit 65. Unit 45 provides a destination flag to unit 65 for each packet identified by unit 45 for decoding. The flags indicate the individual unit 60 destination locations for the identified packets and are stored by control unit 65 in an internal memory table. Control unit 65 determines a series of read and write pointers associated with packets stored in buffer 60 based on the First-In-First-Out (FIFO) principle. The write pointers in conjunction with the destination flags permit sequential storage of an identified packet from units 45 or 50 in the next empty location within the appropriate destination buffer in unit 60. The read pointers permit sequential reading of packets from the appropriate unit 60 destination buffers by controller 115 and application interface 70.
The non-encrypted and decrypted packets provided by units 45 and 50 to decoder 55 contain a transport header as defined by section 2.4.3.2 of the MPEG systems standard. Decoder 55 determines from the transport header whether the non-encrypted and decrypted packets contain an adaptation field (per the MPEG systems standard). The adaptation field contains timing information including, for example, Program Clock References (PCRs) that permit synchronization and decoding of content packets. Upon detection of a timing information packet, that is a packet containing an adaptation field, decoder 55 signals controller 115, via an interrupt mechanism by setting a system interrupt, that the packet has been received. In addition, decoder 55 changes the timing packet destination flag in unit 65 and provides the packet to unit 60. By changing the unit 65 destination flag, unit 65 diverts the timing information packet provided by decoder 55 to the unit 60 buffer location assigned to hold data for use by controller 115, instead of an application buffer location.
Upon receiving the system interrupt set by decoder 55, controller 115 reads the timing information and PCR value and stores it in internal memory. PCR values of successive timing information packets are used by controller 115 to adjust the system 25 master clock (27 MHz). The difference between PCR based and master clock based estimates of the time interval between the receipt of successive timing packets, generated by controller 115, is used to adjust the system 25 master clock. Controller 115 achieves this by applying the derived time estimate difference to adjust the input control voltage of a voltage controlled oscillator used to generate the master clock. Controller 115 resets the system interrupt after storing the timing information in internal memory.
Packets received by decoder 55 from units 45 and 50 that contain program content including audio, video, caption, and other information, are directed by unit 65 from decoder 55 to the designated application device buffers in packet buffer 60. Application control unit 70 sequentially retrieves the audio, video, caption and other data from the designated buffers in buffer 60 and provides the data to corresponding application devices 75, 80 and 85. The application devices comprise audio and video decoders 80 and 85 and high speed data port 75. For example, packet data corresponding to a composite program guide generated by the controller 115 as described above, may be transported to the video decoder 85 for formatting into video signal suitable for display on a display monitor 89 connected to the video decoder 85. Also, for example, data port 75 may be used to provide high speed data such as computer programs, for example, to a computer. Alternatively, port 75 may be used to output or receive data to and from an HDTV to display or process images corresponding to a selected program or a program guide, for example.
Packets that contain PSI information are recognized by unit 45 as destined for the controller 115 buffer in unit 60. The PSI packets are directed to this buffer by unit 65 via units 45, 50 and 55 in a similar manner to that described for packets containing program content. Controller 115 reads the PSI from unit 60 and stores it in internal memory.
Controller 115 also generates condensed PSI (CPSI) from the stored PSI and incorporates the CPSI in a packetized data stream suitable for storage on a selectable storage medium. The packet identification and direction is governed by controller 115 in conjunction with the unit 45 and unit 47 PID, destination and encryption key look-up tables and control unit 65 functions in the manner previously described.
In addition, controller 115 is coupled to a communication interface unit 116. Unit 116 provides the capability to upload and download information to and from the Internet. Communication interface unit 116 includes, for example, communication circuitry for connecting to an Internet service provider, e.g., via a wired or wireless connection such as an Ethernet, WiFi connection, or via cable, fiber or telephone line. The communication capability allows the system shown in
A number of different detection techniques exist to detect and separate different audio component types in an audio source. In general, the existing techniques analyze tones in the received audio signal to determine whether or not music, speech, or another component is present. Most, if not all, of these tone-based detection techniques may be separated into two basic categories: (i) stochastic model-based techniques and (ii) deterministic model-based techniques. A discussion of stochastic model-based techniques may be found in, for example, Compure Company, “Music and Speech Detection System Based on Hidden Markov Models and Gaussian Mixture Models,” a Public White Paper, http://www.compure.com. A discussion of deterministic model-based techniques may be found, for example, in U.S. Pat. No. 7,130,795. The teachings of both are incorporated herein by reference in their entirety.
In addition, U.S. Pat. No. 8,554,560, entitled “Voice Activity Detection” provides a further discussion of speech detection. Also, US Pat. App. Pub. No. 20120155655 A1, entitled “Music Detection based on Pause Analysis”, provides a general discussion of audio component types detection and separation techniques, and also proposes another music detection technique based on pause and/or tone analysis. The teachings of both are also incorporated herein by reference in their entirety.
Based on existing techniques and teachings, one skilled in the art can readily appreciate and recognize that the audio component types separator 80 may be implemented in many different ways and employing many different techniques. The separator may be implemented in hardware, software, or a combination of hardware and software, for example, by using software-based filters, hardware-based filters, Digital Signal Processor (DSP), Programmable Logic Arrays (PLA), discrete components, software logics, etc., or a combination of any of the above. In addition, the input to the separator 80 may be in analog form (after D/A 222) or in digital form (before D/A 222). Both possibilities are shown in
In addition, one person in the skilled can readily appreciate and recognize that
According to one aspect of the present invention, a user is able to select one or more, or remove one or more audio component types using, for example, remote control unit 125. In response to the user selection of the audio component types, the system controller 115 generates the corresponding control signal to control a mixer 224 in
In one embodiment, the mixer 224 has another set of audio component inputs associated with a program from another channel, from the Internet, or from a storage. In this embodiment, the system shown in
A user can then select which audio component types he would like to listen to by, e.g., highlighting the respective selection icons 341-343. After the audio component types have been selected, system controller 115 sends the corresponding control signal via control signal path 72 to audio decoder 80 to provide the modified audio signal via mixer 224 and equalizer 225 to speakers 88 as described previously.
Similar to what has been shown and described in connection with
In addition,
The exemplary control program, when executed, facilitates display of user interface shown, for example, in
At step 601 of
In addition, at step 615 of
At step 621 of
While this invention has been described as having a preferred design, the present is intended to cover any variations, uses, or adaptations of the invention using its general principles. Further, this application is intended to cover such departures from the present disclosure as come within known or customary practice in the art to which this invention pertains and which fall within the limits of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
14305724.8 | May 2014 | EP | regional |