The present disclosure relates generally to electronic devices. More specifically, the present disclosure relates to devices for encoding and decoding a watermarked signal.
In the last several decades, the use of electronic devices has become common. In particular, advances in electronic technology have reduced the cost of increasingly complex and useful electronic devices. Cost reduction and consumer demand have proliferated the use of electronic devices such that they are practically ubiquitous in modern society. As the use of electronic devices has expanded, so has the demand for new and improved features of electronic devices. More specifically, electronic devices that perform functions faster, more efficiently or with higher quality are often sought after.
Some electronic devices (e.g., cellular phones, smart phones, computers, etc.) use audio or speech signals. These electronic devices may encode speech signals for storage or transmission. For example, a cellular phone captures a user's voice or speech using a microphone. For instance, the cellular phone converts an acoustic signal into an electronic signal using the microphone. This electronic signal may then be formatted for transmission to another device (e.g., cellular phone, smart phone, computer, etc.) or for storage.
Improved quality or additional capacity in a communicated signal is often sought for. For example, cellular phone users may desire greater quality in a communicated speech signal. However, improved quality or additional capacity may often require greater bandwidth resources and/or new network infrastructure. As can be observed from this discussion, systems and methods that allow efficient signal communication may be beneficial.
An electronic device configured for encoding a watermarked signal is disclosed. The electronic device includes modeler circuitry. The modeler circuitry determines parameters based on a first signal and a first-pass coded signal. The electronic device also includes coder circuitry coupled to the modeler circuitry. The coder circuitry performs a first-pass coding on a second signal to obtain the first-pass coded signal and performs a second-pass coding based on the parameters to obtain a watermarked signal. The electronic device may also include a transmitter for sending the watermarked signal. The first-pass coded signal may be a first-pass coded excitation. The modeler circuitry may determine the parameters based on high band coding. The watermarked signal may be decodable to recover a version of the second signal without information from the first signal.
The electronic device may include an analysis filter bank for dividing a signal into the first signal and the second signal. The first signal may be a higher frequency component signal and the second signal may be a lower frequency component signal.
The coder circuitry may include an adaptive multi-rate narrowband (AMR-NB) coder. The coder circuitry may perform the second-pass coding using a watermarking codebook. The second-pass coding may use a set of linear predictive coding coefficients obtained from the first-pass coding.
An electronic device configured for decoding a watermarked signal is also disclosed. The electronic device includes modeler circuitry that produces a decoded first signal based on a decoded second signal and a watermarked bitstream. The electronic device also includes decoder circuitry coupled to the modeler circuitry that provides the decoded second signal based on the watermarked bitstream. The decoded first signal may include a higher frequency component signal and the decoded second signal may include a lower frequency component signal.
The electronic device may include combining circuitry that combines the decoded first signal and the decoded second signal. The combining circuitry may include a synthesis filter bank.
A method for encoding a watermarked signal on an electronic device is also disclosed. The method includes obtaining a first signal and a second signal. The method also includes performing a first-pass coding on the second signal to obtain a first-pass coded signal. The method further includes determining parameters based on the first signal and the first-pass coded signal. The method additionally includes performing a second-pass coding based on the parameters to obtain a watermarked signal.
A method for decoding a watermarked signal on an electronic device is also disclosed. The method includes decoding a watermarked bitstream to obtain a decoded second signal. The method also includes decoding the watermarked bitstream based on the decoded second signal to obtain a decoded first signal.
A computer-program product for encoding a watermarked signal is also disclosed. The computer-program product includes a non-transitory tangible computer-readable medium with instructions. The instructions include code for causing an electronic device to obtain a first signal and a second signal. The instructions also include code for causing the electronic device to perform a first-pass coding on the second signal to obtain a first-pass coded signal. The instructions further include code for causing the electronic device to determine parameters based on the first signal and the first-pass coded signal. The instructions additionally include code for causing the electronic device to perform a second-pass coding based on the parameters to obtain a watermarked signal.
A computer-program product for decoding a watermarked signal is also disclosed. The computer-program product includes a non-transitory tangible computer-readable medium with instructions. The instructions include code for causing an electronic device to decode a watermarked bitstream to obtain a decoded second signal. The instructions also include code for causing the electronic device to decode the watermarked bitstream based on the decoded second signal to obtain a decoded first signal.
An apparatus for encoding a watermarked signal is also disclosed. The apparatus includes means for obtaining a first signal and a second signal. The apparatus also includes means for performing a first-pass coding on the second signal to obtain a first-pass coded signal. The apparatus further includes means for determining parameters based on the first signal and the first-pass coded signal. The apparatus additionally includes means for performing a second-pass coding based on the parameters to obtain a watermarked signal.
An apparatus for decoding a watermarked signal is also disclosed. The apparatus includes means for decoding a watermarked bitstream to obtain a decoded second signal. The apparatus further includes means for decoding the watermarked bitstream based on the decoded second signal to obtain a decoded first signal.
The systems and methods disclosed herein may be applied to a variety of electronic devices. Examples of electronic devices include voice recorders, video cameras, audio players (e.g., Moving Picture Experts Group-1 (MPEG-1) or MPEG-2 Audio Layer 3 (MP3) players), video players, audio recorders, desktop computers, laptop computers, personal digital assistants (PDAs), gaming systems, etc. One kind of electronic device is a communication device, which may communicate with another device. Examples of communication devices include telephones, laptop computers, desktop computers, cellular phones, smartphones, wireless or wired modems, e-readers, tablet devices, gaming systems, cellular telephone base stations or nodes, access points, wireless gateways and wireless routers.
An electronic device or communication device may operate in accordance with certain industry standards, such as International Telecommunication Union (ITU) standards and/or Institute of Electrical and Electronics Engineers (IEEE) standards (e.g., Wireless Fidelity or “Wi-Fi” standards such as 802.11a, 802.11b, 802.11g, 802.11n and/or 802.11ac). Other examples of standards that a communication device may comply with include IEEE 802.16 (e.g., Worldwide Interoperability for Microwave Access or “WiMAX”), Third Generation Partnership Project (3GPP), 3GPP Long Term Evolution (LTE), Global System for Mobile Telecommunications (GSM) and others (where a communication device may be referred to as a User Equipment (UE), Node B, evolved Node B (eNB), mobile device, mobile station, subscriber station, remote station, access terminal, mobile terminal, terminal, user terminal, subscriber unit, etc., for example). While some of the systems and methods disclosed herein may be described in terms of one or more standards, this should not limit the scope of the disclosure, as the systems and methods may be applicable to many systems and/or standards.
It should be noted that some communication devices may communicate wirelessly and/or may communicate using a wired connection or link. For example, some communication devices may communicate with other devices using an Ethernet protocol. The systems and methods disclosed herein may be applied to communication devices that communicate wirelessly and/or that communicate using a wired connection or link. In one configuration, the systems and methods disclosed herein may be applied to a communication device that communicates with another device using a satellite.
One configuration of the systems and methods may be used for the extension of code-excited linear prediction (CELP) speech coders using watermarking techniques to embed data that is dependent on the original carrier bit stream. More simply, the systems and methods disclosed herein may provide watermarking for the extension of CELP codecs.
Wideband (e.g., 0-7 kilohertz (kHz)) coding of speech provides superior quality to narrowband (e.g., 0-4 kHz) coding of speech. However, the majority of existing mobile communication networks support narrowband coding only (e.g., adaptive multi-rate narrowband (AMR-NB)). Deploying wideband coders (e.g., adaptive multi-rate wideband (AMR-WB)) may require substantial and costly changes to infrastructure and service deployment.
Furthermore, the next generation of services may support wideband coders (e.g., AMR-WB), while super-wideband (e.g., 0-14 kHz) coders are being developed and standardized. Again, operators may eventually face the costs of deploying yet another codec to move customers to super-wideband.
One configuration of the systems and methods disclosed herein may use an advanced model that can encode additional bandwidth very efficiently and hide this information in a bitstream already supported by existing network infrastructure. The information hiding may be performed by watermarking the bitstream. One example of this technique watermarks the fixed codebook of a CELP coder. For example, the upper band of a wideband input (e.g., 4-7 kHz) may be encoded and carried as a watermark in a narrowband coder's bitstream. In another example, the upper band of a super-wideband input (e.g., 7-14 kHz) may be encoded and carried as a watermark in a wideband coder's bitstream. Other secondary bitstreams, perhaps unrelated to bandwidth extension, may be carried as well. One example that faces similar challenges is the inclusion of parametric stereo data embedded in a monophonic stream. This technique allows the encoder to produce a bitstream compatible with existing infrastructures. A legacy decoder may produce a narrowband output with a quality similar to standard encoded speech (without the watermark, for example), while a decoder that is aware of the watermark may produce wideband speech.
Several technical obstacles in watermarking the information for bandwidth extension have remained that have impeded development of practical systems. Importantly, an encoding model that is sufficiently efficient and a means for applying it to the problem have not been readily available or obvious.
In order to increase or maximize quality, the watermarked information should be as small as possible in order to minimize its impact on the quality of the original bitstream (e.g., a “carrier” bitstream containing the low band). This can be achieved using an advanced model for the high-band, such as the efficient non-linear extension model used in the enhanced variable rate wideband codec (EVRC-WB). However, this model relies on the low band excitation for generating the high-band speech parameters, and consequently the high-band bits. However, the low band excitation is affected by the high-band bits through the watermarking process. Therefore, approximations may be made to escape this loop.
In accordance with the systems and methods disclosed herein, a first-pass of the carrier encoder may be conducted, with no watermark. The resultant signal (e.g., excitation, residual, etc.) is used for calculating the embedded parameters (e.g., the high-band model parameters or other data such as parametric stereo). Then, a second-pass of the carrier encoder is performed, with the watermark (from the embedded parameters) applied to the low-band encoding process. In this way, the cyclical dependency is broken. Running two passes of the encoder may not be an issue, as the complexity of the legacy narrower bandwidth codec is generally quite small compared to current state-of-the art codecs that encode wider bandwidths.
One alternative to this approach would be to use the linear predictive coding (LPC) residual instead of the coded first-pass residual from the carrier encoder as the input to the high-band model. However, this degrades quality, as there may be a bigger mismatch between the signal used to calculate the high-band parameters, and the signal that will be used eventually at the decoder.
Any other solutions to the cyclical dependency problem are currently unknown. However, one alternative would be to use a high-band encoding technique that does not depend on the low band. However, it is unlikely that such a technique would be as efficient as one which leverages the low-band to extrapolate the high band. With this inefficiency, the quality impact of the watermark on the low-band carrier bitstream would likely be more significant.
Various configurations are now described with reference to the Figures, where like element names may indicate functionally similar elements. The systems and methods as generally described and illustrated in the Figures herein could be arranged and designed in a wide variety of different configurations. Thus, the following more detailed description of several configurations, as represented in the Figures, is not intended to limit scope, as claimed, but is merely representative of the systems and methods.
Electronic device A 102 may include an encoder block/module 110 and/or a communication interface 124. The encoder block/module 110 may be used to encode and watermark a signal. The communication interface 124 may transmit one or more signals to another device (e.g., electronic device B 134).
Electronic device A 102 may obtain one or more signals A 104, such as audio or speech signals. For example, electronic device A 102 may capture signal A 104 using a microphone or may receive signal A 104 from another device (e.g., a Bluetooth headset). In some configurations, signal A 104 may be divided into different component signals (e.g., a higher frequency component signal and a lower frequency component signal, a monophonic signal and a stereo signal, etc.). In other configurations, unrelated signals A 104 may be obtained. Signal(s) A 104 may be provided to modeler circuitry 112 and coder circuitry 118 in an encoder 110. For example, a first signal (e.g., signal component) 106 may be provided to the modeler circuitry 112, while a second signal (e.g., another signal component) 108 is provided to the coder circuitry 118.
It should be noted that one or more of the elements 110, 112, 118, 124 included in electronic device A 102 may be implemented in hardware, software or a combination of both. For instance, the term “circuitry” as used herein may indicate that an element may be implemented using one or more circuit components, including processing blocks and/or memory cells. Thus, one or more of the elements 110, 112, 118, 124 included in electronic device A 102 may be implemented as one or more integrated circuits, application specific integrated circuits (ASICs), etc., and/or using a processor and instructions. It should also be noted that the term “block/module” may be used to indicate that an element may be implemented in hardware, software or a combination of both.
The coder circuitry 118 may perform coding on the second signal 108. For example, the coder circuitry 118 may perform adaptive multi-rate (AMR) coding on the second signal 108. The modeler circuitry 112 may determine or calculate parameters or data 116 that may be embedded into the second signal (e.g., “carrier” signal) 108. For example, the coder circuitry 118 may produce a coded bitstream that watermark bits may be embedded into. In another example, the modeler circuitry 112 may separately encode the first signal 106 into bits 116 that can be embedded into the coded bitstream. In some configurations, the modeler circuitry 112 may determine the parameters or data 116 based on high band coding. For instance, the modeler circuitry 112 may use a high band part of the enhanced variable rate wideband (EVRC-WB) codec. Other high band coding techniques may be used. The coded second signal 108 with the embedded watermark signal may be referred to as a watermarked second signal 122.
The coder circuitry 118 may perform a first-pass coding on the second signal 108. This first-pass coding may produce data 114 (e.g., a first-pass coded signal, a first-pass coded excitation 114, etc.), which may be provided to the modeler circuitry 112. In one configuration, the modeler circuitry 112 may use an EVRC-WB model to model higher frequency components (from the first signal 106) that relies on lower frequency components (from the second signal 108) that may be encoded by the coder circuitry 118. Thus, the first-pass coded excitation 114 may be provided to the modeler circuitry 112 for use in modeling the higher frequency components. The resulting higher frequency component parameters or bits 116 may then be embedded into the second signal 108 in a second-pass coding, thereby producing the watermarked second signal 122. For example, the second-pass coding may involve the use of a watermarking codebook (e.g., fixed codebook or FCB) 120 to embed high-band bits 116 into a coded second signal 108 to produce the watermarked second signal (e.g., a watermarked bitstream) 122.
It should be noted that the watermarking process may alter some of the bits of an encoded second signal 108. For example, the second signal 108 may be referred to as a “carrier” signal or bitstream. In the watermarking process, some of the bits that make up the encoded second signal 108 may be altered in order to embed or insert the data or bits 116 derived from the first signal 106 into the second signal 108 to produce the watermarked second signal 122. In some cases, this may be a source of degradation in the encoded second signal 108. However, this approach may be advantageous since decoders that are not designed to extract the watermarked information may still recover a version of the second signal 108, without the extra information provided by the first signal 106. Thus, “legacy” devices and infrastructure may still function regardless of the watermarking. This approach further allows other decoders (that are designed to extract the watermarked information) to be used to extract the additional watermark information provided by the first signal 106.
The watermarked second signal (e.g., bitstream) 122 may be provided to the communication interface 124. Examples of the communication interface 124 may include transceivers, network cards, wireless modems, etc. The communication interface 124 may be used to communicate (e.g., transmit) the watermarked second signal 122 to another device, such as electronic device B 134 over a network 128. For example, the communication interface 124 may be based on wired and/or wireless technology. Some operations performed by the communication interface 124 may include modulation, formatting (e.g., packetizing, interleaving, scrambling, etc.), upconversion, amplification, etc. Thus, electronic device A 102 may transmit a signal 126 that comprises the watermarked second signal 122.
The signal 126 (including the watermarked second signal 122) may be sent to one or more network devices 130. For example, a network 128 may include the one or more network devices 130 and/or transmission mediums for communicating signals between devices (e.g., between electronic device A 102 and electronic device B 134). In the configuration illustrated in
In some cases, one or more network devices 130 may transcode the signal 126 (that includes the watermarked second signal 122). Transcoding may include decoding the transmitted signal 126 and re-encoding it (into another format, for example). In some cases, transcoding the signal 126 may destroy the watermark information embedded in the signal 126. In such a case, electronic device B 134 may receive a signal that no longer contains the watermark information. Other network devices 130 may not use any transcoding. For instance, if a network 128 uses devices that do not transcode signals, the network may provide tandem-free/transcoder-free operation (TFO/TrFO). In this case, the watermark information embedded in the watermarked second signal 122 may be preserved as it is sent to another device (e.g., electronic device B 134).
Electronic device B 134 may receive a signal 132 (via the network 128), such as a signal 132 having watermark information preserved or a signal 132 without watermark information. For instance, electronic device B 134 may receive a signal 132 using a communication interface 136. Examples of the communication interface 136 may include transceivers, network cards, wireless modems, etc. The communication interface 136 may perform operations such as downconversion, synchronization, de-formatting (e.g., de-packetizing, unscrambling, de-interleaving, etc.) on the signal 132. The resulting signal 138 (e.g., a bitstream from the received signal 132) may be provided to a decoder block/module 140. For example, the signal 138 may be provided to modeler circuitry 142 and to decoder circuitry 150.
If watermarked information is embedded on the signal 138, the modeler circuitry 142 may model and/or decode the watermark information (e.g., watermark bits) embedded on the signal (e.g., bitstream) 138. For example, the decoder 140 may extract watermark bits from the signal 138. The modeler circuitry 142 may decode these watermark bits to produce a decoded first signal 154, 144.
The decoder circuitry 150 may decode the signal 138. In some configurations, the decoder circuitry 150 may use a “legacy” decoder (e.g., a standard narrowband decoder) or decoding procedure that decodes the signal 138 regardless of any watermark information that may be included in the signal 138. The decoder circuitry 150 may produce a decoded second signal 148, 152, 158. Thus, for example, if no watermark information is included in the signal 138, the decoder circuitry 150 may still recover a version of the second signal 108, which is the decoded second signal 158.
In some configurations, the operations performed by the modeler circuitry 142 may depend on operations performed by the decoder circuitry 150. For example, a model (e.g., EVRC-WB) used for a higher frequency band may depend on a decoded narrowband signal 152 (decoded using AMR-NB, for example). In this case, the decoded narrowband signal 152 may be provided to the modeler circuitry 142.
In some configurations, a decoded second signal 148 may be combined with a decoded first signal 144 by a combining block/module 146 (e.g., combining circuitry 146) to produce a combined signal 156. In other configurations, the watermark bits from the signal 138 and the signal (itself) 138 may be decoded separately to produce the decoded first signal 154 and the decoded second signal 158. Thus, one or more signals B 160 may include a decoded first signal 154 and a separate decoded second signal 158 and/or may include a combined signal 156. It should be noted that the decoded first signal 154, 144 may be a decoded version of the first signal 106 encoded by electronic device A 102. Additionally or alternatively, the decoded second signal 148, 152, 158 may be a decoded version of the second signal 108 encoded by electronic device A 102.
If no watermarked information is embedded in the received signal 132, the decoder circuitry 150 may decode the signal 138 (in a legacy mode, for example) to produce the decoded second signal 158. This may provide a decoded second signal 158, without the additional information provided by the first signal 106. This may occur, for example, if the watermark information (from the first signal 106, for example) is destroyed in a transcoding process in the network 128.
In some configurations, electronic device B 134 may be incapable of decoding the watermark signal or bits embedded in the received signal 132. For example, electronic device B 134 may not include modeler circuitry 142 for extracting the embedded watermark signal in some configurations. In such a case, electronic device B 134 may simply decode the signal 138 to produce the decoded second signal 158.
It should be noted that one or more of the elements 140, 142, 146, 150, 136 included in electronic device B 134 may be implemented in hardware (e.g., circuitry), software or a combination of both. For instance, one or more of the elements 140, 142, 146, 150, 136 included in electronic device B 134 may be implemented as one or more integrated circuits, application specific integrated circuits (ASICs), etc., and/or using a processor and instructions.
The electronic device 102 may perform 204 a first-pass coding on the second signal 108 to obtain a first-pass coded signal 114. For example, the electronic device may perform AMR-NB encoding on the second signal 108 to obtain the first-pass coded signal 114. In some configurations, the first-pass coded signal 114 may be an excitation signal, while in other configurations (e.g., embedding parametric stereo), the first-pass coded signal 114 may not be an excitation signal. In the first pass, a full encoding may be performed in some configurations. In the case of bandwidth extension, for instance, the first-pass coded signal 114 that is used by a non-linear model (e.g., the modeler circuitry 112) is an excitation. In the case of parametric stereo, for example, the first-pass coded signal 114 may be an actual coded speech signal. It should also be noted that the electronic device 102 may generate linear predictive coding (LPC) coefficients in the first-pass coding that may be used in a second-pass coding (in some configurations).
The electronic device 102 may determine 206 parameters (e.g., parameters, data, bits, etc.) 116 based on the first signal 106 and the first-pass coded signal 114. For example, in the case where the additional information that is to be embedded on the carrier signal (e.g., second signal 108) contains higher frequency components of a speech signal, the electronic device 102 may model or determine the parameters 116 for the higher frequency components (e.g., the first signal 106) based on a first-pass coded excitation 114. In some configurations, the electronic device 102 may determine 206 the parameters based on high band coding. For instance, the electronic device 102 may use EVRC-WB (e.g., a high band part of the EVRC-WB codec) modeling of the first signal 106 (e.g., higher frequency component signal) to generate the parameters 116. Other high band coding techniques may be used.
The electronic device 102 may then perform 208 a second-pass coding based on the parameters 116 to obtain a watermarked second signal 122. For example, the electronic device 102 may use the modeling parameters 116 in conjunction with a watermarking codebook 120 to generate the watermarked second signal 122 (e.g., embed the watermark information). In some configurations, the second pass may also use LPC coefficients (e.g., line spectral frequencies (LSFs) or line spectral pairs (LSPs)) generated from the first-pass coding to produce the watermarked second signal 122.
The electronic device 102 may send 210 the watermarked second signal 122. For example, the electronic device 102 may transmit a signal 126 comprising the watermarked second signal 122 to another device (e.g., electronic device B 134) via a network 128.
The electronic device 134 may obtain 304 a watermarked bitstream 138 from the signal 132. For example, the electronic device 134 may perform one or more operations to extract the watermarked bitstream 138 from the received signal 132. For instance, the electronic device 134 may downconvert, amplify, channel decode, demodulate, de-format (e.g., de-interleave, unscramble, etc.), etc., the received signal 132 in order to obtain 304 the watermarked bitstream 138.
The electronic device 134 may decode 306 the watermarked bitstream 138 in order to obtain a decoded second signal 148, 152, 158. For example, the electronic device 134 may decode 306 the watermarked bitstream 138 using a “legacy” decoder. For instance, the electronic device 134 may use an adaptive multi-rate (AMR) narrowband (NB) decoder to obtain the decoded second signal 152.
The electronic device 134 may decode 308 the watermarked bitstream 138 based on the decoded second signal 152 to obtain a decoded first signal 144, 154. In some configurations, for example, a model (e.g., EVRC-WB) used for a higher frequency band may depend on a decoded narrowband signal 152 (decoded using AMR-NB, for example). In this case, the electronic device 134 may use the decoded second signal 152 to model or decode the watermarked bitstream 138 (e.g., extracted watermark bits) to obtain a decoded first signal 154, 144.
The electronic device 134 may combine 310 the decoded first signal 144 and the decoded second signal 148. In some configurations, for example, the electronic device 134 may combine 310 the decoded first signal 144 and the decoded second signal 148 using a synthesis filter bank, which may produce a combined signal 156.
Wireless communication device A 402 may include a microphone 462, an audio encoder 410, a channel encoder 466, a modulator 468, a transmitter 472 and one or more antennas 474a-n. The audio encoder 410 may be used for encoding and watermarking audio. The channel encoder 466, modulator 468, transmitter 472 and one or more antennas 474a-n may be used to prepare and transmit one or more signals to another device (e.g., wireless communication device B 434).
Wireless communication device A 402 may obtain an audio signal 404. For example, wireless communication device A 402 may capture the audio signal 404 (e.g., speech) using a microphone 462. The microphone 462 may convert an acoustic signal (e.g., sounds, speech, etc.) into the electrical or electronic audio signal 404. The audio signal 404 may be provided to the audio encoder 410, which may include an analysis filter bank 464, a high-band modeling block/module 412 and a coding with watermarking block/module 418.
The audio signal 404 may be provided to the analysis filter bank 464. The analysis filter bank 464 may divide the audio signal 404 into a first signal 406 and a second signal 408. For example, the first signal 406 may be a higher frequency component signal and the second signal 408 may be a lower frequency component signal. The first signal 406 may be provided to the high-band modeling block/module 412. The second signal 408 may be provided to the coding with watermarking block/module 418.
It should be noted that one or more of the elements 410, 412, 418, 464, 466, 468, 472 included in wireless communication device A 402 may be implemented in hardware, software or a combination of both. For instance, one or more of the elements 410, 412, 418, 464, 466, 468, 472 included in wireless communication device A 402 may be implemented as one or more integrated circuits, application specific integrated circuits (ASICs), etc., and/or using a processor and instructions. It should also be noted that the term “block/module” may also be used to indicate that an element may be implemented in hardware, software or a combination of both.
The coding with watermarking block/module 418 may perform coding on the second signal 408. For example, the coding with watermarking block/module 418 may perform adaptive multi-rate (AMR) coding on the second signal 408. The high-band modeling block/module 412 may determine or calculate parameters or data 416 that may be embedded into the second signal (e.g., “carrier” signal) 408. For example, the coding with watermarking block/module 418 may produce a coded bitstream that watermark bits may be embedded into. The coded second signal 408 with the embedded watermark signal may be referred to as a watermarked second signal 422.
The coding with watermarking block/module 418 may perform a first-pass coding on the second signal 408. This first-pass coding may produce a first-pass coded excitation 414, for example, which may be provided to the high-band modeling block/module 412. In one configuration, the high-band modeling block/module 412 may use an EVRC-WB model to model higher frequency components (from the first signal 406) that relies on lower frequency components (from the second signal 408) that may be encoded by the coding with watermarking block/module 418. Thus, the first-pass coded excitation 414 may be provided to the high-band modeling block/module 412 for use in modeling the higher frequency components. The resulting higher frequency component parameters or bits 416 may then be embedded into the second signal 408 in a second-pass coding, thereby producing the watermarked second signal 422. For example, the second-pass coding may involve the use of a watermarking codebook (e.g., fixed codebook or FCB) 420 to embed high-band bits 416 into a coded second signal 408 to produce the watermarked second signal (e.g., a watermarked bitstream) 422.
The watermarked second signal (e.g., bitstream) 422 may be provided to the channel encoder 466. The channel encoder 466 may encode the watermarked second signal 422 to produce a channel-encoded signal 468. For example, the channel encoder 466 may add error detection coding (e.g., a cyclic redundancy check (CRC)) and/or error correction coding (e.g., forward error correction (FEC) coding) to the watermarked second signal 422.
The channel-encoded signal 468 may be provided to the modulator 468. The modulator 468 may modulate the channel-encoded signal 468 to produce a modulated signal 470. For example, the modulator 468 may map bits in the channel-encoded signal 468 to constellation points. For instance, the modulator 468 may apply a modulation scheme to the channel-encoded signal 468 such as binary phase-shift keying (BPSK), quadrature amplitude modulation (QAM), frequency-shift keying (FSK), etc., to produce the modulated signal 470.
The modulated signal 470 may be provided to the transmitter 472. The transmitter 472 may transmit the modulated signal 470 using the one or more antennas 474a-n. For example, the transmitter 472 may upconvert, amplify and transmit the modulated signal 470 using the one or more antennas 474a-n.
The modulated signal 470 that includes the watermarked second signal 422 (e.g., a “transmitted signal”) may be transmitted from wireless communication device A 402 to another device (e.g., wireless communication device B 434) over a network 428. The network 428 may include the one or more network 428 devices and/or transmission mediums for communicating signals between devices (e.g., between wireless communication device A 402 and wireless communication device B 434). For example, the network 428 may include one or more base stations, routers, servers, bridges, gateways, etc.
In some cases, one or more network 428 devices may transcode the transmitted signal (that includes the watermarked second signal 422). Transcoding may include decoding the transmitted signal and re-encoding it (into another format, for example). In some cases, transcoding may destroy the watermark information embedded in the transmitted signal. In such a case, wireless communication device B 434 may receive a signal that no longer contains the watermark information. Other network 428 devices may not use any transcoding. For instance, if a network 428 uses devices that do not transcode signals, the network may provide tandem-free/transcoder-free operation (TFO/TrFO). In this case, the watermark information embedded in the watermarked second signal 422 may be preserved as it is sent to another device (e.g., wireless communication device B 434).
Wireless communication device B 434 may receive a signal (via the network 428), such as a signal having watermark information preserved or a signal without watermark information. For instance, wireless communication device B 434 may receive a signal using one or more antennas 476a-n and a receiver 478. In one configuration, the receiver 478 may downconvert and digitize the signal to produce a received signal 480.
The received signal 480 may be provided to a demodulator 482. The demodulator 482 may demodulate the received signal 480 to produce a demodulated signal 484, which may be provided to a channel decoder 486. The channel decoder 486 may decode the signal (e.g., detect and/or correct errors using error detection and/or correction codes) to produce a (decoded) signal 438.
The signal 438 (e.g., a bitstream) may be provided to an audio decoder 440. For example, the signal 438 may be provided to a high-band modeling block/module 442 and to a decoding block/module 450.
If watermarked information is embedded on the signal 438 (e.g., if the watermarked information has not been lost in transmission), the high-band modeling block/module 442 may model and/or decode the watermark information (e.g., watermark bits) embedded on the signal (e.g., bitstream) 438. For example, the audio decoder 440 may extract watermark bits from the signal 438. The high-band modeling block/module 442 may decode these watermark bits to produce a decoded first signal 444.
The decoding block/module 450 may decode the signal 438. In some configurations, the decoding block/module 450 may use a “legacy” decoder (e.g., a standard narrowband decoder) or decoding procedure that decodes the signal 438 regardless of any watermark information that may be included in the signal 438. The decoding block/module 450 may produce a decoded second signal 448, 452. Thus, for example, if no watermark information is included in the signal 438, the decoding block/module 450 may still recover a version of the second signal 408, which is the decoded second signal 448.
The operations performed by the high-band modeling block/module 442 may depend on operations performed by the decoding block/module 450. For example, a model (e.g., EVRC-WB) used for a higher frequency band may depend on a decoded narrowband signal 452 (decoded using AMR-NB, for example). In this case, the decoded narrowband signal 452 may be provided to the high-band modeling block/module 442.
In some configurations, a decoded second signal 448 may be combined with a decoded first signal 444 by a synthesis filter bank 446 to produce a combined signal 456. For example, the decoded first signal 444 may include higher frequency audio information, while the decoded second signal 448 may include lower frequency audio information. It should be noted that the decoded first signal 444 may be a decoded version of the first signal 406 encoded by wireless communication device A 402. Additionally or alternatively, the decoded second signal 448 may be a decoded version of the second signal 408 encoded by wireless communication device A 402. The synthesis filter bank 446 may combine the decoded first signal 444 and the decoded second signal 448 to produce the combined signal 456, which may be a wide-band audio signal.
The combined signal 456 may be provided to a speaker 488. The speaker 488 may be a transducer that converts electrical or electronic signals into acoustic signals. For instance, the speaker 488 may convert an electronic wide-band audio signal (e.g., the combined signal 456) into an acoustic wide-band audio signal.
If no watermarked information is embedded in the signal 438, the audio decoding block/module 450 may decode the signal 438 (in a legacy mode, for example) to produce the decoded second signal 448. In this case, the synthesis filter bank 446 may be bypassed to provide the decoded second signal 448, without the additional information provided by the first signal 406. This may occur, for example, if the watermark information (from the first signal 406, for example) is destroyed in a transcoding process in the network 428.
It should be noted that one or more of the elements 440, 446, 442, 450, 486, 482, 478 included in wireless communication device B 434 may be implemented in hardware, software or a combination of both. For instance, one or more of the elements 440, 446, 442, 450, 486, 482, 478 included in wireless communication device B 434 may be implemented as one or more integrated circuits, application specific integrated circuits (ASICs), etc., and/or using a processor and instructions.
The second signal 508 or lower frequency component (e.g., 0-4 kHz) may be provided to a modified narrowband encoder (e.g., AMR-NB 12.2 with a fixed codebook (FCB) watermark). The modified narrowband coder 518 may perform a first-pass coding on the second signal 508 (e.g., lower frequency components) to produce a first-pass coded excitation 514 that is provided to the high-band modeling block/module 512.
The first signal 506 or higher frequency component may also be provided to the high-band modeling block/module 512 (that uses a high band part of the EVRC-WB codec, for example). The high-band modeling block/module 512 may encode or model the first signal 506 (e.g., higher frequency component) based on the first-pass coded excitation 514 provided by the modified narrowband coder 518. The encoding or modeling performed by the high-band modeling block/module 512 may produce high-band bits 516 that are provided to the modified narrowband coder 518.
The modified narrowband coder 518 may embed the high-band bits 516 as a watermark on the second signal 508. For example, the modified narrowband coder 518 may perform a second-pass coding, where the second signal 508 is encoded and the high-band bits 516 are embedded onto the encoded second signal 508 using a watermarking fixed codebook (FCB). Performing the second-pass coding may produce the watermarked second signal 522 or bitstream. It should be noted that the watermarked second signal 522 (e.g., bitstream) may be decodable by a standard (e.g., conventional) decoder, such as standard AMR. However, if a decoder does not include watermark decoding functionality, it may only be able to decode a version of the second signal 508 (e.g., lower frequency component).
The high-band modeling block/module 642 may extract and/or model watermark information embedded in the watermarked second signal 638 using the lower frequency component signal 652) to obtain a decoded first signal 644 (e.g., a higher frequency component signal ranging from 4-8 kHz). The decoded first signal 644 and the decoded second signal 648 may be combined by a synthesis filter bank 646 to obtain a wideband (e.g., 0-8 kHz, 16 kHz sampled) output speech signal 656. However, in a “legacy” case or a case that a received bitstream does not contain the watermark signal or bits (instead of the watermarked second signal 638), the watermarking decoder 640 may produce a narrowband (e.g., 0-4 kHz) speech output signal (e.g., the decoded second signal 648).
First-pass coding 790 may be performed on a second signal 708, such as a signal in a lower frequency band ranging from 0-4 kHz, for example. In the first-pass coding 790, a linear predictive coding (LPC) operation 792, a first long term prediction (LTP) operation (e.g., LTP A) 794a and a fixed codebook (FCB) operation 796 may be performed on the second signal 708 to obtain a first-pass coded excitation 714. LPC coefficients 703 from the first-pass coding 790 may be provided (e.g., stored) for the second-pass coding 707.
The first-pass coded excitation 714 may be provided to an EVRC-WB high-band modeling block/module 712, which models a first signal 706, such as a higher frequency component signal ranging from 4-8 kHz to produce high-band bits 705. The second-pass coding 707 may be performed using the LPC coefficients 703 from the first-pass coding 790. For example, a second LTP operation (e.g., LTP B) 794b is performed on the LPC coefficients 703 from the first-pass coding 790. The high-band bits 705 and the output of the second LTP operation 794b are used in a watermarked FCB operation 798 to generate the watermarked second signal 722 (e.g., a coded and watermarked bitstream). For example, the watermarked FCB 798 may be used to embed the high-band bits 705 into the carrier (e.g., second signal 708) bitstream to produce the watermarked second signal 722.
The audio codec 819 may be an electronic device (e.g., integrated circuit) used for coding and/or decoding audio signals. The audio codec 819 may be coupled to one or more speakers 811, an earpiece 813, an output jack 815 and/or one or more microphones 817. The speakers 811 may include one or more electro-acoustic transducers that convert electrical or electronic signals into acoustic signals. For example, the speakers 811 may be used to play music or output a speakerphone conversation, etc. The earpiece 813 may be another speaker or electro-acoustic transducer that can be used to output acoustic signals (e.g., speech signals) to a user. For example, the earpiece 813 may be used such that only a user may reliably hear the acoustic signal. The output jack 815 may be used for coupling other devices to the wireless communication device 809 for outputting audio, such as headphones. The speakers 811, earpiece 813 and/or output jack 815 may generally be used for outputting an audio signal from the audio codec 819. The one or more microphones 817 may be one or more acousto-electric transducers that convert an acoustic signal (such as a user's voice) into electrical or electronic signals that are provided to the audio codec 819.
The audio codec 819 may include a watermarking encoder 821. The encoders 110, 410, 510 described above may be examples of the watermarking encoder 821. The watermarking encoder 821 may be used to perform the methods 200 described above in connection with
The audio codec 819 may additionally or alternatively include a decoder 823. The decoders 140, 440, 640 described above may be examples of the decoder 823. The decoder 823 may perform the method 300 described above in connection with
The application processor 825 may also be coupled to a power management circuit 835. One example of the power management circuit 835 is a power management integrated circuit (PMIC), which may be used to manage the electrical power consumption of the wireless communication device 809. The power management circuit 835 may be coupled to a battery 837. The battery 837 may generally provide electrical power to the wireless communication device 809.
The application processor 825 may be coupled to one or more input devices 839 for receiving input. Examples of input devices 839 include infrared sensors, image sensors, accelerometers, touch sensors, keypads, etc. The input devices 839 may allow user interaction with the wireless communication device 809. The application processor 825 may also be coupled to one or more output devices 841. Examples of output devices 841 include printers, projectors, screens, haptic devices, etc. The output devices 841 may allow the wireless communication device 809 to produce output that may be experienced by a user.
The application processor 825 may be coupled to application memory 843. The application memory 843 may be any electronic device that is capable of storing electronic information. Examples of application memory 843 include double data rate synchronous dynamic random access memory (DDRAM), synchronous dynamic random access memory (SDRAM), flash memory, etc. The application memory 843 may provide storage for the application processor 825. For instance, the application memory 843 may store data and/or instructions for the functioning of programs that are run on the application processor 825.
The application processor 825 may be coupled to a display controller 845, which in turn may be coupled to a display 847. The display controller 845 may be a hardware block that is used to generate images on the display 847. For example, the display controller 845 may translate instructions and/or data from the application processor 825 into images that can be presented on the display 847. Examples of the display 847 include liquid crystal display (LCD) panels, light emitting diode (LED) panels, cathode ray tube (CRT) displays, plasma displays, etc.
The application processor 825 may be coupled to a baseband processor 827. The baseband processor 827 generally processes communication signals. For example, the baseband processor 827 may demodulate and/or decode received signals. Additionally or alternatively, the baseband processor 827 may encode and/or modulate signals in preparation for transmission.
The baseband processor 827 may be coupled to baseband memory 849. The baseband memory 849 may be any electronic device capable of storing electronic information, such as SDRAM, DDRAM, flash memory, etc. The baseband processor 827 may read information (e.g., instructions and/or data) from and/or write information to the baseband memory 849. Additionally or alternatively, the baseband processor 827 may use instructions and/or data stored in the baseband memory 849 to perform communication operations.
The baseband processor 827 may be coupled to a radio frequency (RF) transceiver 829. The RF transceiver 829 may be coupled to a power amplifier 831 and one or more antennas 833. The RF transceiver 829 may transmit and/or receive radio frequency signals. For example, the RF transceiver 829 may transmit an RF signal using a power amplifier 831 and one or more antennas 833. The RF transceiver 829 may also receive RF signals using the one or more antennas 833. The wireless communication device 809 may be one example of an electronic device 102, 134, or wireless communication device 402, 434 as described herein.
The electronic device 951 also includes memory 953 in electronic communication with the processor 959. That is, the processor 959 can read information from and/or write information to the memory 953. The memory 953 may be any electronic component capable of storing electronic information. The memory 953 may be random access memory (RAM), read-only memory (ROM), magnetic disk storage media, optical storage media, flash memory devices in RAM, on-board memory included with the processor, programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically erasable PROM (EEPROM), registers, and so forth, including combinations thereof.
Data 957a and instructions 955a may be stored in the memory 953. The instructions 955a may include one or more programs, routines, sub-routines, functions, procedures, etc. The instructions 955a may include a single computer-readable statement or many computer-readable statements. The instructions 955a may be executable by the processor 959 to implement one or more of the methods 200, 300 described above. Executing the instructions 955a may involve the use of the data 957a that is stored in the memory 953.
The electronic device 951 may also include one or more communication interfaces 963 for communicating with other electronic devices. The communication interfaces 963 may be based on wired communication technology, wireless communication technology, or both. Examples of different types of communication interfaces 963 include a serial port, a parallel port, a Universal Serial Bus (USB), an Ethernet adapter, an IEEE 1394 bus interface, a small computer system interface (SCSI) bus interface, an infrared (IR) communication port, a Bluetooth wireless communication adapter, and so forth.
The electronic device 951 may also include one or more input devices 965 and one or more output devices 969. Examples of different kinds of input devices 965 include a keyboard, mouse, microphone, remote control device, button, joystick, trackball, touchpad, lightpen, etc. For instance, the electronic device 951 may include one or more microphones 967 for capturing acoustic signals. In one configuration, a microphone 967 may be a transducer that converts acoustic signals (e.g., voice, speech) into electrical or electronic signals. Examples of different kinds of output devices 969 include a speaker, printer, etc. For instance, the electronic device 951 may include one or more speakers 971. In one configuration, a speaker 971 may be a transducer that converts electrical or electronic signals into acoustic signals. One specific type of output device which may be typically included in an electronic device 951 is a display device 973. Display devices 973 used with configurations disclosed herein may utilize any suitable image projection technology, such as a cathode ray tube (CRT), liquid crystal display (LCD), light-emitting diode (LED), gas plasma, electroluminescence, or the like. A display controller 975 may also be provided, for converting data stored in the memory 953 into text, graphics, and/or moving images (as appropriate) shown on the display device 973.
The various components of the electronic device 951 may be coupled together by one or more buses, which may include a power bus, a control signal bus, a status signal bus, a data bus, etc. For simplicity, the various buses are illustrated in
The wireless communication device 1077 includes a processor 1097. The processor 1097 may be a general purpose single- or multi-chip microprocessor (e.g., an ARM), a special purpose microprocessor (e.g., a digital signal processor (DSP)), a microcontroller, a programmable gate array, etc. The processor 1097 may be referred to as a central processing unit (CPU). Although just a single processor 1097 is shown in the wireless communication device 1077 of
The wireless communication device 1077 also includes memory 1079 in electronic communication with the processor 1097 (i.e., the processor 1097 can read information from and/or write information to the memory 1079). The memory 1079 may be any electronic component capable of storing electronic information. The memory 1079 may be random access memory (RAM), read-only memory (ROM), magnetic disk storage media, optical storage media, flash memory devices in RAM, on-board memory included with the processor, programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically erasable PROM (EEPROM), registers, and so forth, including combinations thereof.
Data 1081a and instructions 1083a may be stored in the memory 1079. The instructions 1083a may include one or more programs, routines, sub-routines, functions, procedures, code, etc. The instructions 1083a may include a single computer-readable statement or many computer-readable statements. The instructions 1083a may be executable by the processor 1097 to implement one or more of the methods 200, 300 described above. Executing the instructions 1083a may involve the use of the data 1081a that is stored in the memory 1079.
The wireless communication device 1077 may also include a transmitter 1093 and a receiver 1095 to allow transmission and reception of signals between the wireless communication device 1077 and a remote location (e.g., another electronic device, wireless communication device, etc.). The transmitter 1093 and receiver 1095 may be collectively referred to as a transceiver 1091. An antenna 1099 may be electrically coupled to the transceiver 1091. The wireless communication device 1077 may also include (not shown) multiple transmitters, multiple receivers, multiple transceivers and/or multiple antenna.
In some configurations, the wireless communication device 1077 may include one or more microphones 1085 for capturing acoustic signals. In one configuration, a microphone 1085 may be a transducer that converts acoustic signals (e.g., voice, speech) into electrical or electronic signals. Additionally or alternatively, the wireless communication device 1077 may include one or more speakers 1087. In one configuration, a speaker 1087 may be a transducer that converts electrical or electronic signals into acoustic signals.
The various components of the wireless communication device 1077 may be coupled together by one or more buses, which may include a power bus, a control signal bus, a status signal bus, a data bus, etc. For simplicity, the various buses are illustrated in
In the above description, reference numbers have sometimes been used in connection with various terms. Where a term is used in connection with a reference number, this may be meant to refer to a specific element that is shown in one or more of the Figures. Where a term is used without a reference number, this may be meant to refer generally to the term without limitation to any particular Figure.
The term “determining” encompasses a wide variety of actions and, therefore, “determining” can include calculating, computing, processing, deriving, investigating, looking up (e.g., looking up in a table, a database or another data structure), ascertaining and the like. Also, “determining” can include receiving (e.g., receiving information), accessing (e.g., accessing data in a memory) and the like. Also, “determining” can include resolving, selecting, choosing, establishing and the like.
The phrase “based on” does not mean “based only on,” unless expressly specified otherwise. In other words, the phrase “based on” describes both “based only on” and “based at least on.”
The functions described herein may be stored as one or more instructions on a processor-readable or computer-readable medium. The term “computer-readable medium” refers to any available medium that can be accessed by a computer or processor. By way of example, and not limitation, such a medium may comprise RAM, ROM, EEPROM, flash memory, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer or processor. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray® disc where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. It should be noted that a computer-readable medium may be tangible and non-transitory. The term “computer-program product” refers to a computing device or processor in combination with code or instructions (e.g., a “program”) that may be executed, processed or computed by the computing device or processor. As used herein, the term “code” may refer to software, instructions, code or data that is/are executable by a computing device or processor.
Software or instructions may also be transmitted over a transmission medium. For example, if the software is transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of transmission medium.
The methods disclosed herein comprise one or more steps or actions for achieving the described method. The method steps and/or actions may be interchanged with one another without departing from the scope of the claims. In other words, unless a specific order of steps or actions is required for proper operation of the method that is being described, the order and/or use of specific steps and/or actions may be modified without departing from the scope of the claims.
It is to be understood that the claims are not limited to the precise configuration and components illustrated above. Various modifications, changes and variations may be made in the arrangement, operation and details of the systems, methods, and apparatus described herein without departing from the scope of the claims.
This application is related to and claims priority from U.S. Provisional Patent Application Ser. No. 61/440,338 filed Feb. 7, 2011, for “WATERMARKING FOR CODEC EXTENSION.”
Number | Name | Date | Kind |
---|---|---|---|
6061793 | Tewfik et al. | May 2000 | A |
6330672 | Shur | Dec 2001 | B1 |
6332030 | Manjunath et al. | Dec 2001 | B1 |
6418424 | Hoffberg et al. | Jul 2002 | B1 |
6442285 | Rhoads et al. | Aug 2002 | B2 |
6493457 | Quackenbush et al. | Dec 2002 | B1 |
7330814 | McCree | Feb 2008 | B2 |
7389420 | Tian | Jun 2008 | B2 |
7596492 | Sung et al. | Sep 2009 | B2 |
7860992 | Robinson | Dec 2010 | B2 |
7974846 | Tsuchinaga et al. | Jul 2011 | B2 |
8151113 | Rhoads | Apr 2012 | B2 |
8326641 | Song | Dec 2012 | B2 |
8768690 | Gupta | Jul 2014 | B2 |
20010040977 | Nakano | Nov 2001 | A1 |
20010049788 | Shur | Dec 2001 | A1 |
20020128839 | Lindgren et al. | Sep 2002 | A1 |
20030088327 | Taori et al. | May 2003 | A1 |
20030154073 | Ota et al. | Aug 2003 | A1 |
20030163305 | Cheng et al. | Aug 2003 | A1 |
20040068399 | Ding | Apr 2004 | A1 |
20040068404 | Tanaka et al. | Apr 2004 | A1 |
20040186735 | Ferris et al. | Sep 2004 | A1 |
20040252700 | Anandakumar | Dec 2004 | A1 |
20050002526 | Choi et al. | Jan 2005 | A1 |
20050065785 | Bessette | Mar 2005 | A1 |
20050094848 | Carr et al. | May 2005 | A1 |
20050177364 | Jelinek | Aug 2005 | A1 |
20060020450 | Miseki | Jan 2006 | A1 |
20060045309 | Suthaharan | Mar 2006 | A1 |
20060075241 | Deguillaume et al. | Apr 2006 | A1 |
20060133624 | Waserblat | Jun 2006 | A1 |
20060173677 | Sato et al. | Aug 2006 | A1 |
20060239502 | Petrovic et al. | Oct 2006 | A1 |
20060271355 | Wang | Nov 2006 | A1 |
20060282263 | Vos et al. | Dec 2006 | A1 |
20070061577 | Van De Kerkhof | Mar 2007 | A1 |
20070094009 | Ryu et al. | Apr 2007 | A1 |
20070217626 | Sharma | Sep 2007 | A1 |
20080027711 | Rajendran et al. | Jan 2008 | A1 |
20080250091 | Lacy et al. | Oct 2008 | A1 |
20080263359 | Radzishevsky | Oct 2008 | A1 |
20080288263 | Jung et al. | Nov 2008 | A1 |
20090094264 | Nurminen et al. | Apr 2009 | A1 |
20090240491 | Reznik | Sep 2009 | A1 |
20090326961 | Petrovic et al. | Dec 2009 | A1 |
20100106269 | Garudadri et al. | Apr 2010 | A1 |
20100174537 | Vos | Jul 2010 | A1 |
20100228541 | Oshikiri | Sep 2010 | A1 |
20110033081 | Davidson et al. | Feb 2011 | A1 |
20110119054 | Lee | May 2011 | A1 |
20110125505 | Vaillancourt | May 2011 | A1 |
20110131047 | Geiser et al. | Jun 2011 | A1 |
20110208514 | Tsuchinaga et al. | Aug 2011 | A1 |
20110224994 | Norvell | Sep 2011 | A1 |
20110307248 | Yamanashi | Dec 2011 | A1 |
20120022879 | Srinivasan | Jan 2012 | A1 |
20120053950 | Oshikiri | Mar 2012 | A1 |
20120185256 | Virette | Jul 2012 | A1 |
20120203556 | Villette et al. | Aug 2012 | A1 |
20120203561 | Villette et al. | Aug 2012 | A1 |
20120221342 | Oshikiri | Aug 2012 | A1 |
Number | Date | Country |
---|---|---|
1437169 | Aug 2003 | CN |
1575492 | Feb 2005 | CN |
101262418 | Sep 2008 | CN |
101271690 | Sep 2008 | CN |
101345054 | Jan 2009 | CN |
101577605 | Nov 2009 | CN |
1503369 | Feb 2005 | EP |
1959432 | Aug 2008 | EP |
H10513571 | Dec 1998 | JP |
2003295879 | Oct 2003 | JP |
2004069963 | Mar 2004 | JP |
2004310117 | Nov 2004 | JP |
2004348120 | Dec 2004 | JP |
2005049794 | Feb 2005 | JP |
2005506584 | Mar 2005 | JP |
2005513543 | May 2005 | JP |
2006524358 | Oct 2006 | JP |
2007048340 | Feb 2007 | JP |
2008539671 | Nov 2008 | JP |
2009518945 | May 2009 | JP |
2010020333 | Jan 2010 | JP |
200943899 | Oct 2009 | TW |
WO-9624925 | Aug 1996 | WO |
WO-0062501 | Oct 2000 | WO |
WO03036624 | May 2003 | WO |
WO03047138 | Jun 2003 | WO |
WO-03053064 | Jun 2003 | WO |
WO-2004090886 | Oct 2004 | WO |
WO-2006116394 | Nov 2006 | WO |
WO2007043811 | Apr 2007 | WO |
2007051124 | May 2007 | WO |
WO-2007063913 | Jun 2007 | WO |
WO2008016947 | Feb 2008 | WO |
WO-2010059342 | May 2010 | WO |
Entry |
---|
Backwards compatible wideband telephony in mobile networks: CELP watermarking and bandwidth extension by Bernd Geiser and Peter Vary, IEEE 2007. |
Taiwan Search Report—TW101100829—TIPO—Mar. 7, 2014. |
Bernd Geiser., et al., “High rate data hiding in ACELP speech codecs”, Acoustics, Speech and Signal Processing, 2008, ICASSP 2008, IEEE International Conference on, IEEE, Piscataway, NJ, USA, Mar. 31, 2008 (Mar. 31, 2008), pp. 4005-4008, XP031251474, ISBN : 978-1-4244-1483-3 paragraph [03.2] ; table 2. |
International Search Report and Written Opinion—PCT/US2012/020773—ISA/EPO—Apr. 3, 2012. |
Ding, H.,“Wideband Audio Over Narrowband Low-Resolution Media,” IEEE International Conference on Acoustics, Speech, and Signal Processing, 2004. (ICASSP '04). vol. 1, May 17-21, 2004, pp. 1-489 to 1-492. |
Geiser, B., et al., “Artificial Bandwidth Extension of Speech Supported by Watermark-Transmitted Side Information”, Inter Speech, Institute of Communication Systems and Data Processing, Sep. 4-8, 2005, pp. 1497-1500. |
Geiser, B., et al., “Backwards Compatible wideband Telephony in Mobile Networks: CELP Watermarking and Bandwidth Extension”, IEEE, Institute of Communication Systems and Data Processing, 2007, pp. IV 533-536. |
Geiser, B., et al., “Robust Wideband Enhancement of Speech by Combined Coding and Artificial Bandwidth Extension”, Institute of Communication Systems and Data Processing, Jan. 2005, pp. 21-24. |
Loo, P., et al., “Watermark Detection Based on the Properties of Error Control Codes”, IEE Proc.-Vis. Image Signal Process., vol. 150, No. 2, Apr. 2003, pp. 115-121. |
Luo, Z. et al., “Wideband Audio Over Narrow Band Based on Digital Watermarking”, IEEE, School of Information Technology and Engineering, Jun. 2010, pp. 697-701. |
Nishimura, A. et al., “Data Hiding in Pitch Delay Data of the Adaptive Multi-Rate Narrow-band Speech Codec”, IEEE, Fifth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, 2009, pp. 483-486. |
Vary, P. et al., “Steganographic Wideband Telephony Using Narrowband Speech Codecs”, Conference Record of the Forty-First Asilomar Conference on Signals, Systems and Computers, 2007, pp. 1475-1479, Nov. 4-7, 2007. doi: 10.1109/ACSSC.2007.4487475. |
Number | Date | Country | |
---|---|---|---|
20120203555 A1 | Aug 2012 | US |
Number | Date | Country | |
---|---|---|---|
61440338 | Feb 2011 | US |