The present disclosure is generally related to signal processing.
Advances in technology have resulted in smaller and more powerful computing devices. For example, there currently exist a variety of portable personal computing devices, including wireless computing devices, such as portable wireless telephones, personal digital assistants (PDAs), and paging devices that are small, lightweight, and easily carried by users. More specifically, portable wireless telephones, such as cellular telephones and Internet Protocol (IP) telephones, can communicate voice and data packets over wireless networks. Further, many such wireless telephones include other types of devices that are incorporated therein. For example, a wireless telephone can also include a digital still camera, a digital video camera, a digital recorder, and an audio file player.
In traditional telephone systems (e.g., public switched telephone networks (PSTNs)), signal bandwidth is limited to the frequency range of 300 Hertz (Hz) to 3.4 kiloHertz (kHz). In wideband (WB) applications, such as cellular telephony and voice over internet protocol (VoIP), signal bandwidth may span the frequency range from 50 Hz to 7 kHz. Super wideband (SWB) coding techniques support bandwidth that extends up to around 16 kHz. Extending signal bandwidth from narrowband telephony at 3.4 kHz to SWB telephony of 16 kHz may improve speech intelligibility and naturalness.
SWB coding techniques typically involve encoding and transmitting the lower frequency portion of the signal (e.g., 50 Hz to 7 kHz, also called the “low-band”). For example, the low-band may be represented using filter parameters and/or a low-band excitation signal. However, in order to improve coding efficiency, the higher frequency portion of the signal (e.g., 7 kHz to 16 kHz, also called the “high-band”) may be encoded using signal modeling techniques to predict the high-band. In some implementations, data associated with the high-band may be provided to the receiver to assist in the prediction. Such data may be referred to as “side information,” and may include gain information, line spectral frequencies (LSFs, also referred to as line spectral pairs (LSPs)), etc. The gain information may include gain shape information determined based on sub-frame energies of both the high-band signal and the modeled high-band signal. The gain shape information may have a wider dynamic range (e.g., large swings) due to differences in the original high-band signal relative to the modeled high-band signal. The wider dynamic range may reduce efficiency of an encoder used to encode/transmit the gain shape information.
Systems and methods of performing audio signal encoding are disclosed. In a particular embodiment, an audio signal is encoded into a bit stream or data stream that includes a low-band bit stream (representing a low-band portion of the audio signal) and high-band side information (representing a high-band portion of the audio signal). The high-band side information may be generated using the low-band portion of the audio signal. For example, a low-band excitation signal may be extended to generate a high-band excitation signal. The high-band excitation signal may be used to generate (e.g., synthesize) a first modeled high-band signal. Energy differences between the high-band signal and the modeled high-band signal may be used to determine scaling factors (e.g., a first set of one or more scaling factors). The scaling factors (or a second set of scaling factors determined based on the first set of scaling factors) may be applied to the high-band excitation signal to generate (e.g., synthesize) a second modeled high-band signal. The second modeled high-band signal may be used to determine the high-band side information. Since the second modeled high-band signal is scaled to account for energy differences with respect to the high-band signal, the high-band side information based on the second modeled high-band signal may have a reduced dynamic range relative to high-band side information determined without scaling to account for energy differences.
In a particular embodiment, a method includes determining a first modeled high-band signal based on a low-band excitation signal of an audio signal. The audio signal includes a high-band portion and a low-band portion. The method also includes determining scaling factors based on energy of sub-frames of the first modeled high-band signal and energy of corresponding sub-frames of the high-band portion of the audio signal. The method includes applying the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal and determining a second modeled high-band signal based on the scaled high-band excitation signal. The method also includes determining gain information based on the second modeled high-band signal and the high-band portion of the audio signal.
In another particular embodiment, an apparatus includes a first synthesis filter configured to determine a first modeled high-band signal based on a low-band excitation signal of an audio signal, where the audio signal includes a high-band portion and a low-band portion. The apparatus also includes a scaling module configured to determine scaling factors based on energy of sub-frames of the first modeled high-band signal and energy of corresponding sub-frames of the high-band portion of the audio signal and to apply the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal. The apparatus also includes a second synthesis filter configured to determine a second modeled high-band signal based on the scaled high-band excitation signal. The apparatus also includes a gain estimator configured to determine gain information based on the second modeled high-band signal and the high-band portion of the audio signal.
In another particular embodiment, a device includes means for determining a first modeled high-band signal based on a low-band excitation signal of an audio signal, where the audio signal includes a high-band portion and a low-band portion. The device also includes means for determining scaling factors based on energy of sub-frames of the first modeled high-band signal and energy of corresponding sub-frames of the high-band portion of the audio signal. The device also includes means for applying the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal. The device also includes means for determining a second modeled high-band signal based on the scaled high-band excitation signal. The device also includes means for determining gain information based on the second modeled high-band signal and the high-band portion of the audio signal.
In another particular embodiment, a non-transitory computer-readable medium includes instructions that, when executed by a computer, cause the computer to perform operations including determining a first modeled high-band signal based on a low-band excitation signal of an audio signal, where the audio signal includes a high-band portion and a low-band portion. The operations also include determining scaling factors based on energy of sub-frames of the first modeled high-band signal and energy of corresponding sub-frames of the high-band portion of the audio signal. The operations also include applying the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal. The operations also include determining a second modeled high-band signal based on the scaled high-band excitation signal. The operations also include determining gain parameters based on the second modeled high-band signal and the high-band portion of the audio signal.
Particular advantages provided by at least one of the disclosed embodiments include reducing a dynamic range of gain information provided to an encoder by scaling a modeled high-band excitation signal that is used to calculate the gain information. For example, the modeled high-band excitation signal may be scaled based on energies of sub-frames of a modeled high-band signal and corresponding sub-frames of a high-band portion of an audio signal. Scaling the modeled high-band excitation signal in this manner may capture variations in the temporal characteristics from sub-frame-to-sub-frame and reduce dependence of the gain shape information on temporal changes in the high-band portion of an audio signal. Other aspects, advantages, and features of the present disclosure will become apparent after review of the entire application, including the following sections: Brief Description of the Drawings, Detailed Description, and the Claims.
In the following description, various functions performed by the system 100 of
The system 100 includes an analysis filter bank 110 that is configured to receive an audio signal 102. For example, the audio signal 102 may be provided by a microphone or other input device. In a particular embodiment, the input audio signal 102 may include speech. The audio signal 102 may be a SWB signal that includes data in the frequency range from approximately 50 hertz (Hz) to approximately 16 kilohertz (kHz). The analysis filter bank 110 may filter the input audio signal 102 into multiple portions based on frequency. For example, the analysis filter bank 110 may generate a low-band signal 122 and a high-band signal 124. The low-band signal 122 and the high-band signal 124 may have equal or unequal bandwidths, and may be overlapping or non-overlapping. In an alternate embodiment, the analysis filter bank 110 may generate more than two outputs.
In the example of
Although the description of
The system 100 may include a low-band analysis module 130 (also referred to as a low-band encoder) configured to receive the low-band signal 122. In a particular embodiment, the low-band analysis module 130 may represent an embodiment of a code excited linear prediction (CELP) encoder. The low-band analysis module 130 may include a linear prediction (LP) analysis and coding module 132, a linear prediction coefficient (LPC) to line spectral pair (LSP) transform module 134, and a quantizer 136. LSPs may also be referred to as line spectral frequencies (LSFs), and the two terms may be used interchangeably herein. The LP analysis and coding module 132 may encode a spectral envelope of the low-band signal 122 as a set of LPCs. LPCs may be generated for each frame of audio (e.g., 20 milliseconds (ms) of audio, corresponding to 320 samples at a sampling rate of 16 kHz), each sub-frame of audio (e.g., 5 ms of audio), or any combination thereof. The number of LPCs generated for each frame or sub-frame may be determined by the “order” of the LP analysis performed. In a particular embodiment, the LP analysis and coding module 132 may generate a set of eleven LPCs corresponding to a tenth-order LP analysis.
The LPC to LSP transform module 134 may transform the set of LPCs generated by the LP analysis and coding module 132 into a corresponding set of LSPs (e.g., using a one-to-one transform). Alternately, the set of LPCs may be one-to-one transformed into a corresponding set of parcor coefficients, log-area-ratio values, immittance spectral pairs (ISPs), or immittance spectral frequencies (ISFs). The transform between the set of LPCs and the set of LSPs may be reversible without error.
The quantizer 136 may quantize the set of LSPs generated by the transform module 134. For example, the quantizer 136 may include or may be coupled to multiple codebooks (not shown) that include multiple entries (e.g., vectors). To quantize the set of LSPs, the quantizer 136 may identify entries of codebooks that are “closest to” (e.g., based on a distortion measure such as least squares or mean square error) the set of LSPs. The quantizer 136 may output an index value or series of index values corresponding to the location of the identified entries in the codebook. The output of the quantizer 136 may represent low-band filter parameters that are included in a low-band bit stream 142. The low-band bit stream 142 may thus include linear prediction code data representing the low-band portion of the audio signal 102.
The low-band analysis module 130 may also generate a low-band excitation signal 144. For example, the low-band excitation signal 144 may be an encoded signal that is generated by quantizing a LP residual signal that is generated during the LP process performed by the low-band analysis module 130. The LP residual signal may represent prediction error.
The system 100 may further include a high-band analysis module 150 configured to receive the high-band signal 124 from the analysis filter bank 110 and the low-band excitation signal 144 from the low-band analysis module 130. The high-band analysis module 150 may generate high-band side information 172 based on the high-band signal 124 and the low-band excitation signal 144. For example, the high-band side information 172 may include data representing high-band LSPs, data representing gain information (e.g., based on at least a ratio of high-band energy to low-band energy), data representing scaling factors, or a combination thereof.
The high-band analysis module 150 may include a high-band excitation generator 152. The high-band excitation generator 152 may generate a high-band excitation signal (such as high-band excitation signal 202 of
High-band excitation=(α*transformed low-band excitation)+((1×α)*modulated noise)
A ratio at which the transformed low-band excitation signal and the modulated noise are mixed may impact high-band reconstruction quality at a receiver. For voiced speech signals, the mixing may be biased towards the transformed low-band excitation (e.g., the mixing factor α may be in the range of 0.5 to 1.0). For unvoiced signals, the mixing may be biased towards the modulated noise (e.g., the mixing factor α may be in the range of 0.0 to 0.5).
The high-band excitation signal may be used to determine one or more high-band gain parameters that are included in the high-band side information 172. In a particular embodiment, the high-band excitation signal and the high-band signal 124 may be used to determine scaling information (e.g., scaling factors) that are applied to the high-band excitation signal to determine a scaled high-band excitation signal. The scaled high-band excitation signal may be used to determine the high-band gain parameters. For example, as described further with reference to FIGS. 2 and 5-7, the energy estimator 154 may determine estimated energy of frames or sub-frames of the high-band signal and of corresponding frames or sub-frames of a first modeled high band signal. The first modeled high band signal may be determined by applying memoryless linear prediction synthesis on the high-band excitation signal. The scaling module 156 may determine scaling factors (e.g., a first set of scaling factors) based on the estimated energy of frames or sub-frames of the high-band signal 124 and the estimated energy of the corresponding frames or sub-frames of a first modeled high band signal. For example, each scaling factor may correspond to a ratio Ei/Ei′, where Ei is an estimated energy of a sub-frame, i, of the high-band signal and Ei′ is an estimated energy of a corresponding sub-frame, i, of the first modeled high band signal. The scaling module 156 may also apply the scaling factors (or a second set of scaling factors determined based on the first set of scaling factors, e.g., by averaging gains over several subframes of the first set of scaling factors), on a sub-frame-by-sub-frame basis, to the high-band excitation signal to determine the scaled high-band excitation signal.
As illustrated, the high-band analysis module 150 may also include an LP analysis and coding module 158, a LPC to LSP transform module 160, and a quantizer 162. Each of the LP analysis and coding module 158, the transform module 160, and the quantizer 162 may function as described above with reference to corresponding components of the low-band analysis module 130, but at a comparatively reduced resolution (e.g., using fewer bits for each coefficient, LSP, etc.). The LP analysis and coding module 158 may generate a set of LPCs that are transformed to LSPs by the transform module 160 and quantized by the quantizer 162 based on a codebook 166. For example, the LP analysis and coding module 158, the transform module 160, and the quantizer 162 may use the high-band signal 124 to determine high-band filter information (e.g., high-band LSPs) that is included in the high-band side information 172. In a particular embodiment, the high-band side information 172 may include high-band LSPs, high-band gain information, the scaling factors, or a combination thereof. As explained above, the high-band gain information may be determined based on a scaled high-band excitation signal.
The low-band bit stream 142 and the high-band side information 172 may be multiplexed by a multiplexer (MUX) 180 to generate an output data stream or output bit stream 192. The output bit stream 192 may represent an encoded audio signal corresponding to the input audio signal 102. For example, the output bit stream 192 may be transmitted (e.g., over a wired, wireless, or optical channel) and/or stored. At a receiver, reverse operations may be performed by a demultiplexer (DEMUX), a low-band decoder, a high-band decoder, and a filter bank to generate an audio signal (e.g., a reconstructed version of the input audio signal 102 that is provided to a speaker or other output device). The number of bits used to represent the low-band bit stream 142 may be substantially larger than the number of bits used to represent the high-band side information 172. Thus, most of the bits in the output bit stream 192 may represent low-band data. The high-band side information 172 may be used at a receiver to regenerate the high-band excitation signal from the low-band data in accordance with a signal model. For example, the signal model may represent an expected set of relationships or correlations between low-band data (e.g., the low-band signal 122) and high-band data (e.g., the high-band signal 124). Thus, different signal models may be used for different kinds of audio data (e.g., speech, music, etc.), and the particular signal model that is in use may be negotiated by a transmitter and a receiver (or defined by an industry standard) prior to communication of encoded audio data. Using the signal model, the high-band analysis module 150 at a transmitter may be able to generate the high-band side information 172 such that a corresponding high-band analysis module at a receiver is able to use the signal model to reconstruct the high-band signal 124 from the output bit stream 192.
Filter parameters 204 may be applied to the high-band excitation signal 202 using an all-pole LP synthesis filter 206 (e.g., a synthesis filter) to determine a first modeled high-band signal 208. The filter parameters 204 may correspond to the feedback memory of the all-pole LP synthesis filter 206. For purposes of determining the scaling factors, the filter parameters 204 may be memoryless. In particular, the filter memory or filter states that are associated with the i-th subframe LP synthesis filter, 1/Ai(z) are reset to zero before carrying out the all-pole LP synthesis filter 206.
The first modeled high-band signal 208 may be applied to an energy estimator 210 to determine sub-frame energy 212 of each frame or sub-frame of the first modeled high-band signal 208. The high-band signal 124 may also be applied to an energy estimator 222 to determine energy 224 of each frame or sub-frame of the high-band signal 124. The sub-frame energy 212 of the first modeled high-band signal 208 and the energy 224 of the high-band signal 124 may be used to determine scaling factors 230. The scaling factors 230 may quantify energy differences between frames or sub-frames of the first modeled high-band signal 208 and corresponding frames or sub-frames of the high-band signal 124. For example, the scaling factors 230 may be determined as a ratio of energy 224 of the high-band signal 124 and the estimated sub-frame energy 212 of the first modeled high-band signal 208. In a particular embodiment, the scaling factors 230 are determined on a sub-frame-by-sub-frame basis, where each frame includes four sub-frames. In this embodiment, one scaling factor is determined for each set of sub-frames including a sub-frame of the first modeled high-band signal 208 and a corresponding sub-frame of the high-band signal 124.
To determine the gain information, each sub-frame of the high-band excitation signal 202 may be compensated (e.g., multiplied) with a corresponding scaling factor 230 to generate a scaled high-band excitation signal 240. Filter parameters 242 may be applied to the scaled high-band excitation signal 240 using an all-pole filter 244 to determine a second modeled high-band signal 246. The filter parameters 242 may correspond to parameters of a linear prediction analysis and coding module, such as the LP analysis and coding module 158 of
The second modeled high-band signal 246 may be applied to a gain shape estimator 248 along with the high-band signal 124 to determine gain parameters 250. The gain parameters 250, the second modeled high-band signal 246 and the high-band signal 124 may be applied to a gain frame estimator 252 to determine a frame gain 254. The gain parameters 250 and the frame gain 254 together form the gain information. The gain information may have reduced dynamic range relative to gain information determined without applying the scaling factors 230 since the scaling factors account for some of the energy differences between the high-band signal 124 and the second modeled high-band signal 246 determined based on the high-band excitation signal 202.
A plurality of sub-frame LSPs for the Nth Frame 304 may be determined by interpolation using LSP values of a preceding frame (e.g., the N−1th Frame 302) and a current frame (e.g., the Nth Frame 304). For example, weighting factors may be applied to values of a preceding LSP (e.g., the N−1th LSP 310) and to values of a current LSP (e.g., the Nth LSP 312). In the example illustrated in
The sub-frame LSPs (320-326) may be used to perform the LP synthesis without filter memory updates to estimate the first modeled high band signal 208. The first modeled high band signal 208 is then used to estimate sub-frame energy Ei′ 212. The energy estimator 154 may provide sub-frame energy estimates for the first modeled high-band signal 208 and for the high-band signal 124 to the scaling module 156, which may determine sub-frame-by-sub-frame scaling factors 230. The scaling factors may be used to adjust an energy level of the high-band excitation signal 202 to generate a scaled high-band excitation signal 240, which may be used by the LP analysis and coding module 158 to generate a second modeled (or synthesized) high-band signal 246. The second modeled high-band signal 246 may be used to generate gain information (such as the gain parameters 250 and/or the frame gain 254). For example, the second modeled high-band signal 246 may be provided to the gain estimator 164, which may determine the gain parameters 250 and frame gain 254.
A plurality of sub-frame LSPs for the Nth Frame 404 may be determined by interpolation using one or more of the LSP values of a preceding frame (e.g., the LSP—1 408 and/or the LSP—2 410 of the N−1th Frame 402) and one or more of the LSP values of a current frame (e.g., the Nth Frame 404). While the LSP windows (e.g., dashed lines 412, 414 asymmetric LSP windows for Frame N 404) shown in
The sub-frame LSPs (420-426) may be used to perform the LP synthesis without filter memory updates to estimate the first modeled high band signal 208. The first modeled high band signal 208 is then used to estimate sub-frame energy Ei′ 212. The energy estimator 154 may provide sub-frame energy estimates for the first modeled high-band signal 208 and for the high-band signal 124 to the scaling module 156, which may determine sub-frame-by-sub-frame scaling factors 230. The scaling factors may be used to adjust an energy level of the high-band excitation signal 202 to generate a scaled high-band excitation signal 240, which may be used by the LP analysis and coding module 158 to generate a second modeled (or synthesized) high-band signal 246. The second modeled high-band signal 246 may be used to generate gain information (such as the gain parameters 250 and/or the frame gain 254). For example, the second modeled high-band signal 246 may be provided to the gain estimator 164, which may determine the gain parameters 250 and frame gain 254.
The high-band signal 502 may also be received at a windowing module 520. The windowing module 520 may generate linear prediction coefficients (LPCs) for each pair of frames of the high-band signal 502. For example, the windowing module 520 may generate a first LPC 522 (e.g., LPC—1). The windowing module 520 may also generate a second LPC 524 (e.g., LPC—2). The first LPC 522 and the second LPC 524 may each be transformed to LSPs using LSP transform modules 526 and 528. For example, the first LPC 522 may be transformed to a first LSP 530 (e.g. LSP—1), and the second LPC 524 may be transformed to a second LSP 532 (e.g. LSP—2). The first and second LSPs 530, 532 may be provided to a coder 538, which may encode the LSPs 530, 532 to form high-band LSP indices 540.
The first and second LSPs 530, 532 and a third LSP 534 (e.g., LSP—2old) may be provided to an interpolator 536. The third LSP 534 may correspond to a previously processed frame, such as the N−1th Frame 302 of
The sub-frame LSPs 542, 544, 546, and 548 may be provided to an LSP-to-LPC transformation module 550 to determine sub-frame LPCs and filter parameters 552, 554, 556, and 558.
As also illustrated in
Referring to
The sub-frames 622, 624, 626, 628 of the first modeled high-band signal may be provided to energy estimators 632, 634, 636, and 638. The energy estimators 632, 634, 636, and 638 may generate energy estimates 642, 644, 646, 648 (Ei′, where i is an index of a particular sub-frame) of the sub-frames 622, 624, 626, 628 of the first modeled high-band signal.
The energy estimates 652, 654, 656, and 658 of the high-band signal 502 of
Referring to
The sub-frames 702, 704, 706, and 708 of the scaled high-band excitation signal may be applied to all-pole filters 712, 714, 716, 718 (e.g., synthesis filters) to determine sub-frames 742, 744, 746, 748 of a second modeled (or synthesized) high-band signal. For example, the first sub-frame 702 of the scaled high-band excitation signal may be applied to a first all-pole filter 712, along with first filter parameters 722, to determine a first sub-frame 742 of the second modeled high-band signal. Filter parameters 722, 724, 726, and 728 applied to the all-pole filters 712, 714, 716, 718 may include information related to previously processed frames (or sub-frames). For example, each all-pole filter 712, 714, 716 may output filter state update information 732, 734, 736 that is provided to another of the all-pole filters 714, 716, 718. The filter state update 738 from the all-pole filter 718 may be used in the next frame (i.e., first sub-frame) to update the filter memory.
The sub-frames 742, 744, 746, 748 of the second modeled high-band signal may be combined, at a framing module 750, to generate a frame 752 of the second modeled high-band signal. The frame 752 of the second modeled high-band signal may be applied to a gain shape estimator 754 along with the high-band signal 502 to determine gain parameters 756. The gain parameters 756, the frame 752 of the second modeled high-band signal, and the high-band signal 502 may be applied to a gain frame estimator 758 to determine a frame gain 760. The gain parameters 756 and the frame gain 760 together form gain information. The gain information may have reduced dynamic range relative to gain information determined without applying the scaling factors 672, 674, 676, 678 since the scaling factors 672, 674, 676, 678 account for some of the energy differences between the high-band signal 502 and a signal modeled using the high-band excitation signal 560.
The method 800 also includes, at 804, determining scaling factors based on energy of sub-frames of the first modeled high-band signal and energy of corresponding sub-frames of the high-band portion of the audio signal. For example, the scaling factors 230 of
The method 800 includes, at 806, applying the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal. For example, the scaling factor 230 of
The method 800 includes, at 808, determining a second modeled high-band signal based on the scaled high-band excitation signal. To illustrate, linear prediction analysis of the scaled high-band excitation signal may be performed. For example, the scaled high-band excitation signal 240 of
The method 800 includes, at 810, determining gain parameters based on the second modeled high-band signal and the high-band portion of the audio signal. For example, the second modeled high-band signal 246 and the high-band signal 124 may be provided to the gain shape estimator 248 of
Referring to
For example, the instructions 960 may include or correspond to a low-band analysis module 976 and a high-band analysis module 978. In a particular embodiment, the low-band analysis module 976 corresponds to the low-band analysis module 130 of
In various embodiments, the low-band analysis module 976, the high-band analysis module 978, or both, may be implemented via dedicated hardware (e.g., circuitry), by a processor (e.g., the processor 912) executing the instructions 960 or instructions 961 in a memory 980 to perform one or more tasks, or a combination thereof. As an example, the memory 932 or the memory 980 may include or correspond to a memory device, such as a random access memory (RAM), magnetoresistive random access memory (MRAM), spin-torque transfer MRAM (STT-MRAM), flash memory, read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), registers, hard disk, a removable disk, or a compact disc read-only memory (CD-ROM). The memory device may include instructions (e.g., the instructions 960 or the instructions 961) that, when executed by a computer (e.g., the processor 910 and/or the processor 912), may cause the computer to determine scaling factors based on energy of sub-frames of a first modeled high-band signal and energy of corresponding sub-frames of a high-band portion of an audio signal, apply the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal, determine a second modeled high-band signal based on the scaled high-band excitation signal, and determine gain parameters based on the second modeled high-band signal and the high-band portion of the audio signal. As an example, the memory 932 or the memory 980 may be a non-transitory computer-readable medium that includes instructions that, when executed by a computer (e.g., the processor 910 and/or the processor 912), cause the computer perform at least a portion of the method 800 of
In a particular embodiment, the processor 910, the processor 912, the display controller 926, the memory 932, the CODEC 934, and the wireless controller 940 are included in a system-in-package or system-on-chip device 922. In a particular embodiment, an input device 930, such as a touch screen and/or keypad, and a power supply 944 are coupled to the system-on-chip device 922. Moreover, in a particular embodiment, as illustrated in
In conjunction with the described embodiments, an apparatus is disclosed that includes means for determining a first modeled high-band signal based on a low-band excitation signal of an audio signal, where the audio signal includes a high-band portion and a low-band portion. For example, the high-band analysis module 150 (or a component thereof, such as the LP analysis and coding module 158) may determine the first modeled high-band signal based on the low-band excitation signal 144 of the audio signal 102. As another example, a first synthesis filter, such as the all-pole LP synthesis filter 206 of
The apparatus also includes means for determining scaling factors based on energy of sub-frames of the first modeled high-band signal and energy of corresponding sub-frames of the high-band portion of the audio signal. For example, the energy estimator 154 and the scaling module 156 of
The apparatus also includes means for applying the scaling factors to a modeled high-band excitation signal to determine a scaled high-band excitation signal. For example, the scaling module 156 of
The device also includes means for determining a second modeled high-band signal based on the scaled high-band excitation signal. For example, the high-band analysis module 150 (or a component thereof, such as the LP analysis and coding module 158) may determine the second modeled high-band signal based on the scaled high-band excitation signal. As another example, a second synthesis filter, such as the all-pole filter 244 of
The apparatus also includes means for determining gain parameters based on the second modeled high-band signal and the high-band portion of the audio signal. For example, the gain estimator 164 of
Those of skill would further appreciate that the various illustrative logical blocks, configurations, modules, circuits, and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software executed by a processing device such as a hardware processor, or combinations of both. Various illustrative components, blocks, configurations, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or executable software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present disclosure.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in a memory device, such as RAM, MRAM, STT-MRAM, flash memory, ROM, PROM, EPROM, EEPROM, registers, hard disk, a removable disk, or a CD-ROM. An exemplary memory device is coupled to the processor such that the processor can read information from, and write information to, the memory device. In the alternative, the memory device may be integral to the processor. The processor and the storage medium may reside in an ASIC. The ASIC may reside in a computing device or a user terminal. In the alternative, the processor and the storage medium may reside as discrete components in a computing device or a user terminal.
The previous description of the disclosed embodiments is provided to enable a person skilled in the art to make or use the disclosed embodiments. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the principles defined herein may be applied to other embodiments without departing from the scope of the disclosure. Thus, the present disclosure is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope possible consistent with the principles and novel features as defined by the following claims.
The present application claims priority from U.S. Provisional Patent Application No. 61/890,812, entitled “SYSTEMS AND METHODS OF ENERGY-SCALED SIGNAL PROCESSING,” filed Oct. 14, 2013, the contents of which is incorporated by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
61890812 | Oct 2013 | US |