Efficient and scalable parametric stereo coding for low bitrate audio coding applications

Information

  • Patent Grant
  • 9799340
  • Patent Number
    9,799,340
  • Date Filed
    Tuesday, March 14, 2017
    7 years ago
  • Date Issued
    Tuesday, October 24, 2017
    6 years ago
Abstract
The present invention provides improvements to prior art audio codecs that generate a stereo-illusion through post-processing of a received mono signal. These improvements are accomplished by extraction of stereo-image describing parameters at the encoder side, which are transmitted and subsequently used for control of a stereo generator at the decoder side. Furthermore, the invention bridges the gap between simple pseudo-stereo methods, and current methods of true stereo-coding, by using a new form of parametric stereo coding. A stereo-balance parameter is introduced, which enables more advanced stereo modes, and in addition forms the basis of a new method of stereo-coding of spectral envelopes, of particular use in systems where guided HFR (High Frequency Reconstruction) is employed. As a special case, the application of this stereo-coding scheme in scalable HFR-based codecs is described.
Description
BACKGROUND OF THE INVENTION

Technical Field


The present invention relates to low bitrate audio source coding systems. Different parametric representations of stereo properties of an input signal are introduced, and the application thereof at the decoder side is explained, ranging from pseudo-stereo to full stereo coding of spectral envelopes, the latter of which is especially suited for HFR based codecs.


Description of the Related Art


Audio source coding techniques can be divided into two classes: natural audio coding and speech coding. At medium to high bitrates, natural audio coding is commonly used for speech and music signals, and stereo transmission and reproduction is possible. In applications where only low bitrates are available, e.g. Internet streaming audio targeted at users with slow telephone modem connections, or in the emerging digital AM broadcasting systems, mono coding of the audio program material is unavoidable. However, a stereo impression is still desirable, in particular when listening with headphones, in which case a pure mono signal is perceived as originating from “within the head”, which can be an unpleasant experience.


One approach to address this problem is to synthesize a stereo signal at the decoder side from a received pure mono signal. Throughout the years, several different “pseudo-stereo” generators have been proposed. For example in [U.S. Pat. No. 5,883,962], enhancement of mono signals by means of adding delayed/phase shifted versions of a signal to the unprocessed signal, thereby creating a stereo illusion, is described. Hereby the processed signal is added to the original signal for each of the two outputs at equal levels but with opposite signs, ensuring that the enhancement signals cancel if the two channels are added later on in the signal path. In [PCT WO 98/57436] a similar system is shown, albeit without the above mono-compatibility of the enhanced signal. Prior art methods have in common that they are applied as pure post-processes. In other words, no information on the degree of stereo-width, let alone position in the stereo sound stage, is available to the decoder. Thus, the pseudo-stereo signal may or may not have a resemblance of the stereo character of the original signal. A particular situation where prior art systems fall short, is when the original signal is a pure mono signal, which often is the case for speech recordings. This mono signal is blindly converted to a synthetic stereo signal at the decoder, which in the speech case often causes annoying artifacts, and may reduce the clarity and speech intelligibility.


Other prior art systems, aiming at true stereo transmission at low bitrates, typically employ a sum and difference coding scheme. Thus, the original left (L) and right (R) signals are converted to a sum signal, S=(L+R)/2, and a difference signal, D=(L−R)/2, and subsequently encoded and transmitted. The receiver decodes the S and D signals, whereupon the original L/R-signal is recreated through the operations L=S+D, and R=S−D. The advantage of this, is that very often a redundancy between L and R is at hand, whereby the information in D to be encoded is less, requiring fewer bits, than in S. Clearly, the extreme case is a pure mono signal, i.e. L and R are identical. A traditional L/R-codec encodes this mono signal twice, whereas a S/D codec detects this redundancy, and the D signal does (ideally) not require any bits at all. Another extreme is represented by the situation where R=−L, corresponding to “out of phase” signals. Now, the S signal is zero, whereas the D signal computes to L. Again, the S/D-scheme has a clear advantage to standard L/R-coding. However, consider the situation where e.g. R=0 during a passage, which was not uncommon in the early days of stereo recordings. Both S and D equal L/2, and the S/D-scheme does not offer any advantage. On the contrary, L/R-coding handles this very well: The R signal does not require any bits. For this reason, prior art codecs employ adaptive switching between those two coding schemes, depending on what method that is most beneficial to use at a given moment. The above examples are merely theoretical (except for the dual mono case, which is common in speech only programs). Thus, real world stereo program material contains significant amounts of stereo information, and even if the above switching is implemented, the resulting bitrate is often still too high for many applications. Furthermore, as can be seen from the resynthesis relations above, very coarse quantization of the D signal in an attempt to further reduce the bitrate is not feasible, since the quantization errors translate to non-neglectable level errors in the L and R signals.


SUMMARY OF THE INVENTION

The present invention employs detection of signal stereo properties prior to coding and transmission. In the simplest form, a detector measures the amount of stereo perspective that is present in the input stereo signal. This amount is then transmitted as a stereo width parameter, together with an encoded mono sum of the original signal. The receiver decodes the mono signal, and applies the proper amount of stereo-width, using a pseudo-stereo generator, which is controlled by said parameter. As a special case, a mono input signal is signaled as zero stereo width, and correspondingly no stereo synthesis is applied in the decoder. According to the invention, useful measures of the stereo-width can be derived e.g. from the difference signal or from the cross-correlation of the original left and right channel. The value of such computations can be mapped to a small number of states, which are transmitted at an appropriate fixed rate in time, or on an as-needed basis. The invention also teaches how to filter the synthesized stereo components, in order to reduce the risk of unmasking coding artifacts which typically are associated with low bitrate coded signals.


Alternatively, the overall stereo-balance or localization in the stereo field is detected in the encoder. This information, optionally together with the above width-parameter, is efficiently transmitted as a balance-parameter, along with the encoded mono signal. Thus, displacements to either side of the sound stage can be recreated at the decoder, by correspondingly altering the gains of the two output channels. According to the invention, this stereo-balance parameter can be derived from the quotient of the left and right signal powers. The transmission of both types of parameters requires very few bits compared to full stereo coding, whereby the total bitrate demand is kept low. In a more elaborate version of the invention, which offers a more accurate parametric stereo depiction, several balance and stereo-width parameters are used, each one representing separate frequency bands.


The balance-parameter generalized to a per frequency-band operation, together with a corresponding per band operation of a level-parameter, calculated as the sum of the left and right signal powers, enables a new, arbitrary detailed, representation of the power spectral density of a stereo signal. A particular benefit of this representation, in addition to the benefits from stereo redundancy that also S/D-systems take advantage of, is that the balance-signal can be quantized with less precision than the level ditto, since the quantization error, when converting back to a stereo spectral envelope, causes an “error in space”, i.e. perceived localization in the stereo panorama, rather than an error in level. Analogous to a traditional switched L/R- and S/D-system, the level/balance-scheme can be adaptively switched off, in favor of a levelL/levelR-signal, which is more efficient when the overall signal is heavily offset towards either channel. The above spectral envelope coding scheme can be used whenever an efficient coding of power spectral envelopes is required, and can be incorporated as a tool in new stereo source codecs. A particularly interesting application is in HFR systems that are guided by information about the original signal highband envelope. In such a system, the lowband is coded and decoded by means of an arbitrary codec, and the highband is regenerated at the decoder using the decoded lowband signal and the transmitted highband envelope information [PCT WO 98/57436]. Furthermore, the possibility to build a scalable HFR-based stereo codec is offered, by locking the envelope coding to level/balance operation. Hereby the level values are fed into the primary bitstream, which, depending on the implementation, typically decodes to a mono signal. The balance values are fed into the secondary bitstream, which in addition to the primary bitstream is available to receivers close to the transmitter, taking an IBOC (In-Band On-Channel) digital AM-broadcasting system as an example. When the two bitstreams are combined, the decoder produces a stereo output signal. In addition to the level values, the primary bitstream can contain stereo parameters, e.g. a width parameter. Thus, decoding of this bitstream alone already yields a stereo output, which is improved when both bitstreams are available.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will now be described by way of illustrative examples, not limiting the scope or spirit of the invention, with reference to the accompanying drawings, in which:



FIG. 1 illustrates a source coding system containing an encoder enhanced by a parametric stereo encoder module, and a decoder enhanced by a parametric stereo decoder module.



FIG. 2a is a block schematic of a parametric stereo decoder module,



FIG. 2b is a block schematic of a pseudo-stereo generator with control parameter inputs,



FIG. 2c is a block schematic of a balance adjuster with control parameter inputs,



FIG. 3 is a block schematic of a parametric stereo decoder module using multiband pseudo-stereo generation combined with multiband balance adjustment,



FIG. 4a is a block schematic of the encoder side of a scalable HFR-based stereo codec, employing level/balance-coding of the spectral envelope,



FIG. 4b is a block schematic of the corresponding decoder side.





DESCRIPTION OF PREFERRED EMBODIMENTS

The below-described embodiments are merely illustrative for the principles of the present invention. It is understood that modifications and variations of the arrangements and the details described herein will be apparent to others skilled in the art. It is the intent therefore, to be limited only by the scope of the impending patent claims, and not by the specific details presented by way of description and explanation of the embodiments herein. For the sake of clarity, all below examples assume two channel systems, but apparent to others skilled in the art, the methods can be applied to multichannel systems, such as a 5.1 system.



FIG. 1 shows how an arbitrary source coding system comprising of an encoder, 107, and a decoder, 115, where encoder and decoder operate in monaural mode, can be enhanced by parametric stereo coding according to the invention. Let L and R denote the left and right analog input signals, which are fed to an AD-converter, 101. The output from the AD-converter is converted to mono, 105, and the mono signal is encoded, 107. In addition, the stereo signal is routed to a parametric stereo encoder, 103, which calculates one or several stereo parameters to be described below. Those parameters are combined with the encoded mono signal by means of a multiplexer, 109, forming a bitstream, 111. The bitstream is stored or transmitted, and subsequently extracted at the decoder side by means of a demultiplexer, 113. The mono signal is decoded, 115, and converted to a stereo signal by a parametric stereo decoder, 119, which uses the stereo parameter(s), 117, as control signal(s). Finally, the stereo signal is routed to the DA-converter, 121, which feeds the analog outputs, L′ and R′. The topology according to FIG. 1 is common to a set of parametric stereo coding methods which will be described in detail, starting with the less complex versions.


One method of parameterization of stereo properties according to the present invention, is to determine the original signal stereo-width at the encoder side. A first approximation of the stereo-width is the difference signal, D=L−R, since, roughly put, a high degree of similarity between L and R computes to a small value of D, and vice versa. A special case is dual mono, where L=R and thus D=0. Thus, even this simple algorithm is capable of detecting the type of mono input signal commonly associated with news broadcasts, in which case pseudo-stereo is not desired. However, a mono signal that is fed to L and R at different levels does not yield a zero D signal, even though the perceived width is zero. Thus, in practice more elaborate detectors might be required, employing for example cross-correlation methods. One should make sure that the value describing the left-right difference or correlation in some way is normalized with the total signal level, in order to achieve a level independent detector. A problem with the aforementioned detector is the case when mono speech is mixed with a much weaker stereo signal e.g. stereo noise or background music during speech-to-music/music-to-speech transitions. At the speech pauses the detector will then indicate a wide stereo signal. This is solved by normalizing the stereo-width value with a signal containing information of previous total energy level e.g., a peak decay signal of the total energy. Furthermore, to prevent the stereo-width detector from being trigged by high frequency noise or channel different high frequency distortion, the detector signals should be pre-filtered by a low-pass filter, typically with a cutoff frequency somewhere above a voice's second format, and optionally also by a high-pass filter to avoid unbalanced signal-offsets or hum. Regardless of detector type, the calculated stereo-width is mapped to a finite set of values, covering the entire range, from mono to wide stereo.



FIG. 2a gives an example of the contents of the parametric stereo decoder introduced in FIG. 1. The block denoted ‘balance’, 211, controlled by parameter B, will be described later, and should be regarded as bypassed for now. The block denoted ‘width’, 205, takes a mono input signal, and synthetically recreates the impression of stereo width, where the amount of width is controlled by the parameter W. The optional parameters S and D will be described later. According to the invention, a subjectively better sound quality can often be achieved by incorporating a crossover filter comprising of a low-pass filter, 203, and a high-pass filter, 201, in order to keep the low frequency range “tight” and unaffected. Hereby only the output from the high-pass filter is routed to the width block. The stereo output from the width block is added to the mono output from the low-pass filter by means of 207 and 209, forming the stereo output signal.


Any prior art pseudo-stereo generator can be used for the width block, such as those mentioned in the background section, or a Schroeder-type early reflection simulating unit (multitap delay) or reverberator. FIG. 2b gives an example of a pseudo-stereo generator, fed by a mono signal M. The amount of stereo-width is determined by the gain of 215, and this gain is a function of the stereo-width parameter, W. The higher the gain, the wider the stereo-impression, a zero gain corresponds to pure mono reproduction. The output from 215 is delayed, 221, and added, 223 and 225, to the two direct signal instances, using opposite signs. In order not to significantly alter the overall reproduction level when changing the stereo-width, a compensating attenuation of the direct signal can be incorporated, 213. For example, if the gain of the delayed signal is G, the gain of the direct signal can be selected as sqrt(1−G2). According to the invention, a high frequency roll-off can be incorporated in the delay signal path, 217, which helps avoiding pseudo-stereo caused unmasking of coding artifacts. Optionally, crossover filter, roll-off filter and delay parameters can be sent in the bitstream, offering more possibilities to mimic the stereo properties of the original signal, as also shown in FIGS. 2a and 2b as the signals X, S and D. If a reverberation unit is used for generating a stereo signal, the reverberation decay might sometimes be unwanted after the very end of a sound. These unwanted reverb-tails can however easily be attenuated or completely removed by just altering the gain of the reverb signal. A detector designed for finding sound endings can be used for that purpose. If the reverberation unit generates artifacts at some specific signals e.g., transients, a detector for those signals can also be used for attenuating the same.


An alternative method of detecting stereo-properties according to the invention, is described as follows. Again, let L and R denote the left and right input signals. The corresponding signal powers are then given by PL˜L2 and PR˜R2. Now, a measure of the stereo-balance can be calculated as the quotient of the two signal powers, or more specifically as B=(PL+e)/(PR+e), where e is an arbitrary, very small number, which eliminates division by zero. The balance parameter, B, can be expressed in dB given by the relation BdB==10 log10(B). As an example, the three cases PL=10PR, PL=PR, and PL=0.1PR correspond to balance values of +10 dB, 0 dB, and −10 dB respectively. Clearly, those values map to the locations “left”, “center”, and “right”. Experiments have shown that the span of the balance parameter can be limited to for example +/−40 dB, since those extreme values are already perceived as if the sound originates entirely from one of the two loudspeakers or headphone drivers. This limitation reduces the signal space to cover in the transmission, thus offering bitrate reduction. Furthermore, a progressive quantization scheme can be used, whereby smaller quantization steps are used around zero, and larger steps towards the outer limits, which further reduces the bitrate. Often the balance is constant over time for extended passages. Thus, a last step to significantly reduce the number of average bits needed can be taken: After transmission of an initial balance value, only the differences between consecutive balance values are transmitted, whereby entropy coding is employed. Very commonly, this difference is zero, which thus is signaled by the shortest possible codeword. Clearly, in applications where bit errors are possible, this delta coding must be reset at an appropriate time interval, in order to eliminate uncontrolled error propagation.


The most rudimental decoder usage of the balance parameter, is simply to offset the mono signal towards either of the two reproduction channels, by feeding the mono signal to both outputs and adjusting the gains correspondingly, as illustrated in FIG. 2c, blocks 227 and 229, with the control signal B. This is analogous to turning the “panorama” knob on a mixing desk, synthetically “moving” a mono signal between the two stereo speakers.


The balance parameter can be sent in addition to the above described width parameter, offering the possibility to both position and spread the sound image in the sound-stage in a controlled manner, offering flexibility when mimicking the original stereo impression. One problem with combining pseudo stereo generation, as mentioned in a previous section, and parameter controlled balance, is unwanted signal contribution from the pseudo stereo generator at balance positions far from center position. This is solved by applying a mono favoring function on the stereo-width value, resulting in a greater attenuation of the stereo-width value at balance positions at extreme side position and less or no attenuation at balance positions close to the center position.


The methods described so far, are intended for very low bitrate applications. In applications where higher bitrates are available, it is possible to use more elaborate versions of the above width and balance methods. Stereo-width detection can be made in several frequency bands, resulting in individual stereo-width values for each frequency band. Similarly, balance calculation can operate in a multiband fashion, which is equivalent to applying different filter-curves to two channels that are fed by a mono signal. FIG. 3 shows an example of a parametric stereo decoder using a set of N pseudo-stereo generators according to FIG. 2b, represented by blocks 307, 317 and 327, combined with multiband balance adjustment, represented by blocks 309, 319 and 329, as described in FIG. 2c. The individual passbands are obtained by feeding the mono input signal, M, to a set of bandpass filters, 305, 315 and 325. The bandpass stereo outputs from the balance adjusters are added, 311, 321, 313, 323, forming the stereo output signal, L and R. The formerly scalar width- and balance parameters are now replaced by the arrays W(k) and B(k). In FIG. 3, every pseudo-stereo generator and balance adjuster has unique stereo parameters. However, in order to reduce the total amount of data to be transmitted or stored, parameters from several frequency bands can be averaged in groups at the encoder, and this smaller number of parameters be mapped to the corresponding groups of width and balance blocks at the decoder. Clearly, different grouping schemes and lengths can be used for the arrays W(k) and B(k). S(k) represents the gains of the delay signal paths in the width blocks, and D(k) represents the delay parameters. Again, S(k) and D(k) are optional in the bitstream.


The parametric balance coding method can, especially for lower frequency bands, give a somewhat unstable behavior, due to lack of frequency resolution, or due to too many sound events occurring in one frequency band at the same time but at different balance positions. Those balance-glitches are usually characterized by a deviant balance value during just a short period of time, typically one or a few consecutive values calculated, dependent on the update rate. In order to avoid disturbing balance-glitches, a stabilization process can be applied on the balance data. This process may use a number of balance values before and after current time position, to calculate the median value of those. The median value can subsequently be used as a limiter value for the current balance value i.e., the current balance value should not be allowed to go beyond the median value. The current value is then limited by the range between the last value and the median value. Optionally, the current balance value can be allowed to pass the limited values by a certain overshoot factor. Furthermore, the overshoot factor, as well as the number of balance values used for calculating the median, should be seen as frequency dependent properties and hence be individual for each frequency band.


At low update ratios of the balance information, the lack of time resolution can cause failure in synchronization between motions of the stereo image and the actual sound events. To improve this behavior in terms of synchronization, an interpolation scheme based on identifying sound events can be used. Interpolation here refers to interpolations between two, in time consecutive balance values. By studying the mono signal at the receiver side, information about beginnings and ends of different sound events can be obtained. One way is to detect a sudden increase or decrease of signal energy in a particular frequency band. The interpolation should after guidance from that energy envelope in time make sure that the changes in balance position should be performed preferably during time segments containing little signal energy. Since human ear is more sensitive to entries than trailing parts of a sound, the interpolation scheme benefits from finding the beginning of a sound by e.g., applying peak-hold to the energy and then let the balance value increments be a function of the peak-holded energy, where a small energy value gives a large increment and vice versa. For time segments containing uniformly distributed energy in time i.e., as for some stationary signals, this interpolation method equals linear interpolation between the two balance values. If the balance values are quotients of left and right energies, logarithmic balance values are preferred, for left-right symmetry reasons. Another advantage of applying the whole interpolation algorithm in the logarithmic domain is the human ear's tendency of relating levels to a logarithmic scale.


Also, for low update ratios of the stereo-width gain values, interpolation can be applied to the same. A simple way is to interpolate linearly between two in time consecutive stereo-width values. More stable behavior of the stereo-width can be achieved by smoothing the stereo-width gain values over a longer time segment containing several stereo-width parameters. By utilizing smoothing with different attack and release time constants, a system well suited for program material containing mixed or interleaved speech and music is achieved. An appropriate design of such smoothing filter is made using a short attack time constant, to get a short rise-time and hence an immediate response to music entries in stereo, and a long release time, to get a long fall-time. To be able to fast switch from a wide stereo mode to mono, which can be desirable for sudden speech entries, there is a possibility to bypass or reset the smoothing filter by signaling this event. Furthermore, attack time constants, release time constants and other smoothing filter characteristics can also be signaled by an encoder.


For signals containing masked distortion from a psycho-acoustical codec, one common problem with introducing stereo information based on the coded mono signal is an unmasking effect of the distortion. This phenomenon usually referred as “stereo-unmasking” is the result of non-centered sounds that do not fulfill the masking criterion. The problem with stereo-unmasking might be solved or partly solved by, at the decoder side, introducing a detector aimed for such situations. Known technologies for measuring signal to mask ratios can be used to detect potential stereo-unmasking. Once detected, it can be explicitly signaled or the stereo parameters can just simply be decreased.


At the encoder side, one option, as taught by the invention, is to employ a Hilbert transformer to the input signal, i.e. a 90 degree phase shift between the two channels is introduced. When subsequently forming the mono signal by addition of the two signals, a better balance between a center-panned mono signal and “true” stereo signals is achieved, since the Hilbert transformation introduces a 3 dB attenuation for center information. In practice, this improves mono coding of e.g. contemporary pop music, where for instance the lead vocals and the bass guitar commonly is recorded using a single mono source.


The multiband balance-parameter method is not limited to the type of application described in FIG. 1. It can be advantageously used whenever the objective is to efficiently encode the power spectral envelope of a stereo signal. Thus, it can be used as tool in stereo codecs, where in addition to the stereo spectral envelope a corresponding stereo residual is coded. Let the total power P, be defined by P=PL+PR, where PL and PR are signal powers as described above. Note that this definition does not take left to right phase relations into account. (E.g. identical left and right signals but of opposite signs, does not yield a zero total power.) Analogous to B, P can be expressed in dB as PdB=10 log10(P/Pref), where Pref is an arbitrary reference power, and the delta values be entropy coded. As opposed to the balance case, no progressive quantization is employed for P. In order to represent the spectral envelope of a stereo signal, P and B are calculated for a set of frequency bands, typically, but not necessarily, with bandwidths that are related to the critical bands of human hearing. For example those bands may be formed by grouping of channels in a constant bandwidth filterbank, whereby PL and PR are calculated as the time and frequency averages of the squares of the subband samples corresponding to respective band and period in time. The sets P0, P1, P2, . . . , PN-1 and B0, B1, B2, BN-1, where the subscripts denote the frequency band in an N band representation, are delta and Huffman coded, transmitted or stored, and finally decoded into the quantized values that were calculated in the encoder. The last step is to convert P and B back to PL and PR. As easily seen form the definitions of P and B, the reverse relations are (when neglecting e in the definition of B) PL=BP/(B+1), and PR=P/(B+1).


One particularly interesting application of the above envelope coding method is coding of highband spectral envelopes for HFR-based codecs. In this case no highband residual signal is transmitted. Instead this residual is derived from the lowband. Thus, there is no strict relation between residual and envelope representation, and envelope quantization is more crucial. In order to study the effects of quantization, let Pq and Bq denote the quantized values of P and B respectively. Pq and Bq are then inserted into the above relations, and the sum is formed: PLq+PR q=BqPq/(Bq+1)+Pq/(Bq+1)=Pq(Bq+1)/(Bq+1)=Pq. The interesting feature here is that Bq is eliminated, and the error in total power is solely determined by the quantization error in P. This implies that even though B is heavily quantized, the perceived level is correct, assuming that sufficient precision in the quantization of P is used. In other words, distortion in B maps to distortion in space, rather than in level. As long as the sound sources are stationary in the space over time, this distortion in the stereo perspective is also stationary, and hard to notice. As already stated, the quantization of the stereo-balance can also be coarser towards the outer extremes, since a given error in dB corresponds to a smaller error in perceived angle when the angle to the centerline is large, due to properties of human hearing.


When quantizing frequency dependent data e.g., multi band stereo-width gain values or multi band balance values, resolution and range of the quantization method can advantageously be selected to match the properties of a perceptual scale. If such scale is made frequency dependent, different quantization methods, or so called quantization classes, can be chosen for the different frequency bands. The encoded parameter values representing the different frequency bands, should then in some cases, even if having identical values, be interpreted in different ways i.e., be decoded into different values.


Analogous to a switched L/R- to S/D-coding scheme, the P and B signals may be adaptively substituted by the PL and PR signals, in order to better cope with extreme signals. As taught by [PCT/SE00/00158], delta coding of envelope samples can be switched from delta-in-time to delta-in-frequency, depending on what direction is most efficient in terms of number of bits at a particular moment. The balance parameter can also take advantage of this scheme: Consider for example a source that moves in stereo field over time. Clearly, this corresponds to a successive change of balance values over time, which depending on the speed of the source versus the update rate of the parameters, may correspond to large delta-in-time values, corresponding to large codewords when employing entropy coding. However, assuming that the source has uniform sound radiation versus frequency, the delta-in-frequency values of the balance parameter are zero at every point in time, again corresponding to small codewords. Thus, a lower bitrate is achieved in this case, when using the frequency delta coding direction. Another example is a source that is stationary in the room, but has a non-uniform radiation. Now the delta-in-frequency values are large, and delta-in-time is the preferred choice.


The P/B-coding scheme offers the possibility to build a scalable HFR-codec, see FIG. 4. A scalable codec is characterized in that the bitstream is split into two or more parts, where the reception and decoding of higher order parts is optional. The example assumes two bitstream parts, hereinafter referred to as primary, 419, and secondary, 417, but extension to a higher number of parts is clearly possible. The encoder side, FIG. 4a, comprises of an arbitrary stereo lowband encoder, 403, which operates on the stereo input signal, IN (the trivial steps of AD-respective DA-conversion are not shown in the figure), a parametric stereo encoder, which estimates the highband spectral envelope, and optionally additional stereo parameters, 401, which also operates on the stereo input signal, and two multiplexers, 415 and 413, for the primary and secondary bitstreams respectively. In this application, the highband envelope coding is locked to P/B-operation, and the P signal, 407, is sent to the primary bitstream by means of 415, whereas the B signal, 405, is sent to the secondary bitstream, by means of 413.


For the lowband codec different possibilities exist: It may constantly operate in S/D-mode, and the S and D signals be sent to primary and secondary bitstreams respectively. In this case, a decoding of the primary bitstream results in a full band mono signal. Of course, this mono signal can be enhanced by parametric stereo methods according to the invention, in which case the stereo-parameter(s) also must be located in the primary bitstream. Another possibility is to feed a stereo coded lowband signal to the primary bitstream, optionally together with highband width- and balance-parameters. Now decoding of the primary bitstream results in true stereo for the lowband, and very realistic pseudo-stereo for the highband, since the stereo properties of the lowband are reflected in the high frequency reconstruction. Stated in another way: Even though the available highband envelope representation or spectral coarse structure is in mono, the synthesized highband residual or spectral fine structure is not. In this type of implementation, the secondary bitstream may contain more lowband information, which when combined with that of the primary bitstream, yields a higher quality lowband reproduction. The topology of FIG. 4 illustrates both cases, since the primary and secondary lowband encoder output signals, 411, and 409, connected to 415 and 417 respectively, may contain either of the above described signal types.


The bitstreams are transmitted or stored, and either only 419 or both 419 and 417 are fed to the decoder, FIG. 4b. The primary bitstream is demultiplexed by 423, into the lowband core decoder primary signal, 429 and the P signal, 431. Similarly, the secondary bitstream is demultiplexed by 421, into the lowband core decoder secondary signal, 427, and the B signal, 425. The lowband signal(s) is(are) routed to the lowband decoder, 433, which produces an output, 435, which again, in case of decoding of the primary bitstream only, may be of either type described above (mono or stereo). The signal 435 feeds the HFR-unit, 437, wherein a synthetic highband is generated, and adjusted according to P, which also is connected to the HFR-unit. The decoded lowband is combined with the highband in the HFR-unit, and the lowband and/or highband is optionally enhanced by a pseudo-stereo generator (also situated in the HFR-unit), before finally being fed to the system outputs, forming the output signal, OUT. When the secondary bitstream, 417, is present, the HFR-unit also gets the B signal as an input signal, 425, and 435 is in stereo, whereby the system produces a full stereo output signal, and pseudo-stereo generators if any, are bypassed.


Stated in other words, a method for coding of stereo properties of an input signal, includes at an encoder, the step of calculating a width-parameter that signals a stereo-width of said input signal, and at a decoder, a step of generating a stereo output signal, using said width-parameter to control a stereo-width of said output signal. The method further comprises at said encoder, forming a mono signal from said input signal, wherein, at said decoder, said generation implies a pseudo-stereo method operating on said mono signal. The method further implies splitting of said mono signal into two signals as well as addition of delayed version(s) of said mono signal to said two signals, at level(s) controlled by said width-parameter. The method further includes that said delayed version(s) are high-pass filtered and progressively attenuated at higher frequencies prior to being added to said two signals. The method further includes that said width-parameter is a vector, and the elements of said vector correspond to separate frequency bands. The method further includes that if said input signal is of type dual mono, said output signal is also of type dual mono.


A method for coding of stereo properties of an input signal, includes at an encoder, calculating a balance parameter that signals a stereo-balance of said input signal, and at a decoder, generate a stereo output signal, using said balance-parameter to control a stereo-balance of said output signal.


In this method, at said encoder, a mono signal from said input signal is formed, and at said decoder, said generation implies splitting of said mono signal into two signals, and said control implies adjustment of levels of said two signals. The method further includes that a power for each channel of said input signal is calculated, and said balance-parameter is calculated from a quotient between said powers. The method further includes that said powers and said balance-parameter are vectors where every element corresponds to a specific frequency band. The method further includes that at said decoder it is interpolated between two in time consecutive values of said balance-parameters in a way that the momentary value of the corresponding power of said mono signal controls how steep the momentary interpolation should be. The method further includes that said interpolation method is performed on balance values represented as logarithmic values. The method further includes that said values of balance parameters are limited to a range between a previous balance value, and a balance value extracted from other balance values by a median filter or other filter process, where said range can be further extended by moving the borders of said range by a certain factor. The method further includes that said method of extracting limiting borders for balance values, is, for a multi band system, frequency dependent. The method further includes that an additional level-parameter is calculated as a vector sum of said powers and sent to said decoder, thereby providing said decoder a representation of a spectral envelope of said input signal. The method further includes that said level-parameter and said balance-parameter adaptively are replaced by said powers. The method further includes that said spectral envelope is used to control a HFR-process in a decoder. The method further includes that said level-parameter is fed into a primary bitstream of a scalable HFR-based stereo codec, and said balance-parameter is fed into a secondary bitstream of said codec. Said mono signal and said width-parameter are fed into said primary bitstream. Furthermore, said width-parameters are processed by a function that gives smaller values for a balance value that corresponds to a balance position further from the center position. The method further includes that a quantization of said balance-parameter employs smaller quantization steps around a center position and larger steps towards outer positions. The method further includes that said width-parameters and said balance-parameters are quantized using a quantization method in terms of resolution and range which, for a multiband system, is frequency dependent. The method further includes that said balance parameter adaptively is delta-coded either in time or in frequency. The method further includes that said input signal is passed though a Hilbert transformer prior to forming said mono signal.


An apparatus for parametric stereo coding, includes, at an encoder, means for calculation of a width-parameter that signals a stereo-width of an input signal, and means for forming a mono signal from said input signal, and, at a decoder, means for generating a stereo output signal from said mono signal, using said width-parameter to control a stereo-width of said output signal.

Claims
  • 1. A decoder configured to decode an encoded bitstream, the decoder comprising: a demultiplexer for demultiplexing the encoded bitstream for obtaining a lowband core decoder signal, level parameters, and balance parameters;a lowband core decoder for producing a lowband output signal, the lowband output signal having a lowband mono signal or a lowband stereo signal;a high-frequency reconstruction device for generating a synthetic highband using the lowband output signal, the level parameters, and the balance parameters and for combining the synthetic highband and the lowband output signal to form a combined signal, andan output interface for outputting the combined signal,wherein the level parameters represent a total power in a frequency band of a signal having two channels,wherein the total power represents a sum of an energy of a left channel and an energy of a right channel for a given time segment and frequency band,wherein the balance parameters represent a quotient of an energy of the left channel and an energy of the right channel,wherein the level parameters are delta coded in frequency.
  • 2. A method for decoding an encoded bitstream, the method comprising: demultiplexing, by a demultiplexer, the encoded bitstream for obtaining a lowband core decoder signal, level parameters, and balance parameters;producing, by a lowband decoder, a lowband output signal, the lowband output signal having a lowband mono signal or a lowband stereo signal;generating, by a high-frequency reconstruction device, a synthetic highband using the lowband output signal, the level parameters, and the balance parameters;combining the synthetic highband and the lowband output signal to form a combined signal, andoutputting the combined signal,wherein the level parameters represent a total power in a frequency band of a signal having two channels,wherein the total power represents a sum of an energy of a left channel and an energy of a right channel for a given time segment and frequency band,wherein the balance parameters represent a quotient of an energy of the left channel and an energy of the right channel,wherein the level parameters are delta coded in frequency.
  • 3. A decoder configured to decode an encoded bitstream, the decoder comprising: a demultiplexer for demultiplexing the encoded bitstream for obtaining a lowband core decoder signal, level parameters, and balance parameters;a lowband core decoder for producing a lowband output signal, the lowband output signal having a lowband mono signal or a lowband stereo signal;a high-frequency reconstruction device for generating a synthetic highband using the lowband output signal, the level parameters, and the balance parameters and for combining the synthetic highband and the lowband output signal to form a combined signal,a parametric stereo decoder for generating a stereo audio signal from the combined signal; andan output interface for outputting the stereo signal,wherein the level parameters represent a total power in a frequency band of a signal having two channels,wherein the total power represents a sum of an energy of a left channel and an energy of a right channel for a given time segment and frequency band,wherein the balance parameters represent a quotient of an energy of the left channel and an energy of the right channel,wherein the level parameters are delta coded in frequency.
  • 4. A method for decoding an encoded bitstream, the method comprising: demultiplexing, by a demultiplexer, the encoded bitstream for obtaining a lowband core decoder signal, level parameters, and balance parameters;producing, by a lowband decoder, a lowband output signal, the lowband output signal having a lowband mono signal or a lowband stereo signal;generating, by a high-frequency reconstruction device, a synthetic highband using the lowband output signal, the level parameters, and the balance parameters;combining the synthetic highband and the lowband output signal to form a combined signal;generating a stereo signal from the combined signal; andoutputting the stereo signal,wherein the level parameters represent a total power in a frequency band of a signal having two channels,wherein the total power represents a sum of an energy of a left channel and an energy of a right channel for a given time segment and frequency band,wherein the balance parameters represent a quotient of an energy of the left channel and an energy of the right channel,wherein the level parameters are delta coded in frequency.
Priority Claims (3)
Number Date Country Kind
0102481 Jul 2001 SE national
0200796 Mar 2002 SE national
0202159 Jul 2002 SE national
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/078,456 filed on Nov. 12, 2013 which is a continuation of U.S. patent application Ser. No. 12/610,186 filed on Oct. 30, 2009, which issued on Dec. 10, 2013 as U.S. Pat. No. 8,605,911, which is a divisional of U.S. patent application Ser. No. 11/238,982 filed on Sep. 28, 2005, which issued on Feb. 14, 2012 as U.S. Pat. No. 8,116,460, and which is a divisional of U.S. patent application Ser. No. 10/483,453 filed on Jan. 8, 2004, which issued on Jun. 3, 2008 as U.S. Pat. No. 7,382,886, which claims priority to PCT/SE02/01372, filed Jul. 10, 2002, which claims priority to Swedish Application Serial No. 0102481-9, filed Jul. 10, 2001, Swedish Application Serial No. 0200796-1, filed Mar. 15, 2002, and Swedish Application Serial No. 0202159-0, filed Jul. 9, 2002, each of which is herein incorporated by reference.

US Referenced Citations (100)
Number Name Date Kind
3947827 Dautremont, Jr. et al. Mar 1976 A
4053711 DeFreitas et al. Oct 1977 A
4166924 Berkley et al. Sep 1979 A
4216354 Esteban et al. Aug 1980 A
4330689 Kang et al. May 1982 A
4569075 Nussbaumer Feb 1986 A
4667340 Arjmand et al. May 1987 A
4672670 Wang et al. Jun 1987 A
4700362 Todd et al. Oct 1987 A
4700390 Machida Oct 1987 A
4706287 Blackmer et al. Nov 1987 A
4776014 Zinser, Jr. Oct 1988 A
4969040 Gharavi Nov 1990 A
5001758 Galand et al. Mar 1991 A
5054072 McAulay et al. Oct 1991 A
5093863 Galand et al. Mar 1992 A
5127054 Hong et al. Jun 1992 A
5235420 Gharavi Aug 1993 A
5261027 Taniguchi et al. Nov 1993 A
5285520 Matsumoto et al. Feb 1994 A
5293449 Tzeng Mar 1994 A
5309526 Pappas et al. May 1994 A
5321793 Drogo De Iacovo et al. Jun 1994 A
5396237 Ohta Mar 1995 A
5455888 Iyengar et al. Oct 1995 A
5463424 Dressler Oct 1995 A
5490233 Kovacevic Feb 1996 A
5517581 Johnston et al. May 1996 A
5559891 Kuusama et al. Sep 1996 A
5579434 Kudo et al. Nov 1996 A
5581562 Lin et al. Dec 1996 A
5581652 Abe et al. Dec 1996 A
5581653 Todd Dec 1996 A
5604810 Yanagawa Feb 1997 A
5613035 Kim Mar 1997 A
5632005 Davis et al. May 1997 A
5671287 Gerzon Sep 1997 A
5677985 Ozawa Oct 1997 A
5687191 Lee et al. Nov 1997 A
5701346 Herre et al. Dec 1997 A
5701390 Griffin et al. Dec 1997 A
5757938 Akagiri et al. May 1998 A
5787387 Aguilar Jul 1998 A
5848164 Levine Dec 1998 A
5862228 Davis Jan 1999 A
5873065 Akagiri et al. Feb 1999 A
5875122 Acharya Feb 1999 A
5878388 Nishiguchi et al. Mar 1999 A
5883962 Hawks Mar 1999 A
5889857 Boudy et al. Mar 1999 A
5890108 Yeldener Mar 1999 A
5890125 Davis et al. Mar 1999 A
5915235 DeJaco et al. Jun 1999 A
5950153 Ohmori et al. Sep 1999 A
5951235 Young et al. Sep 1999 A
RE36478 McAulay et al. Dec 1999 E
6014619 Wuppermann et al. Jan 2000 A
6144937 Ali Nov 2000 A
6226325 Nakamura May 2001 B1
6233551 Cho et al. May 2001 B1
6298361 Suzuki Oct 2001 B1
6389006 Bialik May 2002 B1
6456657 Yeap et al. Sep 2002 B1
6507658 Abel et al. Jan 2003 B1
6611800 Nishiguchi et al. Aug 2003 B1
6680972 Liljeryd et al. Jan 2004 B1
6771777 Gbur et al. Aug 2004 B1
6772114 Sluijter et al. Aug 2004 B1
6853682 Min Feb 2005 B2
6871106 Ishikawa et al. Mar 2005 B1
6879955 Rao Apr 2005 B2
6895375 Malah et al. May 2005 B2
6988066 Malah Jan 2006 B2
7050972 Henn et al. May 2006 B2
7095907 Berkner et al. Aug 2006 B1
7151802 Bessette et al. Dec 2006 B1
7191123 Bessette et al. Mar 2007 B1
7191136 Sinha et al. Mar 2007 B2
7200561 Moriya et al. Apr 2007 B2
7205910 Honma et al. Apr 2007 B2
7216074 Malah et al. May 2007 B2
7260521 Bessette et al. Aug 2007 B1
7283967 Nishio et al. Oct 2007 B2
7328160 Nishio et al. Feb 2008 B2
7382886 Henn et al. Jun 2008 B2
7720676 Philippe et al. May 2010 B2
20020010577 Matsumoto et al. Jan 2002 A1
20020037086 Irwan et al. Mar 2002 A1
20020040299 Makino et al. Apr 2002 A1
20020103637 Henn et al. Aug 2002 A1
20020123975 Poluzzi et al. Sep 2002 A1
20030063759 Brennan et al. Apr 2003 A1
20030088423 Nishio et al. May 2003 A1
20030093278 Malah May 2003 A1
20030206624 Domer et al. Nov 2003 A1
20030215013 Budnikov Nov 2003 A1
20040117177 Kjorling et al. Jun 2004 A1
20040252772 Renfors et al. Dec 2004 A1
20050074127 Herre et al. Apr 2005 A1
20050187759 Malah et al. Aug 2005 A1
Foreign Referenced Citations (61)
Number Date Country
19947098 Nov 2000 DE
0478096 Jan 1987 EP
0273567 Jul 1988 EP
0485444 May 1992 EP
501690 Jan 1997 EP
0858067 Aug 1998 EP
0918407 May 1999 EP
0989543 Mar 2000 EP
1119911 Jul 2000 EP
1107232 Jun 2001 EP
2100430 Dec 1982 GB
2344036 Jan 2004 GB
02012299 Jan 1990 JP
02177782 Jul 1990 JP
03214956 Sep 1991 JP
04301688 Oct 1992 JP
5-191885 Jul 1993 JP
05165500 Jul 1993 JP
06-85607 Mar 1994 JP
06090209 Mar 1994 JP
6-118995 Apr 1994 JP
06202629 Jul 1994 JP
06215482 Aug 1994 JP
H08-123495 May 1996 JP
08254994 Oct 1996 JP
08305398 Nov 1996 JP
H08-263096 Nov 1996 JP
09-500252 Jan 1997 JP
09-046233 Feb 1997 JP
09-055778 Feb 1997 JP
09-501286 Feb 1997 JP
09-090992 Apr 1997 JP
09-101798 Apr 1997 JP
09505193 May 1997 JP
09261064 Oct 1997 JP
H09-282793 Oct 1997 JP
H10-504170 Apr 1998 JP
11262100 Sep 1999 JP
11317672 Nov 1999 JP
2000083014 Mar 2000 JP
2000505266 Apr 2000 JP
2000-267699 Sep 2000 JP
2001184090 Jul 2001 JP
2001-521648 Nov 2001 JP
2004535145 Nov 2004 JP
96003455 Mar 1996 KR
960012475 Sep 1996 KR
WO-9504442 Feb 1995 WO
WO-9516333 Jun 1995 WO
WO-9700594 Jan 1997 WO
WO-9730438 Aug 1997 WO
WO-9803036 Jan 1998 WO
WO-9803037 Jan 1998 WO
WO-9857436 Dec 1998 WO
WO-9857436 Dec 1998 WO
WO0045379 Aug 2000 WO
WO-0045379 Aug 2000 WO
WO-0045378 Aug 2000 WO
WO-0079520 Dec 2000 WO
WO-03007656 Jan 2003 WO
WO2004027368 Apr 2004 WO
Non-Patent Literature Citations (40)
Entry
Bauer, D., “Examinations Regarding the Similarity of Digital Stereo Signals in High Quality Music Reproduction”, University of Erlangen-Neumberg, 1991, 1-30.
Chen, S., “A Survey of Smoothing Techniques for ME Models”, IEEE, R. Rosenfeld (Additional Author), Jan. 2000, 37-50.
Cheng, Yan M. et al., “Statistical Recovery of Wideband Speech from Narrowband Speech”, IEEE Trans. Speech and Audio Processing, vol. 2, No. 4, Oct. 1994, 544-548.
Chennoukh, S. et al., “Speech Enhancement Via Frequency Bandwidth Extension Using Line Spectral Frequencies”, IEEE Conference on Acoustics, Speech, and Signal Processing Proceedings (ICASSP), 2001, 665-668.
Chouinard, et al., “Wideband communications in the high frequency band using direct sequence spread spectrum with error control coding”, IEEE Military Communications Conference, Nov. 5, 1995, pp. 560-567.
Depalle, et al., “Extraction of Spectral Peak Parameters Using a Short-time Fourier Transform Modeling and No Sidelobe Windows”, IEEE ASSP Workshop on Volume, Oct. 1997, 4 pages.
Dutilleux, Pierre, “Filters, Delays, Modulations and Demodulations: A Tutorial”, Retrieved from internet address: http://on1.akm.de/skm/Institute/Musik/SKMusik/veroeffentlicht/PD.sub.—Fi- lters, No publication date can be found. Retrieved on Feb. 19, 2009, Total of 13 pages.
Enbom, Niklas et al., “Bandwidth Expansion of Speech Based on Vector Quantization of the Mel Frequency Cepstral Coefficients”, Proc. IEEE Speech Coding Workshop (SCW), 1999, 171-173.
Epps, Julien , “Wideband Extension of Narrowband Speech for Enhancement and Coding”, School of Electical Engineering and Telecommunications, The University of New South Wales, Sep. 2000, 1-155.
George, et al., “Analysis-by-Synthesis/Overlap-Add Sinusoidal Modeling Applied to the Analysis and Synthesis of Musical Tones”, Journal of Audio Engineering Society, vol. 40, No. 6, Jun. 1992, 497-516.
Herre, Jurgen et al., “Intensity Stereo Coding”, Preprints of Papers Presented at the Audio Engineering Society Convention, vol. 96, No. 3799, XP009025131, Feb. 26, 1994, 1-10.
Holger, C et al., “Bandwidth Enhancement of Narrow-Band Speech Signals”, Signal Processing VII Theories and Applications, Proc. of EUSIPCO-94, Seventh European Signal Processing Conference; European Association for Signal Processing Sep. 13-16, 1994, 1178-1181.
Kubin, Gernot, “Synthesis and Coding of Continuous Speech With the Nonlinear Oscillator Model”, Institute of Communications and High-Frequency Engineering, Vienna University of Technology, Vienna, Austria, IEEE, 1996, 267-270.
Makhoul, et al., “High-Frequency Regeneration in Speech Coding Systems”, Proc. Intl. Conf. Acoustic: Speech, Signal Processing, Apr. 1979, pp. 428-431.
McNally, G.W., “Dynamic Range Control of Digital Audio Signals”, Journal of Audio Engineering Society, vol. 32, No. 5, May 1984, 316-327.
Princen, John P. et al., “Analysis/Synthesis Filter Bank Design Based on Time Domain Aliasing Cancellation”, IEEE Trans. on Acoustics, Speech, and Signal Processing, vol. ASSP-34, No. 5, Oct. 5, 1986, 1153-1161.
Proakis, “Digital Signal Processing”, Sampling and Reconstrction of Signals, Chapter 9, Monolakic (Additional Author) Submitted with a Declaration 1, 1996, 771-773.
Schroeder, Manfred R., “An Artificial Stereophonic Effect Obtained from Using a Single Signal”, 9th Annual Meeting, Audio Engineering Society, Oct. 8-12, 1957, 1-5.
Taddei, et al., “A Scalable Three Bit-rates 8-14.1-24 kbit/s Audio Coder”, vol. 55, Sep. 2000, pp. 483-492.
Vaidyanathan, P. P., “Multirate Digital Filters, Filter Banks,Polyphase Networks, and Applications: A Tutorial”, Proceedings of the IEEE, vol. 78, No. 1, Jan. 1990, 56-93.
Valin, et al., “Bandwidth Extension of Narrowband Speech for Low Bit-Rate Wideband Coding”, IEEE Workshop Speech Coding Proceedings, Sep. 2000, pp. 130-132.
Yasukawa, Hiroshi , “Restoration of Wide Band Signal from Telephone Speech Using Linear Prediction Error Processing”, Conf. Spoken Language Processing (ICSLP), 1996, 901-904.
Zolzer Udo, “Digital Audio Signal Processing”, John Wiley & Sons Ltd., England, 1997, 207-247.
Brandenburg “Introductions to Perceptual Coding”, Published by Audio Engineering Society in “Collected Papers on Digital Audio Bit-Rate Reduction”, Manuscript received on Mar. 13, 1996, 1996, Total of 11 pages.
Britanak, et al., “A new fast algorithm for the unified forward and inverse MDCT/MDST Computation”, Signal Processing, vol. 82, Mar. 2002, pp. 433-459.
Cruz-Roldan, et al., “Alternating Analysis and Synthesis Filters: A New Pseudo-QMF Bank”, Oct. 2001.
Ekstrand, Per , “Bandwidth extension of audio signals by spectral band replication”, Proc.1st IEEE Benelux Workshop on Model Based Processing and Coding of Audio, Leuven, Belgium, Nov. 15, 2002, pp. 53-58.
Gilchrist, N. et al., “Collected Papers on Digital Audio Bit-Rate Reduction”, Audio-Engineering Society, No. 3, 1996, Total of 11 pages.
Gilloire, et al., “Adaptive Filtering in Subbands with Critical Sampling: Analysis, Experiments, and Application to Acoustic Echo”, 1992.
Gilloire, et al., “Adaptive Filtering in Subbands with Critical Sampling: Analysis, Experiments, and Application to Acoustic Echo Cancellation”, IEEE Transaction on Signal Processing, vol. 40, No. 8, Aug. 1992, 1862-1875.
Harteneck, et al., “Filterbank design for oversampled filter banks without aliasing in the subbands”, Electronic Letters, vol. 33, No. 18, Sug. 28, 1997, pp. 1538-1539.
Holger, C et al., “Bandwidth Enhancement of Narrow-Band Speech Signals”, Signal Processing VII Theories and Applications, Proc. of EUSIPC0-94, Seventh European Signal Processing Conference; European Association for Signal Processing, Sep. 13-16, 1994, 1178-1181.
Koilpillai, et al., “A Spectral Factorization Approach to Pseudo-QMF Desig”, IEEE Transactions on Signal Processing, Jan. 1993, 82-92.
Kok, et al., “Multirate filter banks and transform coding gain”, IEEE Transactions on Signal Processing, vol. 46 (7), Jul. 1998,2041-2044.
Nguyen “Near-Perfect-Reconstruction Pseudo-QMF Banks”, IEEE Transaction on Signal Processing, vol. 42, No. 1, Jan. 1994, 65-76.
Ramstad, T.A. et al., “Cosine-modulated analysis-syntheses filter bank with critical sampling and perfect reconstruction”, IEEE Int'l. Conf. ASSP, Toronto, Canada, May 1991, 1789-1792.
Tam, et al., “Highly Oversampled Subband Adaptive Filters for Noise Cancellation on a Low-Resource DSP System”, ICSLP, Sep. 2002, Total of 4 pages.
Weiss, S. et al., “Efficient implementations of complex and real valued filter banks for comparative subband processing with an application to adaptive filtering”, Proc. Int'l Symposium Communication Systems & Digital Signal Processing, vol. 1, Sheffield, UK, Apr. 1998, 4 pages.
Ziegler, et al., “Enhancing mp3 with SBR: Fetaures and Capabilities of the new mp3PRO Algorithm”, AES 112th Convention, Munich, Germany, May 2002, Total of 7 pages.
Zolzer, Udo, “Digital Audio Signal Processing”, John Wiley & Sons Ltd., England, 1997, pp. 207-247.
Related Publications (1)
Number Date Country
20170186435 A1 Jun 2017 US
Divisions (3)
Number Date Country
Parent 12610186 Oct 2009 US
Child 14078456 US
Parent 11238982 Sep 2005 US
Child 12610186 US
Parent 10483453 US
Child 11238982 US
Continuations (1)
Number Date Country
Parent 14078456 Nov 2013 US
Child 15458135 US