The application relates to audio coding, in particular to stereo audio coding combining parametric and waveform based coding techniques.
Joint coding of the left (L) and right (R) channels of a stereo signal enables more efficient coding compared to independent coding of L and R. A common approach for joint stereo coding is mid/side (M/S) coding. Here, a mid (M) signal is formed by adding the L and R signals, e.g. the M signal may have the form
Also, a side (S) signal is formed by subtracting the two channels L and R, e.g. the S signal may have the form
In case of M/S coding, the M and S signals are coded instead of the L and R signals.
In the MPEG (Moving Picture Experts Group) AAC (Advanced Audio Coding) standard (see standard document ISO/IEC 13818-7), L/R stereo coding and M/S stereo coding can be chosen in a time-variant and frequency-variant manner Thus, the stereo encoder can apply L/R coding for some frequency bands of the stereo signal, whereas M/S coding is used for encoding other frequency bands of the stereo signal (frequency variant). Moreover, the encoder can switch over time between L/R and M/S coding (time-variant). In MPEG AAC, the stereo encoding is carried out in the frequency domain, more particularly in the MDCT (modified discrete cosine transform) domain. This allows to adaptive choose either L/R or M/S coding in a frequency and also time variant manner. The decision between L/R and M/S stereo encoding may be based by evaluating the side signal: when the energy of the side signal is low, M/S stereo encoding is more efficient and should be used. Alternatively, for deciding between both stereo coding schemes, both coding schemes may be tried out and the selection may be based on the resulting quantization efforts, i.e., the observed perceptual entropy.
An alternative approach to joint stereo coding is parametric stereo (PS) coding. Here, the stereo signal is conveyed as a mono downmix signal after encoding the downmix signal with a conventional audio encoder such as an AAC encoder. The downmix signal is a superposition of the L and R channels. The mono downmix signal is conveyed in combination with additional time-variant and frequency-variant PS parameters, such as the inter-channel (i.e. between L and R) intensity difference (IID) and the inter-channel cross-correlation (ICC). In the decoder, based on the decoded downmix signal and the parametric stereo parameters a stereo signal is reconstructed that approximates the perceptual stereo image of the original stereo signal. For reconstructing, a decorrelated version of the downmix signal is generated by a decorrelator. Such decorrelator may be realized by an appropriate all-pass filter. PS encoding and decoding is described in the paper “Low Complexity Parametric Stereo Coding in MPEG-4”, H. Purnhagen, Proc. Of the 7th Int. Conference on Digital Audio Effects (DAFx′04), Naples, Italy, Oct. 5-8, 2004, pages 163-168. The disclosure of this document is hereby incorporated by reference.
The MPEG Surround standard (see document ISO/IEC 23003-1) makes use of the concept of PS coding. In an MPEG Surround decoder a plurality of output channels is created based on fewer input channels and control parameters. MPEG Surround decoders and encoders are constructed by cascading parametric stereo modules, which in MPEG Surround are referred to as OTT modules (One-To-Two modules) for the decoder and R-OTT modules (Reverse-One-To-Two modules) for the encoder. An OTT module determines two output channels by means of a single input channel (downmix signal) accompanied by PS parameters. An OTT module corresponds to a PS decoder and an R-OTT module corresponds to a PS encoder. Parametric stereo can be realized by using MPEG Surround with a single OTT module at the decoder side and a single R-OTT module at the encoder side; this is also referred to as “MPEG Surround 2-1-2” mode. The bitstream syntax may differ, but the underlying theory and signal processing are the same. Therefore, in the following all the references to PS also include “MPEG Surround 2-1-2” or MPEG Surround based parametric stereo.
In a PS encoder (e.g. in a MPEG Surround PS encoder) a residual signal (RES) may be determined and transmitted in addition to the downmix signal. Such residual signal indicates the error associated with representing original channels by their downmix and PS parameters. In the decoder the residual signal may be used instead of the decorrelated version of the downmix signal. This allows to better reconstruct the waveforms of the original channels L and R. The use of an additional residual signal is e.g. described in the MPEG Surround standard (see document ISO/IEC 23003-1) and in the paper “MPEG Surround—The ISO/MPEG Standard for Efficient and Compatible Multi-Channel Audio Coding, J. Herre et al., Audio Engineering Convention Paper 7084, 122nd Convention, May 5-8, 2007. The disclosure of both documents, in particular the remarks to the residual signal therein, is herewith incorporated by reference.
PS coding with residual is a more general approach to joint stereo coding than M/S coding: M/S coding performs a signal rotation when transforming L/R signals into M/S signals. Also, PS coding with residual performs a signal rotation when transforming the L/R signals into downmix and residual signals. However, in the latter case the signal rotation is variable and depends on the PS parameters. Due to the more general approach of PS coding with residual, PS coding with residual allows a more efficient coding of certain types of signals like a paned mono signal than M/S coding. Thus, the proposed coder allows to efficiently combine parametric stereo coding techniques with waveform based stereo coding techniques.
Often, perceptual stereo encoders, such as an MPEG AAC perceptual stereo encoder, can decide between L/R stereo encoding and M/S stereo encoding, where in the latter case a mid/side signal is generated based on the stereo signal. Such selection may be frequency-variant, i.e. for some frequency bands L/R stereo encoding may be used, whereas for other frequency bands M/S stereo encoding may be used.
In a situation where the L and R channels are basically independent signals, such perceptual stereo encoder would typically not use M/S stereo encoding since in this situation such encoding scheme does not offer any coding gain in comparison to L/R stereo encoding. The encoder would fall back to plain L/R stereo encoding, basically processing L and R independently.
In the same situation, a PS encoder system would create a downmix signal that contains both the L and R channels, which prevents independent processing of the L and R channels. For PS coding with a residual signal, this can imply less efficient coding compared to stereo encoding, where L/R stereo encoding or M/S stereo encoding is adaptively selectable.
Thus, there are situations where a PS coder outperforms a perceptual stereo coder with adaptive selection between L/R stereo encoding and M/S stereo encoding, whereas in other situations the latter coder outperforms the PS coder.
The present application describes an audio encoder system and an encoding method that are based on the idea of combing PS coding using a residual with adaptive L/R or M/S perceptual stereo coding (e.g. AAC perceptual joint stereo coding in the MDCT domain). This allows to combine the advantages of adaptive L/R or M/S stereo coding (e.g. used in MPEG AAC) and the advantages of PS coding with a residual signal (e.g. used in MPEG Surround). Moreover, the application describes a corresponding audio decoder system and a decoding method.
A first aspect of the application relates to an encoder system for encoding a stereo signal to a bitstream signal. According to an embodiment of the encoder system, the encoder system comprises a downmix stage for generating a downmix signal and a residual signal based on the stereo signal. The residual signal may cover all or only a part of the used audio frequency range. In addition, the encoder system comprises a parameter determining stage for determining PS parameters such as an inter-channel intensity difference and an inter-channel cross-correlation. Preferably, the PS parameters are frequency-variant. Such downmix stage and the parameter determining stage are typically part of a PS encoder.
In addition, the encoder system comprises perceptual encoding means downstream of the downmix stage, wherein two encoding schemes are selectable:
It should be noted that in case encoding is based on the downmix signal and the residual signal, the downmix signal and the residual signal may be encoded or signals proportional thereto may be encoded. In case encoding is based on a sum and on a difference, the sum and difference may be encoded or signals proportional thereto may be encoded.
The selection may be frequency-variant (and time-variant), i.e. for a first frequency band it may be selected that the encoding is based on a sum signal and a difference signal, whereas for a second frequency band it may be selected that the encoding is based on the downmix signal and based on the residual signal.
Such encoder system has the advantage that is allows to switch between L/R stereo coding and PS coding with residual (preferably in a frequency-variant manner): If the perceptual encoding means select (for a particular band or for the whole used frequency range) encoding based on downmix and residual signals, the encoding system behaves like a system using standard PS coding with residual. However, if the perceptual encoding means select (for a particular band or for the whole used frequency range) encoding based on a sum signal of the downmix signal and the residual signal and based on a difference signal of the downmix signal and the residual signal, under certain circumstances the sum and difference operations essentially compensate the prior downmix operation (except for a possibly different gain factor) such that the overall system can actually perform L/R encoding of the overall stereo signal or for a frequency band thereof. E.g. such circumstances occur when the L and R channels of the stereo signal are independent and have the same level as will be explained in detail later on.
Preferably, the adaption of the encoding scheme is time and frequency dependent. Thus, preferably some frequency bands of the stereo signal are encoded by a L/R encoding scheme, whereas other frequency bands of the stereo signal are encoded by a PS coding scheme with residual.
It should be noted that in case the encoding is based on the downmix signal and based on the residual signal as discussed above, the actual signal which is input to the core encoder may be formed by two serial operations on the downmix signal and residual signal which are inverse (except for a possibly different gain factor). E.g. a downmix signal and a residual signal are fed to an M/S to L/R transform stage and then the output of the transform stage is fed to a L/R to M/S transform stage. The resulting signal (which is then used for encoding) corresponds to the downmix signal and the residual signal (expect for a possibly different gain factor).
The following embodiment makes use of this idea. According to an embodiment of the encoder system, the encoder system comprises a downmix stage and a parameter determining stage as discussed above. Moreover, the encoder system comprises a transform stage (e.g. as part of the encoding means discussed above). The transform stage generates a pseudo L/R stereo signal by performing a transform of the downmix signal and the residual signal. The transform stage preferably performs a sum and difference transform, where the downmix signal and the residual signals are summed to generate one channel of the pseudo stereo signal (possibly, the sum is also multiplied by a factor) and subtracted from each other to generate the other channel of the pseudo stereo signal (possibly, the difference is also multiplied by a factor). Preferably, a first channel (e.g. the pseudo left channel) of the pseudo stereo signal is proportional to the sum of the downmix and residual signals, where a second channel (e.g. the pseudo right channel) is proportional to the difference of the downmix and residual signals. Thus, the downmix signal DMX and residual signal RES from the PS encoder may be converted into a pseudo stereo signal Lp, Rp according to the following equations:
Lp=g(DMX+RES)
Rp=g(DMX−RES)
In the above equations the gain normalization factor g has e.g. a value of g=√{square root over (½)}.
The pseudo stereo signal is preferably processed by a perceptual stereo encoder (e.g. as part of the encoding means). For encoding, L/R stereo encoding or M/S stereo encoding is selectable. The adaptive L/R or M/S perceptual stereo encoder may be an AAC based encoder. Preferably, the selection between L/R stereo encoding and M/S stereo encoding is frequency-variant; thus, the selection may vary for different frequency bands as discussed above. Also, the selection between L/R encoding and M/S encoding is preferably time-variant. The decision between L/R encoding and M/S encoding is preferably made by the perceptual stereo encoder.
Such perceptual encoder having the option for M/S encoding can internally compute (pseudo) M and S signals (in the time domain or in selected frequency bands) based on the pseudo stereo L/R signal. Such pseudo M and S signals correspond to the downmix and residual signals (except for a possibly different gain factor). Hence, if the perceptual stereo encoder selects M/S encoding, it actually encodes the downmix and residual signals (which correspond to the pseudo M and S signals) as it would be done in a system using standard PS coding with residual.
Moreover, under special circumstances the transform stage essentially compensates the prior downmix operation (except for a possibly different gain factor) such that the overall encoder system can actually perform L/R encoding of the overall stereo signal or for a frequency band thereof (if L/R encoding is selected in the perceptual encoder). This is e.g. the case when the L and R channels of the stereo signal are independent and have the same level as will be explained in detail later on. Thus, for a given frequency band the pseudo stereo signal essentially corresponds or is proportional to the stereo signal, if—for the frequency band—the left and right channels of the stereo signal are essentially independent and have essentially the same level.
Thus, the encoder system actually allows to switch between L/R stereo coding and PS coding with residual, in order to be able to adapt to the properties of the given stereo input signal. Preferably, the adaption of the encoding scheme is time and frequency dependent. Thus, preferably some frequency bands of the stereo signal are encoded by a L/R encoding scheme, whereas other frequency bands of the stereo signal are encoded by a PS coding scheme with residual. It should be noted that M/S coding is basically a special case of PS coding with residual (since the L/R to M/S transform is a special case of the PS downmix operation) and thus the encoder system may also perform overall M/S coding.
Said embodiment having the transform stage downstream of the PS encoder and upstream of the L/R or M/S perceptual stereo encoder has the advantage that a conventional PS encoder and a conventional perceptual encoder can be used. Nevertheless, the PS encoder or the perceptual encoder may be adapted due to the special use here.
The new concept improves the performance of stereo coding by enabling an efficient combination of PS coding and joint stereo coding.
According to an alternative embodiment, the encoding means as discussed above comprise a transform stage for performing a sum and difference transform based on the downmix signal and the residual signal for one or more frequency bands (e.g. for the whole used frequency range or only for one frequency range). The transform may be performed in a frequency domain or in a time domain. The transform stage generates a pseudo left/right stereo signal for the one or more frequency bands. One channel of the pseudo stereo signal corresponds to the sum and the other channel corresponds to the difference.
Thus, in case encoding is based on the sum and difference signals the output of the transform stage may be used for encoding, whereas in case encoding is based on the downmix signal and the residual signal the signals upstream of the encoding stage may be used for encoding. Thus, this embodiment does not use two serial sum and difference transforms on the downmix signal and residual signal, resulting in the downmix signal and residual signal (except for a possibly different gain factor).
When selecting encoding based on the downmix signal and residual signal, parametric stereo encoding of the stereo signal is selected. When selecting encoding based on the sum and difference (i.e. encoding based on the pseudo stereo signal) L/R encoding of the stereo signal is selected.
The transform stage may be a L/R to M/S transform stage as part of a perceptual encoder with adaptive selection between L/R and M/S stereo encoding (possibly the gain factor is different in comparison to a conventional L/R to M/S transform stage). It should be noted that the decision between L/R and M/S stereo encoding should be inverted. Thus, encoding based on the downmix signal and residual signal is selected (i.e. the encoded signal did not pass the transform stage) when the decision means decide M/S perceptual decoding, and encoding based on the pseudo stereo signal as generated by the transform stage is selected (i.e. the encoded signal passed the transform stage) when the decision means decide L/R perceptual decoding.
The encoder system according to any of the embodiments discussed above may comprise an additional SBR (spectral band replication) encoder. SBR is a form of HFR (High Frequency Reconstruction). An SBR encoder determines side information for the reconstruction of the higher frequency range of the audio signal in the decoder. Only the lower frequency range is encoded by the perceptual encoder, thereby reducing the bitrate. Preferably, the SBR encoder is connected upstream of the PS encoder. Thus, the SBR encoder may be in the stereo domain and generates SBR parameters for a stereo signal. This will be discussed in detail in connection with the drawings.
Preferably, the PS encoder (i.e. the downmix stage and the parameter determining stage) operates in an oversampled frequency domain (also the PS decoder as discussed below preferably operates in an oversampled frequency domain). For time-to-frequency transform e.g. a complex valued hybrid filter bank having a QMF (quadrature mirror filter) and a Nyquist filter may be used upstream of the PS encoder as described in MPEG Surround standard (see document ISO/IEC 23003-1). This allows for time and frequency adaptive signal processing without audible aliasing artifacts. The adaptive L/R or M/S encoding, on the other hand, is preferably carried out in the critically sampled MDCT domain (e.g. as described in AAC) in order to ensure an efficient quantized signal representation.
The conversion between downmix and residual signals and the pseudo L/R stereo signal may be carried out in the time domain since the PS encoder and the perceptual stereo encoder are typically connected in the time domain anyway. Thus, the transform stage for generating the pseudo L/R signal may operate in the time domain.
In other embodiments as discussed in connection with the drawings, the transform stage operates in an oversampled frequency domain or in a critically sampled MDCT domain.
A second aspect of the application relates to a decoder system for decoding a bitstream signal as generated by the encoder system discussed above.
According to an embodiment of the decoder system, the decoder system comprises perceptual decoding means for decoding based on the bitstream signal. The decoding means are configured to generate by decoding an (internal) first signal and an (internal) second signal and to output a downmix signal and a residual signal. The downmix signal and the residual signal is selectively
As discussed above in connection with the encoder system, also here the selection may be frequency-variant or frequency-invariant.
Moreover, the system comprises an upmix stage for generating the stereo signal based on the downmix signal and the residual signal, with the upmix operation of the upmix stage being dependent on the one or more parametric stereo parameters.
Analogously to the encoder system, the decoder system allows to actually switch between L/R decoding and PS decoding with residual, preferably in a time and frequency variant manner.
According to another embodiment, the decoder system comprises a perceptual stereo decoder (e.g. as part of the decoding means) for decoding the bitstream signal, with the decoder generating a pseudo stereo signal. The perceptual decoder may be an AAC based decoder. For the perceptual stereo decoder, L/R perceptual decoding or M/S perceptual decoding is selectable in a frequency-variant or frequency-invariant manner (the actual selection is preferably controlled by the decision in the encoder which is conveyed as side-information in the bitstream). The decoder selects the decoding scheme based on the encoding scheme used for encoding. The used encoding scheme may be indicated to the decoder by information contained in the received bitstream.
Moreover, a transform stage is provided for generating a downmix signal and a residual signal by performing a transform of the pseudo stereo signal. In other words: The pseudo stereo signal as obtained from the perceptual decoder is converted back to the downmix and residual signals. Such transform is a sum and difference transform: The resulting downmix signal is proportional to the sum of a left channel and a right channel of the pseudo stereo signal. The resulting residual signal is proportional to the difference of the left channel and the right channel of the pseudo stereo signal. Thus, quasi an L/R to M/S transform was carried out. The pseudo stereo signal with the two channels LP, RP may be converted to the downmix and residual signals according to the following equations:
In the above equations the gain normalization factor g may have e.g. a value of g=√{square root over (½)}. The residual signal RES used in the decoder may cover the whole used audio frequency range or only a part of the used audio frequency range.
The downmix and residual signals are then processed by an upmix stage of a PS decoder to obtain the final stereo output signal. The upmixing of the downmix and residual signals to the stereo signal is dependent on the received PS parameters. According to an alternative embodiment, the perceptual decoding means may comprise a sum and difference transform stage for performing a transform based on the first signal and the second signal for one or more frequency bands (e.g. for the whole used frequency range). Thus, the transform stage generates the downmix signal and the residual signal for the case that the downmix signal and the residual signal are based on the sum of the first signal and of the second signal and based on the difference of the first signal and of the second signal. The transform stage may operate in the time domain or in a frequency domain.
As similarly discussed in connection with the encoder system, the transform stage may be a M/S to L/R transform stage as part of a perceptual decoder with adaptive selection between L/R and M/S stereo decoding (possibly the gain factor is different in comparison to a conventional M/S to L/R transform stage). It should be noted that the selection between L/R and M/S stereo decoding should be inverted.
The decoder system according to any of the preceding embodiments may comprise an additional SBR decoder for decoding the side information from the SBR encoder and generating a high frequency component of the audio signal. Preferably, the SBR decoder is located downstream of the PS decoder. This will be discussed in detail in connection with drawings.
Preferably, the upmix stage operates in an oversampled frequency domain, e.g. a hybrid filter bank as discussed above may be used upstream of the PS decoder.
The L/R to M/S transform may be carried out in the time domain since the perceptual decoder and the PS decoder (including the upmix stage) are typically connected in the time domain.
In other embodiments as discussed in connection with the drawings, the L/R to M/S transform is carried out in an oversampled frequency domain (e.g., QMF), or in a critically sampled frequency domain (e.g., MDCT).
A third aspect of the application relates to a method for encoding a stereo signal to a bitstream signal. The method operates analogously to the encoder system discussed above. Thus, the above remarks related to the encoder system are basically also applicable to encoding method.
A fourth aspect of the invention relates to a method for decoding a bitstream signal including PS parameters to generate a stereo signal. The method operates in the same way as the decoder system discussed above. Thus, the above remarks related to the decoder system are basically also applicable to decoding method.
The invention is explained below by way of illustrative examples with reference to the accompanying drawings, wherein
Typically, the matrix H−1 is frequency-variant and time-variant, i.e. the elements of the matrix H−1 vary over frequency and vary from time slot to time slot. The matrix H−1 may be updated every frame (e.g. every 21 or 42 ms) and may have a frequency resolution of a plurality of bands, e.g. 28, 20, or 10 bands (named “parameter bands”) on a perceptually oriented (Bark-like) frequency scale.
The elements of the matrix H−1 depend on the time- and frequency-variant PS parameters IID (inter-channel intensity difference; also called CLD—channel level difference) and ICC (inter-channel cross-correlation). For determining PS parameters 5, e.g. IID and ICC, the PS encoder 1 comprises a parameter determining stage. An example for computing the matrix elements of the inverse matrix H is given by the following and described in the MPEG Surround specification document ISO/IEC 23003-1, subclause 6.5.3.2 which is hereby incorporated by reference:
where
and where
and where ρ=ICC.
Moreover, the encoder system comprises a transform stage 2 that converts the downmix signal DMX and residual signal RES from the PS encoder 1 into a pseudo stereo signal Lp, Rp, e.g. according to the following equations:
Lp=g(DMX+RES)
Rp=g(DMX−RES)
In the above equations the gain normalization factor g has e.g. a value of g=√{square root over (½)}. For g=√{square root over (½)}, the two equations for pseudo stereo signal Lp, Rp can be rewritten as:
The pseudo stereo signal Lp, Rp is then fed to a perceptual stereo encoder 3, which adaptively selects either L/R or M/S stereo encoding. M/S encoding is a form of joint stereo coding. L/R encoding may be also based on joint encoding aspects, e.g. bits may be allocated jointly for the L and R channels from a common bit reservoir.
The selection between L/R or M/S stereo encoding is preferably frequency-variant, i.e. some frequency bands may be L/R encoded, whereas other frequency bands may be M/S encoded. An embodiment for implementing the selection between L/R or M/S stereo encoding is described in the document “Sum-Difference Stereo Transform Coding”, J. D. Johnston et al., IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP) 1992, pages 569-572. The discussion of the selection between L/R or M/S stereo encoding therein, in particular sections 5.1 and 5.2, is hereby incorporated by reference.
Based on the pseudo stereo signal Lp, Rp, the perceptual encoder 3 can internally compute (pseudo) mid/side signals Mp, Sp. Such signals basically correspond to the downmix signal DMX and residual signal RES (except for a possibly different gain factor). Hence, if the perceptual encoder 3 selects M/S encoding for a frequency band, the perceptual encoder 3 basically encodes the downmix signal DMX and residual signal RES for that frequency band (except for a possibly different gain factor) as it also would be done in a conventional perceptual encoder system using conventional PS coding with residual. The PS parameters 5 and the output bitstream 4 of the perceptual encoder 3 are multiplexed into a single bitstream 6 by a multiplexer 7.
In addition to PS encoding of the stereo signal, the encoder system in
However, preferably, the right column of the 2.2 matrix H should instead be modified to
The left column is preferably computed as given in the MPEG Surround specification.
Modifying the right column of the upmix matrix H ensures that for IID=0 dB and ICC=0 (i.e. the case where for the respective band the stereo channels L and R are independent and have the same level) the following upmix matrix H is obtained for the band:
Please note that the upmix matrix H and also the downmix matrix H−1 are typically frequency-variant and time-variant. Thus, the values of the matrices are different for different time/frequency tiles (a tile corresponds to the intersection of a particular frequency band and a particular time period). In the above case the downmix matrix H−1 is identical to the upmix matrix H. Thus, for the band the pseudo stereo signal Lp, Rp can computed by the following equation:
Hence, in this case the PS encoding with residual using the downmix matrix H−1 followed by the generation of the pseudo L/R signal in the transform stage 2 corresponds to the unity matrix and does not change the stereo signal for the respective frequency band at all, i.e.
Lp=L
Rp=R
In other words: the transform stage 2 compensates the downmix matrix H−1 such that the pseudo stereo signal Lp, Rp corresponds to the input stereo signal L, R. This allows to encode the original input stereo signal L, R by the perceptual encoder 3 for the particular band. When L/R encoding is selected by the perceptual encoder 3 for encoding the particular band, the encoder system behaves like a L/R perceptual encoder for encoding the band of the stereo input signal L, R.
The encoder system in
In the above equations, the gain normalization factor g is identical to the gain normalization factor g at the encoder side and has e.g. a value of g=√{square root over (½)}.
The downmix signal DMX and residual signal RES are then processed by the PS decoder 13 to obtain the final L and R output signals. The upmix step in the decoding process for PS coding with a residual can be described by means of the 2.2 upmix matrix H that converts the downmix signal DMX and residual signal RES back to the L and R channels:
The computation of the elements of the upmix matrix H was already discussed above.
The PS encoding and PS decoding process in the PS encoder 1 and the PS decoder 13 is preferably carried out in an oversampled frequency domain. For time-to-frequency transform e.g. a complex valued hybrid filter bank having a QMF (quadrature mirror filter) and a Nyquist filter may be used upstream of the PS encoder, such as the filter bank described in MPEG Surround standard (see document ISO/IEC 23003-1). The complex QMF representation of the signal is oversampled with factor 2 since it is complex-valued and not real-valued. This allows for time and frequency adaptive signal processing without audible aliasing artifacts. Such hybrid filter bank typically provides high frequency resolution (narrow band) at low frequencies, while at high frequency, several QMF bands are grouped into a wider band. The paper “Low Complexity Parametric Stereo Coding in MPEG-4”, H. Purnhagen, Proc. of the 7th Int. Conference on Digital Audio Effects (DAFx '04), Naples, Italy, Oct. 5-8, 2004, pages 163-168 describes an embodiment of a hybrid filter bank (see section 3.2 and
Preferably, the adaptive L/R or M/S encoding, on the other hand, is carried out in the critically sampled MDCT domain (e.g. as described in AAC) in order to ensure an efficient quantized signal representation. The conversion of the downmix signal DMX and residual signal RES to the pseudo stereo signal Lp, Rp in the transform stage 2 may be carried out in the time domain since the PS encoder 1 and the perceptual encoder 3 may be connected in the time domain anyway. Also in the decoding system, the perceptual stereo decoder 11 and the PS decoder 13 are preferably connected in the time domain. Thus, the conversion of the pseudo stereo signal Lp, Rp to the downmix signal DMX and residual signal RES in the transform stage 12 may be also carried out in the time domain.
An adaptive L/R or M/S stereo coder such as shown as the encoder 3 in
The perceptual stereo encoder (such as the encoder 3 in
An alternative approach to optimize psycho-acoustic control is to augment the encoder system with a detector forming a deactivation stage that is able to effectively deactivate PS encoding when appropriate, preferably in a time- and frequency-variant manner Deactivating PS encoding is e.g. appropriate when L/R stereo coding is expected to be beneficial or when the psycho-acoustic control would have problems to encode the pseudo L/R signal efficiently. PS encoding may be effectively deactivated by setting the downmix matrix H−1 in such a way that the downmix matrix H−1 followed by the transform (see stage 2 in
Such detector controlling a PS parameter modification is shown in
In the following figures, the term QMF (quadrature mirror filter or filter bank) also includes a QMF subband filter bank in combination with a Nyquist filter bank, i.e. a hybrid filter bank structure. Furthermore, all values in the description below may be frequency dependent, e.g. different downmix and upmix matrices may be extracted for different frequency ranges. Furthermore, the residual coding may only cover part of the used audio frequency range (i.e. the residual signal is only coded for a part of the used audio frequency range). Aspects of downmix as will be outlined below may for some frequency ranges occur in the QMF domain (e.g. according to prior art), while for other frequency ranges only e.g. phase aspects will be dealt with in the complex QMF domain, whereas amplitude transformation is dealt with in the real-valued MDCT domain.
In
Due to the re-arrangement of the SBR encoder 42, the PS encoder 41 may be configured to operate not on the full bandwidth of the input signal but e.g. only on the frequency range below the SBR crossover frequency. In
This advantage of the embodiment in
In
In
The advantage becomes even more apparent when operating on intermediate bitrates where the residual signal bandwidth approaches or is equal to the core coder bandwidth. In this case, the time frequency representation of
In
In
The encoder 48 is a stereo AAC style MDCT based coder. When the mode decision 73 steers the input signal to use MDCT based coding, the mono input signal or the stereo input signals are coded by the AAC based MDCT coder 48. The MDCT coder 48 does an MDCT analysis of the one or two signals in MDCT stages 74. In case of a stereo signal, further, an M/S or L/R decision on a frequency band basis is performed in a stage 75 prior to quantization and coding. L/R stereo encoding or M/S stereo encoding is selectable in a frequency-variant manner. The stage 75 also performs a L/R to M/S transform. If M/S encoding is decided for a particular frequency band, the stage 75 outputs an M/S signal for this frequency band. Otherwise, the stage 75 outputs a L/R signal for this frequency band.
Hence, when the transform coding mode is used, the full efficiency of the stereo coding functionality of the underlying core coder can be used for stereo.
When the mode decision 73 steers the mono signal to the linear predictive domain coder 71, the mono signal is subsequently analyzed by means of linear predictive analysis in block 72. Subsequently, a decision is made on whether to code the LP residual by means of a time-domain ACELP style coder 76 or a TCX style coder 77 (Transform Coded eXcitation) operating in the MDCT domain. The linear predictive domain coder 71 does not have any inherent stereo coding capability. Hence, to allow coding of stereo signal with the linear predictive domain coder 71, an encoder configuration similar to that shown in
While the mode decision 73 in
When the mode decision 73′ steers the downmix signal DMX to the linear predictive domain coder 71, the downmix signal DMX is subsequently analyzed by means of linear predictive analysis in block 72. Subsequently, a decision is made on whether to code the LP residual by means of a time-domain ACELP style coder 76 or a TCX style coder 77 (Transform Coded eXcitation) operating in the MDCT domain. The linear predictive domain coder 71 does not have any inherent stereo coding capability that can be used for coding the residual signal in addition to the downmix signal DMX. Hence, a dedicated residual coder 78 is employed for encoding the residual signal RES when the downmix signal DMX is encoded by the predictive domain coder 71. E.g. such coder 78 may be a mono AAC coder.
It should be noted that the coder 71 and 78 in
The stage 75 decides between L/R or M/S encoding. Based on the decision, either the pseudo stereo signal Lp, Rp or the pseudo mid/side signal Mp, Sp are selected (see selection switch) and encoded in AAC block 97. It should be noted that also two AAC blocks 97 may be used (not shown in
It should be noted that the switch in
Moreover, it should be noted that all blocks 2, 98 and 99 can be called “sum and difference transform blocks” since all blocks implement a transform matrix in the form of
Merely, the gain factor c may be different in the blocks 2, 98, 99.
In
HCOMPLEX=Hϕ·H,
where
and where
φ1=φopd
φ2=φopd−φipd.
The stage 80 of the PS encoder which operates in the complex QMF domain only takes care of phase dependencies between the channels L, R. The downmix rotation (i.e. the transformation from the L/R domain to the DMX/RES domain which was described by the matrix H−1 above) is taken care of in the MDCT domain as part of the stereo core coder 81. Hence, the phase dependencies between the two channels are extracted in the complex QMF domain, while other, real-valued, waveform dependencies are extracted in the real-valued critically sampled MDCT domain as part of the stereo coding mechanism of the core coder used. This has the advantage that the extraction of linear dependencies between the channels can be tightly integrated in the stereo coding of the core coder (though, to prevent aliasing in the critical sampled MDCT domain, only for the frequency range that is covered by residual coding, possibly minus a “guard band” on the frequency axis).
The phase adjustment stage 80 of the PS encoder in
As discussed before, the downmix rotation part of the PS module is dealt with in the stereo coding module 81 of the core coder in
In
In
Based on the encoding information given in the bitstream, the stage 101 selects either L/R or M/S decoding. When L/R decoding is selected, the output signal of the decoding block 100 is fed to the transform stage 12.
It should be noted that the switch in
Preferably, the selection is time- and frequency-variant.
The encoding means 110 comprises a sum and difference transform stage 111 which generates the sum and difference signals. Further, the encoding means 110 comprise a selection block 112 for selecting encoding based on the sum and difference signals or based on the downmix signal DMX and the residual signal RES. Furthermore, an encoding block 113 is provided. Alternatively, two encoding blocks 113 may be used, with the first encoding block 113 encoding the DMX and RES signals and the second encoding block 113 encoding the sum and difference signals. In this case the selection 112 is downstream of the two encoding blocks 113.
The sum and difference transform in block 111 is of the form
The transform block 111 may correspond to transform block 99 in
The output of the perceptual encoder 110 is combined with the parametric stereo parameters 5 in the multiplexer 7 to form the resulting bitstream 6.
In contrast to the structure in
The downmix signal DMX and the residual signal RES are selectively
Preferably, the selection is time- and frequency-variant. The selection is performed in the selection stage 125.
The decoding means 120 comprise a sum and difference transform stage 124 which generates sum and difference signals.
The sum and difference transform in block 124 is of the form
The transform block 124 may correspond to transform block 105′ in
After selection, the DMX and RES signals are fed to an upmix stage 126 for generating the stereo signal L, R based on the downmix signal DMX and the residual signal RES. The upmix operation is dependent on the PS parameters 5.
Preferably, in
It should be noted that in the above-described embodiments, the signals, parameters and matrices may be frequency-variant or frequency-invariant and/or time-variant or time-invariant. The described computing steps may be carried out frequency-wise or for the complete audio band.
Moreover, it should be noted that the various sum and difference transforms, i.e. the DMX/RES to pseudo L/R transform, the pseudo L/R to DMX/RES transform, the L/R to M/S transform and the M/S to L/R transform, are all of the form
Merely, the gain factor c may be different. Therefore, in principle, each of these transforms may be exchanged by a different transform of these transforms. If the gain is not correct during the encoding processing, this may be compensated in the decoding process. Moreover, when placing two same or two different of the sum and difference transforms is series, the resulting transform corresponds to the identity matrix (possibly, multiplied by a gain factor).
In an encoder system comprising both a PS encoder and a SBR encoder, different PS/SBR configurations are possible. In a first configuration, shown in
Also in a decoder system comprising both a PS decoder and a SBR decoder, different PS/SBR configurations are possible. In a first configuration, shown in
As discussed above, in order to ensure correct decoder operation, there is preferably a mechanism to signal from the encoder to the decoder which configuration is to be used in the decoder. This can be done explicitly (e.g. by means of an dedicated bit or field in the configuration header of the bitstream as discussed below) or implicitly (e.g. by checking whether the SBR data is mono or stereo in case of PS data being present).
As discussed above, to signal the chosen PS/SBR configuration, a dedicated element in the bitstream header of the bitstream conveyed from the encoder to the decoder may be used. Such a bitstream header carries necessary configuration information that is needed to enable the decoder to correctly decode the data in the bitstream. The dedicated element in the bitstream header may be e.g. a one bit flag, a field, or it may be an index pointing to a specific entry in a table that specifies different decoder configurations.
Instead of including in the bitstream header an additional dedicated element for signaling the PS/SBR configuration, information already present in the bitstream may be evaluated at the decoding system for selecting the correct PS/SBR configuration. E.g. the chosen PS/SBR configuration may be derived from bitstream header configuration information for the PS decoder and SBR decoder. This configuration information typically indicates whether the SBR decoder is to be configured for mono operation or stereo operation. If, for example, a PS decoder is enabled and the SBR decoder is configured for mono operation (as indicated in the configuration information), the PS/SBR configuration according to
The above-described embodiments are merely illustrative for the principles of the present application. It is understood that modifications and variations of the arrangements and the details described herein will be apparent to others skilled in the art. It is the intent, therefore, that the scope of the application is not limited by the specific details presented by way of description and explanation of the embodiments herein.
The systems and methods disclosed in the application may be implemented as software, firmware, hardware or a combination thereof. Certain components or all components may be implemented as software running on a digital signal processor or microprocessor, or implemented as hardware and or as application specific integrated circuits.
Typical devices making use of the disclosed systems and methods are portable audioplayers, mobile communication devices, set-top-boxes, TV-sets, AVRs (audio-video receiver), personal computers etc.
This application is a continuation of U.S. patent application Ser. No. 16/369,728 filed Mar. 29, 2019, which is a continuation of U.S. patent application Ser. No. 15/873,083 filed Jan. 17, 2018, which is a continuation of U.S. patent application Ser. No. 14/734,088 filed on Jun. 9, 2015, now issued U.S. Pat. No. 9,905,230 issued on Feb. 27, 2018, which is continuation of U.S. patent application Ser. No. 13/255,143 filed on Sep. 12, 2011, now issued U.S. Pat. No. 9,082,395 issued on Jul. 14, 2015, which is a national application of PCT application PCT/EP2010/052866 filed Mar. 5, 2010, which claims the benefit of the filing date of U.S. Provisional Patent Application Ser. No. 61/219,484 filed on Jun. 23, 2009 and U.S. Provisional Patent Application Ser. No. 61/160,707 filed on Mar. 17, 2009, all of which are hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
4790016 | Mazor | Dec 1988 | A |
4914701 | Zibman | Apr 1990 | A |
5222189 | Fielder | Jun 1993 | A |
5274740 | Davis | Dec 1993 | A |
5357594 | Fielder | Oct 1994 | A |
5394473 | Davidson | Feb 1995 | A |
5583962 | Davis | Dec 1996 | A |
5890125 | Davis | Mar 1999 | A |
5909664 | Davis | Jun 1999 | A |
6021386 | Davis | Feb 2000 | A |
6240388 | Fukuchi | May 2001 | B1 |
6708145 | Liljeryd | Mar 2004 | B1 |
6978236 | Liljeryd | Dec 2005 | B1 |
7003451 | Kjorling | Feb 2006 | B2 |
7050972 | Henn | May 2006 | B2 |
7143030 | Chen | Nov 2006 | B2 |
7181389 | Liljeryd | Feb 2007 | B2 |
7191121 | Liljeryd | Mar 2007 | B2 |
7191136 | Sinha | Mar 2007 | B2 |
7283955 | Liljeryd | Oct 2007 | B2 |
7382886 | Henn | Jun 2008 | B2 |
7433817 | Kjorling | Oct 2008 | B2 |
7469206 | Kjorling | Dec 2008 | B2 |
7483758 | Liljeryd | Jan 2009 | B2 |
7487097 | Engdegard | Feb 2009 | B2 |
7548864 | Kjorling | Jun 2009 | B2 |
7577570 | Kjorling | Aug 2009 | B2 |
7590543 | Kjorling | Sep 2009 | B2 |
7680552 | Liljeryd | Mar 2010 | B2 |
7751572 | Villemoes | Jul 2010 | B2 |
7835918 | Myburg | Nov 2010 | B2 |
20030236583 | Baumgarte | Dec 2003 | A1 |
20040186735 | Ferris | Sep 2004 | A1 |
20050078832 | Van De Par | Apr 2005 | A1 |
20050149322 | Bruhn | Jul 2005 | A1 |
20060190247 | Lindblom | Aug 2006 | A1 |
20060233379 | Villemoes | Oct 2006 | A1 |
20070291951 | Faller | Dec 2007 | A1 |
20080004883 | Vilermo | Jan 2008 | A1 |
20080097763 | Van De Par | Apr 2008 | A1 |
20090326931 | Ragot | Dec 2009 | A1 |
20100153119 | Lee | Jun 2010 | A1 |
Number | Date | Country |
---|---|---|
1209718 | Mar 1999 | CN |
1510662 | Jul 2004 | CN |
1677491 | Oct 2005 | CN |
1276407 | Sep 2006 | CN |
101010985 | Aug 2007 | CN |
101366321 | Feb 2009 | CN |
102388417 | Mar 2012 | CN |
1107232 | Jun 2001 | EP |
0797324 | Nov 2004 | EP |
1906705 | Apr 2008 | EP |
2235719 | Oct 2010 | EP |
2006139082 | May 2008 | RU |
2006048226 | May 2006 | WO |
2006091139 | Aug 2006 | WO |
2006091150 | Aug 2006 | WO |
2006108462 | Oct 2006 | WO |
2008046530 | Apr 2008 | WO |
2008046531 | Apr 2008 | WO |
2008131903 | Nov 2008 | WO |
Entry |
---|
Herre et al., “Combined Stereo Coding”, Oct. 4, 1992, Audio Engineering Society, 93rd Convention of the AES, pp. 1-19. (Year: 1992). |
Herre, J., et al. “The reference model architecture for MPEG spatial audio coding”, May 31, 2005, Audio Engineering Society, 118th Convention, pp. 1-13. (Year: 2005). |
Derrien, et al., “A New Model-Based Algorithm for Optimizing the MPEG-AAC in MS-Stereo” IEEE Transactions on Audio, Speech, and Language Processing, vol. 16, No. 8, Nov. 2008, pp. 1373-1382. |
Heinen, et al., “Transactions Papers, Source-Optimized Channel Coding for Digital Transmission Channels” IEEE Transactions on Communications, vol. 53, No. 4, Apr. 2005, pp. 592-600. |
Herre, et al., “MPEG Surround—The ISO/MPEG Standard for Efficient and Compatible Multi-Channel Audio Coding” AES presentd at the 122nd Convention, May 5-8, 2007, Vienna, Austria, pp. 1-23. |
Herre, et al., “MPEG-4 High-Efficiency AAC Coding” IEEE Signal Processing Magazine, IEEE Service Center, Piscataway, NJ, vol. 25, No. 3, May 1, 2008, pp. 137-142. |
Meltzer, et al., “MPEG-4, HE-AAC V2-Audio Coding for Today's Digital Media World” EBU Technical Review, Jan. 31, 2006, pp. 1-12. |
MPEG Surround Standard, ISO/IEC 23003-1. |
MPEG-2 Advanced Audio Coding (AAC) standard, ISO/IEC 13818-7. |
Neuendorf, et al., “Unified Speech and Audio Coding Scheme for High Quality at Low Bitrates” Acoustics, Speech and Signal Processing, 2009. ICASSP 2009, Apr. 19, 2009, pp. 1-4. |
Purnhagen, Heiko, “Low Complexity Parametric Stereo Coding in MPEG-4” Proceedings of the 7th Int. Conference on Digital Audio Effects Naples, Italy, Oct. 5-8, 2004, pp. 163-168. |
Schuijers, et al., “Low Complexity Parametric Stereo Coding” AES Convention Paper 6073, presented at the 116th Convention, May 8-11, 2004, Berlin, Germany. |
Shin, et al., “Designing a Unified Speech/Audio Codec by Adopting a Single Channel Harmonic Source Separation Module” Acoustics, Speech and Signal Processing, 2008. IEEE International Conference on IEEE, Piscataway, NJ, USA, Mar. 31, 2008, pp. 185-188. |
Number | Date | Country | |
---|---|---|---|
20190318748 A1 | Oct 2019 | US |
Number | Date | Country | |
---|---|---|---|
61219484 | Jun 2009 | US | |
61160707 | Mar 2009 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16369728 | Mar 2019 | US |
Child | 16456476 | US | |
Parent | 15873083 | Jan 2018 | US |
Child | 16369728 | US | |
Parent | 14734088 | Jun 2015 | US |
Child | 15873083 | US | |
Parent | 13255143 | US | |
Child | 14734088 | US |