The invention disclosed herein generally relates to audio encoding and decoding. In particular, it relates to an audio encoder and an audio decoder adapted to perform high frequency reconstruction of audio signals.
Audio coding systems use different methodologies for coding of audio, such as pure waveform coding, parametric spatial coding, and high frequency reconstruction algorithms including the Spectral Band Replication (SBR) algorithm. The MPEG-4 standard combines waveform coding and SBR of audio signals. More precisely, an encoder may waveform code an audio signal for spectral bands up to a cross-over frequency and encode the spectral bands above the cross-over frequency using SBR encoding. The waveform-coded part of the audio signal is then transmitted to a decoder together with SBR parameters determined during the SBR encoding. Based on the waveform-coded part of the audio signal and the SBR parameters, the decoder then reconstructs the audio signal in the spectral bands above the cross-over frequency as discussed in the review paper Brinker et al., An overview of the Coding Standard MPEG-4 Audio Amendments 1 and 2: HE-AAC, SSC, and HE-AAC v2, EURASIP Journal on Audio, Speech, and Music Processing, Volume 2009, Article ID 468971.
One problem with this approach is that strong tonal components, i.e. strong harmonic components, or any component in the high spectral bands that is not nicely reconstructed by the SBR algorithm will be missing in the output.
To this end, the SBR algorithm implements a missing harmonics detection procedure. Tonal components that will not be properly regenerated by the SBR high frequency reconstruction are identified at the encoder side. Information of the frequency location of these strong tonal components is transmitted to the decoder where the spectral contents in the spectral bands where the missing tonal components are located are replaced by sinusoids generated in the decoder.
An advantage of the missing harmonics detection provided for in the SBR algorithm is that it is a very low bitrate solution since, somewhat simplified, only the frequency location of the tonal component and its amplitude level needs to be transmitted to the decoder.
A drawback of the missing harmonics detection of the SBR algorithm is that it is a very rough model. Another drawback is that when the transmission rate is low, i.e. when the number of bits that may be transmitted per second is low, and as a consequence thereof the spectral bands are wide, a large frequency range will be replaced by a sinusoid.
Another drawback of the SBR algorithm is that it has a tendency to smear out transients occurring in the audio signal. Typically, there will be a pre-echo and a post-echo of the transient in the SBR reconstructed audio signal. There is thus room for improvements.
In what follows, example embodiments will be described in greater detail and with reference to the accompanying drawings, on which:
All the figures are schematic and generally only show parts which are necessary in order to elucidate the invention, whereas other parts may be omitted or merely suggested. Unless otherwise indicated, like reference numerals refer to like parts in different figures.
In view of the above it is an object to provide an encoder and a decoder and associated methods which provides an improved reconstruction of transients and tonal components in the high frequency bands.
As used herein, an audio signal may be a pure audio signal, an audio part of an audiovisual signal or multimedia signal or any of these in combination with metadata.
According to a first aspect, example embodiments propose decoding methods, decoding devices, and computer program products for decoding. The proposed methods, devices and computer program products may generally have the same features and advantages.
According to example embodiments there is provided a decoding method in an audio processing system comprising: receiving a first waveform-coded signal having a spectral content up to a first cross-over frequency; receiving a second waveform-coded signal having a spectral content corresponding to a subset of the frequency range above the first cross-over frequency; receiving high frequency reconstruction parameters; performing high frequency reconstruction using the first waveform-coded signal and the high frequency reconstruction parameters so as to generate a frequency extended signal having a spectral content above the first cross-over frequency; and interleaving the frequency extended signal with the second waveform-coded signal.
As used herein, a waveform-coded signal is to be interpreted as a signal that has been coded by direct quantization of a representation of the waveform; most preferred a quantization of the lines of a frequency transform of the input waveform signal. This is opposed to a parametric coding, where the signal is represented by variations of a generic model of a signal attribute.
The decoding method thus suggests to use waveform-coded data in a subset of the of the frequency range above the first cross-over frequency and to interleave that with a high frequency reconstructed signal. In this way, important parts of a signal in the frequency band above the first cross-over frequency, such as tonal components or transients which are typically not well reconstructed by parametric high frequency reconstruction algorithms, may be waveform-coded. As a result, the reconstruction of these important parts of a signal in the frequency band above the first cross-over frequency is improved.
According to exemplary embodiments, the subset of the frequency range above the first cross-over frequency is a sparse subset. For example it may comprise a plurality of isolated frequency intervals. This is advantageous in that the number of bits to code the second waveform-coded signal is low. Still, by having a plurality of isolated frequency intervals tonal components, e.g. single harmonics, of the audio signal may be well captured by the second waveform-coded signal. As a result, an improvement of the reconstruction of tonal components for high frequency bands is achieved at a low bit cost.
As used herein, a missing harmonics or a single harmonics means any arbitrary strong tonal part of the spectrum. In particular, it is to be understood that a missing harmonics or a single harmonics is not limited to a harmonics of a harmonic series.
According to exemplary embodiments, the second waveform-coded signal may represent a transient in the audio signal to be reconstructed. A transient is typically limited to a short temporal range, such as approximately hundred temporal samples at a sampling rate of 48 kHz, e.g. a temporal range in the order of 5 to 10 milliseconds, but may have a wide frequency range. To capture the transient, the subset of the frequency range above the first cross-over frequency may therefore comprise a frequency interval extending between the first cross-over frequency and a second cross-over frequency. This is advantageous in that an improved reconstruction of transients may be achieved.
According to exemplary embodiments, the second cross-over frequency varies as a function of time. For example, the second cross-over frequency may vary within a time frame set by the audio processing system. In this way, the short temporal range of transients may be accounted for.
According to exemplary embodiments, the step of performing high frequency reconstruction comprises performing spectral band replication, SBR. High frequency reconstruction is typically performed in a frequency domain, such as a pseudo Quadrature Mirror Filters, QMF, domain of e.g. 64 sub-bands.
According to exemplary embodiments, the step of interleaving the frequency extended signal with the second waveform-coded signal is performed in a frequency domain, such as a QMF, domain. Typically, for ease of implementation and better control over the time- and frequency-characteristics of the two signals, the interleaving is performed in the same frequency domain as the high frequency reconstruction.
According to exemplary embodiments, the first and the second waveform-coded signal as received are coded using the same Modified Discrete Cosine Transform, MDCT.
According to exemplary embodiments, the decoding method may comprise adjusting the spectral content of the frequency extended signal in accordance with the high frequency reconstruction parameters so as to adjust the spectral envelope of the frequency extended signal.
According to exemplary embodiments, the interleaving may comprise adding the second waveform-coded signal to the frequency extended signal. This is the preferred option if the second waveform-coded signal represents tonal components, such as when the subset of the frequency range above the first cross-over frequency comprises a plurality of isolated frequency intervals. Adding the second waveform-coded signal to the frequency extended signal mimics the parametric addition of harmonics as known from SBR, and allows the SBR copy-up signal to be used to avoid large frequency ranges to be replaced by a single tonal component by mixing it in at a suitable level.
According to exemplary embodiments, the interleaving comprises replacing the spectral content of the frequency extended signal by the spectral content of the second waveform-coded signal in the subset of the frequency range above the first cross-over frequency which corresponds to the spectral content of the second waveform-coded signal. This is the preferred option when the second waveform-coded signal represents a transient, for example when the subset of the frequency range above the first cross-over frequency may therefore comprise a frequency interval extending between the first cross-over frequency and a second cross-over frequency. The replacement is typically only performed for a time range covered by the second waveform-coded signal. In this way, as little as possible may be replaced while still enough to replace a transient and potential time smear present in the frequency extended signal, and the interleaving is thus not limited to a time-segment specified by the SBR envelope time-grid.
According to exemplary embodiments, the first and the second waveform-coded signal may be separate signals, meaning that they have been coded separately. Alternatively, the first waveform-coded signal and the second waveform-coded signal form first and second signal portions of a common, jointly coded signal. The latter alternative is more attractive from an implementation point of view.
According to exemplary embodiments, the decoding method may comprise receiving a control signal comprising data relating to one or more time ranges and one or more frequency ranges above the first cross-over frequency for which the second waveform-coded signal is available, wherein the step of interleaving the frequency extended signal with the second waveform-coded signal is based on the control signal. This is advantageous in that it provides an efficient way of controlling the interleaving.
According to exemplary embodiments, the control signal comprises at least one of a second vector indicating the one or more frequency ranges above the first cross-over frequency for which the second waveform-coded signal is available for interleaving with the frequency extended signal, and a third vector indicating the one or more time ranges for which the second waveform-coded signal is available for interleaving with the frequency extended signal. This is a convenient way of implementing the control signal.
According to exemplary embodiments, the control signal comprises a first vector indicating one or more frequency ranges above the first cross-over frequency to be parametrically reconstructed based on the high frequency reconstruction parameters. In this way, the frequency extended signal may be given precedence over the second waveform-coded signal for certain frequency bands.
According to exemplary embodiments, there is also provided a computer program product comprising a computer-readable medium with instructions for performing any decoding method of the first aspect.
According to exemplary embodiments, there is also provided a decoder for an audio processing system, comprising: a receiving stage configured to receive a first waveform-coded signal having a spectral content up to a first cross-over frequency, a second waveform-coded signal having a spectral content corresponding to a subset of the frequency range above the first cross-over frequency, and high frequency reconstruction parameters; a high frequency reconstructing stage configured to receive the first waveform-decoded signal and the high frequency reconstruction parameters from the receiving stage and to perform high frequency reconstruction using the first waveform-coded signal and the high frequency reconstruction parameters so as to generate a frequency extended signal having a spectral content above the first cross-over frequency; and an interleaving stage configured to receive the frequency extended signal from the high frequency reconstruction stage and the second waveform-coded signal from the receiving stage, and to interleave the frequency extended signal with the second waveform-coded signal.
According to exemplary embodiments, the decoder may be configured to perform any decoding method disclosed herein.
According to a second aspect, example embodiments propose encoding methods, encoding devices, and computer program products for encoding. The proposed methods, devices and computer program products may generally have the same features and advantages.
Advantages regarding features and setups as presented in the overview of the decoder above may generally be valid for the corresponding features and setups for the encoder
According to example embodiments, there is provided an encoding method in an audio processing system, comprising the steps of: receiving an audio signal to be encoded; calculating, based on the received audio signal, high frequency reconstruction parameters enabling high frequency reconstruction of the received audio signal above the first cross-over frequency; identifying, based on the received audio signal, a subset of the frequency range above the first cross-over frequency for which the spectral content of the received audio signal is to be waveform-coded and subsequently, in a decoder, be interleaved with a high frequency reconstruction of the audio signal; generating a first waveform-coded signal by waveform-coding the received audio signal for spectral bands up to a first cross-over frequency; and a second waveform-coded signal by waveform-coding the received audio signal for spectral bands corresponding to the identified subset of the frequency range above the first cross-over frequency.
According to example embodiments, the subset of the frequency range above the first cross-over frequency may comprise a plurality of isolated frequency intervals.
According to example embodiments, the subset of the frequency range above the first cross-over frequency may comprise a frequency interval extending between the first cross-over frequency and a second cross-over frequency.
According to example embodiments, the second cross-over frequency may vary as a function of time.
According to example embodiments, the high frequency reconstruction parameters are calculated using spectral band replication, SBR, encoding.
According to example embodiments, the encoding method may further comprise adjusting spectral envelope levels comprised in the high frequency reconstruction parameters so as to compensate for addition of a high frequency reconstruction of the received audio signal with the second waveform-coded signal in a decoder. As the second waveform-coded signal is added to a high frequency reconstructed signal in the decoder, the spectral envelope levels of the combined signal is different from the spectral envelope levels of the high frequency reconstructed signal. This change in spectral envelope levels may be accounted for in the encoder, so that the combined signal in the decoder gets a target spectral envelope. By performing the adjustment on the encoder side, the intelligence needed on the decoder side may be reduced, or put differently; the need for defining specific rules in the decoder for how to handle the situation is removed by specific signaling from the encoder to the decoder. This allows for future optimizations of the system by future optimizations of the encoder without having to update potentially widely deployed decoders.
According to example embodiments, the step of adjusting the high frequency reconstruction parameters may comprise: measuring an energy of the second waveform-coded signal; and adjusting the spectral envelope levels, as intended to control the spectral envelope of the High Frequency Reconstructed signal, by subtracting the measured energy of the second waveform-coded signal from the spectral envelope levels for spectral bands corresponding to the spectral contents of the second waveform-coded signal.
According to exemplary embodiments, there is also provided a computer program product comprising a computer-readable medium with instructions for performing any encoding method of the second aspect.
According to example embodiments, there is provided and encoder for an audio processing system, comprising: a receiving stage configured to receive an audio signal to be encoded; a high frequency encoding stage configured to receive the audio signal from the receiving stage and to calculate, based on the received audio signal, high frequency reconstruction parameters enabling high frequency reconstruction of the received audio signal above the first cross-over frequency; an interleave coding detection stage configured to identify, based on the received audio signal, a subset of the frequency range above the first cross-over frequency for which the spectral content of the received audio signal is to be waveform-coded and subsequently, in a decoder, be interleaved with a high frequency reconstruction of the audio signal; and a waveform encoding stage configured to receive the audio signal from the receiving stage and to generate a first waveform-coded signal by waveform-coding the received audio signal for spectral bands up to a first cross-over frequency; and to receive the identified subset of the frequency range above the first cross-over frequency from the interleave coding detection stage and to generate a second waveform-coded signal by waveform-coding the received audio signal for spectral bands corresponding to the received identified subset of the frequency range.
According to example embodiments, the encoder may further comprise an envelope adjusting stage configured to receive the high frequency reconstruction parameters from the high frequency encoding stage and the identified subset of the frequency range above the first cross-over frequency from the interleave coding detection stage, and, based on the received data, to adjust the high frequency reconstruction parameters so as to compensate for the subsequent interleaving of a high frequency reconstruction of the received audio signal with the second waveform coded signal in the decoder.
According to example embodiments, the decoder may be configured to perform any decoding method disclosed herein.
The operation of the decoder 100 will now be explained in more detail with reference to the example embodiment of
The receiving stage 110 receives, in step D04, a second waveform-coded signal 202. The second waveform-coded signal 202 has a spectral content which corresponds to a subset of the frequency range above the first cross-over frequency fc. In the illustrated example of
The receiving stage 110 may receive the first and the second waveform-coded signal 201 and 202 as two separate signals. Alternatively, the first and the second waveform-coded signal 201 and 202 may form first and second signal portions of a common signal received by the receiving stage 110. In other words, the first and the second waveform-coded signals may be jointly coded, for example using the same MDCT transform.
Typically, the first waveform-coded signal 201 and the second waveform-coded signal 202 as received by the receiving stage 110 are coded using an overlapping windowed transform, such as a MDCT transform. The receiving stage may comprise a waveform decoding stage 240 configured to transform the first and the second waveform-coded signals 201 and 202 to the time domain. The waveform decoding stage 240 typically comprises a MDCT filter bank configured to perform inverse MDCT transform of the first and the second waveform-coded signal 201 and 202.
The receiving stage 110 further receives, in step D06, high frequency reconstruction parameters which are used by the high frequency reconstruction stage 120 as will be disclosed in the following.
The first waveform-coded signal 201 and the high frequency parameters received by the receiving stage 110 are then input to the high frequency reconstructing stage 120. The high frequency reconstruction stage 120 typically operates on signals in a frequency domain, preferably a QMF domain. Prior to being input to the high frequency reconstruction stage 120, the first waveform-coded signal 201 is therefore preferably transformed into the frequency domain, preferably the QMF domain, by a QMF analysis stage 250. The QMF analysis stage 250 typically comprises a QMF filter bank configured to perform a QMF transform of the first waveform-coded signal 201.
Based on the first waveform-coded signal 201 and the high frequency reconstructing parameters, the high frequency reconstruction stage 120, in step D08, extends the first waveform-coded signal 201 to frequencies above the first cross-over frequency fc. More specifically, the high frequency reconstructing stage 120 generates a frequency extended signal 203 which has a spectral content above the first cross-over frequency fc. The frequency extended signal 203 is thus a high-band signal.
The high frequency reconstructing stage 120 may operate according to any known algorithm for performing high frequency reconstruction. In particular, the high frequency reconstructing stage 120 may be configured to perform SBR as disclosed in the review paper Brinker et al., An overview of the Coding Standard MPEG-4 Audio Amendments 1 and 2: HE-AAC, SSC, and HE-AAC v2, EURASIP Journal on Audio, Speech, and Music Processing, Volume 2009, Article ID 468971. As such, the high frequency reconstructing stage may comprise a number of sub-stages configured to generate the frequency extended signal 203 in a number of steps. For example, the high frequency reconstructing stage 120 may comprise a high frequency generating stage 221, a parametric high frequency components adding stage 222, and an envelope adjusting stage 223.
In brief, the high frequency generating stage 221, in a first sub-step D08a, extends the first waveform-coded signal 201 to the frequency range above the cross-over frequency fc in order to generate the frequency extended signal 203. The generation is performed by selecting sub-band portions of the first waveform-coded signal 201 and according to specific rules, guided by the high frequency reconstruction parameters, mirror or copy the selected sub-band portions of the first waveform-coded signal 201 to selected sub-band portions of the frequency range above the first cross-over frequency fc.
The high frequency reconstruction parameters may further comprise missing harmonics parameters for adding missing harmonics to the frequency extended signal 203. As discussed above, a missing harmonics is to be interpreted as any arbitrary strong tonal part of the spectrum. For example, the missing harmonics parameters may comprise parameters relating to the frequency and amplitude of the missing harmonics. Based on the missing harmonics parameters, the parametric high frequency components adding stage 222 generates, in sub-step D08b, sinusoid components and adds the sinusoid components to the frequency extended signal 203.
The high frequency reconstruction parameters may further comprise spectral envelope parameters describing the target energy levels of the frequency extended signal 203. Based on the spectral envelope parameters, the envelope adjusting stage 223 may in sub-step D08c adjust the spectral content of the frequency extended signal 203, i.e. the spectral coefficients of the frequency extended signal 203, so that the energy levels of the frequency extended signal 203 corresponds to the target energy levels described by the spectral envelope parameters.
The frequency extended signal 203 from the high frequency reconstructing stage 120 and the second waveform-coded signal from the receiving stage 110 are then input to the interleaving stage 130. The interleaving stage 130 typically operates in the same frequency domain, preferably the QMF domain, as the high frequency reconstructing stage 120. Thus, the second waveform-coded signal 202 is typically input to the interleaving stage via the QMF analysis stage 250. Further, the second waveform-coded signal 202 is typically delayed, by a delay stage 260, to compensate for the time it takes for the high frequency reconstructing stage 120 to perform the high frequency reconstruction. In this way, the second wave-form coded signal 202 and the frequency extended signal 203 will be aligned such that the interleaving stage 130 operates on signals corresponding to the same time frame.
The interleaving stage 130, in step D10, then interleaves, i.e., combines the second waveform-coded signal 202 with the frequency extended signal 203 in order to generate an interleaved signal 204. Different approaches may be used to interleave the second waveform-coded signal 202 with the frequency extended signal 203.
According to one example embodiment, the interleaving stage 130 interleaves the frequency extended signal 203 with the second waveform-coded signal 202 by adding the frequency extended signal 203 and the second waveform-coded signal 202. The spectral contents of the second waveform-coded signal 202 overlaps the spectral contents of the frequency extended signal 203 in the subset of the frequency range corresponding to the spectral contents of the second waveform-coded signal 202. By adding the frequency extended signal 203 and the second waveform-coded signal 202 the interleaved signal 204 thus comprises the spectral contents of the frequency extended signal 203 as well as the spectral contents of the second waveform-coded signal 202 for the overlapping frequencies. As a result of the addition, the spectral envelope levels of the interleaved signal 204 increases for the overlapping frequencies. Preferably, and as will be disclosed later, the increase in spectral envelope levels due to the addition is accounted for on the encoder side when determining energy envelope levels comprised in the high frequency reconstruction parameters. For example, the spectral envelope levels for the overlapping frequencies may be decreased on the encoder side by an amount corresponding to the increase in spectral envelope levels due to interleaving on the decoder side.
Alternatively, the increase in spectral envelope levels due to addition may be accounted for on the decoder side. For example, there may be an energy measuring stage which measures the energy of the second waveform-coded signal 202, compares the measured energy to the target energy levels described by the spectral envelope parameters, and adjusts the extended frequency signal 203 such that the spectral envelope levels for the interleaved signal 204 equals the target energy levels.
According to another example embodiment, the interleaving stage 130 interleaves the frequency extended signal 203 with the second waveform-coded signal 202 by replacing the spectral contents of the frequency extended signal 203 by the spectral contents of the second waveform-coded signal 202 for those frequencies where the frequency extended signal 203 and the second waveform-coded signal 202 overlaps. In example embodiments where the frequency extended signal 203 is replaced by the second waveform-coded signal 202 it is not necessary to adjust the spectral envelope levels to compensate for the interleaving of the frequency extended signal 203 and the second waveform-coded signal 202.
The high frequency reconstruction stage 120 preferably operates with a sampling rate which equals the sampling rate of the underlying core encoder that was used to encode the first wave-form coded signal 201. In this way, the same overlapping windowed transform, such as the same MDCT, may be used to code the second waveform-coded signal 202 as was used to code the first waveform-coded signal 202.
The interleaving stage 130 may further be configured to receive the first waveform-coded signal 201 from the receiving stage, preferably via the waveform decoding stage 240, the QMF analysis stage 250, and the delay stage 260, and to combine the interleaved signal 204 with the first waveform-coded signal 201 in order to generate a combined signal 205 having a spectral content for frequencies below as well as above the first cross-over frequency.
The output signal from the interleaving stage 130, i.e. the interleaved signal 204 or the combined signal 205, may subsequently, by a QMF synthesis stage 270, be transformed back to the time domain.
Preferably, the QMF analysis stage 250 and the QMF synthesis stage 270 have the same number of sub-bands, meaning that the sampling rate of the signal being input to the QMF analysis stage 250 is equal to the sampling rate of the signal being output of the QMF synthesis stage 270. As a consequence, the waveform-coder (using MDCT) that was used to waveform-code the first and the second waveform-coded signals may operate on the same sampling rate as the output signal. Thus the first and the second waveform-coded signal can efficiently and structurally easily be coded by using the same MDCT transform. This is opposed to prior art where the sampling rate of the waveform coder typically was limited to half of that of the output signal, and the subsequent high frequency reconstruction module did an up-sampling as well as a high frequency reconstruction. This limits the ability to waveform code frequencies covering the entire output frequency range.
During the second time portion the receiving stage 110 receives a first waveform-coded signal 401b having a spectral content up to the first cross-over frequency fc1, and a second waveform-coded signal 402b having a spectral content which corresponds to a subset of the frequency range above the first cross-over frequency fc1. In the illustrated example of
During the third time portion the receiving stage 110 receives a first waveform-coded signal 401c having a spectral content up to the first cross-over frequency fc1. No second waveform-coded signal is received for the third time portion.
For the first and the third illustrated time portions there are no second waveform-coded signals. For these time portions the decoder will operate according to a conventional decoder configured to perform high frequency reconstruction, such as a conventional SBR decoder. The high frequency reconstruction stage 120 will generate frequency extended signals 403a and 403c based on the first waveform-coded signals 401a and 401c, respectively. However, since there are no second waveform-coded signals, no interleaving will be carried out by the interleaving stage 130.
For the second illustrated time portion there is a second waveform-coded signal 402b. For the second time portion the decoder 400 will operate in the same manner as described with respect to
In the example above, there is no second waveform-coded signal for the first and the third time portions. For these time portions the second cross-over frequency is equal to the first cross-over frequency, and no interleaving is performed. For the second time frame the second cross-over frequency is larger than the first cross-over frequency, and interleaving is performed. Generally, the second cross-over frequency may thus vary as a function of time. Particularly, the second cross-over frequency may vary within a time frame. Interleaving will be carried out when the second cross-over frequency is larger than the first cross-over frequency and smaller than a maximum frequency represented by the decoder. The case where the second cross-over frequency equals the maximum frequency corresponds to pure waveform coding and no high frequency reconstruction is needed.
It is to be noted that the embodiments described with respect to
In the present example, two strong tonal components in frequency bands 710 and 720 have been identified in the audio signal on the encoder side. The frequency bands 710 and 720 may be of the same bandwidth as e.g. SBR envelope bands, i.e. the same frequency resolution as is used for representing the spectral envelope. These tonal components in bands 710 and 720 have a time range corresponding to the full time frame, i.e. the time range of the tonal components includes the time ranges T1 to T3. On an encoder side, it has been decided to waveform-code the tonal components of 710 and 720 during the first time range T1, illustrated by the tonal component 710a and 720 being dashed during the first time range T1. Further it has been decided on an encoder side that during the second and third time ranges T2 and T3, the first tonal component 710 is to be parametrically reconstructed in the decoder by including a sinusoid as explained in connection to the parametric high frequency components stage 222 of
Additionally, a transient 730 has been identified in the audio signal on the encoder side. The transient 730 has a time duration corresponding to the second time range T2, and corresponds to a frequency interval between the first cross-over frequency fc1 and a second cross-over frequency fc2. On an encoder side it has been decided to waveform-code the time-frequency portion of the audio signal corresponding to the location of the transient. In this embodiment the interleaving of the waveform-coded transient is done by replacement.
A signalling scheme is set up to signal this information to the decoder. The signalling scheme comprises information relating to in which time ranges and/or in which frequency ranges above the first cross-over frequency fc1 a second waveform-coded signal are available. The signalling scheme may also be associated with rules relating to how the interleaving is to be performed, i.e. if the interleaving is by means of addition or replacement. The signalling scheme may also be associated with rules defining the order of priority of adding or replacing the different signals as will be explained below.
The signalling scheme includes a first vector 740, labelled “additional sinusoid”, indicating for each frequency sub-band if a sinusoid should be parametrically added or not. In
The signalling scheme further includes a second vector 750, labelled “waveform coding”. The second vector 750 indicates for each frequency sub-band if a waveform-coded signal is available for interleaving with a high frequency reconstruction of the audio signal. In
The signalling scheme further includes a third vector 760, labelled “waveform coding”. The third vector 760 indicates for each time slot if a waveform-coded signal is available for interleaving with a high frequency reconstruction of the audio signal. In
There are many alternatives for how to embody the first, the second and the third vector 740, 750, 760. In some embodiments, the vectors 740, 750, 760 are binary vectors which use a logic zero or a logic one to provide their indications. In other embodiments, the vectors 740, 750, 760 may take different forms. For example, a first value such as “0” in the vector may indicate that no waveform-coded data is available for the specific frequency band or time slot. A second value such as “1” in the vector may indicate that interleaving is to be performed by way of addition for the specific frequency band or time slot. A third value such as “2” in the vector may indicate that interleaving is to be performed by way of replacement for the specific frequency band or time slot.
The above exemplary signalling scheme may also be associated with an order of priority which may be applied in case of conflict. By way of example, the third vector 760, representing interleaving of a transient by way of replacement may take precedence over the first and second vectors 740 and 750. Further, the first vector 740 may take precedence over the second vector 750. It is understood that any order of priority between the vectors 740, 750, 760 may be defined.
The decision logic component 1302 is shown in more detail in
For some time/frequency tiles there will be a conflict between the information from the first vector 740, the second vector 750 and the third vector 760, meaning that more than one of the vectors 740-760 indicates a number different from zero, such as a “1”, for the same time/frequency tile of the time/frequency matrix 870. In such situation, the prioritizing component 13022 needs to make a decision on how to prioritize the information from the vectors in order to remove the conflicts in the time/frequency matrix 870. More precisely, the prioritizing component 13022 decides whether the output signal 804 is to be based on the frequency extended signal 803 (thereby giving priority to the first vector 740), by interleaving of the second wave-form coded signal 802 in a frequency direction (thereby giving priority to the second vector 750), or by interleaving of the second wave-form coded signal 802 in a time direction (thereby giving priority to the third vector 750).
For this purpose the prioritizing component 13022 comprises predefined rules relating to an order of priority of the vectors 740-760. The prioritizing component 13022 may also comprise predefined rules relating to how the interleaving is to be performed, i.e. if the interleaving is to be performed by way of addition or replacement.
Preferably, these rules are as follows:
The operation of the encoder 500 will be described in the following with reference to
The received audio signal is input to the high frequency encoding stage 530. Based on the received audio signal, the high frequency encoding stage 530, and in particular the high frequency reconstruction parameters calculating stage 530a, calculates in step E04 high frequency reconstruction parameters enabling high frequency reconstruction of the received audio signal above the first cross-over frequency fc. The high frequency reconstruction parameters calculating stage 530a may use any known technique for calculating the high frequency reconstruction parameters, such as SBR encoding. The high frequency encoding stage 530 typically operates in a QMF domain. Thus, prior to calculating the high frequency reconstruction parameters, the high frequency encoding stage 530 may perform QMF analysis of the received audio signal. As a result, the high frequency reconstruction parameters are defined with respect to a QMF domain.
The calculated high frequency reconstruction parameters may comprise a number of parameters relating to high frequency reconstruction. For example, the high frequency reconstruction parameters may comprise parameters relating to how to mirror or copy the audio signal from sub-band portions of the frequency range below the first cross-over frequency fc to sub-band portions of the frequency range above the first cross-over frequency fc. Such parameters are sometimes referred to as parameters describing the patching structure.
The high frequency reconstruction parameters may further comprise spectral envelope parameters describing the target energy levels of sub-band portions of the frequency range above the first cross-over frequency.
The high frequency reconstruction parameters may further comprise missing harmonics parameters indicating harmonics, or strong tonal components that will be missing if the audio signal is reconstructed in the frequency range above the first cross-over frequency using the parameters describing the patching structure.
The interleave coding detection stage 540 then, in step E06, identifies a subset of the frequency range above the first cross-over frequency fc for which the spectral content of the received audio signal is to be waveform-coded. In other words, the role of the interleave coding detection stage 540 is to identify frequencies above the first cross-over frequency for which the high frequency reconstruction does not give a desirable result.
The interleave coding detection stage 540 may take different approaches to identify a relevant subset of the frequency range above the first cross-over frequency fc. For example, the interleave coding detection stage 540 may identify strong tonal components which will not be well reconstructed by the high frequency reconstruction. Identification of strong tonal components may be based on the received audio signal, for example, by determining the energy of the audio signal as a function of frequency and identifying the frequencies having a high energy as comprising strong tonal components. Further, the identification may be based on knowledge about how the received audio signal will be reconstructed in the decoder. In particular, such identification may be based on tonality quotas being the ratio of a tonality measure of the received audio signal and the tonality measure of a reconstruction of the received audio signal for frequency bands above the first cross-over frequency. A high tonality quota indicates that the audio signal will not be well reconstructed for the frequency corresponding to the tonality quota.
The interleave coding detection stage 540 may also detect transients in the received audio signal which will not be well reconstructed by the high frequency reconstruction. Such identification may be the result of a time-frequency analysis of the received audio signal. For example, a time-frequency interval where a transient occurs may be detected from a spectrogram of the received audio signal. Such time-frequency interval typically has a time range which is shorter than a time frame of the received audio signal. The corresponding frequency range typically corresponds to a frequency interval which extends to a second cross-over frequency. The subset of the frequency range above the first cross-over frequency may therefore be identified by the interleave coding detection stage 540 as an interval extending from the first cross-over frequency to a second cross-over frequency.
The interleave coding detection stage 540 may further receive high frequency reconstruction parameters from the high frequency reconstruction parameters calculating stage 530a. Based on the missing harmonics parameters from the high frequency reconstruction parameters, the interleave coding detection stage 540 may identify frequencies of missing harmonics and decide to include at least some of the frequencies of the missing harmonics in the identified subset of the frequency range above the first cross-over frequency fc. Such an approach may be advantageous if there are strong tonal component in the audio signal which cannot be correctly modelled within the limits of the parametric model.
The received audio signal is also input to the waveform encoding stage 520. The waveform encoding stage 520, in step E08, performs waveform encoding of the received audio signal. In particular, the waveform encoding stage 520 generates a first waveform-coded signal by waveform-coding the audio signal for spectral bands up to the first cross-over frequency fc Further, the waveform encoding stage 520 receives the identified subset from the interleave coding detection stage 540. The waveform encoding stage 520 then generates a second waveform-coded signal by waveform-coding the received audio signal for spectral bands corresponding to the identified subset of the frequency range above the first cross-over frequency. The second waveform-coded signal will hence have a spectral content corresponding to the identified subset of the frequency range above the first cross-over frequency fc.
According to example embodiments, the waveform encoding stage 520 may generate the first and the second waveform-coded signals by first waveform-coding the received audio signal for all spectral bands and then, remove the spectral content of the so waveform-coded signal for frequencies corresponding to the identified subset of frequencies above the first cross-over frequency fc.
The waveform encoding stage may for example perform waveform coding using an overlapping windowed transform filter bank, such as a MDCT filter bank. Such overlapping windowed transform filter banks use windows having a certain temporal length, causing the values of the transformed signal in one time frame to be influenced by values of the signal in the previous and the following time frame. In order to reduce the effect of this fact it may be advantageous to perform a certain amount of temporal over-coding, meaning that the waveform-coding stage 520 not only waveform-codes the current time frame of the received audio signal but also the previous and the following time frame of the received audio signal. Similarly, also the high frequency encoding stage 530 may encode not only the current time frame of the received audio signal but also the previous and the following time frame of the received audio signal. In this way, an improved cross-fade between the second waveform-coded signal and a high frequency reconstruction of the audio signal can be achieved in the QMF domain. Further, this reduces the need for adjustment of the spectral envelope data borders.
It is to be noted that the first and the second waveform-coded signals may be separate signals. However, preferably they form first and second waveform-coded signal portions of a common signal. If so, they may be generated by performing a single waveform-encoding operation on the received audio signal, such as applying a single MDCT transform to the received audio signal.
The high frequency encoding stage 530, and in particular the high frequency reconstruction parameters adjusting stage 530b, may also receive the identified subset of the frequency range above the first cross-over frequency fc. Based on the received data the high frequency reconstruction parameters adjusting stage 530b may in step E10 adjust the high frequency reconstruction parameters. In particular, the high frequency reconstruction parameters adjusting stage 530b may adjust the high frequency reconstruction parameters corresponding to spectral bands comprised in the identified subset.
For example, the high frequency reconstruction parameters adjusting stage 530b may adjust the spectral envelope parameters describing the target energy levels of sub-band portions of the frequency range above the first cross-over frequency. This is particularly relevant if the second waveform-coded signal is to be added with a high frequency reconstruction of the audio signal in a decoder, since then the energy of the second waveform-coded signal will be added to the energy of the high frequency reconstruction. In order to compensate for such addition, the high frequency reconstruction parameters adjusting stage 530b may adjust the energy envelope parameters by subtracting a measured energy of the second waveform-coded signal from the target energy levels for spectral bands corresponding to the identified subset of the frequency range above the first cross-over frequency fc. In this way, the total signal energy will be preserved when the second waveform-coded signal and the high frequency reconstruction are added in the decoder. The energy of the second wave-form coded signal may for example be measured by the interleave coding detection stage 540.
The high frequency reconstruction parameters adjusting stage 530b may also adjust the missing harmonics parameters. More particularly, if a sub-band comprising a missing harmonics as indicated by the missing harmonics parameters is part of the identified subset of the frequency range above the first cross-over frequency fc, that sub-band will be waveform coded by the waveform encoding stage 520. Thus, the high frequency reconstruction parameters adjusting stage 530b may remove such missing harmonics from the missing harmonics parameters, since such missing harmonics need not be parametrically reconstructed at the decoder side.
The transmission stage 550 then receives the first and the second waveform-coded signal from the waveform encoding stage 520 and the high frequency reconstruction parameters from the high frequency encoding stage 530. The transmission stage 550 formats the received data into a bit stream for transmission to a decoder.
The interleave coding detection stage 540 may further signal information to the transmission stage 550 for inclusion in the bit stream. In particular, the interleave coding detection stage 540 may signal how the second waveform-coded signal is to be interleaved with a high frequency reconstruction of the audio signal, such as whether the interleaving is to be performed by addition of the signals or by replacement of one of the signals with the other, and for what frequency range and what time interval the waveform coded signals should be interleaved. For example, the signalling may be carried out using the signalling scheme discussed with reference to
Equivalents, Extensions, Alternatives and Miscellaneous
Further embodiments of the present disclosure will become apparent to a person skilled in the art after studying the description above. Even though the present description and drawings disclose embodiments and examples, the disclosure is not restricted to these specific examples. Numerous modifications and variations can be made without departing from the scope of the present disclosure, which is defined by the accompanying claims. Any reference signs appearing in the claims are not to be understood as limiting their scope.
Additionally, variations to the disclosed embodiments can be understood and effected by the skilled person in practicing the disclosure, from a study of the drawings, the disclosure, and the appended claims. In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measured cannot be used to advantage.
The systems and methods disclosed hereinabove may be implemented as software, firmware, hardware or a combination thereof. In a hardware implementation, the division of tasks between functional units referred to in the above description does not necessarily correspond to the division into physical units; to the contrary, one physical component may have multiple functionalities, and one task may be carried out by several physical components in cooperation. Certain components or all components may be implemented as software executed by a digital signal processor or microprocessor, or be implemented as hardware or as an application-specific integrated circuit. Such software may be distributed on computer readable media, which may comprise computer storage media (or non-transitory media) and communication media (or transitory media). As is well known to a person skilled in the art, the term computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by a computer. Further, it is well known to the skilled person that communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
The present application is a continuation of U.S. patent application Ser. No. 16/169,964 filed Oct. 24, 2018, which is a continuation of U.S. patent application Ser. No. 15/279,365 filed Sep. 28, 2016, now U.S. Pat. No. 10,121,479 issued Nov. 6, 2018, which is a continuation of U.S. patent application Ser. No. 14/781,891 filed Oct. 1, 2015, now U.S. Pat. No. 9,514,761 issued Dec. 6, 2016, which is the U.S. national stage of International Patent Application No. PCT/EP2014/056856 filed Apr. 4, 2014, which claims priority to U.S. Provisional Patent Application No. 61/808,687 filed Apr. 5, 2013, all of which are incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
5280561 | Satoh | Jan 1994 | A |
5581618 | Satoshi | Dec 1996 | A |
5598478 | Tanaka | Jan 1997 | A |
5796843 | Inanaga | Aug 1998 | A |
5970443 | Fujii | Oct 1999 | A |
6442275 | Diethorn | Aug 2002 | B1 |
7191136 | Sinha et al. | Mar 2007 | B2 |
7519538 | Villemoes | Apr 2009 | B2 |
7668722 | Villemoes | Feb 2010 | B2 |
7684981 | Thumpudi | Mar 2010 | B2 |
7693709 | Thumpudi | Apr 2010 | B2 |
8015368 | Sharma | Sep 2011 | B2 |
8046214 | Mehrotra | Oct 2011 | B2 |
8190425 | Mehrotra et al. | May 2012 | B2 |
8194754 | Oh | Jun 2012 | B2 |
8199827 | Oh | Jun 2012 | B2 |
8199828 | Oh | Jun 2012 | B2 |
8255231 | Villemoes | Aug 2012 | B2 |
8363842 | Kenmochi | Jan 2013 | B2 |
20020103637 | Henn | Aug 2002 | A1 |
20040225505 | Andersen et al. | Nov 2004 | A1 |
20050047624 | Kleen | Mar 2005 | A1 |
20050096917 | Kjörling | May 2005 | A1 |
20060031075 | Oh | Feb 2006 | A1 |
20080259014 | Inoue | Oct 2008 | A1 |
20080260048 | Oomen et al. | Oct 2008 | A1 |
20080270124 | Son et al. | Oct 2008 | A1 |
20080319739 | Mehrotra | Dec 2008 | A1 |
20090132261 | Kjorling | May 2009 | A1 |
20100023322 | Schnell | Jan 2010 | A1 |
20100063802 | Gao | Mar 2010 | A1 |
20100223052 | Nilsson | Sep 2010 | A1 |
20100262420 | Herre et al. | Oct 2010 | A1 |
20110019838 | Kaulberg | Jan 2011 | A1 |
20110099018 | Neuendorf et al. | Apr 2011 | A1 |
20110202352 | Neuendorf | Aug 2011 | A1 |
20110202355 | Grill et al. | Aug 2011 | A1 |
20110264454 | Ullberg | Oct 2011 | A1 |
20110288873 | Nagel | Nov 2011 | A1 |
20120022676 | Ishikawa | Jan 2012 | A1 |
20120065983 | Ekstrand | Mar 2012 | A1 |
20120078640 | Shirakawa | Mar 2012 | A1 |
20120201388 | Seo | Aug 2012 | A1 |
20120243526 | Yamamoto | Sep 2012 | A1 |
20120275607 | Kjoerling | Nov 2012 | A1 |
20120328124 | Kjoerling | Dec 2012 | A1 |
20130090929 | Ishikawa et al. | Apr 2013 | A1 |
20130096912 | Resch | Apr 2013 | A1 |
20130226597 | Kjoerling | Aug 2013 | A1 |
20140088973 | Gibbs | Mar 2014 | A1 |
20150003632 | Thesing | Jan 2015 | A1 |
Number | Date | Country |
---|---|---|
1177433 | Oct 2003 | CN |
1947174 | Apr 2007 | CN |
101048935 | Oct 2007 | CN |
101086845 | Dec 2007 | CN |
101129090 | Feb 2008 | CN |
101236745 | Aug 2008 | CN |
102089758 | Jun 2011 | CN |
102089814 | Jun 2011 | CN |
102177545 | Sep 2011 | CN |
102246231 | Nov 2011 | CN |
102598121 | Jul 2012 | CN |
103000177 | Mar 2013 | CN |
1682567 | Jun 2014 | CN |
104869335 | Aug 2015 | CN |
10338694 | Mar 2005 | DE |
1035732 | Sep 2000 | EP |
1158494 | Nov 2001 | EP |
1919259 | May 2008 | EP |
2104096 | Sep 2009 | EP |
2008096567 | Apr 2008 | JP |
2008139844 | Jun 2008 | JP |
2010041579 | Feb 2010 | JP |
2013125187 | Jun 2013 | JP |
6859394 | Apr 2021 | JP |
1020070118173 | Dec 2007 | KR |
1020070119722 | Dec 2007 | KR |
2335809 | Oct 2008 | RU |
2470384 | Dec 2012 | RU |
2003046891 | Jun 2003 | WO |
2008084688 | Jul 2008 | WO |
2012158333 | Nov 2012 | WO |
Entry |
---|
A.C. Den Brinker et al. “An Overview of the Coding Standard MPEG-4 Audio Amendments 1 and 2: HE-AAC, SSC and HE-AAC v2” EURASIP Journal on Audio, Speech, and Music Processing, vol. 2009. |
Ehret, A. et al “aacPlus, Only a Low-Bitrate Codec?” AES 117 Convention, Oct. 2004. |
Ekstrand, Per, “Bandwidth Extension of Audio Signals by Spectral Band Replication” Proc 1st IEEE Benelux Workshop on Model Based Processing and Coding of Audio, Leuven, Belgium, Nov. 15, 2002. |
Geiser, B. et al “Bandwidth Extension for Hierarchical Speech and Audio Coding in ITU-T Rec. G.729-1” IEEE Transactions on Audio, Speech, and Language Processing, vol. 15, No. 8, Nov. 2007, pp. 2496-2509. |
Kim et al “Quality Improvement Using a Sinusoidal Model in HE-AAC”, AES 123rd Convention, New York, NY, USA Oct. 1, 2007. |
Kovesi, B. et al “A Scalable Speech and Audio Coding Scheme with Continuous Bitrate Flexibility” IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1, May 17, 2004, pp. 273-276. |
Ragot, S. et al “ITU-T G.729.1: An 8-32 KBIT/S Scalable Coder Interoperable with G.729 for Wideband Telephony and Voice Over IP” IEEE International Conference on Acoustics, Speech, and Signal Processing, Apr. 15-20, 2007, Honolulu, HI, USA. |
Zernicki, T. et al “Improved Coding of Tonal Components in Audio Techniques Utilizing the SBR Tool” MPEG Meeting ISO/IEC JTC1/SC29/WG11, Coding of Moving Pictures and Audio, Jul. 22, 2010, Geneva, Switzerland. |
Nagel, F. et al “A Harmonic Bandwidth Extension Method for Audio Codecs” ICASSP 2009, pp. 145-148. |
Liu Dong-Bing, et al “Overview of Spectral Band Replication in Audio Coding” Digital Engineering Center, Communication University of China, Beijing, Dec. 2011. |
Number | Date | Country | |
---|---|---|---|
20220101865 A1 | Mar 2022 | US |
Number | Date | Country | |
---|---|---|---|
61808687 | Apr 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16169964 | Oct 2018 | US |
Child | 17495184 | US | |
Parent | 15279365 | Sep 2016 | US |
Child | 16169964 | US | |
Parent | 14781891 | US | |
Child | 15279365 | US |