This application is based upon and claims the benefit of priority from the prior Japanese Patent Application No. 2008-315150 filed on Dec. 11, 2008, the entire contents of which are incorporated herein by reference.
The embodiment to be discussed herein relates to an encoding technique for compressing and decompressing an audio signal. The embodiment is also related to an audio encoding and decoding technique, in accordance with which a decoder side reproduces an original audio signal based on a decoded audio signal and a decoded auxiliary signal. For example, the audio encoding and decoding technique includes a parametric stereophonic encoding technique for generating a pseudo-stereophonic signal from a monophonic signal.
The parametric stereophonic encoding technique is adopted in the high-efficiency advanced audio coding (HE-AAC) version 2 standard (hereinafter referred to as “HE-AAC v2”), as one of the MPEG-4 Audio standards. The parametric stereophonic encoding technique as an audio compression technique substantially improves a codec efficiency of a low-bit rate stereophonic signal, and is optimum for applications in mobile devices, broadcasting, and the Internet.
l(t)=c1x(t)+c2h(t)*x(t) (1)
r(t)=c3x(t)+c4h(t)*x(t) (2)
Since a HE-AAC v2 decoder cannot obtain a signal equivalent to the sound source x(t) illustrated in
l′(t)=c′1s(t)+c′2h′(t)*s(t) (3)
r′(t)=c′3s(t)+c′4h′(t)*s(t) (4)
A variety of production methods of the reverberation component are available. For example, a parametric stereophonic (hereinafter referred to as PS) decoder complying with the HE-AAC v2 standard decorrelates (orthogonalizes) a monophonic signal s(t) in order to generate a reverberation signal d(t) and generates a stereophonic signal in accordance with the following equations:
l′(t)=c′1s(t)+c′2d(t) (5)
r′(t)=c′3s(t)+c′4d(t) (6)
For convenience of explanation, the process described above is performed in the time domain. The PS decoder performs a pseudo-stereophonic operation in the time-frequency domain (quadrature mirror filter bank (QMF) coefficient domain). Equations (5) and (6) are thus represented by the following equations (7) and (8) respectively:
l′(b,t)=h11s(b,t)+h12d(b,t) (7)
r′(b,t)=h21s(b,t)+h22d(b,t) (8)
where b is an index representing frequency, and t is an index representing time.
A method of producing a reverberation signal d(b,t) from a monophonic signal s(b,t) is described below. A variety of techniques are available to generate the reverberation signal d(b,t). The PS decoder complying with the HE-AAC v2 standard decorrelates (orthogonalizes) the monophonic signal s(b,t) as illustrated in
For simplicity of explanation, the lengths of L and R are equal to each other in
A method of the decoder of generating a stereophonic signal from the monophonic signal s(b,t) and the reverberation signal d(b,t) is described below. Referring to
Equations (9) and (10) are combined as equations (11) and (12):
A parametric stereophonic decoding apparatus operating on the above-described principle is described below.
A core decoder 2002 decodes the encoded core data and outputs a monophonic audio signal S(b,t). Here, b represents an index of a frequency band. The core decoder 2002 may be based on a known audio encoding and decoding technique such as an advanced audio coding (AAC) system or a spectral band replication (SBR) system.
The monophonic audio signal S(b,t) and the PS data are input to a parametric stereophonic (PS) decoder 2003. The PS decoder 2003 converts the monophonic audio signal S(b,t) into stereophonic decoded signals L(b,t) and R(b,t) in the frequency domain in accordance with the information of the PS data.
Frequency-time converters 2004(L) and 2004(R) convert an L channel frequency-domain decoded signal L(b,t) and an R channel frequency-domain decoded signal R(b,t) into an L channel time-domain decoded signal L(t) and an R channel time-domain decoded signal R(t), respectively.
A PS analyzer 2103 analyzes the PS data, thereby extracting a similarity and an intensity difference from the PS data. As previously discussed with reference to
A coefficient calculator 2104 calculates a coefficient matrix H from the similarity and the intensity difference in accordance with the above-described equation (12). A stereophonic signal generator 2105 generates the stereophonic signals L(b,t) and R(b,t) based on the monophonic audio signal S(b,t), the reverberation signal D(b,t), and the coefficient matrix H in accordance with the above-described equations (11) and (12). Time suffix t is omitted in
L(b)=h11S(b)+h12D(b)
R(b)=h21S(b)+h22D(b) (13)
In one case, the above-described parametric stereophonic system of the related art may receive audio signals having no substantial correlation between an L channel input signal and an R channel input signal, such as two different language voices in encoded form.
In the parametric stereophonic system, a stereophonic signal is generated from a monophonic signal S on a decoder side. As understood from the above-described equation (13), the property of the monophonic signal S affects the output signals L′ and R′.
For example, if an original L channel input signal is completely different from an original R channel input signal (with the similarity being zero), the output audio signal from the PS decoder 2003 of
L′(b)=h11S(b)
R′(b)=h21S(b) (14)
In other words, a component of the monophonic signal S appears in the output signals L′ and R′.
The parametric stereophonic decoding apparatus of the related art emits similar sounds from the left and right if the output signals L′ and R′ are heard at the same time. The user may hear the similar sound as an echo, with the sound quality degraded.
An audio decoding method includes: acquiring, from encoded audio data, a reception audio signal and first auxiliary decoded audio information; calculating coefficient information from the first auxiliary decoded audio information; generating a decoded output audio signal based on the coefficient information and the reception audio signal; decoding to result in a decoded audio signal based on the first auxiliary decoded audio signal and the reception audio signal; calculating, from the decoded audio signal, second auxiliary decoded audio information corresponding to the first auxiliary decoded audio information; detecting a distortion caused in a decoding operation of the decoded audio signal by comparing the second auxiliary decoded audio information with the first auxiliary decoded audio information; correcting the coefficient information in response to the detected distortion; and supplying the corrected coefficient information as the coefficient information when generating the decoded output audio signal.
The object and advantages of the invention will be realized and attained by means of the elements and combinations particularly pointed out in the claims.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are not restrictive of the invention, as claimed.
The best mode embodiments are described below with reference to the drawings.
A reception processor 101 acquires, from encoded audio data, a reception audio signal and auxiliary decoded audio information. More specifically, the reception processor 101 acquires from parametric stereophonic encoded audio data a monophonic audio signal, a reverberation audio signal, and parametric stereophonic parameter information.
A coefficient calculator 102 calculates coefficient information from first auxiliary decoded audio information. More specifically, the coefficient calculator 102 acquires the coefficient information from the parametric stereophonic parameter information.
A decoded audio analyzer 104 decodes an audio signal to generate a decoded audio signal in accordance with the first auxiliary decoded audio information, and the reception audio signal, and calculates, from the decoded audio signal, second auxiliary decoded audio information corresponding to the first auxiliary decoded audio information. More specifically, the decoded audio analyzer 104 decodes the audio signal to generate the decoded audio signal in accordance with parametric stereophonic parameter information as first parametric stereophonic parameter information, a monophonic decoded audio signal, and a reverberation audio signal. The decoded audio analyzer 104 calculates, from the decoded audio signal, second parametric stereophonic parameter information corresponding to the first parametric stereophonic parameter information.
A distortion detector 105 detects distortion caused in the decoding process by comparing the second auxiliary decoded audio information with the first auxiliary decoded audio information. More specifically, the distortion detector 105 detects the distortion caused in the decoding process by comparing the second parametric stereophonic parameter information with the first parametric stereophonic parameter information.
A coefficient corrector 106 corrects the coefficient information in response to the distortion detected by the distortion detector 105, and supplies the corrected coefficient information to an output signal generator 103. The output signal generator 103 generates an output audio signal in a decoded form in response to the corrected coefficient information and the reception audio signal. More specifically, the output signal generator 103 generates an output stereophonic decoded audio signal based on the corrected coefficient information, the monophonic audio signal, and the reverberation audio signal.
In the above-described arrangement, the parametric stereophonic parameter information contains similarity information between stereophonic audio channels and intensity difference information indicating an intensity difference between signals of the stereophonic audio channels. The decoded audio analyzer 104 calculates second similarity information and second intensity difference information, corresponding to first similarity information, as the first parametric stereophonic parameter information, and first intensity difference information, respectively.
The distortion detector 105 compares the second similarity information and the second intensity difference information with the first similarity information and the first intensity difference information, respectively, for each frequency band. The distortion detector 105 thus detects distortion, caused in the decoding process, and an audio channel causing the distortion for each frequency band and for each stereophonic audio channel.
The coefficient corrector 106 corrects the coefficient information of the audio channel detected by the distortion detector 105 in response to the distortion detected by the distortion detector 105 for each frequency band and for each stereophonic audio channel.
A pseudo-stereophonic operation or the like is performed on a monophonic decoded audio signal in accordance with the first parametric stereophonic parameter information. A stereophonic decoded audio signal is thus produced. In such a system, the second parametric stereophonic parameter information corresponding to the first parametric stereophonic parameter information is generated from the stereophonic decoded audio signal. The first parametric stereophonic parameter information is thus compared with the second parametric stereophonic parameter information in order to detect the distortion in the decoding process for the pseudo-stereophonic operation.
A coefficient correction operation to remove echoing may be applied to the stereophonic decoded audio signal. Sound degradation on the decoded audio signal is thus controlled.
A data separator 201, a SBR decoder 203, an AAC decoder 202, a delay adder 205, a decorrelator 206, and a parametric stereophonic (PS) analyzer 207 in
The data separator 201 illustrated in
The AAC decoder 202 illustrated in
The monophonic audio signal S(b,t) and the PS data are input to the parametric stereophonic (PS) decoder 204. The PS decoder 204 illustrated in
The parametric stereophonic (PS) analyzer 207 illustrated in
The coefficient calculator 208 illustrated in
The distortion detector 210 illustrated in
The coefficient corrector 211 illustrated in
The stereophonic signal generator 212 generates stereophonic signals L(b,t) and R(b,t) based on the monophonic audio signal S(b,t), the reverberation signal D(b,t), and the corrected coefficient matrix H′(b) (step S310 illustrated in
Frequency-time converters 213(L) and 213(R) convert an L channel frequency-domain decoded signal and an R channel frequency-domain decoded signal, spectrum corrected in accordance with the corrected coefficient matrix H′(b), into an L channel time-domain decoded signal L(t) and an R channel time-domain decoded signal R(t), and then outputs the L channel time-domain decoded signal L(t) and the R channel time-domain decoded signal R(t) (step S311 illustrated in
The input stereophonic sound may be jazz, which is typically free from echoing, as illustrated in
The input stereophonic sound may be two languages (for example, L channel: German, and R channel: Japanese) with echoing as illustrated in
In accordance with the second embodiment illustrated in
If the input stereophonic sound is two languages (for example, L channel: German, and R channel: Japanese) as illustrated in
The decoded audio analyzer 209, the distortion detector 210, and the coefficient corrector 211 illustrated in
The first intensity difference iid(b) and the first similarity icc(b) at a frequency band b, transmitted from a parametric stereophonic encoding apparatus and then extracted by a parametric stereophonic decoding apparatus, are calculated in accordance with the following equations (15):
where N represents a frame length (see
From the equations (15), the first intensity difference iid(b) is the logarithm of the power ratio of the mean power eL(b) at the L channel signal L(b,t) to the mean power eR(b) at the R channel signal R(b,t) at a current frame (0≦t≦N−1) at the frequency band b, and the first similarity icc(b) is a correlation between the L channel signal L(b,t) and the R channel signal R(b,t).
The relationship illustrated in
icc(b)=cos(2α) (16)
The norm ratio of the L channel signal L(b,t) to the R channel signal R(b,t) is defined as the first intensity difference iid(b). As illustrated in
The coefficient calculator 208 illustrated in
α=1/2 arccos(icc(b)) (17)
Scale factors Cl and Cr in equation (12) are calculated based on the first intensity difference iid(b) output from the PS analyzer 207 illustrated in
The decoded audio analyzer 209 illustrated in
The decoded audio analyzer 209 calculates the second intensity difference iid′(b) and the second similarity icc′(b) at the frequency band b in accordance with the following equations (19), based on the decoded L channel signal L′(b,t) and the decoded R channel signal R′(b,t) as in the same manner as with equations (15):
In the same manner as with equations (15), the relationship illustrated in
icc′(b)=cos(2α′) (20)
The norm ratio of the decoded L channel signal L′(b,t) to the decoded R channel signal R′(b,t) is defined as the second intensity difference iid′(b).
The L channel signal L(b,t), the R channel signal R(b,t), the first similarity icc(b), and the first intensity difference iid(b), prior to the parametric stereophonic operation, are related to each other as illustrated in
(1) The L channel signal L(b,t) and the decoded L channel signal L′(b,t) are different from each other by an angle of θl related to a difference between angles α and α′. The R channel signal R(b,t) and the decoded R channel signal R′(b,t) are different from each other by an angle of θr related to the difference between the angles α and α′. Let a distortion 1 represent the difference. In practice, the assumption of the distortion 1=θ=θl=θr holds without any problem.
(2) The L channel signal L(b,t) and the decoded L channel signal L′(b,t) are different from each other by an amplitude Xl. The R channel signal R(b,t) and the decoded R channel signal R′(b,t) are also different from each other by an amplitude Xr. Let a distortion 2 represent the difference. In practice, the assumption of the distortion 2=X=Xl=Xr holds without any problem.
From the above understanding, the distortion detector 210 illustrated in
A specific detection method of the distortion detector 210 detecting the distortion 1=θ is described below. The angle α′ (see
α′=1/2 arccos(icc′(b)) (21)
The angle α (see
The distortion 1=θ (=θ(b)) at the frequency band b (see
θ=α−α′=1/2{ arccos(icc(b))−arccos(icc′(b))} (22)
More specifically, the distortion detector 210 performs equation (22) based on the first similarity icc(b) at the frequency band b calculated by the PS analyzer 207, and the second similarity icc′(b) at the frequency band b calculated by the decoded audio analyzer 209. As a result, the distortion 1=θ(=θ(b)) at the frequency band b is calculated.
The distortion 1=θ may also be calculated in the manner described below. The distortion detector 210 calculates a difference A(b) between the similarities at the frequency band b from the first similarity icc(b) and the second similarity icc′(b) at the frequency band b in accordance with the following equation (23):
A(b)=icc′(b)−icc(b) (23)
The distortion detector 210 calculates the distortion 1=θ=θ(b) for the similarity difference A(b) calculated in accordance with equation (23) based on a conversion table relating to a pre-calculated similarity difference to the distortion 1. The distortion detector 210 continuously stores a graph (relationship) on which the conversion table is based as illustrated in
The detection method of the distortion detector 210 detecting the distortion 2=X (see
The distortion detector 210 converts the distortion 2=γ(b) in accordance with the following equation (24), and outputs the resulting physical quantity X as the distortion 2 in order to perform the spectrum power correction as a correction to the coefficient matrix H(b):
The correction process of the coefficient corrector 211 correcting the coefficient matrix H(b) is described below.
The coefficient corrector 211 calculates the corrected coefficient matrix H′(b) for the coefficient matrix H(b) calculated by the coefficient calculator 208 in accordance with the following equations (25) in view of equations (12), (17), and (18).
where an angle α is the angle α calculated by the coefficient calculator 208 in accordance with equation (17), and scale factors Cl and Cr are the scale factors Cl and Cr calculated by the coefficient calculator 208 in accordance with equation (18). The angle correction values θ=θl=θr and the power correction values X=Xl=Xr are respectively the distortion 1 and the distortion 2 output by the distortion detector 210.
In accordance with the following equation (26), the stereophonic signal generator 212 decodes the L channel signal L(b,t) and the R channel signal R(b,t) based on the monophonic audio signal S(b,t) output from the SBR decoder 203 and the reverberation signal D(b,t) output from the decorrelator 206. Equation (26) is based on the corrected coefficient matrix H′(b) calculated by the coefficient corrector 211:
The parametric stereophonic decoding apparatus performs the above-described operations in every frequency band b while determining whether to perform the correction or not. In such operations, the operations of the distortion detector 210 and the coefficient corrector 211 is described further in detail.
The distortion detector 210 and coefficient corrector 211 set a frequency band number to zero in step S1001. The distortion detector 210 and coefficient corrector 211 perform a series of process steps from step S1001 to step S1013 at each frequency band b with the frequency band number in step S1015 incremented by 1 until it is determined in step S1014 whether the frequency band number exceeds a maximum value NB−1.
The distortion detector 210 calculates the similarity difference A(b) in accordance with equation (23) (step S1002). The distortion detector 210 compares the similarity difference A(b) with a threshold value Th1 (step S1003). Referring to
If the similarity difference A(b) is equal to or smaller than the threshold value Th1, the distortion detector 210 determines that no distortion exists. The distortion detector 210 then sets, to a variable ch(b) indicating a channel suffering from distortion at the frequency band b, a value zero meaning that none of the channels are to be corrected. Processing proceeds to step S1013 (step S1003→step S1010→step S1013).
If the similarity difference A(b) is larger than the threshold value Th1, the distortion detector 210 determines that a distortion exists, and then performs steps S1004-S1009.
In accordance with the following equation (27), the distortion detector 210 subtracts the value of the first intensity difference iid(b) output from the PS analyzer 207 of
B(b)=iid′(b)−iid(b) (27)
As a result, a difference B(b) between the intensity differences at the frequency band b is calculated (step S1004).
The distortion detector 210 compares the difference B(b) between the intensity differences with a threshold value Th2 and a threshold value −Th2 (steps S1005 and 1006). If the intensity difference B(b) is larger than the threshold value Th2 as illustrated in
A larger value of the first intensity difference iid(b) in the calculation of the first intensity difference iid(b) in accordance with equation (15) shows that the power of the L channel is stronger. If this tendency is more pronounced on the decoder side than on the encoder side, i.e., if the difference B(b) is above the threshold value Th2, a stronger distortion component is superimposed on the L channel. Conversely, a smaller value of the first intensity difference iid(b) means that the power of the R channel is higher. If this tendency is more pronounced on the decoder side than on the encoder side, i.e., if the difference B(b) is below the threshold value −Th2, a stronger distortion component is superimposed on the R channel.
In other words, if the difference B(b) is larger the threshold value Th2, the distortion detector 210 determines that the L channel suffers from distortion. The distortion detector 210 thus sets a value L to the distortion-affected channel ch(b), and then proceeds to step S1011 (step S1005→step S1009→step S1011).
If the difference B(b) is equal to or smaller than the threshold value −Th2, the distortion detector 210 determines that the R channel suffers from distortion. The distortion detector 210 thus sets a value R to the distortion-affected channel ch(b), and then proceeds to step S1011 (step S1005→step S1006→step S1008→step S1011).
If the difference B(b) is larger the threshold value −Th2 but equal to or smaller than the threshold value Th2, the distortion detector 210 determines that both channels suffer from distortion. The distortion detector 210 thus sets a value LR to the distortion-affected channel ch(b), and then proceeds to step S1011 (step S1005→step S1006→step S1007→step S1011).
Subsequent to any one of steps S1007-S1009, the distortion detector 210 calculates the distortion 1. As previously discussed, the distortion detector 210 calculates equation (22) based on the first similarity icc(b) at the frequency band b calculated by the PS analyzer 207 and the second similarity icc′(b) at the frequency band b calculated by the decoded audio analyzer 209. As a result, the distortion 1=θ (=θ(b)) at the frequency band b is calculated.
The distortion detector 210 then calculates the distortion 2. As previously discussed, the distortion detector 210 calculates the physical quantity γ(b) for the similarity difference A(b) calculated in step S1002 based on the relationship of the pre-calculated similarity difference and the distortion 2. The distortion detector 210 further calculates the distortion 2=X for the physical quantity γ(b) in accordance with equation (24).
In this way, the distortion detector 210 detects the distortion-affected channel ch(b), the distortion 1 and the distortion 2 at the frequency band b. These pieces of information are then transferred to the coefficient corrector 211 (step S1011→step S1012→step S1013).
If the value LR is set to the distortion-affected channel, the coefficient corrector 211 calculates the corrected coefficient matrix H′(b) based on the angular correction values θl=θr=θ (distortion 1) and the power correction values Xl=Xr=X (distortion 2) in accordance with equation (25).
If the value R is set to the distortion-affected channel, the coefficient corrector 211 calculates the corrected coefficient matrix H′(b) based on the angular correction values θr=θ (distortion 1) and θl=θ, and the power correction values Xl=X (distortion 2) and Xr=1 in accordance with equation (25).
If the value L is set to the distortion-affected channel, the coefficient corrector 211 calculates the corrected coefficient matrix H′(b) based on the angular correction values θl=θ (distortion 1) and θr=θ and the power correction values Xl=X (distortion 2) and Xr=1 in accordance with equation (25).
If the value zero is set to the distortion-affected channel, the coefficient corrector 211 calculates the corrected coefficient matrix H′(b) based on the angular correction values θl=θr=0 and the power correction values Xl=Xr=1 in accordance with equation (25).
The input data mainly includes an ADTS header 1201, AAC data 1202 as monophonic audio AAC encoded data, and an extension data region (FILL element) 1203.
SBR data 1204 as monophonic audio SBR encoded data and SBR extension data (sbr_extension) 1205 are included in the FILL element 1203.
Parametric stereophonic PS data 1206 is stored in sbr_extension 1205. Parameters needed for a PS decoding operation, such as the first similarity icc(b) and the first intensity difference iid(b), are contained in the PS data 1206.
A third embodiment is described below. The third embodiment is different in the operation of the coefficient corrector 211 from the second embodiment illustrated in
In accordance with the second embodiment, the relationship used by the coefficient corrector 211 in the determination of γ(b) from the similarity difference A(b) is fixed. In accordance with the third embodiment, an appropriate relationship may be used in response to the power of a decoded audio signal.
If the power of the decoded audio signal is high as illustrated in
The “power of the decoded audio signal” refers to the power of the decoded L channel signal L′(b,t) or the decoded R channel signal R′(b,t), calculated by the decoded audio analyzer 209, at the frequency band b of the channel to be corrected.
A fourth embodiment is described.
Referring to
Every discrete time t, the coefficient storage unit 1401 successively stores a corrected coefficient matrix (hereinafter referred to as H′(b,t)) output from the coefficient corrector 211 while outputting, to the coefficient smoother 1402, a corrected coefficient matrix (hereinafter referred to as H′(b,t−1)) at time (t−1) one discrete time unit before.
Using the corrected coefficient matrix H′(b,t) at discrete time t output from the coefficient corrector 211, the coefficient smoother 1402 smoothes each coefficient (see equation (25)) forming the corrected coefficient matrix H′(b,t−1) at time (t−1) one discrete time unit before input from the coefficient storage unit 1401. The coefficient smoother 1402 thus outputs the resulting matrix to the stereophonic signal generator 212 as the corrected coefficient matrix H″(b,t−1).
A smoothing technique of the coefficient smoother 1402 is not limited to any particular one. For example, a technique of weighted summing the output from the coefficient storage unit 1401 and the output from the coefficient corrector 211 at each coefficient may be used.
Alternatively, a plurality of past frames output from the coefficient corrector 211 may be stored on the coefficient storage unit 1401, and the plurality of past frames and the output from the coefficient corrector 211 may be weighted summed for smoothing.
The smoothing operation is not limited to the time axis. The smoothing operation may be performed on the output from the coefficient corrector 211 in the direction of the frequency band b. More specifically, the weighted summing operation for smoothing may be performed on the coefficients forming the corrected coefficient matrix H′(b,t) at the frequency band b output from the coefficient corrector 211, the coefficients at the frequency band b−1 and the coefficients at the frequency band b+1. When the weighted summing operation is performed, the corrected coefficient matrices output from the coefficient corrector 211 at a plurality of adjacent frequency bands may be used.
Supplementary to First Through Fourth Embodiments
The computer illustrated in
The CPU 1501 generally controls the computer. When programs are executed or data is updated, the memory 1502 such as a RAM or the like stores a program stored on the external storage device 1505 (or the removable recording medium 1509) or data. The CPU 1501 reads the program onto the memory 1502 and executes the read program, thereby generally controlling the computer.
The input unit 1503 includes a keyboard, a mouse, etc. and interfaces thereof. The input unit 1503 detects an input operation performed on the keyboard, the mouse, etc. by a user, and notifies the CPU 1501 of the detection results.
The output unit 1504 includes a display, a printer, etc., and interfaces thereof. The output unit 1504 outputs data supplied under the control of the CPU 1501 to the display or the printer.
The external storage device 1505 may be a hard disk storage, for example and may be mainly used to store a variety of data and programs. The removable recording medium driver 1506 receives the removable recording medium 1509 such as an optical disk, a synchronous dynamic random access memory (SDRAM), or a Compact Flash (registered trademark). The removable recording medium driver 1506 serves as an auxiliary unit to the external storage device 1505.
The network interface device 1507 connects to a local-area network (LAN) or a wide-area network (WAN). The parametric stereophonic decoding system according to of the first through fourth embodiments is implemented by the CPU 1501 that executes the program incorporating the functions as described above. The program may be distributed in the external storage device 1505 or the removable recording medium 1509 or may be acquired via the network by the network interface device 1507.
In the first through fourth embodiments, the present invention is applied to the parametric stereophonic decoding apparatus. The present invention is not limited to the parametric stereophonic apparatus. The present invention may be applicable to a variety of systems including a surround system in which the decoding process is performed with audio decoded auxiliary information combined with the decoded audio signal.
All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the invention and the concepts contributed by the inventor to furthering the art, and are to be constructed as being without limitation to such specifically recited examples and conditions, nor does the organization of such examples in the specification relate to a showing of the superiority and inferiority of the invention. Although the embodiments of the present inventions have been described in detail, it should be understood that the various changes, substitutions, and alterations could be made hereto without departing from the spirit and scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
2008-315150 | Dec 2008 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
7200561 | Moriya et al. | Apr 2007 | B2 |
7382886 | Henn et al. | Jun 2008 | B2 |
7447317 | Herre et al. | Nov 2008 | B2 |
7555434 | Nomura et al. | Jun 2009 | B2 |
7822617 | Taleb et al. | Oct 2010 | B2 |
7848931 | Miyasaka et al. | Dec 2010 | B2 |
8073687 | Suzuki et al. | Dec 2011 | B2 |
8108220 | Saunders et al. | Jan 2012 | B2 |
8170882 | Davis | May 2012 | B2 |
20050053242 | Henn et al. | Mar 2005 | A1 |
20050149322 | Bruhn et al. | Jul 2005 | A1 |
20050226426 | Oomen et al. | Oct 2005 | A1 |
20050254446 | Breebaart | Nov 2005 | A1 |
20060023888 | Henn et al. | Feb 2006 | A1 |
20060023891 | Henn et al. | Feb 2006 | A1 |
20060023895 | Henn et al. | Feb 2006 | A1 |
20060029231 | Henn et al. | Feb 2006 | A1 |
20070127585 | Suzuki et al. | Jun 2007 | A1 |
20080071549 | Chong et al. | Mar 2008 | A1 |
20080097750 | Seefeldt et al. | Apr 2008 | A1 |
20080170711 | Breebaart et al. | Jul 2008 | A1 |
20080192941 | Oh et al. | Aug 2008 | A1 |
20080205658 | Breebaart | Aug 2008 | A1 |
20080255860 | Osada | Oct 2008 | A1 |
20080260170 | Nakano | Oct 2008 | A1 |
20090010440 | Jung et al. | Jan 2009 | A1 |
20090083040 | Myburg et al. | Mar 2009 | A1 |
20090129601 | Ojala et al. | May 2009 | A1 |
20090234656 | Oh et al. | Sep 2009 | A1 |
20090287495 | Breebaart et al. | Nov 2009 | A1 |
20100080397 | Suzuki et al. | Apr 2010 | A1 |
Number | Date | Country |
---|---|---|
1355428 | Jan 2002 | EP |
2004-535145 | Nov 2004 | JP |
2005-523624 | Aug 2005 | JP |
2007-79487 | Mar 2007 | JP |
2008-26914 | Feb 2008 | JP |
2008-519306 | Jun 2008 | JP |
2002-223167 | Aug 2012 | JP |
03007656 | Jan 2003 | WO |
2006003891 | Jan 2006 | WO |
Entry |
---|
Japanese Office Action issued on Sep. 25, 2012 in corresponding Japanese Patent Application No. 2008-315150. |
Number | Date | Country | |
---|---|---|---|
20100153120 A1 | Jun 2010 | US |