The present invention relates to a distance rendering method used for audio signal output and an audio signal output apparatus using the same.
Recently, with advances in information technology (IT), a variety of smart devices has been developed. Particularly, smart devices basically provide audio output having various effects. Therefore, various methods for more realistic audio output have been attempted in a virtual reality (VR) environment. Thereamong, a distance rendering method provides more realistic audio in a VR environment by adjusting sense of distance caused by the distance between a sound image and a user through a direct-to-reverberation (D/R) ratio. A conventional general distance rendering method will now be described.
For example, in order to reproduce, through a headphone or an earphone, an effect of localizing a sound image from an arbitrary direction in a VR space, information about the VR space and information about the direction of a sound source are needed. Generally, the information about the direction of the sound source is provided by a user. Accordingly, if a database (DB) of head-related transfer functions (HRTFs) measured in all directions is present, a sound image may be localized in a desired direction with reference to the provided direction information. Further, if a head tracker is used together with the DB, the sound image may be localized in a fixed direction regardless of movement of the head of a user. Generally, information about a space may be acquired by measuring a room impulse response (simply, “room response”) measured in the corresponding space. However, since the room response cannot be actually measured in a “VR space”, the room response may be artificially generated using a room impulse response modeling (hereinafter, simplified to “room modeling”) method. If the generated room response is filtered together with an input signal, a user may perceive a spatial sense or reverberation for the measured space.
A user (listener) in an arbitrary space may move and the distance between a sound image and the user differs according to a moved location. This means that reverberant sound as well as the magnitude of a sound source to which the listener listens becomes different. Accordingly, a room response characteristic varies according to movement of a user.
A conventional distance rendering method adjusts sense of distance through signal processing by modifying a room response. Generally, the conventional distance rendering method uses a D/R ratio. The D/R ratio is expressed by Equation (1).
In Equation (1), h(t) denotes a room response and t1 denotes time until a response of the direct part 11 is received. As indicated by Equation (1), the D/R ratio is controlled by the ratio of energy of the direct part 11 to energy of the early reflection part 12 and the late reverberation part 13.
For reference,
For example, if it is desired that a relative distance between the user and the sound image be short, a gain of a direct part 11b of the room response may be adjusted to be increased and a gain of an early reflection and late reverberation part 14b may be adjusted to be decreased as illustrated in
On the contrary, if it is desired that a relative distance between the user and the sound image be increased, a gain of a direct part 11c may be adjusted to be decreased and a gain of an early reflection and late reverberation part 14c may be adjusted to be increased, as illustrated in
However, if it is desired that the distance between the sound image and the user be very close, a gain of the direct part is remarkably increased relative to a gain of the early reflection and a gain of the late reverberation part. In addition, if a remarkably modified gain is applied to a room response, the user may not feel reverberation even though the response has been measured in a space having reverberation. That is, a spatial sense is decreased as if the response is measured in an anechoic room (or dead room) and thus a spatial characteristic varies.
Moreover, a correlation between left and right ears (simply, binaural correlation) according to change in distance cannot be adjusted only by adjustment of the D/R ratio. Accordingly, if the distance between the sound image and the user is longer, a gain of the sound image is decreased and the binaural correlation is also decreased.
Accordingly, a new method capable of maintaining a spatial characteristic even when a relative distance between the user and the sound image is changed is needed. In addition, a new method capable of maintaining a binaural correlation characteristic based on the spatial characteristic even when a relative distance between the user and the sound image is changed is needed. That is, a binaural correlation value should be changed according to change in distance in order to maintain a spatial characteristic.
Accordingly, the present invention is directed to a distance rendering method used for audio signal output and an audio signal output apparatus using the same that substantially obviate one or more problems due to limitations and disadvantages of the related art.
An object of the present invention is to provide a new distance rendering method for maintaining a spatial characteristic during audio output in an arbitrary space even when a relative distance between a user and a sound image is changed.
Another object of the present invention is to provide a new distance rendering method for changing a correlation value so as to maintain a binaural correlation characteristic according to an initial spatial characteristic during audio output in an arbitrary space even when a relative distance between a user and a sound image is changed.
Another object of the present invention is to provide an audio output apparatus for providing audio output using a new distance rendering method.
Another object of the present invention is to provide an audio output system for providing audio output using a new distance rendering method.
Additional advantages, objects, and features of the invention will be set forth in part in the description which follows and in part will become apparent to those having ordinary skill in the art upon examination of the following or may be learned from practice of the invention. The objectives and other advantages of the invention may be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
To achieve these objects and other advantages and in accordance with the purpose of the invention, as embodied and broadly described herein, an audio distance rendering method used for audio output includes receiving a decoded audio signal, checking whether a distance between a user and a sound image is changed, if the distance is changed, calculating a gain parameter gnew and a period parameter τnew, applied to room response modeling, based on the changed distance, so as to maintain a space characteristic, and generating room response modeling for maintaining the space characteristic even when the distance is changed, using the calculated period parameter τnew and gain parameter gnew.
In accordance with another aspect of the present invention, an audio output apparatus includes an audio decoder configured to decode an input audio stream and output the decoded audio signal, and a renderer configured to render the decoded audio signal into an audio signal suitable for a space characteristic, wherein the renderer performs a process of checking whether a distance between a user and a sound image is changed, a process of calculating, when the distance is changed, a gain parameter gnew and a period parameter τnew, applied to room response modeling, based on the changed distance, so as to maintain the space characteristic, and a process of generating room response modeling for maintaining the space characteristic even when the distance is changed, using the calculated period parameter τnew and gain parameter gnew.
The gain parameter gnew may be calculated by adjusting a gain of a room response by the same ratio and the period parameter τnew may be calculated by changing a generation period of a late reverberation part among room response characteristics.
When the distance between the user and the sound image becomes shorter, a gain of a direct part, a gain of an early reflection part, and a gain of a late reverberation part among elements constituting the room response may be adjusted to be increased by the same ratio, and a generation period of the late reverberation part among the elements constituting the room response may be adjusted to be long, so that less late reverberation is generated in comparison with the amount of late reverberation when the distance is not changed.
When the distance between the user and the sound image becomes shorter, the gain parameter gnew and the period parameter τnew may be generated to have values larger than a gain parameter g and a period parameter τ when the distance is not changed.
When the distance between the user and the sound image becomes longer, a gain of a direct part, a gain of an early reflection part, and a gain of a late reverberation part among elements constituting the room response may be adjusted to be decreased by the same ratio, and a generation period of the late reverberation part among the elements constituting the room response may be adjusted to be short, so that a greater amount of late reverberation is generated in comparison with the amount of late reverberation when the distance is not changed.
When the distance between the user and the sound image becomes longer, the gain parameter gnew, and the period parameter τnew may be generated to have values smaller than a gain parameter g and a period parameter τ when the distance is not changed.
In accordance with another aspect of the present invention, an audio distance rendering method used for audio output includes receiving a decoded audio signal, checking whether a distance between a user and the sound image is changed, and if the distance is changed, adjusting a binaural correlation based on the changed distance.
The adjusting the binaural correlation may include adjusting a correlation of a late reverberation part of a binaural room impulse response (BRIR).
The adjusting the binaural correlation may further include extracting the late reverberation part of the BRIR, calculating a parameter IACCLate corresponding to an interaural cross correlation (IACC) of the extracted late reverberation part, estimating a parameter IACCLate_Δd corresponding to an IACC at a changed location with reference to information about a distance variation Δd, modeling the late reverberation part of the BRIR so as to have a correlation corresponding to the estimated parameter IACCLate_Δd and replacing an existing late reveration part with the modeled late reverberation part, and changing a gain of the BRIR based on the modeled late reverberation part according to change in distance.
It is to be understood that both the foregoing general description and the following detailed description of the present invention are exemplary and explanatory and are intended to provide further explanation of the invention as claimed.
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the principle of the invention. In the drawings:
Reference will now be made in detail to the embodiments of the present invention, examples of which are illustrated in the accompanying drawings. In the drawings, the same or similar elements are denoted by the same reference numerals even though they are depicted in different drawings, and a detailed description of the same or similar elements will be omitted. The suffixes “module” and “unit” used in the description below are given or used together only in consideration of ease in preparation of the specification and do not have distinctive meanings or functions. In addition, in the following description of the embodiments disclosed herein, a detailed description of related known technologies will be omitted when it may make the subject matter of the embodiments disclosed herein rather unclear. In addition, the accompanying drawings have been made only for a better understanding of the embodiments disclosed herein and are not intended to limit technical ideas disclosed herein, and it should be understood that the accompanying drawings are intended to encompass all modifications, equivalents and substitutions within the sprit and scope of the present invention.
The audio output apparatus of the present invention includes an audio decoder 100, a renderer 200, a user control interface 300, digital-to-analog (D/A) converters 401 and 402, amplifiers 501 and 502, and transducers 601 and 602.
A bitstream input to the audio decoder 100 may be transmitted by an encoder (not shown) and may be input in the form of a compressed audio file (e.g., .mp3 or .aac). The audio decoder 100 decodes the input audio bitstream according to a coded format to output the decoded audio signal 1001. Generally, the decoded audio signal 1001 may be pulse code modulation (PCM) data (.pcm) but the present invention is not limited thereto. The decoded audio signal 1001 is input to the renderer 200. The renderer 200 outputs left and right signals 2001 by applying a new distance rendering method related to first and second embodiments of the present invention. The left and right output signals 2001 are reproduced through left and right transducers 601 and 602 of headphones (or earphones) via D/A converters 401 and 402 and amplifiers 501 and 502, respectively. Hereinafter, a detailed configuration and operation of the renderer 200 to which the new distance rendering method is applied will be described in detail.
The distance information provider 204 provides information about the distance d between a user (listener) and a sound image. The room information provider 203 provides information about a room type (e.g., a rectangular room structure, a circular room structure, a partially open room structure, etc.) and information about a room size, corresponding to an arbitrary space in which a user is located and a VR space. The tracking information provider 201 provides a localization point (θ, Φ) of the sound image.
The room response generator 205 receives the information (e.g., the distance information, room type information, and tracking information (θ, Φ)) provided by the distance information provider 204, the room information provider 203, and the tracking information provider 201 and generates a room impulse response signal according to the new distance rendering method.
The audio renderer 200 further includes the HRTF selector 202. The tracking information provider 201 provides the localization point (θ, Φ) of a sound image desired to be currently output to the HRTF selector 202. The tracking information provider 201 may be a head tracker for tracking movement of a user or the user may directly provide related information through the user control interface 300. For example, the sound localization point (θ, Φ) provided by the tracking information provider 201 is information about an azimuth angle θ and an altitude angle Φ. The HRTF selector 202 extracts HRTF data corresponding to the sound localization point (θ, Φ) provided by the tracking information provider 201 from an HRTF (DB) 2021. If the HRTF data corresponding to the sound localization point (θ, Φ) is not present in the HRTF DB 2021, the HRTF selector 202 may generate new HRTF data by interpolating data present in the HRTF DB 2021.
The audio renderer 200 further includes the filter 206. The filter 206 receives the room impulse response signal generated by the room response generator 205 and the HRTF data HL and HR determined by the HRTF selector 202 and performs filtering using a preset filtering method to output an audio signal.
Particularly, new room modeling in which a period of generating a late reverberant part is adjusted is applied to the audio renderer 200 according to the first embodiment of the present invention as the new distance rendering method. Hereinafter, the audio renderer 200 and the distance rendering method according to the first embodiment of the present invention will be described in detail with reference to
The above-described HRTF selector 202 extracts HRTF data hL and hR corresponding to tracking information (θ, φ) of a sound image provided by the tracking information provider 201 from the HRTF DB 2021. In addition, as described above, in a VR space environment, a user should directly model a space or a room. Therefore, the user may select preset values similar to values of a room that is desired to be made or directly input room characteristic information RT (room shape, room size, etc.), through the room information provider 203. The distance information provider 204 provides information about a relative distance between a user and a sound image.
A distance determiner 2051 determines whether the location of a user is changed, from the information about the distance provided by the distance information provider 204. That is, the distance determiner 2051 controls a switching means 2054 to select an operation according to a path “y” when it is determined that the relative location of the user is changed. Meanwhile, the distance determiner 2051 controls the switching means 2054 to select an operation according to a path “n” when it is determined that the relative location of the user is not changed.
First, the operation according to the path “n” through the switching means 2054 will now be described. That is, the operation is performed in the case in which a distance variation Δd out of the distance information provided by the distance information provider 204 is 0, i.e., when the relative location of the user is not changed. In this case, a switch flag of the switching means 2054 is selected as the path “n”.
A room response may generally enable individual modeling of an early reflection part and a late reverberation part. In
An early reflection part rearly modeled by the early reflection modeling unit 2052 is filtered together with the extracted HRTF data hL and hR and an input audio signal s by a first filter 206a and the first filter 206a generates left and right early reflection signals srearly_L and srearly_R.
A late reverberation part rlate modeled by the late reverberation modeling unit 2057 is also filtered together with the input audio signal s by a second filter 206b and the second filter 206b generates a late reverberation signal srlate. The late reverberation signal srlate signal and the early reflection signals srearly_L and srearly_R are added by adders 206c and 206d and then output as rendered audio signals and OutputR.
Next, the operation according to the path “y” through the switching means 2054 will now be described. That is, the operation is performed in the case in which the distance variation Δd out of the distance information provided by the distance information provider 204 is not 0, i.e., when the relative location of the user is changed. In this case, the switch flag of the switching means 2054 is selected as the path “y”.
If the user moves (i.e., Δd≠0), a first gain modification unit 2055 of
In addition, if the user moves (i.e., Δd≠0), a τ-modification unit 2056 of
The second filter 206b performs filtering on an output signal of the second gain modification unit 2059 and the input audio signal s and generates the late reverberation signal srlate.
The adders 206c and 206d add the output signals srlate of the second filter 206b to the output signals srearly_L and srearly_R of the first gain modification unit 2055 and generate the rendered audio signals and OutputR.
The early reflection modeling unit 2052 may model only parts from the direct part to the early reflection part of the room response and use conventional normal modeling methods. The late reverberation part of the room response is modeled using an artificial reverberator in each of the late reverberation modeling units 2057 and 2058. The artificial reverberator has the structure of a feedback delay network (FDN) including a plurality of comb filters as illustrated in
The parameter P indicates the number of comb filters in
The parameters g=[g1, g2, . . . , gP] and c=[c1, c2, . . . , cP] are values given from the user and are set to values less than 1. When the late reverberation part is modeled using the FDN structure of
Referring to
Referring to
A bitstream is received (S100). The bitstream is decoded and output as a decoded signal (S101). It is determined through a distance tracking process (S201) whether there is change in distance (S202).
If there is change in distance, the next process (path “y”) is performed. An HRTF is extracted from an HRTF DB with reference to head tracking information (S206). A new gain parameter gnew and a new period parameter τnew are calculated by calculating the amount of change in distance (S207). A new late reverberation part is modeled using the calculated period parameter τnew (S208). An early reflection part is modeled according to a conventional scheme (S209). The calculated gain parameter gnew is applied to the HRTF, the modeled early reflection part, and the modeled late reverberation part (S210). The HRTF, early reflection part, and late reverberation part are filtered together with the decoded signal in step S101 (S300). As a result of filtering, a rendered audio signal is output (S400).
Meanwhile, if it is determined in step S202 that there is no change in distance, the next process (path “n”) is performed. The HRTF is extracted from the HRTF DB with reference to the head tracking information (S203). An early reverberation part is modeled (S204). A late reverberation part is modeled (S205). The HRTF, early reflection part, and late reverberation part are filtered together with the decoded signal in step S101 (S300). As a result of filtering, the rendered audio signal is output (S400).
In step S207, a process of generating the new gain parameter gnew and the new period parameter τnew is as follows. If a distance variation Δd is not 0, since it is determined that a user moves, a parameter τ associated with a late reverberation generation period is modified. The new parameter τnew is calculated as indicated by Equation (2)
In Equation (2), d denotes the distance (an original distance) between a user and a sound image. In consideration of an inverse square law indicating that the magnitude of a sound pressure is inversely proportional to the square of distance, the square of d is applied. Therefore, if the modified period τnew is used as a new parameter of an artificial reverberator, since the value of τnew becomes small as the distance between the user and the sound source becomes long (Δd>0), a greater amount of reverberation is generated in a unit time. Meanwhile, since the value of τnew is increased as the distance between the user and the sound source becomes short (Δd<0), less reverberation is generated in a unit time.
The amount of variation in gain of a room response according to change in distance is calculated in step S207 and the gain modification units 2055 and 2059 of
Similar to Equation (2), in consideration of an inverse square law indicating that the magnitude of a sound pressure is inversely proportional to the square of distance, the square of d is applied. The modified gain parameter gnew is applied to the extracted HRTF responses hL and hR and the modeled early reflection part rearly in the gain modification unit 2055. The modified gain parameter gnew is also applied to the late reverberation part r′late in the gain modification unit 2059. That is, the modified gain parameter gnew is identically applied to all room response signals, thereby preventing gain from being excessively generated. For example, since the value of gnew is decreased as the distance between the user and the sound source becomes long (Δd>0), a total gain value is decreased by the same ratio as illustrated in
Referring to
In accordance with the second embodiment of the present invention, the audio renderer 200 includes a distance information provider 701, a BRIR modifier 704, a tracking information provider 702, a BRIR selector 703, and a filter 705.
The distance information provider 701 provides information regarding the distance d between a user (listener) and a sound image. The tracking information provider 702 provides a localization point (θ, Φ) of the sound image to the BRIR selector 703. The tracking information provider 702 may be a head tracker for tracking movement of the user or the user may directly provide related information through the user control interface (300 in
The BRIR modifier 704 receives information (e.g. the distance information, room type information, and BRIR data BL and BR) from the distance information provider 701 and the BRIR selector 703 and generates output signals BL′ and BR′ according to a new distance rendering method. The output signals BL′ and BR′ of the BRIR modifier 704 may be obtained by extracting the BRIR data BL and BR corresponding to the sound localization point (θ, Φ) from the BRIR DB 7021 and then modifying the extracted BRIR data BL and BR with reference to information about a distance so as to maintain a correlation characteristic caused by change in distance
The audio renderer 200 according to the second embodiment of the present invention adjusts a correction value using the new distance rendering method so as to maintain a binaural correlation characteristic. Hereinafter, a detailed configuration of the audio renderer 200 and the distance rendering method according to the second embodiment of the present invention will be described with reference to
Referring to
The distance information provider 701 provides information about the distance d between a user and a sound image and a distance variation Δd. A distance determiner 7041 determines whether the location of the user is changed, using the information about the distance d and the distance variation Δd provided by the distance information provider 701. That is, if it is determined that a relative location of the user has been changed, the distance determiner 7041 controls the switching means 7042 to select an operation according to a path “y”. If it is determined that the relative location of the user has not been changed, the distance determiner 7041 controls the switching means 7042 to select an operation according to a path “n”.
First, the operation according to the path “n” through the switching means 7042 will now be described. That is, the operation is performed in the case in which the distance variation Δd provided by the distance information provider 7014 is 0, i.e., when the relative location of the user is not changed. In this case, a switch flag of the switching means 7042 is selected as the path “n”. The filter 705 performs filtering on the BRIR data bL and bR provided by the BRIR selector 703 together with an input audio signal s and outputs rendered audio signals OutputL and OutputR.
Next, the operation according to the path “y” through the switching means 7042 will now be described. That is, the operation is performed in the case in which the distance variation Δd provided by the distance information provider 701 is not 0 (i.e., Δd≠0), i.e., when the relative location of the user is changed. In this case, the switch flag of the switching means 7042 is selected as the path “y”.
If the user moves (i.e., Δd≠0) in an arbitrary space, a late reverberation extraction unit 7043 extracts late reverberation parts bL,late and bR,late of a BRIR. Next, two processes of using interaural cross correlation (IACC) information are performed. First, an IACC at a changed location, i.e., IACCLate_Δd, is estimated. Second, a response of the late reverberation part is generated so as to satisfy IACCLate_Δd.
In the first process, an IACCLate calculation unit 7044 calculates an IACC of the extracted responses bL,late and bR,late, i.e., IACCLate. An IACCLate_Δd calculation unit 7045 estimates an IACC at a changed location, i.e., IACCLate_Δd, with reference to the existing information about the distance d and the distance variation Δd.
In the second process, an adder 7043a adds the two extracted responses bL,late and bR,late to output one response bLR,late. Next, a transformer 7046 generates new responses bL′,late and bR′,late so that the response signals bL′,late and bR′,late have the same correlation with IACCLate_Δd estimated in the first process.
Next, a BRIR modification unit 7047 outputs new BRIR responses bL′ and bR′ by adjusting a gain of the generated responses bL′,late and bR′,late in proportion to a distance variation of a user. Lastly, the filter 705 performs filtering on the input signal s and the BRIR signals bL′ and bR′ and outputs the rendered audio output signals OutputL and OutputR.
Meanwhile, if it is determined that there is no change in distance in step S602, the next process (path “n”) is performed. A BRIR is extracted (modeled) from the BRIR DB with reference to the head tracking information (S603). Next, the extracted BRIR is filtered together with the decoded signal output in step S501 (S700). As a result of filtering, the rendered audio signal is output (S800).
A process from steps S605 to S610 in
If a user moves (Δd≠0), the late reverberation extraction unit 7043 extracts only a late reverberation part of a BRIR. There are many methods of extracting the late reverberation part of the BRIR. Generally, since components after about 50 ms to 80 ms of a response are regarded as the late reverberation part, in the present invention, only response components after about 80 ms are extracted as indicated by Equation (4).
Next, IACCLate is calculated using the extracted responses bL,late and bR,late. IACC represents a correlation between two sound pressures measured at both ears and is calculated as indicated by Equation (5).
In Equation (5), pL and pR denote sound pressures measured at the left ear and the right ear, respectively. In the second embodiment of the present, pL and pR are regarded as a BRIR (bL, bR). As two response characteristics measured at both ears become similar, the IACC approximates to 1 and, if the two response characteristics have no relation, the IACC approximates to 0. In other words, the IACC differs according to the distance between a user and a sound image. If the distance between the user and the sound image is short, the IACC is calculated as a value approximating to 1 and, if the distance between the user and the sound image is long, the IACC is calculated as a value approximating to 0. This is because, as the relative distance becomes longer, a spatial reverberation part rather than a direct part of the sound image is greatly influenced. If a gain of a direct part of a BRIR is changed to be small and therefore a D/R ratio is reduced, the user may feel as though the sound image is becoming distant. However, since IACCLate is not changed, a sound image characteristic recognized at both ears is not changed. Therefore, a response should be changed such that the sound image characteristic may differ according to change in location of the user. The second embodiment of the present invention performs the following process in order to change a response characteristic according to change in location of the user.
IACCLate_Δd is estimated from IACCLate in proportion to a moved distance and is indicated by Equation (6).
Referring to
Next, in the transformer 7046, an FFT unit 7046a transforms the response bLR.late into a frequency-axis signal to be divided into a magnitude component |bLRlate| and a phase component ∠bLR.late. A random number generator 7046b generates a random number ∠brand in the range of [−π, π]. However, since the random number ∠brand is uncorrelated with ∠bLR.late, ∠bLR.late is multiplied with IACCLate_Δd and the multiplied result is added to ∠brand to generate ∠b′rand (i.e., ∠b′rand=(IACCLate_Δd*∠bLR.late)+∠brand) in an adder 7046c. As a result, a correlation of ∠bLR.late and ∠b′rand is identical to IACCLate_Δd. Therefore, an IFFT unit 7046d synthesizes |bLR.late| and ∠bLR.late and transforms the synthesized result into a time-axis signal bL′,late (=bLR,late). At the same time, an IFFT unit 7046e synthesizes |bLR.late| and ∠b′rand and transforms the synthesized result into a time-axis signal bR′,late. Then a correlation of the two signals bL′,late and bR′,late is identical to IACCLate_Δd.
The two new response signals bL′,late and bR′,late output by the transformer 7046 are input to the BRIR modification unit 7047 and are replaced with a late reverberation part of a BRIR as indicated by Equation (7).
Additionally, in order to reflect characteristics of a D/R ratio in the modified BRIR, a gain of a direct part of the BRIR in proportion to Δd is changed as indicated by Equation (8).
In Equation (8), d denotes the distance between a user and a sound image and the square of d and the square of (d+Δd) are applied in consideration of an inverse square law indicating that the magnitude of sound pressure is inversely proportional to the square of the distance. gch,d denotes a gain value of a direct part and may be obtained by searching for a maximum value of bch(n) of the BRIR and nd denotes a timing at which the BRIR has a maximum value. Accordingly, finally modified responses of the BRIR modification unit 7047, i.e., output responses bL′ and bR′, may be obtained by replacing the components of a late reverberation part with reference to Equations (7), (8), and (9) and modifying a gain of the direct part.
A distance rendering method and apparatus in accordance with an embodiment of the present invention have the following effects.
First, a room response characteristic can be maintained in an arbitrary space and, therefore, more realistic audio can be provided.
Second, since a room response characteristic can be maintained even in a VR space, a natural and realistic effect can be provided even in any virtual space by configuring a VR space characteristic.
Third, since a binaural correlation is also adjusted, more realistic audio can be provided during configuration of a binaural room impulse response (BRIR).
The present invention may be implemented as computer-readable code that can be written on a computer-readable medium in which a program is recorded. The computer-readable medium may be any type of recording device in which data that can be read by a computer system is stored. Examples of the computer-readable medium include a hard disk drive (HDD), a solid state drive (SSD), a silicon disk drive (SDD), a read only memory (ROM), a random access memory (RAM), a compact disc (CD)-ROM, a magnetic tape, a floppy disk, an optical data storage, and a carrier wave (e.g., data transmission over the Internet). The computer may include an audio decoder and a renderer. It will be apparent to those skilled in the art that various modifications and variations can be made in the present invention without departing from the spirit or scope of the invention. Thus, the present invention is intended to cover the modifications and variations of this invention provided they come within the scope of the appended claims and their equivalents.
Pursuant to 35 U.S.C. § 119(e), this application claims the benefit of U.S. provisional applications Nos. 62/393,659 filed on Sep. 13, 2016 and 62/418,807 filed on Nov. 8, 2016, which are hereby incorporated by reference as if fully set forth herein.
Number | Date | Country | |
---|---|---|---|
62393659 | Sep 2016 | US | |
62418807 | Nov 2016 | US |