This application is a National Stage application under 35 U.S.C. § 371 of International Application No. PCT/JP2019/007754, having an International Filing Date of Feb. 28, 2019, which claims priority to Japanese Application Serial No. 2018-036186, filed on Mar. 1, 2018. The disclosure of the prior application is considered part of the disclosure of this application, and is incorporated in its entirety into this application.
The present invention relates to an acoustic signal processing device, an acoustic signal processing method, and an acoustic signal processing program for converting an input acoustic signal into output acoustic signals for a plurality of speakers in a speaker array formed by arranging the speakers for creating a virtual sound source.
In public viewings and concerts, voice, music, and the like are reproduced from a plurality of speakers installed at the screening site. In recent years, efforts have been made to implement acoustic reproduction with a more live feeling than ever by creating a virtual sound source in the screening space. For example, a high live feeling is achieved in particular by using a speaker array formed by linearly arranging a number of speakers to generate a virtual sound source that protrudes forward of the speakers and is closer to the audience.
Also, generally, the power of sound or voice emitted from a musical instrument or a human body differs from one direction to another. Thus, by reproducing the direction-specific difference (directivity) in the power of an acoustic signal when a virtual sound source is generated in a screening space, an acoustic content with an even higher live feeling can be expected to be created.
There is a technique called wave field reconstruction (Patent document 1) as opposed to the acoustic reproduction technique that creates a virtual sound source in a screening space. In the method based on Patent document 1, acoustic signals at an acoustic signal recording point are recorded with microphones installed at a plurality of points. Then, the incoming directions of the top, bottom, left, and right acoustic signals are analyzed, and a plurality of speakers installed in the screening space are used to physically reconstruct the acoustic signals in the recording site.
There is a technique which assumes a suction-type sound source (acoustic sink) as a virtual sound source to be implemented, and applies a drive signal derived from the first Rayleigh integral to a speaker array to generate a virtual sound image forward the speakers (Non-patent document 1). There is also a technique that can implement primitive directivity such as a dipole with a virtual sound source to be generated in a screening space using a linear speaker array (Non-patent document 2).
There is a multipole sound source as means for controlling the directivity of sound emitted from speakers (Non-patent document 3). A multipole sound source is means for expressing the directivity of sound with a combination of primitive directivities such as a dipole or a quadrupole, and each primitive directivity is implemented by combining non-directional point sound sources (monopole sound sources) that are close in distance to each other and have different polarities. Non-patent document 3 discloses that primitive directivities with different intensities are superimposed to rotate the direction of directivity.
However, none of the documents mentions a technique to implement any directional characteristics via superposition of multipoles. Hence, with any of the documents, it is difficult to model the directivity of a sound source such as a musical instrument by using multipoles.
It is therefore an objective of the present invention to provide an acoustic signal processing device, an acoustic signal processing method, and an acoustic signal processing program that implement any directional characteristics by superimposing multipoles.
In order to solve the above problems, a first aspect of the present invention is related to an acoustic signal processing device for converting an input acoustic signal into output acoustic signals for a plurality of speakers in a speaker array formed by arranging the speakers for creating a virtual sound source. The first aspect of the present invention includes a focal point position determination unit that obtains a plurality of sets of initial focal point coordinates, coordinates of the virtual sound source, and a direction of directivity thereof, and for a pair of sets of initial focal point coordinates with different polarities among the plurality of sets of initial focal point coordinates, multiplies the sets of initial focal point coordinates by a rotation matrix based on the coordinates of the virtual sound source to thereby determine sets of focal point coordinates, the rotation matrix being specified from the direction of the directivity, a circular harmonic coefficient conversion unit that calculates weights to be applied to multipoles including the sets of focal point coordinates from a circular harmonic coefficient, a filter coefficient computation unit that, for each of the speakers in the speaker array, computes a weighted driving function to be applied to the speaker from the sets of focal point coordinates, polarities of the sets of focal point coordinates, and the weights to be applied to the multipoles, and a convolutional operation unit that, for each of the speakers in the speaker array, convolves the weighted driving function for the speaker into the input acoustic signal to output the output acoustic signal for the speaker.
The circular harmonic coefficient conversion unit may calculate the weight to be applied to the multipole with equation (1).
[Math. 1]
where
dm,n: the weight to be applied to a multipole pm,n,
m,n: orders of partial differentiations of an acoustic field in an x-axis direction and a y-axis direction,
Š(2)(m+n): the circular harmonic coefficient,
Hm+n(2)(k): a Hankel function of a second kind of (m+n)-th order, and
k: a wavenumber (k=ω/c).
The filter coefficient computation unit may calculate driving functions by respectively using the sets of focal point coordinates and compute the weighted driving function to be applied to the speaker from composite driving functions calculated respectively for the multipoles and the weights to be applied to the multipoles, the composite driving functions being calculated from the polarities of the sets of focal point coordinates forming the multipoles and the driving functions.
The filter coefficient computation unit may calculate each of the composite driving functions for the multipoles by adding together functions which are obtained respectively for the sets of focal point coordinates included in the multipole and in each of which the polarity of the set of focal point coordinates and the corresponding driving function are multiplied.
The filter coefficient computation unit may calculate the weighted driving function by multiplying the composite driving functions calculated for the multipoles by the weights to be applied to the multipoles and adding the multiplied composite driving functions together.
A second aspect of the present invention is related to an acoustic signal processing method for converting an input acoustic signal into output acoustic signals for a plurality of speakers in a speaker array formed by arraying the speakers for creating a virtual sound source. The second aspect of the present invention includes obtaining a plurality of sets of initial focal point coordinates, coordinates of the virtual sound source, and a direction of directivity thereof, for a pair of sets of initial focal point coordinates with different polarities among the plurality of sets of initial focal point coordinates, multiplying the sets of initial focal point coordinates by a rotation matrix based on the coordinates of the virtual sound source to thereby determine sets of focal point coordinates, the rotation matrix being specified from the direction of the directivity, calculating weights to be applied to multipoles including the sets of focal point coordinates from a circular harmonic coefficient, for each of the speakers in the speaker array, computing a weighted drive function to be applied to the speaker from the sets of focal point coordinates, polarities of the sets of focal point coordinates, and the weights to be applied to the multipoles, and for each of the speakers in the speaker array, convolving the weighted drive function for the speaker into the input acoustic signal to output the output acoustic signal for the speaker.
A third aspect of the present invention is related to a non-transitory computer readable medium having stored thereon an acoustic signal processing program that causes a computer to function as the acoustic signal processing device according to the first aspect.
According to the present invention, it is possible to provide an acoustic signal processing device, an acoustic signal processing method, and an acoustic signal processing program that implement any directional characteristics by superimposing multipoles.
Next, an embodiment of the present invention will be described with reference to the drawings. In the description of the following drawings, the same or similar parts are denoted by the same or similar references.
(Acoustic Signal Processing Device)
An acoustic signal processing device 1 according to an embodiment of the present invention will be described with reference to see
The acoustic signal processing device 1 according to the embodiment of the present invention uses a linear speaker array as illustrated in
In the embodiment of the present invention, in order to create the virtual sound source, two or more focal point sound sources with different polarities are generated at positions close to each other to create a multipole sound source. The focal point sound sources are a combination of omnidirectional point sound sources (monopole sound sources) with different polarities. In the embodiment of the present invention, a description will be given of a case where the focal point sound sources include two multipoles, and one of the multipoles is formed of a single monopole sound source while the other multipole is formed of two monopole sound sources with different polarities. However, the focal point sound sources are not limited to these.
In the embodiment of the present invention, a multipole M1 and a multipole M2 illustrated in
In order to create such a virtual sound source, the acoustic signal processing device 1 converts an input acoustic signal I into output acoustic signals O for the speakers in the linear speaker array.
As illustrated in
The memory 10 stores focal point data 11. In the focal point data 11, the coordinates of a plurality of focal points for creating the virtual sound source and the polarities of the focal points are associated with each other. In the embodiment of the present invention, the focal points stored in the focal point data 11 will be referred to as initial focal points, and the coordinates of the initial focal points will be referred to as initial focal point coordinates.
The focal point position determination unit 12 receives information on the position of the virtual sound source, information on the direction of its directivity, and information on target frequencies, and outputs the coordinates of a necessary number of focal points taking the directivity into account. The focal point position determination unit 12 obtains the plurality of sets of initial focal point coordinates and the coordinates and directivity of the virtual sound source. Then, for a pair of sets of initial focal point coordinates with different polarities among the plurality of sets of initial focal point coordinates, the focal point position determination unit 12 multiplies each set of initial focal point coordinates by a rotation matrix specified from the direction of the directivity based on the coordinates of the virtual sound source to thereby determine a set of focal point coordinates. The focal point position determination unit 12 multiplies the relative coordinates of each set of initial focal point coordinates relative to the coordinates of the virtual sound source by the rotation matrix, and adds the coordinates of the virtual sound source to the set of coordinates obtained by the multiplication by the rotation matrix to thereby determine a set of focal point coordinates taking the directivity into account. Note that the virtual sound source is in the center among these sets of focal point coordinates.
The focal point position determination unit 12 determines the sets of initial focal point coordinates among the plurality of sets of initial focal point coordinates that do not form a pair as sets of focal point coordinates without performing any conversion on these sets of initial focal point coordinates. In the example illustrated in
The focal point position determination unit 12 obtains one or more pairs of sets of initial focal point coordinates with difference polarities from the memory 10 and also obtains the coordinates of the virtual sound source and the direction of its directivity as the characteristics to be implemented by the acoustic signal processing device 1 in response to an external input or the like. The focal point position determination unit 12 specifies a direction θ of the rotation of the sets of initial focal point coordinates from the obtained direction of the directivity.
Let a pair of sets of initial focal point coordinates be
x1=(δ,0), and x2=(−δ,0). [Math. 1]
Then, if the direction θ is designated with respect to the X-axis direction, a rotation matrix G that can be specified from this direction can be figured out with equation (1). Hence, the focal point position determination unit 12 can determine the coordinates of the monopoles after rotation with equation (2).
For the one or more pairs of sets of initial focal point coordinates corresponding to the desired characteristics and read from the memory, the focal point position determination unit 12 multiplies each set of coordinates by the rotation matrix that can be specified from the direction of the directivity, and adds the coordinates of the virtual sound source to each set of coordinates to thereby calculate all sets of focal point coordinates.
The focal point position determination unit 12 outputs identifiers of the multipoles, the sets of focal point coordinates forming these multipoles, and the polarities of these sets of focal point coordinates in association with each other.
In the case of a multipole sound source formed of more than two monopole sound sources, such as a quadrupole sound source, the focal point position determination unit 12 calculates the additional sets of coordinates via rotation with a rotation matrix to calculate the monopole sound sources corresponding to the rotation of the directivity.
The focal point position determination process by the focal point position determination unit 12 according to the embodiment of the present invention will be described with reference to
First, in step S11, the focal point position determination unit 12 obtains information on the coordinates of the virtual sound source and the direction of its directivity. In step S12, the focal point position determination unit 12 reads information on one or more initial focal points corresponding to the desired characteristics from the memory.
Thereafter, the focal point position determination unit 12 iterates processes of steps S13 and S14 for each initial focal point read in step S12. In step S13, the focal point position determination unit 12 multiplies the target set of focal point coordinates to be processed by a rotation matrix specified from the direction of the directivity obtained in step S11. The target set of focal point coordinates used here is a set of relative coordinates relative to the virtual sound source. In step S14, the focal point position determination unit 12 adds the set of coordinates multiplied by the rotation matrix in step S13 to the coordinates of the virtual sound source to thereby determine a set of focal point coordinates taking the directivity into account.
The focal point position determination unit 12 terminates the process when the processes of steps S13 and S14 are finished for each initial focal point read in step S12.
Note that the processes of steps S13 and S14 only need to be performed on each focal point and may be performed in any order.
The result of a simulation of the process by the focal point position determination unit 12 will be described with reference to
The focal point position determination unit 12 multiplies each of these sets of initial focal point coordinates by the rotation matrix specified by equation (1). As illustrated in
After a set of focal point coordinates taking the directivity into account is calculated by the focal point position determination unit 12 for each initial focal point, the set of focal point coordinates is processed by the filter coefficient computation unit 14.
The circular harmonic coefficient conversion unit 13 calculates weights to be applied to the multipoles including the sets of focal point coordinates by using circular harmonic coefficients.
The circular harmonic coefficient conversion unit 13 analytically converts a circular harmonic series to determine the weights to be applied to the focal point sound sources, and enables creation of a virtual sound image having the directional characteristics of a sound source that exists in reality. The circular harmonic coefficient conversion unit 13 calculates the weights to be applied to the multipoles including the sets of focal point coordinates outputted by the focal point position determination unit 12.
The circular harmonic coefficient conversion unit 13 calculates the weights to be applied to the multipoles with equation (3).
In equation (3), m and n are the orders of partial differentiations of the acoustic field in the x-axis direction and the y-axis direction, respectively. Since combinations of m and n do not overlap, they may be used as mere indexes.
The circular harmonic coefficient conversion unit 13 obtains each circular harmonic coefficient as appropriate. For example, the circular harmonic coefficient may be received from an external program, or the circular harmonic coefficient may be obtained via observation with a plurality of microphones disposed in a circle centered on the sound source whose directivity is to be measured. Also, the circular harmonic coefficient may be stored beforehand in a separately provided memory and read out when necessary by the circular harmonic coefficient conversion unit 13.
Here, the derivation of equation (3) for outputting the weight for each multipole from the circular harmonic coefficient will be described. First, a sound source having any directivity is assumed to be present at the origin in the xy plane, and the acoustic field generated by this sound source is S(x). When this acoustic field is Taylor-expanded at the origin, the acoustic field at a point x=(cos α, sin α) in a unit circle is given as the following equation.
S(x): The acoustic field generated by the sound source having any directivity at the origin in the xy plane
x: A point in a unit circle and x=(cos α,sin α)
Meanwhile, any acoustic field can be expressed by equation (5) via circular harmonic expansion.
[Math. 5]
S(x,ω)=ΣW=−∞∞Š(2)(v,ω)Hv(2)(kr)ejvα, Equation (5)
ejvα: Complex sinusoidal wave
v: Order
ω: Angular frequency
Euler's formula is applied for the complex sinusoidal wave, and then binomial expansion is performed for v to perform transformation as the following equation.
Further, the coefficients in equations (4) and (6) are compared. As a result, a weight coefficient can be calculated as in equation (3).
The circular harmonic coefficient conversion process by the circular harmonic coefficient conversion unit 13 will be described with reference to
The circular harmonic coefficient conversion unit 13 performs a process of step S21 for each multipole outputted by the focal point position determination unit 12. In step S21, the circular harmonic coefficient conversion unit 13 calculates the weight for the multipole from the circular harmonic coefficient in accordance with equation (3).
For each speaker in the speaker array, the filter coefficient computation unit 14 computes a weighted driving function to be applied to the speaker from the sets of focal point coordinates, the polarities of the sets of focal point coordinates, and the weights to be applied to the multipoles. For each speaker in the linear speaker array, the filter coefficient computation unit 14 calculates a weighted driving function to be convolved into the input acoustic signal I from each set of focal point coordinates determined by the focal point position determination unit 12. The filter coefficient computation unit 14 calculates driving functions by respectively using the sets of focal point coordinates and computes a weighted driving function to be applied to the speaker from composite driving functions calculated respectively for the multipoles and the weights to be applied to the multipoles, the composite driving functions being calculated from the polarities of the sets of focal point coordinates forming the multipoles and the driving functions. Here, the filter coefficient computation unit 14 calculates each of the composite driving functions for the multipoles by adding together functions which are obtained respectively for the sets of focal point coordinates included in the multipole and in each of which the polarity of the set of focal point coordinates and the corresponding driving function are multiplied. Also, the filter coefficient computation unit 14 calculates the weighted driving function by multiplying the composite driving functions calculated for the multipoles by the weights to be applied to the multipoles and adding the multiplied composite driving functions together.
Firstly, when calculating a weighted driving function for a predetermined speaker, the filter coefficient computation unit 14 calculates a driving function for each focal point with equation (7).
The position of the virtual sound source: xs=(xs,ys)
The position of the i-th speaker: xi=(xi,yi)
Then, the filter coefficient computation unit 14 calculates a composite driving function for a predetermined multipole with equation (8) from the polarity of the focal point sound source belonging to this multipole and the driving function for each focal point calculated with equation (7).
[Math. 8]
Dm,n(x0)=Σi=0N−1gs(i)·D(x0,xs(1)), Equation (8))
xs(i)∈Xm,n: The coordinates of a focal point included in the multipole pm,n
gs(i)∈Gm,n: The polarity of the focal point xs(i)
N: The number of focal points included in the multipole pm,n
For each multipole, the filter coefficient computation unit 14 applies the weight calculated by the circular harmonic coefficient conversion unit 13 to the composite driving function calculated with equation (8), and calculates a weighted driving function with equation (9).
[Math. 9]
D(x0)=Σm,ndm,n·Dm,n(xD), Equation (9)
Next, the filter coefficient computation process by the filter coefficient computation unit 14 will be described with reference to
First, in step S31, the filter coefficient computation unit 14 obtains each set of focal point coordinates determined in the focal point position determination process. In doing so, the filter coefficient computation unit 14 additionally obtains the polarities of the focal points and the relationship between the sets of focal point coordinates forming the multipoles.
The filter coefficient computation unit 14 iterates processes of steps S32 to S37 to calculate a weighted driving function for each speaker. In step S32, the filter coefficient computation unit 14 initializes the weighted driving function for the target speaker with zero.
The filter coefficient computation unit 14 iterates the process of step S33 for each focal point. In step S33, the filter coefficient computation unit 14 calculates a driving function by using the coordinates of the target focal point. In the example illustrated in
The filter coefficient computation unit 14 iterates the processes of steps S34 to S36 for each multipole to thereby calculate a composite driving function for each multipole. In step S34, the filter coefficient computation unit 14 initializes the composite driving function for the processing target multipole.
The filter coefficient computation unit 14 performs the process of step S35 for each focal point included in the processing target multipole. In step S35, using the polarity of the target focal point, the filter coefficient computation unit 14 adds the drive function for the target focal point calculated in step S33 to the composite drive function. In the example illustrated in
In step S36, the filter coefficient computation unit 14 applies the weights calculated by the circular harmonic coefficient conversion unit 13 to the composite driving functions calculated in step S35 to calculate a weighted driving function. In the example illustrated in
In step S37, the filter coefficient computation unit 14 outputs the weighted driving function obtained after the calculation for each multipole as a weighted driving function to be applied to the target speaker.
After the filter coefficient computation unit 14 calculates a weighted driving function for each speaker in the linear speaker array, the convolutional operation unit 15 convolves the weighted driving function into the input acoustic signal I to thereby calculate the output acoustic signal O to be applied to the speaker.
For each speaker in the linear speaker array, the convolutional operation unit 15 convolves the weighted driving function for the speaker into the input acoustic signal I to output the output acoustic signal O for the speaker. For a predetermined speaker, the convolutional operation unit 15 obtains the output acoustic signal O for this speaker by convolving the weighted driving function for this speaker into the input acoustic signal I. The convolutional operation unit 15 iterates similar processes for each speaker to obtain the output acoustic signal O for the speaker.
The convolutional computation process by the convolutional operation unit 15 will be described with reference to
The convolutional operation unit 15 iterates processes of steps S41 and S42 for each speaker in the linear speaker array. In step S41, the convolutional operation unit 15 obtains the weighted drive function for the target speaker to be processed from the filter coefficient computation unit 14. In step S42, the convolutional operation unit 15 convolves the weighted drive function obtained in step S41 into the input acoustic signal I to obtain the output acoustic signal O.
The convolutional operation unit 15 terminates the process when the processes of steps S41 and S42 are finished for each speaker. Note that the processes of steps S41 and S42 only need to be performed on each speaker and may be performed in any order.
The acoustic signal processing device 1 according to the embodiment of the present invention rotates sets of initial focal point coordinates to calculate sets of focal point coordinates for implementing desired directivity in advance and, for these sets of focal point coordinates, calculates a weighted driving function corresponding to each speaker. The acoustic signal processing device 1 convolves the weighted driving function corresponding to each speaker into the input acoustic signal I to thereby obtain the output acoustic signal O for the speaker. This weighted driving function is given weights converted from circular harmonic coefficients for respective multipoles. Thus, by setting each circular harmonic coefficient as appropriate, the output acoustic signal O for each speaker can be adjusted as desired. As described above, the acoustic signal processing device 1 according to the embodiment of the present invention is capable of modeling the directivity of a sound source such as a musical instrument and implementing any directional characteristics by superimposing multipoles.
As described above, a description has been by using the embodiment of the present invention. However, it should not be understood that the description and drawings which constitute part of this disclosure limit the invention. From this disclosure, various alternative embodiments, examples, and operation techniques will be easily found by those skilled in the art.
The present invention naturally includes various embodiments which are not described herein. Accordingly, the technical scope of the present invention should be determined only by the matters to define the invention in the scope of claims regarded as appropriate based on the description.
Number | Date | Country | Kind |
---|---|---|---|
JP2018-036186 | Mar 2018 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2019/007754 | 2/28/2019 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2019/168083 | 9/6/2019 | WO | A |
Number | Date | Country |
---|---|---|
2011-244306 | Dec 2011 | JP |
2012169895 | Sep 2012 | JP |
Entry |
---|
Haneda, Yoichi, et al. “Design of Multipole Loudspeaker Array Based on Spherical Harmonic Expansion.” ICASSP 2011, 2011, pp. 141-144. (Year: 2011). |
Choi, Jung-Woo, and Yang-Hahn Kim. “Sound Field Reproduction of a Virtual Source Inside a Loudspeaker Array with Minimal External Radiation.” IEEE, vol. 21, No. 2, Feb. 2013, pp. 247-259. (Year: 2013). |
Ahrens et al., “Implementation of Directional Sources in Wave Field Synthesis,” 2007 IEEE Workshop on Application of Signal Processing to Audio and Acoustics, Oct. 2007, pp. 66-69. |
Haneda et al., “Directivity Synthesis Using Multipole Sources Based on Spherical Harmonic Expansion,” The Journal of the Acoustical Society of Japan, Nov. 2013, 69(11):577-588, English Abstract. |
Spors et al., “Physical and Perceptual Properties of Focused Sources in Wave Field Synthesis,” Audio Engineering Society: Convention Paper, Oct. 2009, 19 pages. |
Tsutsumi et al., “Directivity Synthesis with Multipoles Comprising a Cluster of Focused Sources Using a Linear Loudspeaker Array,” IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 2018, pp. 496-500. |
Number | Date | Country | |
---|---|---|---|
20210006892 A1 | Jan 2021 | US |