Priority is claimed on Japanese Patent Application No. 2012-203393, filed on Sep. 14, 2012, the contents of which are entirely incorporated herein by reference.
The present invention relates to a sound direction estimation device, a sound direction estimation method, and a sound direction estimation program.
Techniques of recognizing uttered speech from sound signals generated from sound sources have been proposed. In such speech recognition, noise signals are separated from the sound signals or the noise signals are suppressed, and a target sound signal to be recognized is extracted from the sound signals. For example, speech recognition is performed on the extracted sound signal. In such a system, in order to extract speech to be recognized, the directions in which the sound signals are generated should be known or the directions in which the sound signals are generated should be estimated.
For example, in Japanese Unexamined Patent Application, First Publication No. 2012-42465 (Patent Document 1), a type of sound source of an input sound signal is identified based on a sound feature value of the sound signal and a sound direction of the identified type of sound source of the sound signal is estimated. In Patent Document 1, a GEVD (Generalized Eigenvalue Decomposition)-MUSIC method or a GSVD (Generalized Singular Value Decomposition)-MUSIC method is used to estimate information of the sound direction (hereinafter, also referred to as sound source localization). In this way, when the GEVD-MUSIC or the GSVD-MUSIC method is used, computation efficiency of sound source localization in a sound source estimating device is enhanced.
However, in such a sound source estimating device, transfer functions correlated with sound directions are acquired in advance through measurement or calculation in a range in which the sound directions are searched for and are stored in the device. In such a sound source estimating device, spatial spectrums are calculated using the transfer functions stored in the storage unit and sound directions are calculated based on the calculated spatial spectrums. Accordingly, in order to enhance estimation accuracy of a sound direction, transfer functions correlated with multiple sound directions are necessary. Therefore, in the sound source estimating device according to the related art, there is a problem in that a large computational load is required for enhancing the estimation accuracy of a sound direction and thus computation efficiency is low.
The present invention is made in consideration of the above-mentioned circumstances, and an object thereof is to provide a sound direction estimation device, a sound direction estimation method, and a sound direction estimation program which can enhance processing efficiency in estimation of a sound direction.
(1) In order to achieve the above-mentioned object, according to an aspect of the present invention, a sound direction estimation device is provided including: a transfer function storage unit configured to store transfer functions of sound sources in correlation with directions of the sound sources; a calculation unit configured to calculate the number of classes to be searched and a search interval for each class based on a desired search range and a desired spatial resolution for searching for the directions of the sound sources; and a sound source localization unit configured to search the search range for every search interval using the transfer function, to estimate the direction of the sound source based on the search result, to update the search range and the search interval based on the estimated direction of the sound source until the number of classes calculated by the calculation unit is reached, and to estimate the direction of the sound source.
(2) According to another aspect of the present invention, in the sound direction estimation device, the sound source localization unit may be configured to search an n-th class (where n is an integer greater than or equal to 1) in the predetermined search range with the calculated search interval and update at least one search interval in the search range to the search range of an (n+1)-th class based on the search result, to update the search interval of the (n+1)-th class to be searched based on the updated search range of the (n+1)-th class and the desired spatial resolution, and to update and estimate the directions of the sound sources using the updated search range of the (n+1)-th class, the updated search interval of the (n+1)-th class, and the transfer function corresponding to an azimuth of the search interval until the number of classes (n+1) reaches the number of classes calculated by the calculation unit.
(3) According to another aspect of the present invention, in the sound direction estimation device, the calculation unit may be configured to calculate the number of classes and the search interval so that a search number for each class is constant in all the classes.
(4) According to another aspect of the present invention, in the sound direction estimation device, the calculation unit may be configured to calculate the number of classes and the search interval so as to minimize the total search number in all the classes.
(5) According to another aspect of the present invention, in the sound direction estimation device, the sound source localization unit may be configured to determine whether the transfer function corresponding to the azimuth of the search interval is stored in the transfer function storage unit, to read the transfer function corresponding to the azimuth of the search interval from the transfer function storage unit when it is determined that the transfer function corresponding to the azimuth of the search interval is stored in the transfer function storage unit, to calculate an interpolated transfer function by interpolating the transfer function corresponding to the azimuth of the search interval when it is determined that the transfer function corresponding to the azimuth of the search interval is not stored in the transfer function storage unit, and to estimate the directions of the sound sources using the read transfer function or the calculated interpolated transfer function.
(6) According to another aspect of the present invention, in the sound direction estimation device, the calculation unit may be configured to calculate the number of classes and the search interval so as to minimize a calculation cost which is a sum of a search cost spent in calculating the search number in the search range and an interpolation cost spent in the interpolation.
(7) According to another aspect of the present invention, a sound direction estimation method in a sound direction estimation device is provided, including: calculating the number of classes to be searched and a search interval for each class based on a desired search range and a desired spatial resolution for searching for a direction of a sound source; searching the search range for every search interval using a transfer function from the sound source stored in a transfer function storage unit in correlation with the direction of the sound source and estimating the direction of the sound source based on the search result; and updating the search range and the search interval based on the estimated direction of the sound source until the calculated number of classes calculation unit is reached and estimating the direction of the sound source.
(8) According to another aspect of the present invention, a sound direction estimation program is provided causing a computer of a sound direction estimation device to perform: calculating the number of classes to be searched and a search interval for each class based on a desired search range and a desired spatial resolution for searching for a direction of a sound source; searching the search range for every search interval using a transfer function from the sound source stored in a transfer function storage unit in correlation with the direction of the sound source and estimating the direction of the sound source based on the search result; and updating the search range and the search interval based on the estimated direction of the sound source until the calculated number of classes calculation unit is reached and estimating the direction of the sound source.
According to Aspects (1), (2), (7), and (8) of the present invention, since the calculation unit calculates the number of classes to be searched and the search interval for each class based on the desired search range and the desired spatial resolution used to search for directions of sound sources, it is possible to reduce the processing time to be spent in estimating the directions of sound sources.
According to Aspects (3) and (4) of the present invention, since the calculation unit can calculate the number of classes to be searched and the search interval for each class with a small computational load, it is possible to reduce the processing time to be spent in estimating the directions of sound sources.
According to Aspect (5) of the present invention, since the sound source localization unit estimates the directions of the sound sources using an interpolated transfer function obtained by interpolating a transfer function when the transfer function corresponding to the azimuth to be searched for is not stored in the transfer function storage unit and using a read transfer function when the transfer function corresponding to the azimuth to be searched for is stored in the transfer function storage unit, it is possible to accurately estimate the directions of the sound sources.
According to Aspect (6) of the present invention, since the sound source localization unit calculates the number of classes and the search interval so that the calculation cost which is the sum of the search cost spent in calculating the search number and the interpolation cost spent in the interpolation, it is possible to reduce the processing time to be spent in estimating the directions of sound sources.
Hereinafter, embodiments of the present invention will be described in detail. The present invention is not limited to the embodiments and can be modified in various forms without departing from the technical spirit thereof.
First, a summary of this embodiment will be described below.
The sound source localization unit 12 includes a sound signal input unit and performs, for example, Fourier transform on sound signals collected by multiple microphones. The sound source localization unit 12 estimates a sound direction in multiple sound signals subjected to Fourier transform (hereinafter, referred to as “sound source localization”). The sound source localization unit 12 outputs information indicating the result of sound source localization to the sound source separation unit 13.
The sound source separation unit 13 separates a target sound and noise based on the information indicating the result of sound source localization which is input from the sound source localization unit 12. The sound source separation unit 13 outputs signals corresponding to the separated sound sources to the acoustic feature extraction unit 14. The target sound is, for example, speeches uttered by multiple speakers. The noise is sounds such as wind noise and a sound generated from equipment located in the sound-collected room.
The acoustic feature extraction unit 14 extracts acoustic features of the signals corresponding to the sound sources input from the sound source separation unit 13 and outputs information indicating the extracted acoustic features to the automatic speech recognition unit 15.
When speech uttered by a person is included in the sources, the automatic speech recognition unit 15 recognizes the speech based on the acoustic features input from the acoustic feature extraction unit 14 and outputs the recognition result to the recognition result output unit 16.
The recognition result output unit 16 is, for example, a display or a sound signal output device. The recognition result output unit 16 displays information based on the recognition result input from the automatic speech recognition unit 15, for example, on a display unit.
The sound processing system 1 or the sound direction estimation unit 11 may be mounted on, for example, a robot, a vehicle, an aircraft (including a helicopter), and a mobile terminal. Examples of the mobile terminal include a mobile phone terminal, a portable information terminal, and a portable game terminal.
In this embodiment, a sound direction is hierarchically estimated to reduce calculation costs while enhancing a spatial resolution of a sound source to improve estimation accuracy of a sound direction. In hierarchical estimation of a sound source, the sound direction estimation unit 11 (
As shown in
The sound input unit 101 includes M (where M is an integer equal to or more than 2) sound input means (for example, microphones) and the sound input means are arranged at different positions. The sound input unit 101 is, for example, a microphone array including M microphones. The sound input unit 101 outputs sound waves received by the sound input means as a sound signal of each channel to the short-time Fourier transform unit 102. The sound input unit 101 may convert a sound signal from an analog signal to a digital signal and may output the sound signal converted into the digital signal to the short-time Fourier transform unit 102.
The short-time Fourier transform unit 102 performs short-time Fourier transform (STFT) on the sound signal of each channel input from the sound input unit 101 for each frame in the time domain to generate an input signal in the frequency domain. The short-time Fourier transform is an operation of performing a Fourier transform by multiplying a function by a window function while shifting the window function. The frame is a time interval of a predetermined length (frame length) or a signal included in the time interval. The frame length is, for example, 10 [msec].
The short-time Fourier transform unit 102 generates a matrix X(ω, f) for each frequency ω and each frame time f using the input signal subjected to the short-time Fourier transform for each frame, and outputs the generated input vector X(ω, f) of M columns to the first correlation matrix calculation unit 103.
The first correlation matrix calculation unit 103 calculates a spatial correlation matrix R(ω, f) for each frequency ω and each frame time f using the input vector X(ω, f) input from the short-time Fourier transform unit 102 and using Expression (1). The first correlation matrix calculation unit 103 outputs the calculated spatial correlation matrix R(ω, f) to the matrix calculation unit 106. The spatial correlation matrix R(ω, f) is a square matrix of M rows and M columns.
In Expression (1), f represents a current frame time and TR represents a length of a section (number of frames) used to calculate the spatial correlation matrix R(ω, f). This length of a section is called a window length. τ is a parameter indicating a frame time (not limited to the current frame time) and has a value in a range of 0 to TR−1. In addition, * represents a complex conjugate transposition operator of a vector or a matrix. In Expression (1), the spatial correlation matrix R(ω, f) is smoothed with a TR frame so as to improve robustness to noise. That is, Expression (1) is an inter-channel correlation between channel n (where n is an integer greater than or equal to 1) and channel m (where m is an integer greater than or equal to 1 and different from n) and is a value obtained by averaging products of an input signal vector of channel n and a complex conjugate of an input signal vector of channel k over a section with a window length TR−1 from the current frame time 0.
A matrix N(ω, f) of a noise source (hereinafter, referred to as a noise matrix) for each frequency ω and each frame time f is stored in advance in the noise database 104. Noise is, for example, window noise or a sound generated from equipment located in the sound-collected room other than a target sound.
The second correlation matrix calculation unit 105 reads a noise matrix N(ω, f) stored in the noise database 104 and calculates a correlation matrix K(ω, f) of noise (hereinafter, referred to as noise correlation matrix) for each frequency ω and each frame time f using the read N(ω, f) and Expression (2). The noise correlation matrix K(ω, f) is a noise correlation matrix calculated from a signal to be suppressed (whitened) at the time of localization, but is a unit matrix for the purpose of simplification in this embodiment. The second correlation matrix calculation unit 105 outputs the calculated noise correlation matrix K(ω, f) to the matrix calculation unit 106.
In Expression (2), Tk represents a length of a section (number of frames) used to calculate the noise correlation matrix K(ω, f). τk is a parameter indicating a frame time (not limited to a current frame time) and has a value in a range of 0 to Tk−1. N* represents a complex conjugate transposition operator of a matrix N. As shown in Expression (2), the noise correlation matrix K(ω, f) is an inter-channel correlation between a noise signal of channel n and a noise signal of channel m and is a value obtained by averaging products of the noise matrix N of channel n and a complex conjugate of the noise matrix N of channel k over a section with a window length Tk−1 from the current frame time 0.
The matrix calculation unit 106 calculates eigenvectors for each frequency ω and each frame time f in a space in which the spatial correlation matrix R(ω, f) input from the first correlation matrix calculation unit 103 extends using the noise correlation matrix K(ω, f) input from the second correlation matrix calculation unit 105. The matrix calculation unit 106 outputs the calculated eigenvectors to the class number calculation unit 107.
The matrix calculation unit 106 performs a computing operation on a matrix K−1(ω, f)R(ω, f), which is obtained by multiplying an inverse matrix K−1(ω, f) of the noise correlation matrix K(ω, f) by the spatial correlation matrix R(ω, f) from the left side, using a generalized singular-value decomposition (GSVD)-MUSIC method expressed by Expression (3). The matrix calculation unit 106 calculates a vector El(ω, f) and an eigenvalue matrix Λ(ω, f) satisfying the relationship of Expression (3) using the GSVD-MUSIC method. The matrix calculation unit 106 decomposes a space into a partial space of a target sound and a partial space of noise through this process. In this embodiment, it is possible to whiten noise using Expression (3).
K−1(ω,f)R(ω,f)=El(ω,f)Λ(ω,f)Er*(ω,f) (3)
In Expression (3), the vector El(ω, f) is the left-singular vector and the vector Er*(ω, f) is the complex conjugate of the right-singular vector. The vector E(ω, f) is a vector having eigenvectors e1(ω, f), . . . , eM(ω, f) as element values. The eigenvectors e1(ω, f), . . . , eM(ω, f) are eigenvectors corresponding to eigenvalues λ1(ω, f), . . . , λM(ω, f). M represents the number of microphones.
The eigenvalue matrix Λ(ω, f) is expressed by Expression (4). In Expression (4), “diag” represents a diagonal matrix.
Λ(ω,f)=diag(λ1(ω,f), . . . ,λM(ω,f)) (4)
As described later, the class number calculation unit 107 calculates the number of classes in which the first spatial spectrum calculation unit 108 to the STF processing unit 111 perform estimation and a search interval used for the search, and outputs the calculated number of classes and the calculated search interval and the eigenvectors input from the matrix calculation unit 106 to the first spatial spectrum calculation unit 108. This search interval corresponds to a spatial resolution.
The first spatial spectrum calculation unit 108 calculates a spatial spectrum P(ω, ψ, f) before the frequency ω is integrated for each frequency ω and each sound direction ψ for every frame f using the number of classes and the search interval input from the class number calculation unit 107, the eigenvectors, and transfer function A or the interpolated transfer function A^ input from the STF processing unit 111 and using Expression (5). The first spatial spectrum calculation unit 108 performs the computing operation using the interpolated transfer function A^ instead of the transfer function A of Expression (5), when the interpolated transfer function A^ is input from the STF processing unit 111.
The first spatial spectrum calculation unit 108 the calculated spatial spectrum P(ω, ψ, f) and the number of classes input from the class number calculation unit 107 to the second spatial spectrum calculation unit 109.
In Expression (5), | . . . | represents the absolute value. In Expression (5), the spatial spectrum represents a ratio of a component based on noise to the entire transfer function. In Expression (5), em(ω, f) represents the left-singular vector El(ω, f) (=e1(ω, f), . . . eM(ω, f)) of Expression (3). A* represents the complex conjugate of a transfer function A. LS represents the number of sound sources and is an integer equal to or more than 0. A(ω, ψ) represents a known transfer function stored in advance in the transfer function storage unit 112 and is expressed by Expression (6).
A(ω,ψi)=[(Ai(ω,ψi), . . . ,AM(ω,ψi))]T (6)
Here, ψi represents a previously-measured sound source azimuth, that is, the direction of the transfer function A. In addition, i is an integer greater than or equal to 1. M represents the number of microphones. T represents a transpose matrix.
The second spatial spectrum calculation unit 109 calculates an average spatial spectrum P(ψ, f) by averaging the spatial spectrum P(ω, ψ, f) input from the first spatial spectrum calculation unit 108 in the co direction using Expression (7). The second spatial spectrum calculation unit 109 outputs the calculated average spatial spectrum P(ψ, f) and the number of classes and the search interval input from the first spatial spectrum calculation unit 108 to the peak search unit 110.
In Expression (7), ω[k] represents the frequency corresponding to the k-th frequency bin. The frequency bin is discrete frequency. kh and kl represent indices of the frequency bins corresponding to the maximum frequency (upper limit frequency) and the minimum frequency (lower limit frequency) in the frequency domain. In Expression (7), kh−kl+1 is the number of spatial spectrums P(ω, ψ, f) to be subjected to summation (Σ). The reason for addition of 1 in this way is that since the frequencies ω are discrete, both the spatial spectrum P(ω[k], ψ, f) of the upper frequency and the spatial spectrum P(ω[k], ψ, f) of the lower limit frequency which are both ends of the frequency band should be added.
The average spatial spectrum P(ψ, f), the number of classes, and the search interval are input to the peak search unit 110 from the second spatial spectrum calculation unit 109. The peak search unit 110 detects an azimuth ψ[l] (where l is a value in a range of 1 to Ls) which is the peak value of the input average spatial spectrum P(ψ, f).
The peak search unit 110 determines whether the estimation process in the first spatial spectrum calculation unit 108 to the STF processing unit 111 ends by the number of classes input from the second spatial spectrum calculation unit 109. When it is determined that the estimation process ends by the number of classes, the peak search unit 110 outputs the detected peak value ψ[l] as the estimated azimuth to the output unit 113. When it is determined that the estimation process does not end by the number of classes, the peak search unit 110 outputs the detected peak value ψ[l], the number of classes, and the search interval to the STF processing unit 111.
The STF processing unit 111 reads a transfer function A from the transfer function storage unit 112 using the peak value ψ[l], the number of classes, and the search interval input from the peak search unit 110 or calculates an interpolated transfer function A^. Specifically, the STF processing unit determines whether the transfer function A corresponding to an azimuth to be searched for is stored in the transfer function storage unit 112. When it is determined that the transfer function A corresponding to an azimuth to be searched for is stored in the transfer function storage unit 112, the STF processing unit 111 reads the corresponding transfer function A from the transfer function storage unit 112. When it is determined that the transfer function A corresponding to an azimuth to be searched for is not stored in the transfer function storage unit 112, the STF processing unit 111 calculates the corresponding interpolated transfer function A^. The STF processing unit 111 outputs the read transfer function A or the calculated interpolated transfer function A^ to the first spatial spectrum calculation unit 108.
When a predetermined search range (also referred to as a class) is completely searched, the STF processing unit 111 calculates a new search range and a new search interval based on the azimuth in which the peak value is detected as described later. The STF processing unit 111 outputs the calculated search range and the calculated search interval to the first spatial spectrum calculation unit 108.
The output unit 113 outputs the estimated azimuth ψ input from the peak search unit 110, for example, to the sound source separation unit 13 (see
As described above, the sound direction estimation device according to this embodiment includes a transfer function storage unit (112) configured to store transfer functions of sound sources in correlation with the directions of the sound sources, a calculation unit (class number calculation unit 107) configured to calculate the number of classes to be searched and a search interval for each class based on a desired search range and a desired spatial resolution for searching for the directions of the sound sources, and a sound source localization unit (peak search unit 110 and STF processing unit 111) configured to search the search range for every search interval using the transfer function, to estimate the direction of the sound source based on the search result, to update the search range and the search interval based on the estimated direction of the sound source until the number of classes calculated by the calculation unit is reached, and to estimate the direction of the sound source.
With this configuration according to this embodiment, the number of classes to be searched by the class number calculation unit 107 and the search interval are calculated. Then, the sound source localization unit 12 (
The first cost calculation unit 121 calculates a cost function JGC(W) using a sound source separation (GSS (Geometric Source Separation)) method based on constraint conditions and outputs the calculated cost function JGC(W) to the sound source separation processing unit 123.
The first cost calculation unit 121 realizes geometric constraint by designating a transfer function for D of the cost function JGC(W) expressed by Expression (8). The cost function JGC(W) is a degree of geometric constraint and is used to calculate a separation matrix W.
JGC(W)=∥EGC∥2 (8)
In Expression (8), EGC is expressed by Expression (9).
EGC=diag[WD−I] (9)
In Expression (9), W represents a separation matrix and D represents a transfer function matrix. In this embodiment, by using the interpolated transfer function A^ interpolated by the STF processing unit 111 or the read transfer function A for the transfer function matrix D, it is possible to apply geometric constraint in a correct sound direction.
The second cost calculation unit 122 calculates a cost function JHDSS(W) using a high-order decorrelation-based sound separation (HDSS) method which is an extension of independent component analysis, and outputs the calculated cost function JHDSS(W) to the sound source separation processing unit 123.
The sound source separation processing unit 123 calculates a cost function JGHDSS(W) as expressed by Expression (10) using the cost function JGC(W) input from the first cost calculation unit 121 and the cost function JHDSS(W) input from the second cost calculation unit 122. That is, in this embodiment, a method in which the GC method and the HDSS method are combined is used. This combined method is called a GHDSS (Geometric High-order Decorrelation-based Source Separation) method in the present invention.
JGHDSS(W)=αJHDSS(W)+(1−α)JGC(W) (10)
In Expression (10), α is a scalar and is an integer of 0 to 1. In Expression (10), the cost function JHDSS(W) is expressed by Expression (11).
JHDSS(W)=∥E[Eϕ]∥2 (11)
In Expression (11), E[⋅] represents an expected value and a bold character E represents a vector E. Eϕ represents a cost function used instead of the correlation matrix E in DSS (Decorrelation-based Source Separation).
In Expression (11), Eϕ is expressed by Expression (12).
Eϕ=ϕ(y)yH−diag[ϕ(y)yH] (12)
In Expression (12), a bold character y represents a vector. A vector E=yyH−I is defined. Vector I represents a unit matrix. Reference sign H represents a complex conjugate.
ϕ(y) is a nonlinear function and is expressed by Expression (13).
In Expression (13), p(yi) represents a joint probability density function (PDF) of y.
ϕ(yi) can be defined variously, and a hyperbolic-tangent-based function expressed by Expression (14) may be used as an example of ϕ(yi).
ϕ(yi)=tan h(η|yi|)ej·θ(y
In Expression (14), η is a scaling parameter.
The sound source separation processing unit 123 adaptively calculates the separation matrix W using Expression (15) so as to minimize the cost functions JGC and JHDSS. The sound source separation processing unit 123 separates sound components from multiple channels of sound signals input from the sound input unit 101 (see
Wt+1=Wt−μHDSSJ′HDSS(Wt)−μGCJ′GC(Wt) (15)
In Expression (15), t represents time, μHDSS represents a step size used to update an error matrix, μGC represents a step size used to update a geometric error matrix. J′HDSS(Wt) represents an HDSS error matrix and is a numerical sequence obtained by differentiating the matrix JHDSS with respect to each input element. J′GC(Wt) represents a geometric error matrix and is a numerical sequence obtained by differentiating the matrix JGC with respect to each input element.
The step size μGC is expressed by Expression (16) using EGC and the geometric error matrix J′GC(Wt). The step size μHDSS is expressed by Expression (17) using EHDSS and the geometric error matrix J′GHDSS.
As described above, the sound source separation unit 13 in this embodiment separates sound sources by sequentially calculating the separation matrix W using the GHDSS method.
This embodiment describes an example where the sound source separation unit 13 separates sound sources using the GDHSS method, but the sound sources may be separated using a known blind signal source separation (BSS) method or the like.
The classification search process which is performed by the class number calculation unit 107 to the STF processing unit 111 will be described below.
As shown in
It is assumed that the transfer functions A(ω, ψp1) to A(ω, ψp4) corresponding to the azimuths p1, p2, p3, and p4 are stored in advance in the transfer function storage unit 112.
(Step S1) A use of the sound direction estimation device 11 selects a desired spatial resolution. The class number calculation unit 107 calculates the number of classes S and the search interval δi using Expressions (18) and (19) based on the desired spatial resolution and the search range selected by the user of the device, and outputs the calculated number of classes S, the calculated search interval δi, and the eigenvectors input from the matrix calculation unit 106 to the first spatial spectrum calculation unit 108. The calculation of the number of classes S and the search interval δi will be described later. After the process of step S1 ends, the process flow goes to step S2.
In Expression (19), dS represents the search interval of the s-th class. Here, s is an integer of 1 to S.
(Step S2) The STF processing unit 111 determines whether the transfer function A(ω, ψ) corresponding to the first search point (q11 in
(Step S3) When it is determined that the transfer function A(ω, ψ) corresponding to the first search point is stored in the transfer function storage unit 112, the STF processing unit 111 reads the transfer function A(ω, ψ) corresponding to the first search point (q11 in
(Step S4) When it is determined that the transfer function A(ω, ψ) corresponding to the first search point is not stored in the transfer function storage unit 112, the STF processing unit 111 an interpolated transfer function A^ of the first search point through interpolation using the transfer functions corresponding to two measurement points adjacent to the first search point. For example, in case of the search point q12, the interpolated transfer function A^ of the search point q12 is calculated using the transfer functions A of two measurement points p2 and p3 adjacent thereto. The STF processing unit 111 may use, for example, an FDLI method, a TDLI method, and an FTDLI method for the interpolation. These interpolation methods will be described later.
(Step S5) The first spatial spectrum calculation unit 108 calculates a spatial spectrum P(ω, ψ, f) using the transfer function A or the interpolated transfer function A^ input from the STF processing unit 111 and using Expression (5). Then, the second spatial spectrum calculation unit 109 calculates an average spatial spectrum P(ψ, f) using the spatial spectrum P(ω, ψ, f) calculated by the first spatial spectrum calculation unit 108 and using Expression (7) and outputs the calculated average spatial spectrum P(ψ, f) and information indicating the number of the search point to the peak search unit 110. The numbers of the search points are numbers assigned to correspond to the search points q11 to q15. After the process of step S5 ends, the process flow goes to step S6.
(Step S6) The peak search unit 110 determines whether all the search points in the search range are searched. The peak search unit 110 performs the process of step S8 when it is determined that all the search points in the search range are searched (YES in step S6), and performs the process of step S7 when it is determined that all the search points in the search range are not searched (NO in step S6).
(Step S7) When it is determined that all the search points in the search range are not searched, the peak search unit 110 outputs information indicating an instruction to search the next search point to the STF processing unit 111. Regarding the next search point, for example, the search point q12 is the next search point when the process of searching the search point q11 ends in step S5. The search may be performed from 0 to d0 or may be performed from d0 to 0.
(Step S8) When it is determined that all the search points in the search range are searched, the peak search unit 110 extracts an azimuth ψ[i] having the maximum value out of all the average spatial spectrum P(ψ, f) calculated in the search range. After the process of step S8 ends, the process flow goes to step S9.
(Step S9) The peak search unit 110 determines whether all the classes calculated in step S1 are searched. The peak search unit 110 performs the process of step S10 when it is determined that all the classes calculated in step S1 are searched (YES in step S9), and performs the process of step S11 when it is determined that all the classes calculated in step S1 are not searched (NO in step S9).
(Step S10) The peak search unit 110 outputs the extracted azimuth ψ[i] as the estimated azimuth ψ^ to the output unit 113. Then, the output unit 113 outputs the estimated azimuth ψ^ input from the peak search unit 110, for example, to the acoustic feature extraction unit 14 (see
(Step S11) When it is determined that all the classes calculated in step S1 are searched, the STF processing unit 111 selects two search points (ψ[i]−δi) and (ψ[i]+δi) adjacent to the azimuth ψ[i] having the peak value as a section to be searched next. Hereinafter, ψ[i]−δi is written as ψ[i−] and ψ[i]+δi is written as ψ[i+].
For example, when the search point at which the peak value is detected is q13 in
(Step S12) The STF processing unit 111 calculates the search interval d used to search the second class. The calculation of the search interval d in each class will be described later. After the process of step S12 ends, the process flow returns to step S2.
The first spatial spectrum calculation unit 108 to the STF processing unit 111 repeatedly perform the processes of steps S2 to S9 in the second class, calculates the average spatial spectrum P(ψ, f) using the transfer function (or interpolated transfer function) of each interval, and estimates the azimuth ψ having the peak value. After the search of the second class ends, the STF processing unit 111 selects two search points adjacent to the search point having the peak value and calculates the search interval d used for the next class. Thereafter, the first spatial spectrum calculation unit 108 to the STF processing unit 111 repeatedly perform the processes of steps S2 to S12 on all the classes calculated in step S1 to estimate a sound direction.
The interpolation performed by the STF processing unit 111 will be described below. The STF processing unit 111 interpolates the transfer function A to generate an interpolated transfer function A^, for example, using any one of (1) an FDLI method, (2) a TDLI method, and (3) an FTDLI method to be described below.
The transfer functions A at the two measurement points are expressed by Expressions (20) and (21). Such transfer functions A are stored in advance in the transfer function storage unit 112. A transfer function A(ω, ψ1) is a transfer function in the direction ψ1 and a transfer function A(ω, ψ2) is a transfer function in the direction ψ2.
A(ω,ψ1)=[(A1(ω,ψ1), . . . ,AM(ω,ψ1))]T (20)
A(ω,ψ2)=[(A2(ω,ψ2), . . . ,AM(ω,ψ2))]T (21)
(1) Frequency Domain Linear or Bi-Linear Interpolation (FDLI) Method
In the FDLI method, the STF processing unit 111 performs linear interpolation at two measurement points using Expression (22) to calculate the interpolated transfer function A^.
Âm[ψ
In Expression (22), DA represents an interpolation coefficient and has a value in a range of 0 to 1. The FDLI method is characterized in that a phase can be linearly interpolated.
(2) Time Domain Linear Interpolation (TDLI) Method
In the TDLI method, Expression (23) is used.
In Expression (23), dψ^ is expressed by Expression (24).
In Expressions (23) and (24), kψ1 and kψ2 are coefficients which are geometrically calculated, dψ1 and dψ2 are time delays which are geometrically calculated, and am(t, ψ) is an expression of Am(ω, ψ) in the time domain.
Expression (23) is considered as amplitude interpolation, Expression (24) is considered as phase interpolation, and the TDLI method in the frequency domain is expressed by Expression (25).
As described above, in the TDLI method, the STF processing unit 111 calculates the interpolated transfer function A^ by performing linear interpolation on two measurement points using Expression (25). The TDLI method is characterized in that amplitude can be interpolated.
(3) Frequency Time Domain Linear or Bi-Linear Interpolation (FTDLI) Method
The FTDLI method is a linear interpolation method in which the FDLI method and the TDLI method are combined. In the FTDLI method, a phase is calculated through the linear interpolation in the frequency domain and an amplitude is calculated through the linear interpolation in the time domain. In the FTDLI method, the STF processing unit 111 calculates the interpolated transfer function A^ by performing linear interpolation on two measurement points using Expression (26).
Âm[ψ
The process of calculating the interpolated transfer function A^ in the FTDLI method will be described below.
(I) First, an interpolated transfer function is calculated using Expressions (22) and (25). In the following description, the calculated interpolated transfer functions are expressed by Expressions (27) and (28).
Âm[F|ψ
Âm[T|ψ
(II) Then, Expressions (27) and (28) are decomposed into the phase and the amplitude as expressed by Expressions (29) and (30).
Âm[F|ψ
Âm[T|ψ
By using Expressions (29) and (30), the interpolated transfer function A^(ω, ψ^) is expressed by Expression (26). The FTDLI method is characterized in that a phase and an amplitude can be linearly interpolated.
The process of calculating the optimal number of classes and the search interval which is performed by the class number calculation unit 107 in step S1 will be described below.
In the following description, known values are the search range d0 of the first class and the search interval of the S-th class dS. Values to be calculated are the number of classes S in which the search number is minimized and the search interval ds in each class.
First, the interval between the direction ψ1 and the direction ψ2 of two transfer functions is defined as d0. That is, the range from the interval 0 between the direction ψ1 and the direction ψ2 of two transfer functions to the interval d0 is searched with an interval ds, the search number is d0/ds. In order to minimize the search number to reduce calculation costs, the search is hierarchically performed as follows in this embodiment.
As shown in
Number of Classes of 2
As shown in
In Expression (31), only d1 is a variable. Accordingly, when Expression (31) is partially differentiated with the variable d1 to calculate the minimum value of Expression (31), Expression (32) is obtained.
Therefore, the interval d1 in which Expression (32) is 0 is √(d0dS).
With this value of the interval d1, the search number of the first class is expressed by Expression (33) and the search number of the second class is expressed by Expression (34).
By setting the search numbers of the classes to be equal to each other using Expressions (33) and (34), it is possible to minimize the total search number.
Number of Classes of s
Conditions for minimizing the total search number when the number of classes is s will be described below. Even when the number of classes is s, it is possible to minimize the total search number by setting the search numbers of the classes to be equal to each other.
The reason thereof will be described below using a proof-by-contradiction method.
When the number of classes is s, the total search number is d0/d1+d1/d2+ . . . +d(s−1)/ds. It is assumed that the total search number is the minimum and there is a class of d(i−1)/di≠di/d(i+1). Here, i is an integer of 1 to S. In two classes of i−1 to i+1, when the search numbers of the classes are d(i−1)/di=di/d(i+1), (d(i−1)/di)+(di/d(i+1) is minimized as described in the example of two classes and thus this assumption is contradictory. That is, classes of d(i−1)/di≠di/d(i+1) do not exist.
As a result, when the search numbers of the classes are set to be equal to each other regardless of the number of classes, it is possible to minimize the total search number. Therefore, when the number of classes is s, the condition for minimizing the total search number is d0/d1=d1/d2= . . . =d(s−1)/ds.
This conditional expression can be modified to Expression (35).
From Expression (35), the granularity dS−1 in which the total search number is minimized is expressed by Expression (36).
Number of Classes of S
When the number of classes is S, the total search number G(S) is expressed by Expression (37) from d0/d1=d1/d2= . . . =d(S−1)/dS.
When Expression (37) is partially differentiated with S to calculate S of minimizing Expression (37), Expression (38) is obtained.
When Expression (38) is 0, the total search number is minimized. Accordingly, the minimum total search number is S=log(d0/dS). Similarly to Expressions (35) and (36), when the conditional expression is modified and the modified conditional expression is substituted for S=log(d0/dS), the granularity ds of each class can be calculated using Expression (39).
As described above, the class number calculation unit 107 (see
As described above, in searching for a sound source in this embodiment, a wide range is searched with a coarse resolution and a narrow range is searched with a finer interval based on the search result. According to this embodiment, it is possible to reduce the number of times of search by hierarchically performing such a search. Accordingly, it is possible to reduce calculation costs while maintaining the estimation performance of sound source localization.
The first embodiment has described the example where the class number calculation unit 107 calculates the number of classes S and the granularity dS with which the total search number is minimized.
This embodiment describes an example where interpolation is performed between the direction ψ1 and the direction ψ2 of the transfer functions and a search is performed in consideration of the interpolated direction.
In the following description, known values are the search range d0 of the first class and the search interval dS of the S-th class. Values to be calculated are the number of classes S in which the calculation cost is minimized and the search interval ds in each class.
As shown in
In this embodiment, the class number calculation unit 107 calculates the number of classes S with which the calculation cost due to classification is minimized and the granularity ds.
First, the search number of s-th class is ds−1/ds and the search cost is (ds−1/ds)ct. The search cost (ds−1/ds)ct has a fixed value.
The interpolation cost of the s-th class will be described below. In the s-th class, the interval (granularity) is ds and the number of search points is ds−1. It is assumed that IS points out of the search points require interpolation. In
When all the search points do not require any more interpolation, the interpolation cost is 0.
When all the search points require interpolation, the interpolation cost is (ds−1/ds)cI.
Therefore, the range of the interpolation cost IScI of the s-th class ranges from 0 to (ds−1/ds)cI. here, cs is set to a range of 0 to CI and the interpolation cost is set to a fixed value (ds−1/ds)cs.
Therefore, the calculation cost of the s-th class is the sum of the search cost and the interpolation cost and is thus expressed by Expression (40).
Accordingly, the calculation cost G˜(S) in all the classes is expressed by Expression (41).
In Expression (41), i ranges from 1 to S. In Expression (41), when there is only the search cost, this case corresponds to ci=0.
The cost is normalized by setting ct to 1 so that the total search number G(S) of Expression (37) described in the first embodiment is equal to G˜(S) in case of ci=0. By the normalization with ct, ct+ci in the right side of Expression (41) is modified to 1+ci. (1+ci) is defined as a new variable Ci. Here, the variable Ci is greater than or equal to 1. By replacing ct+ci with Ci, Expression (41) is modified to Expression (42).
Number of Classes of 2
A case where the number of classes is 2 will be described below.
The total search number F(d1) in this case is (d0/d1)C1+(d1/dS)CS.
In the total search number F(d1), only d1 is a variable. When the total search number F(d1) is partially differentiated to calculate the minimum value of the total search number F(d1) with d1, Expression (43) is obtained.
Here, d1 with which Expression (43) is 0 is √(C1d0dS/CS).
At this time, the search number of the first class is expressed by Expression (44) and the search number of the second class is expressed by Expression (45).
From Expressions (44) and (45), when the intervals (granularity) weighted with C1 and CS are equal to each other, it is possible to minimize the calculation cost from d0 to ds.
Similarly, when the number of classes is s, the condition for minimizing the calculation cost is that the calculation costs (di−1/di)Ci of the classes weighted by Ci are equal to each other. Here, i ranges from 1 to s.
Number of Classes of S
When the number of classes is S, the calculation cost G˜(S) of Expression (42) is calculated using the condition of (d0/d1)C1=(d1/d2)C2= . . . =(d(S−1)/dS)CS. This conditional expression can be modified to Expression (46).
From Expression (46), the granularity dS−1 with which the calculation cost is minimized is expressed by Expression (47).
From Expression (47), the calculation cost G˜(S) is expressed by Expression (48).
When Expression (48) is partially differentiated with S to calculate S for minimizing Expression (48), Expression (49) is obtained.
S causing Expression (49) to be 0 is expressed by Expression (50).
Therefore, when the number of classes is S, the interval (granularity) ds with which the calculation cost is minimized is expressed by Expression (51).
The class number calculation unit 107 (see
Test Result
A result of a test in which the sound direction estimation device 11 according to the first embodiment or the second embodiment is mounted on a robot auditory software HARK (HRI-JP Audition for Robots with Kyoto University) which is an open source will be described below. In the following test, the sound direction estimation device 11 according to the first embodiment is mounted on a humanoid robot.
The test is carried out with the azimuth ψ1 of the search points fixed to 0 degrees while changing the azimuth ψ2 to 30 degrees, 60 degrees, 90 degrees, and 120 degrees. The interpolation points between the azimuths are set every one degree and an error e−[ψ1, ψ2] from the transfer function every one degree, which is obtained by measurement, is calculated by averaging using Expression (52).
In Expression (52), f(ω[k], ψ^[i]) represents an interpolation error of the frequency bin corresponding to the frequency ω[k] at the interpolation point ψ^[i]. In the test, kl and kh in Expression (52) are selected so that the frequency ω which is a low frequency region used in the sound source localization ranges from 500 [Hz] to 280 [Hz]. In Expression (52), iψ represents the number of interpolation points every one degree in which is in a range larger than ψ1 and smaller than ψ2.
In the following test, a phase estimation error (PEE), a spectrum distortion error (SD), and a signal-to-distortion ratio (SDR) are used as an error index of f(ω[k], ψ^[i]).
The PEE is expressed by Expression (53) and is an index of a phase error.
The SD is expressed by Expression (54) and is an index of an error amplitude.
The SDR is expressed by Expression (55) and is an index of an error of a transfer function itself.
Hereinafter, e1−, e2−, and e3− are e−[ψ1, ψ2] when the PEE, the SD, and the SDR are used, respectively.
Linearity of the interpolation coefficient DA is evaluated using Expression (56).
In Expression (56), the interpolation coefficient DA[i] is a value of DA in which the interpolation error is the minimum. In Expression (56), when the interpolation coefficient DA is close to linearity with respect to the interpolation point, it means that the interpolation coefficient DA can be determined by (ψ^−ψ2)/(ψ1−ψ2) and thus it is practical.
Hereinafter, d1−, d2−, and d3− are d−[ψ1, ψ2] when the PEE, the SD, and the SDR are used, respectively.
In Parts (a) to (f) of
As shown in Parts (a) to (c) of
As shown in Parts (d) to (f) of
The test is carried out by reproducing white noise every 1 degree (where the azimuth ψ ranges from −90 degrees to 90 degrees) and calculating an error between the reproduction direction of the reproduced sound source and the estimation direction thereof. In the test, the measurement interval of the transfer function is changed so that the azimuth ψ is 1 degree, 5 degrees, 10 degrees, 30 degrees, 60 degrees, 90 degrees, and 120 degrees, and the transfer functions for every 1 degree are generated by interpolation.
As shown in
An example of a test on the calculation cost due to hierarchical search will be described below.
In the test, the average processing time spent in calculation for the sound source localization (Expressions (3) to (7)) is calculated when the number of sound sources is changed and 1000 frames are processed. The number of classes to be searched for a sound source is set to 2, the first class is searched at intervals of 10 degrees, and then the second class is searched at intervals of 1 degree. The test is carried out when the number of sound sources is 1, 2, 3, and 4.
As shown in
In this way, even when four sound sources exist and concurrently utter speeches, the process flow ends within a frame period of 10 [msec] or less. As shown in
Speech data of 10 persons×216 words in an ATR speech database are used as speech in the test and speech concurrently uttered by four speakers are recognized. The transfer functions used in the test have the same intervals as shown in
As shown in
As described above, when the sound direction estimation device 11 according to the first embodiment or the second embodiment is applied to the sound processing system 1, it is possible to reduce the computational load on the search while maintaining the same performance in word correct rate as in the case where the transfer functions with a finer spatial resolution of 1 degree are used. As a result, it is possible to accurately estimate a sound direction within an actual time of each frame.
The first embodiment and the second embodiment describes the example where the matrix calculation unit 106 calculates the GSVD matrix and a sound direction is estimated based on the calculated GSVD matrix, but are not limited to this example. The matrix calculation unit 106 may calculate a generalized eigenvalue decomposition (GEVD) matrix and a sound direction may be estimated based on the calculated GEVD matrix. In this case, the sound direction estimation device 11 calculates the number of classes and the search interval (granularity) with which the total search number is minimized or calculates the number of classes and the granularity in which the calculation cost is minimized, as described in the first embodiment or the second embodiment. The sound direction estimation device 11 first performs a search with a coarse search interval and then selects a next search range from the search range, based on the calculated number of classes and the calculated search interval (granularity). The sound direction estimation device 11 narrows the search interval in the selected search range. In the sound direction estimation device 11, it is possible to accurately estimate a sound direction and to reduce calculation costs.
The first embodiment and the second embodiment describe the example where the class number calculation unit 107 calculates both the number of classes used in the search and the search interval, but are not limited to this example. The number of classes may be selected in advance by a user of the device and only the search interval may be calculated.
The first embodiment and the second embodiment describe the example where the FDLI method, the TDLI method, and the TFDLI method are used as the technique of interpolating the transfer functions, but are not limited to this example. The transfer functions may be interpolated using other techniques.
Here, a program for realizing the function of the sound direction estimation device 11 in the present invention may be recorded in a computer-readable recording medium, and a computer system may read the program recorded on the recording medium and may execute the program to estimate the direction of a sound source. The term “computer system” used herein includes an OS and hardware, such as peripherals. The “computer system” also includes a WWW system including a homepage provision environment (or display environment). The “computer-readable recording medium” includes a portable medium such as a flexible disc, a magneto-optical disc, a ROM, or a CD-ROM or a storage device such as a hard disk built in the computer system. Moreover, the “computer-readable recording medium” also includes a device storing a program for a predetermined time, like an internal volatile memory (RAM) of a computer system serving as a server or a client when the programs are transmitted through a network such as the Internet or a communication line such as a telephone line.
The above programs may be transmitted from a computer system having the programs stored in a storage device thereof or the like to another computer system through a transmission medium or by carrier waves in the transmission medium. The “transmission medium” which transmits a program means a medium having a function of transmitting information and examples thereof include a network (communication network) such as the Internet and a communication link (communication line) such as a telephone line. The program may realize some of the above-described functions. The program may realize the above-described functions in combination with a program already recorded in a computer system, that is, the program may be a differential file (differential program).
Although an embodiment of the invention has been described referring to the drawings, a specific configuration is not limited to those described above, and various changes in design and the like may be made within the scope without departing from the scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
2012-203393 | Sep 2012 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
8120993 | Amada | Feb 2012 | B2 |
8693287 | Nakadai | Apr 2014 | B2 |
20040091119 | Duraiswami | May 2004 | A1 |
20070092016 | Cheng | Apr 2007 | A1 |
20110025563 | Ferreol et al. | Feb 2011 | A1 |
Number | Date | Country |
---|---|---|
H11-231033 | Aug 1999 | JP |
2001-091617 | Apr 2001 | JP |
2005-201798 | Jul 2005 | JP |
2006-292660 | Oct 2006 | JP |
2012-042465 | Mar 2012 | JP |
Entry |
---|
Manoochchr Azmoodeh. Abstract Data Types and Algorithms. 2nd Edition. Publisher Springer. ISBN 1349211516, 9781349211517.377 pages. 1990. |
Roy, Deb K. “Speaker indexing using neural network clustering of vowel spectra.” International Journal of Speech Technology 1.2 (1997): 143-149. |
Duraiswami, Rainani, Dmitry Zotkin, and Larry S. Davis. “Active speech source localization by a dual coarse-to-fine search.” Acoustics, Speech, and Signal Processing, 2001. Proceedings.(ICASSP'01). 2001 IEEE International Conference on. vol. 5. IEEE, 2001. |
Parra, Lucas C., and Christopher V. Alvino. “Geometric source separation: Merging convolutive source separation with geometric beamforming.” Speech and Audio Processing, IEEE Transactions on 10.6 (2002): 352-362. |
Ringach, Dario L. “Look at the big picture (details will follow).” nature neuroscience 6.1 (2003): 7-8. |
Zotkin, Dmitry N., and Ramani Duraiswami. “Accelerated speech source localization via a hierarchical search of steered response power.” Speech and Audio Processing, IEEE Transactions on 12.5 (2004): 499-508. |
Valin, Jean-Marc, et al. “Robust recognition of simultaneous speech by a mobile robot.” Robotics, IEEE Transactions on 23.4 (2007): 742-752. |
Yamamoto, Shunnichi, et al. “Design and implementation of a robot audition system for automatic speech recognition of simultaneous speech.” Automatic Speech Recognition & Understanding, 2007. ASRU. IEEE Workshop on. IEEE, 2007. |
Nakajima, Hirofumi, et al. “Adaptive step-size parameter control for real-world blind source separation.” Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on. IEEE, 2008. |
Nakadai, Kazuhiro, et al. “Sound source separation of moving speakers for robot audition.” Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on. IEEE, 2009. |
Nakadai, Kazuhiro, et al. “Design and Implementation of Robot Audition System‘HARK’—Open Source Software for Listening to Three Simultaneous Speakers.” Advanced Robotics 24.5-6 (2010): 739-761. |
Asano, Futoshi, and Hideki Asoh. “Joint estimation of sound source location and noise covariance in spatially colored noise.” Proc. Eusipco 2011 (2011). |
Nakamura, Keisuke, et al. “Intelligent sound source localization and its application to multimodal human tracking.” Intelligent Robots and Systems (IROS), 2011 IEEE/RSJ International Conference on. IEEE, 2011. |
Nakamura, Keisuke, et al. “Correlation matrix interpolation in sound source localization for a robot.” Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on. IEEE, 2011. |
Nakadai, Kazuhiro, et al. “Robot audition for dynamic environments.” Signal Processing, Communication and Computing (ICSPCC), 2012 IEEE International Conference on. IEEE, 2012. |
Nakamura, Keisuke, Kazuhiro Nakadai, and Gökhan Ince. “Real-time super-resolution sound source localization for robots.” Intelligent Robots and Systems (IROS), 2012 IEEE/RSJ International Conference on. IEEE, 2012. |
Okutani, Keita, et al. “Outdoor auditory scene analysis using a moving microphone array embedded in a quadrocopter.” Intelligent Robots and Systems (IROS), 2012 IEEE/RSJ International Conference on. IEEE, 2012. |
Nakamura, Keisuke, Kazuhiro Nakadai, and Hiroshi G. Okuno. “A real-time super-resolution robot audition system that improves the robustness of simultaneous speech recognition.” Advanced Robotics 27.12 (2013): 933-945. |
Ohata, Takuma, et al. “Improvement in outdoor sound source detection using a quadrotor-embedded microphone array.” Intelligent Robots and Systems (IROS 2014), 2014 IEEE/RSJ International Conference on. IEEE, 2014. |
Okuno, Hiroshi G., and Kazuhiro Nakadai. “Robot audition: Its rise and perspectives.” Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on. IEEE, 2015. |
Takeda, Ryu, and Kazunori Komatani. “Performance comparison of Music-based sound localization methods on small humanoid under low SNR conditions.” Humanoid Robots (Humanoids), 2015 IEEE-RAS 15th International Conference on. IEEE, 2015. |
Matsumoto, Mitsuo, Mikio Tohyama, and Hirofumi Yanagawa. “A method of interpolating binaural impulse responses for moving sound images.” Acoustical Science and Technology 24.5 (2003): 284-292. |
Japanese Office Action Notice of Reasons for Rejection application No. 2012-203393 dated Oct. 20, 2015. |
Number | Date | Country | |
---|---|---|---|
20140078867 A1 | Mar 2014 | US |