The present application is based on PCT filing PCT/JP2019/032687, filed Aug. 21, 2019, the entire contents of which are incorporated herein by reference.
The present invention relates to an estimation device, an estimation method, and an estimation program.
Known sound source separation methods are independent component analysis (ICA), which is a scheme for performing a sound source separation method based on statistical independence between sound sources, and independent low-rank matrix analysis (ILRMA) provided by combining ICA and nonnegative matrix factorization (NMF), which is a scheme for performing sound source separation based on a low rank of a power spectrum of a sound source (for example, NPL 1).
NPL 1: D. Kitamura, N. Ono, H. Sawada, H. Kameoka, and H. Saruwatari, “Determined Blind Source Separation Unifying Independent Vector Analysis and Nonnegative Matrix Factorization”, IEEE/ACM Trans. ASLP, vol. 24, no. 9, pp. 1626-1641, 2016.
In models of ILRMA and ICA and NMF serving as a basis thereof described in NPL 1, it is assumed that there is no correlation between time frequency bins of sound source spectra. However, because an actual sound source signal often has some correlation between time frequency bins of sound source spectra, a model of the related art seems to be not suitable for modeling an unsteady signal such as vocal sound. In fact, when models of the related art are used, sound source separation sometimes cannot be performed accurately.
The present invention has been made in view of the above, and an object of the present invention is to provide an estimation device, an estimation method, and an estimation program capable of estimating information on sound source separation filter information that enables sound source separation with better performance than in the related art to be realized.
In order to solve the above-described problem and achieve the object, an estimation device according to the present invention includes an estimation unit configured to estimate a covariance matrix having information on a correlation between sound source spectra and information on a correlation between channels as information on sound source separation filter information for separating an individual sound source signal from a mixed acoustic signal.
Further, an estimation method according to the present invention includes estimating a covariance matrix having information on a correlation between sound source spectra and information on a correlation between channels as information on sound source separation filter information for separating an individual sound source signal from a mixed acoustic signal.
Further, an estimation program according to the present invention causes a computer to execute estimating a covariance matrix having information on a correlation between sound source spectra and information on a correlation between channels as information on sound source separation filter information for separating an individual sound source signal from a mixed acoustic signal.
According to the present invention, it is possible to estimate the information on sound source separation filter information that enables sound source separation with higher performance than in the related art to be realized.
Hereinafter, embodiments of an estimation device, an estimation method, and an estimation program according to the present application will be described in detail based on the drawings. The present invention is not limited to the embodiments to be described hereinafter.
Hereinafter, when “{circumflex over ( )}A” is written for A, which is a vector, matrix, or scalar, {circumflex over ( )}A is assumed to be equivalent to “a symbol in which “{circumflex over ( )}” is written immediately above “A””. When “˜A” is written for A, which is a vector, matrix, or scalar, —A is the same as “a symbol in which “˜” is written immediately above “A””.
The present embodiment proposes a new probabilistic model in which a correlation between sound source spectra has been considered in addition to a correlation between channels. In the present embodiment, sound source separation is performed using a spatial covariance matrix estimated by using the probabilistic model, which enables sound source separation with higher performance than that in the related art. The spatial covariance matrix is information on sound source separation filter information for separating an individual sound source signal from a mixed acoustic signal, and is a parameter for modeling spatial characteristics of each sound source signal. First, a new probabilistic model used in the present embodiment will be described.
Let a mixed acoustic signal, which is an acoustic signal observed by M microphones, be denoted as xf,t∈ CM. In the following equation, “an outlined character C” corresponds to “C”. Here, f ∈ [F] is an index of a frequency bin. t ∈ [T] is an index of a time frame. CM indicates a set of M-dimensional complex vectors. Here, [I]:={1, . . . , I} (I is an integer). In each time frequency bin, the mixed acoustic signal xf,t∈ CM is expressed by a sum of microphone observation signals of N sound sources, and is shown by Equation (1).
[Math. 1]
xƒ,t=z1,ƒ,t+ . . . +zN,ƒ,t,∈M (1)
Let D=FTM, and x and zn, are defined as Expressions (2) and (3) below.
[Math. 2]
x:=(xƒ,t|ƒ∈[F],t ∈[T])∈D (2)
[Math. 3]
zn:=(zn,ƒ,t|ƒ∈[F],t ∈[T])∈D (3)
Here, a sound source separation problem dealt with in the present embodiment is formulated as a problem of estimation of an acoustic signal {zn}n=1N of each sound source from an observed mixed acoustic signal x under the two conditions below (See Equations (4) and (5)).
(Condition 1) Sound source signals are assumed to be independent of each other.
[Math. 4]
p({zn}n=1N)=Πn=1Np(zn) (4)
(Condition 2) For each n ∈ [N], it is assumed that zn follows a complex Gaussian distribution with the following mean 0 and spatial covariance matrix Rn.
[Math. 5]
p(zn)=N(zn|0, Rn) (n ∈[N]) (5)
As the above model shows, when the spatial covariance matrix Rn can be estimated, a signal of each sound source can be estimated using Equations (1), (4), and (5).
Here, ILRMA, which is the related art, is a technology for estimating the spatial covariance matrix Rn on the assumption that there is no correlation between time frequency bins of the sound source spectra, in addition to conditions 1 and 2 above. In ILRMA, estimation is performed on the assumption that Rn satisfies properties shown in Equations (6) to (8) and Relationship (9) below.
[Math. 6]
Rn=⊕f=1F⊕t=1TRn,ƒ,t∈S+FTM (6)
[Math. 7]
WƒHRn,ƒ,tWƒ=λn,ƒ,tEn,n∈S+M (7)
[Math. 8]
λn,ƒ,t=Σk=1Kφn,ƒ,kψn,k,t∈≥ (8)
[Math. 9]
φn,ƒ,k,ψn,k,t∈≥0 (9)
Here, S+D is a set of all semi-fixed Hermitian matrices having a size D×D. En,n is a matrix in which the (n, n) component is 1 and the others are 0. Further, {λn,f,t}f,t⊆R≥0 is a power spectrum of a sound source n, and is obtained by modeling through non-negative matrix factorization (NMF) as shown in Equations (8) and (9). K is the number of bases of NMF. {Φn,f,k}f=1F is a k-th base of the sound source n. {Ψn,k,t}t=1T is an activation for the k-th base of the sound source n.
The present embodiment proposes a model obtained by extending the model ILRMA, which is a method of the related art, so that a correlation between the sound source spectra is considered. Specifically, in the present embodiment, a spatial covariance matrix having information on the correlation between the sound source spectra and information on a correlation between channels is estimated as the information on the sound source separation filter information for separating an individual sound source signal from the mixed acoustic signal. Models in which the correlation between channels and the correlation between the sound source spectra are considered include three patterns including an expression format in which frequency correlation is considered (ILRMA-F), an expression format in which time correlation is considered (ILRMA-T), and an expression format in which both the time correlation and the frequency correlation are considered (ILRMA-FT), and sound source separation can be performed using any of these patterns.
ILRMA-F
First, ILRMA-F, which is a model in which frequency correlation has been considered, will be described. ILRMA-F uses a model in which Equations (10) and (11) below have been assumed instead of Equations (6) and (7) assumed in ILRMA of the related art because correlation between frequency bins is considered.
[Math. 10]
Rn=⊕t=1TRn,t∈S+FTM (10)
[Math. 11]
PHRn,tP=⊕f=1F(λn,ƒtEn,n)∈S+FM (11)
Here, P ∈ GL(FM) is a block matrix having a size F×F, which includes a matrix having a size M×M as an element, and a (f1, f2)-th block thereof is expressed by Expression (12) below.
Here, for each f ∈ [F], it is assumed that Δf⊆Z (Z is a set of all integers) is a set of integers and satisfies 0 ∈Δf. As an example of P satisfying the above properties, P in the case of F=4 and Δf={0,2,3,−1} (f ∈ [F]) is shown in Equation (13) below.
Thus, P is characterized in that P has one or more non-zero components in non-diagonal blocks, in addition to a diagonal block Pf,0 (f ∈ [F]). In P, the diagonal blocks indicate the correlation between the channels, and the non-diagonal blocks indicates the correlation between frequency directions. Further, it is possible to reduce a calculation time required for estimation of the spatial covariance matrix by modeling P in which most of the non-diagonal blocks are 0. Further, in ILRMA-F, Δf⊆Z is designed so that P satisfies Equation (14), making it possible to greatly reduce the calculation time required for estimation of the spatial covariance matrix.
ILRMA-T
Next, ILRMA-T which is a model in which time correlation is considered will be described. Because correlation between time frames is considered, ILRMA-T uses a model in which Equations (15) and (16) below are assumed instead of Equations (6) and (7) assumed in ILRMA of the related art.
[Math. 15]
Rn=⊕ƒ=1FRn,ƒ∈S+FTM (15)
[Math. 16]
PƒHRn,ƒPƒ=⊕t=1T*λn,ƒ,tEn,n)∈S+TM (16)
Here, P ∈ GL (TM) is a block matrix having a size T×T, includes a matrix having a size M×M as an element, and it is assumed that a (t1, t2)-th block thereof is expressed by Expression (17) below.
Here, for each f ∈ [F], it is assumed that Δf⊆Z is a set of integers and satisfies 0 ∈ Δf.
ILRMA-FT
Next, ILRMA-FT, which is a model in which both time correlation and frequency correlation have been considered, will be described. ILRMA-FT uses a model in which Equation (18) below has been assumed instead of Equations (6) and (7) assumed in the ILRMA of the related art is used because the correlation between frequency bins and the correlation between time frames are considered.
[Math. 18]
PHRnP=⊕ƒ=1F⊕t=1T(λn,ƒ,tEn,n)∈S+FTM (18)
Here, P ∈ GL (FTM) is a block matrix having a size FT×FT, which includes a matrix having a size M×M as an element, and a ((f1-1)T+t1, (f2-1)T+t2)-th block is assumed to be expressed by Expression (19) below.
Here, it is assumed that, for each f ∈ [F], Δf⊆Z×Z is a set of pairs of integers and satisfies (0,0) ∈ Δf. As an example of P satisfying the above properties, P ∈ GL (6M) in the case of F=3, T=2, and Δf={(0,0), (0, −1), (−1, ±1), (−2, 0)} (f ∈ [F]) is shown by Expression (20) below.
Thus, P is characterized in that P has one or more non-zero blocks in non-diagonal blocks, in addition to a diagonal blocks Pf,0,0 (f ∈ [F]). The diagonal blocks express correlation between channels and the non-diagonal blocks express correlation between time-frequency bins. Further, it is possible to reduce the calculation time required for estimation of the spatial covariance matrix by modeling P in which most of the non-diagonal blocks are 0. Further, in ILRMA-FT, it is possible to greatly reduce the calculation time required for estimation of the spatial covariance matrix by designing Δf⊆Z×Z so that P satisfies Equation (21).
Thus, the model proposed in the present embodiment estimates the spatial covariance matrix having the information on the correlation between the sound source spectra and the information on the correlation between the channels as the information on the sound source separation filter information for separating an individual sound source signal from a mixed acoustic signal. In the present embodiment, the spatial covariance matrix is estimated by modeling such that the spatial covariance matrices as may as the sound sources are diagonalizable at the same time. In the present embodiment, the spatial covariance matrix is estimated on the assumption that a matrix after simultaneous diagonalization is modeled according to nonnegative matrix factorization.
Thus, in the present embodiment, it is possible to estimate the spatial covariance matrix in consideration of not only inter-channel correlation of the related art but also sound source spectrum correlation that cannot be considered in the related art by estimating the spatial covariance matrix Itr, based on the models ILRMA-F, ILRMA-T, or ILRMA-FT.
Next, the sound source separation filter information estimation device according to embodiment 1 will be described. Here, the information regarding the sound source separation filter is information for separating an individual sound source signal from the mixed acoustic signal, and is the spatial covariance matrix Rn in the ILRMA-F, ILRMA-T, or ILRMA-FT models described above. Because the ILRMA-FT model includes the ILRMA-F and ILRMA-T models in a special case, the sound source separation filter information estimation device to which the ILRMA-FT model has been applied will be described hereinafter.
The initial value setting unit 11 sets Δf⊆Z×Z that determines a non-zero structure of a simultaneous decorrelation matrix P. Here, the initial value setting unit 11 sets Δf⊆Z×Z so that the simultaneous decorrelation matrix P satisfies Equation (22).
Further, in the initial value setting unit 11 sets appropriate initial values for the simultaneous decorrelation matrix P and an NMF parameter {(φn,f,k, Ψn,k,t}n,f,k,t in advance.
The NMF parameter updating unit 12 updates the NMF parameter {φn,f,k, Ψn,k,t}n,f,k,t according to Relationships (23) and (24). Here, as the mixed acoustic signal input to the sound source separation filter information estimation device 10, for example, it is assumed that an acoustic signal obtained by performing short-time Fourier transform on a collected mixed acoustic signal is used.
Here, yn,f,t is Expression (25).
[Math. 25]
n,f,t:=edTPHx∈ (25)
However, d:=fTM+tM+n. ed is a vector in which a d-th element is 1 and the others are 0. The superscript T indicates the transpose of a matrix or vector. The superscript H indicates the Hermitian transpose of a matrix or vector. Further, x is a symbol indicating the input mixed acoustic signal.
The NMF parameter updating unit 12 uses the updated parameter {φn,f,k, Ψn,k,t}n,f,k,t to update the value of λn,f,t according to Equation (8). λn,f,t can be regarded as analogs of the power spectrum.
The simultaneous decorrelation matrix updating unit 13 updates a matrix (a simultaneous decorrelation matrix) P that simultaneously decorrelates the inter-channel correlation and the sound source spectrum correlation from the input mixed acoustic signal according to the following procedure A or B.
Procedure A
The simultaneous decorrelation matrix updating unit 13 updates {circumflex over ( )}pn,f for each n according to Equations (26) and (27).
[Math. 26]
{circumflex over (a)}n=(((P0,0H)−1en)T,0N(|Δ|−1))T∈N|Δ| (26)
[Math. 27]
{circumflex over (p)}n={circumflex over (G)}n−1{circumflex over (a)}n({circumflex over (a)}nH{circumflex over (G)}n−1{circumflex over (a)}n)−1/2e√{square root over (−10)}(θ∈) (27)
Here, {circumflex over ( )}xf,t, {circumflex over ( )}Pf, {circumflex over ( )}pn,f, and {circumflex over ( )}Gn,f are as Expressions (28) to (31) below.
However, in Equations (26) and (27), the frequency bin index f ∈ [F] is omitted. Further, as shown in Expression (30), because {circumflex over ( )}pn,f is information for specifying the simultaneous decorrelation matrix AP, it can be said that updating {circumflex over ( )}pn,f and updating {circumflex over ( )}P are synonymous.
Procedure B
Procedure B is a scheme that can be applied only when the number of sound sources N=2. In step B, the simultaneous decorrelation matrix updating unit 13 updates {circumflex over ( )}Pf according to Equations (32) to (34).
Here, Vn indicates a 2×2 principal minor matrix in the upper left of {circumflex over ( )}Gn−1 (a matrix corresponding to the first 2-by-2 matrix). Further, u1 and u2 are eigenvectors of a generalized eigenvalue problem V1u=λ,V2u. Further, in Equations (32) to (34), the index f ∈ [F] of the frequency bin is omitted.
The simultaneous decorrelation matrix updating unit 13 may use a result of adding εI based on a small ε>0 to {circumflex over ( )}Gn,f shown in Expression (31), as {circumflex over ( )}Gn,f in order to achieve numerical stability in executing procedure A or procedure B.
The iterative control unit 14 alternately and interactively executes the processing of the NMF parameter updating unit 12 and the processing of the simultaneous decorrelation matrix updating unit 13 until a predetermined condition is satisfied. The iterative control unit 14 ends the iterative processing when the predetermined condition is satisfied. The predetermined condition is, for example, that a predetermined number of iterations is reached, that an amount of updating of the NMF parameter and the simultaneous decorrelation matrix is equal to or smaller than a predetermined threshold value, or the like.
The estimation unit 15 applies a parameter P and λn,f,t at the time of ending of the processing of the NMF parameter updating unit 12 and the processing of the simultaneous decorrelation matrix updating unit 13 to Equation (18) to estimate the spatial covariance matrix Rn. The estimation unit 15 outputs the estimated spatial covariance matrix Rn to, for example, the sound source separation device.
When the ILRMA-F model is applied, the estimation unit 15 applies the parameter P and λn,f,t at the time of ending of the processing of the NMF parameter updating unit 12 and the processing of the simultaneous decorrelation matrix updating unit 13 to Equations (10) and (11) to estimate the spatial covariance matrix Rn. Further, when the ILRMA-T model is applied, the estimation unit 15 applies the parameter P and λn,f,t at the time of the ending of the processing of the NMF parameter updating unit 12 and the processing of the simultaneous decorrelation matrix updating unit 13 to Equations (15) and (16) to estimate the spatial covariance matrix Rn.
Next, estimation processing for estimating the information on the sound source separation filter information that is executed by the sound source separation filter information estimation device 10 of
As illustrated in
The NMF parameter updating unit 12 updates the NMF parameter {φn,f,k, Ψn,k,t}n,f,k,t according to Expressions (23) and (24), and uses the updated parameter {φn,f,k, Ψn,k,t}n,f,k,t and Equation (8) to update the value of λn,f,t (step S2). The simultaneous decorrelation matrix updating unit 13 updates the simultaneous decorrelation matrix P from the input mixed acoustic signal according to procedure A or B below (step S3).
The iterative control unit 14 determines whether or not the predetermined condition is satisfied (step S4). When the predetermined condition is not satisfied (step S4: No), the iterative control unit 14 returns to step S2 and causes the processing of the NMF parameter updating unit 12 and the processing of the simultaneous decorrelation matrix updating unit 13 to be executed.
When the predetermined condition is satisfied (step S4: Yes), the estimation unit 15 applies the parameter P and λn,f,t at the time of the ending of the processing of the NMF parameter updating unit 12 and the processing of the simultaneous decorrelation matrix updating unit 13, to the ILRMA-F, ILRMA-T, or ILRMA-T model to estimate the spatial covariance matrix Rn (step S5).
Thus, the sound source separation filter information estimation device 10 according to embodiment 1 estimates the spatial covariance matrix by modeling such that the spatial covariance matrices including information on the correlation between the sound source spectra and information on the correlation between channels as the information on the sound source separation filter information for separating an individual sound source signal from the mixed acoustic signal are diagonalizable at the same time. In other words, the sound source separation filter information estimation device 10 estimates the spatial covariance matrix including the information on the correlation between the sound source spectra and the information on the correlation between channels, unlike the model of the related art in which time-frequency bins of a sound source spectrum are assumed to be uncorrelated. Thus, according to the sound source separation filter information estimation device 10, because a spatial covariance matrix that is more compatible with an actual sound source signal that often has a correlation between the time frequency bins of the sound source spectra is used as the information on the sound source separation filter information, it is possible to realize sound source separation with higher performance than in a model of the related art.
Next, embodiment 2 will be described.
The sound source separation device 20 is implemented by, for example, a predetermined program being read into a computer including a ROM, RAM, CPU, and the like and executed by the CPU. The sound source separation device 20 separates each sound source signal from the mixed acoustic signal by using the spatial covariance matrix estimated by the sound source separation filter information estimation device 10.
Specifically, the sound source separation device 20 uses the spatial covariance matrix Rn output from the sound source separation filter information estimation device 10 to acquire an estimation result ˜zn of each sound source signal according to Equation (35) and output the estimation result ˜zn.
[Math. 35]
{tilde over (z)}n=[zn|x]=Rn(Σn=1NRn)−1x∈D (35)
Alternatively, the sound source separation device 20 uses the simultaneous decorrelation matrix P obtained by the sound source separation filter information estimation device 10 instead of the spatial covariance matrix Rn to acquire the estimation result ˜zn of each sound source signal according to Equation (36), and outputs the estimation result zn.
[Math. 36]
{tilde over (z)}n=(QH)−1(⊕ƒ=1F⊕t=1TEn,n)PHx (36)
Here, Q corresponds to a matrix in which (δF, δT)∈Δf, δF=0, and δT<0 are satisfied in P defined by Expression (19), and replacement with Equation (37) has been performed.
[Math. 37]
Pƒ,δ
Next, sound source separation processing that is executed by the sound source separation system 1 of
As illustrated in
The sound source separation device 20 performs the sound source separation processing for separating an individual sound source signal from the mixed acoustic signal using the spatial covariance matrix estimated by the sound source separation filter information estimation device 10 (step S22).
Thus, the sound source separation system 1 according to embodiment 2 uses the spatial covariance matrix including the information on the correlation between the sound source spectra and the information on the correlation between channels to perform sound source separation, thereby realizing sound source separation with a higher accuracy than in the related art.
An evaluation experiment was conducted to evaluate the separation performance of the ILRMA model of the related art and the ILRMA-F model, ILRMA-T model, or ILRMA-FT model proposed in the present embodiment. In this evaluation experiment, a mixed signal which was created using two microphones and two sound sources from live sound recording data of a data set provided by SiSEC2008 as evaluation data, and separation accuracies were compared. A frame length of 128 ms and 256 ms was used. Results of this evaluation experiment are shown in Table 1.
As shown in Table 1, irrespective of the ILRMA-F, ILRMA-T, and ILRMA-FT models used, results showing higher separation accuracy than in the ILRMA model of the related art were obtained.
Each component of each of the illustrated devices is a functional concept, and is not necessarily physically configured as illustrated in the figures. That is, a specific form of distribution and integration of the respective devices is not limited to the one illustrated in the figure, and all or some of the devices can be configured to be functionally or physically distributed and integrated in arbitrary units according to various loads, use situations, or the like. For example, the sound source separation filter information estimation device 10 and the sound source separation device 20 may be an integrated device. Further, all or some of processing functions performed by the respective devices may be realized by a CPU and a program analyzed and executed by the CPU, or may be realized as hardware by wired logic.
Further, all or some of the processing described as being performed automatically among the respective processing described in the present embodiment can be performed manually, or all or some of the processing described as being performed manually can be performed automatically using a known method. Further, the respective processes described in the present embodiment can not only be executed in chronological order according to the order in the description, but may also be executed in parallel or individually depending on a processing capability of a device that executes the processing or as necessary. In addition, information including the processing procedures, control procedures, specific names, and various types of data or parameters illustrated in the above document or drawings can be arbitrarily changed unless otherwise specified.
Memory 1010 includes a ROM 1011 and a RAM 1012. The ROM 1011 stores, for example, a boot program such as a basic input output system (BIOS). The hard disk drive interface 1030 is connected to a hard disk drive 1031. The disc drive interface 1040 is connected to a disc drive 1041. For example, a removable storage medium such as a magnetic disk or an optical disc is inserted into the disc drive 1041. The serial port interface 1050 is connected to, for example, a mouse 1110 and a keyboard 1120. The video adapter 1060 is connected to, for example, a display 1130.
The hard disk drive 1031 stores, for example, an OS 1091, an application program 1092, a program module 1093, and program data 1094. That is, a program defining each of processing of the sound source separation filter information estimation device 10 and the sound source separation device 20 is implemented by the program module 1093 in which code that can be executed by the computer 1000 is written. The program module 1093 is stored in, for example, the hard disk drive 1031. For example, the program module 1093 for executing the same processing as that of a functional configuration in the sound source separation filter information estimation device 10 or the sound source separation device 20 is stored in the hard disk drive 1031. The hard disk drive 1031 may be replaced with a solid state drive (SSD).
Further, configuration data to be used in the processing of the embodiments described above is stored as the program data 1094 in, for example, the memory 1010 or the hard disk drive 1031. The CPU 1020 reads the program module 1093 or the program data 1094 stored in the memory 1010 or the hard disk drive 1031 into the RAM 1012 as necessary, and executes the program module 1093 or the program data 1094.
The program module 1093 or the program data 1094 is not limited to being stored in the hard disk drive 1031, and may be stored, for example, in a removable storage medium and read by the CPU 1020 via the disc drive 1041 or the like. Alternatively, the program module 1093 and the program data 1094 may be stored in another computer connected via a network (a local area network (LAN), a wide area network (WAN), or the like). The program module 1093 and the program data 1094 may be read from another computer via the network interface 1070 by the CPU 1020.
The embodiments to which the invention made by the present inventor has been applied have been described above, but the present invention is not limited to the description and the drawings, which form a part of the disclosure of the present invention according to the embodiments. That is, all other embodiments, examples, operation techniques, and the like made by those skilled in the art based on the embodiment are included in the scope of the present invention.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2019/032687 | 8/21/2019 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2021/033296 | 2/25/2021 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
9788119 | Vilermo | Oct 2017 | B2 |
10325615 | Koretzky | Jun 2019 | B2 |
10720174 | Ikeshita | Jul 2020 | B2 |
Entry |
---|
Kitamura et al., “Determined Blind Source Separation Unifying Independent Vector Analysis and Nonnegative Matrix Factorization”, IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 24, No. 9, Sep. 2016, pp. 1626-1641. |
Ikegita, “Independent Semi-Positive Constant Tensor Analysis for Multi-Channel Sound Source Separation”, Lectures by the Acoustical Society of Japan, Mar. 2018, 9 pages including English Translation. |
Number | Date | Country | |
---|---|---|---|
20220301570 A1 | Sep 2022 | US |