VOICE ACTIVITY DETECTION METHOD AND SYSTEM, AND VOICE ENHANCEMENT METHOD AND SYSTEM

Information

  • Patent Application
  • 20240046956
  • Publication Number
    20240046956
  • Date Filed
    September 19, 2023
    8 months ago
  • Date Published
    February 08, 2024
    3 months ago
Abstract
Microphone signals output by the microphone array satisfy a first model corresponding to a noise signal or a second model corresponding to a target voice signal mixed with a noise signal. A method and system may optimize the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determine a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model; and determine, by using a statistical hypothesis testing method, whether the microphone signals satisfy the first model or the second model, so as to determine whether the target voice signal is present in the microphone signals, determine a noise covariance matrix of the microphone signals, and further perform voice enhancement on the microphone signals.
Description
TECHNICAL FIELD

This disclosure relates to the field of target voice signal processing technologies, and in particular, to a voice activity detection method and system, and a voice enhancement method and system.


BACKGROUND

In a voice enhancement technology based on a beamforming algorithm, and especially in a minimum variance distortionless response (MVDR) adaptive beamforming algorithm, it is very important to solve a noise covariance matrix—a parameter describing a noise statistical feature relationship between different microphones. A main method in the existing technologies is calculating a noise covariance matrix based on a voice presence probability method, for example, estimating a voice presence probability by using a voice activity detection (VAD) method, and then calculating the noise covariance matrix. However, the accuracy of estimating the voice presence probability in existing technologies is not high enough, resulting in low accuracy in estimating the noise covariance matrix, and further causing a poor voice enhancement effect of the MVDR algorithm. Especially when a quantity of microphones is small, for example, less than 5, the effect deteriorates sharply. Therefore, the MVDR algorithm in the existing technologies is mainly used in a microphone array device having a large quantity of microphones with large spaces therebetween, such as a mobile phone and a smart speaker, but the voice enhancement effect is poor for a device having a small quantity of microphones with small spaces therebetween, such as a head phone.


Therefore, a voice activity detection method and system, and a voice enhancement method and system having higher accuracy need to be provided.


SUMMARY

This disclosure provides a voice activity detection method and system, and a voice enhancement method and system having higher accuracy.


According to a first aspect, this disclosure provides a voice activity detection system, including at least one storage medium storing a set of instructions for voice activity detection; and at least one processor in communication with the at least one storage medium, where during a process of voice activity detection for M microphones distributed in a preset array shape, where M is an integer greater than 1, the at least one processor executes the set of instructions to: obtain microphone signals output by the M microphones, where the microphone signals satisfy a first model corresponding to absence of a target voice signal or a second model corresponding to presence of a target voice signal, optimize the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determine a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model, and determine, based on statistical hypothesis testing, a target model and a noise covariance matrix corresponding to the microphone signals, where the target model includes one of the first model and the second model, and the noise covariance matrix of the microphone signals is a noise covariance matrix of the target model.


According to a second aspect, this disclosure further provides a voice activity detection method, where the method is for M microphones distributed in a preset array shape, and M is an integer greater than 1, the voice activity detection method includes: obtaining microphone signals output by the M microphones, where the microphone signals satisfy a first model corresponding to absence of a target voice signal or a second model corresponding to presence of a target voice signal; optimizing the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determining a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model; and determining, based on statistical hypothesis testing, a target model and a noise covariance matrix corresponding to the microphone signals, where the target model includes one of the first model and the second model, and the noise covariance matrix of the microphone signals is a noise covariance matrix of the target model.


According to a third aspect, this disclosure further provides a voice enhancement system, including at least one storage medium storing a set of instructions for voice enhancement; and at least one processor in communication with the at least one storage medium, where during a process of voice enhancement for M microphones distributed in a preset array shape, where M is an integer greater than 1, the at least one processor executes the set of instructions to: obtain microphone signals output by the M microphones, determine target models of the microphone signals and noise covariance matrices of the microphone signals, where the noise covariance matrices of the microphone signals are noise covariance matrices of the target models, determine, based on an MVDR method and the noise covariance matrices of the microphone signals, filter coefficients corresponding to the microphone signals, and combine the microphone signals based on the filter coefficients, and output a target audio signal.


According to a fourth aspect, this disclosure further provides a voice enhancement method, where the voice enhancement method is for M microphones distributed in a preset array shape, and M is an integer greater than 1, the voice enhancement method includes: obtaining microphone signals output by the M microphones; determining target models of the microphone signals and noise covariance matrices of the microphone signals, where the noise covariance matrices of the microphone signals are noise covariance matrices of the target models; determining, based on an MVDR method and the noise covariance matrices of the microphone signals, filter coefficients corresponding to the microphone signals; and combining the microphone signals based on the filter coefficients, and outputting a target audio signal.


As can be known from the foregoing technical solutions, the voice activity detection method and system, and the voice enhancement method and system provided in this disclosure may be applied to a microphone array including a plurality of microphones. The microphone signals output by the microphone array satisfy a first model corresponding to a noise signal or a second model corresponding to a target voice signal mixed with a noise signal. To determine whether a target voice signal is present in the microphone signals, the method and system may optimize the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determine a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model; and determine, by using a statistical hypothesis testing method, whether the microphone signals satisfy the first model or the second model, so as to determine whether the target voice signal is present in the microphone signals, determine a noise covariance matrix of the microphone signals, and further perform voice enhancement on the microphone signals based on an MVDR method. The method and system may improve accuracy of noise covariance estimation, and further improve a voice enhancement effect.


Other functions of the voice activity detection method and system, and the voice enhancement method and system provided in this disclosure are partially listed in the following description. Based on the description, content described in the following figures and examples would be obvious to a person of ordinary skill in the art. The inventive aspects of the voice activity detection method and system, and the voice enhancement method and system provided in this disclosure may be fully explained by practicing or using the method, apparatus, and a combination thereof in the following detailed examples.





BRIEF DESCRIPTION OF DRAWINGS

To describe the technical solutions in the embodiments of this disclosure more clearly, the following briefly describes the accompanying drawings required for describing the embodiments. Apparently, the accompanying drawings in the following description show merely some exemplary embodiments of this disclosure, and a person of ordinary skill in the art may still derive other drawings from these drawings without creative efforts.



FIG. 1 is a schematic hardware diagram of a voice activity detection system according to some exemplary embodiments of this disclosure;



FIG. 2A is a schematic exploded structural diagram of an electronic device according to some exemplary embodiments of this disclosure;



FIG. 2B is a front view of a first housing according to some exemplary embodiments of this disclosure;



FIG. 2C is a top view of a first housing according to some exemplary embodiments of this disclosure;



FIG. 2D is a front view of a second housing according to some exemplary embodiments of this disclosure;



FIG. 2E is a bottom view of a second housing according to some exemplary embodiments of this disclosure;



FIG. 3 is a flowchart of a voice activity detection method according to some exemplary embodiments of this disclosure;



FIG. 4 is a schematic diagram of a complete observation signal according to some exemplary embodiments of this disclosure;



FIG. 5A is a schematic diagram of an incomplete observation signal according to some exemplary embodiments of this disclosure;



FIG. 5B is a schematic diagram of an incomplete observation signal rearrangement according to some exemplary embodiments of this disclosure;



FIG. 5C is a schematic diagram of an incomplete observation signal rearrangement according to some exemplary embodiments of this disclosure;



FIG. 6 is a flowchart of iterative optimization according to some exemplary embodiments of this disclosure;



FIG. 7 is a flowchart for determining a target model according to some exemplary embodiments of this disclosure; and



FIG. 8 is a flowchart of a voice enhancement method according to some exemplary embodiments of this disclosure.





DETAILED DESCRIPTION

The following description provides specific application scenarios and requirements of this disclosure, to enable a person skilled in the art to make and use content of this disclosure. Various partial modifications to the disclosed exemplary embodiments are obvious to a person skilled in the art. General principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of this disclosure. Therefore, this disclosure is not limited to the illustrated embodiments, but is to be accorded the widest scope consistent with the claims.


The terms used herein are only intended to describe specific exemplary embodiments and are not restrictive. For example, as used herein, singular forms “a”, “an”, and “the” may also include plural forms, unless otherwise explicitly specified in a context. When used in this disclosure, the terms “comprising”, “including”, and/or “containing” indicate presence of associated integers, steps, operations, elements, and/or components, but do not preclude presence of one or more other features, integers, steps, operations, elements, components, and/or groups or addition of other features, integers, steps, operations, elements, components, and/or groups to the system/method.


In view of the following description, these features and other features of this disclosure, operations and functions of related elements of structures, and economic efficiency in combining and manufacturing components may be significantly improved. All of these form a part of this disclosure with reference to the drawings. However, it should be understood that the drawings are only for illustration and description purposes and are not intended to limit the scope of this disclosure. It should also be understood that the drawings are not drawn to scale.


Flowcharts used in this disclosure show operations implemented by the system according to some exemplary embodiments of this disclosure. It should be understood that operations in the flowcharts may be implemented out of the order described herein. The operations may be implemented in a reverse order or simultaneously. In addition, one or more other operations may be added to the flowcharts, and one or more operations may be removed from the flowcharts.


For ease of description, the following first explains terms that will appear in this disclosure.


Statistical hypothesis testing: It is a method for inferring a population from samples based on an assumption in mathematical statistics. A specific method is: making a hypothesis on the population under study based on a requirement of a problem, and marking the hypothesis as an original hypothesis H0; selecting a suitable statistic, where the selection of the statistic should make a distribution thereof known when the original hypothesis H0 is true; and based on the measured samples, calculating a statistic value, performing a test based on a pre-given significance level, and making a decision on rejecting or accepting the original hypothesis H0. Common statistical hypothesis testing methods include u-testing, t-testing, X2 testing (chi-square testing), F-testing, rank sum testing, and the like.


Minimum variance distortionless response (MVDR): It is an adaptive beamforming algorithm based on a maximum signal to interference plus noise ratio (SI NR) criterion. The MVDR algorithm may adaptively minimize power of an array output in a desired direction while maximizing the signal to interference plus noise ratio. Its objective is to minimize a variance of a recorded signal. If a noise signal is uncorrelated with a desired signal, the variance of the recorded signal is a sum of variances of the desired signal and the noise signal. Therefore, the MVDR solution seeks to minimize this sum, thereby mitigating impact of the noise signal. Its principle is to choose an appropriate filter coefficient to minimize average power of the array output under a constraint that the desired signal is distortionless.


Voice activity detection: It is a process of segmenting a target voice signal into a voice period and a non-voice period.


Gaussian distribution: A normal distribution is also known as Gaussian distribution. A normal curve is bell-shaped, low at both ends, high in the middle, and left-right symmetric. Because the curve of the Gaussian distribution is bell-shaped, the curve is also often referred to as a bell-shaped curve. If a random variable X conforms to a with a mathematical expectation μ and a variance σ2, the normal distribution is denoted as N(μ, σ2). A probability density function of the random variable is an expected value μ of the normal distribution, which determines a position of the random variable. A standard deviation a of the random variable determines an amplitude of the distribution. When μ=0 and σ=1, the normal distribution is a standard normal distribution.



FIG. 1 is a schematic hardware diagram of a voice activity detection system according to some exemplary embodiments of this disclosure. The voice activity detection system may be applied to an electronic device 200.


In some exemplary embodiments, the electronic device 200 may be a wireless head phone, a wired head phone, or an intelligent wearable device, for example, a device having an audio processing function such as smart glasses, a smart helmet, or a smart watch. The electronic device 200 may also be a mobile device, a tablet computer, a notebook computer, a built-in apparatus of a motor vehicle, or the like, or any combination thereof. In some exemplary embodiments, the mobile device may include a smart household device, a smart mobile device, or the like, or any combination thereof. For example, the smart mobile device may include a mobile phone, a personal digital assistant, a game device, a navigation device, an ultra-mobile personal computer (UMPC), or the like, or any combination thereof. In some exemplary embodiments, the smart household device may include a smart television, a desktop computer, or the like, or any combination thereof. In some exemplary embodiments, the built-in apparatus of the motor vehicle may include a vehicle-mounted computer, a vehicle-mounted television, or the like.


In this disclosure, an example in which the electronic device 200 is a head phone is used for description. The head phone may be a wireless head phone, or may be a wired head phone. As shown in FIG. 1, the electronic device 200 may include a microphone array 220 and a computing apparatus 240.


The microphone array 222 may be an audio capture device of the electronic device 200. The microphone array 222 may be configured to obtain a local audio, and output microphone signals, that is, an electronic signal carrying audio information. The microphone array 222 may include M microphones 222 distributed in a preset array shape. M is an integer greater than 1. The M microphones 222 may be distributed evenly or unevenly. The M microphones 222 may output microphone signals. The M microphones 222 may output M microphone signals. Each microphone 222 corresponds to one microphone signal. The M microphone signals are collectively referred to as the microphone signals. In some exemplary embodiments, the M microphones 222 may be distributed linearly. In some exemplary embodiments, the M microphones 222 may be distributed in an array of another shape, such as a circular array or a rectangular array. For ease of description, the linear distribution of the M microphones 222 is used as an example for description in the following description. In some exemplary embodiments, M may be any integer greater than 1, such as 2, 3, 4, 5, or even greater. In some exemplary embodiments, due to a space limitation, M may be an integer greater than 1 and not greater than 5, for example, in a product such as a head phone. When the electronic device 200 is a head phone, a spacing between adjacent microphones 222 of the M microphones 222 may be 20 mm to 40 mm. In some exemplary embodiments, the spacing between adjacent microphones 222 may be smaller, for example, 10 mm to 20 mm.


In some exemplary embodiments, the microphone 222 may be a bone conduction microphone that directly captures human body vibration signals. The bone conduction microphone may include a vibration sensor, for example, an optical vibration sensor or an acceleration sensor. The vibration sensor may capture a mechanical vibration signal (for example, a signal generated by a vibration generated by a skin or a bone when a user speaks), and convert the mechanical vibration signal into an electrical signal. Herein, the mechanical vibration signal mainly refers to a vibration propagated by a solid. The bone conduction microphone captures, by touching the skin or bone of the user with the vibration sensor or a vibration component connected to the vibration sensor, a vibration signal generated by the bone or skin when the user generates sound, and converts the vibration signal into an electrical signal. In some exemplary embodiments, the vibration sensor may be an apparatus that is sensitive to a mechanical vibration but insensitive to an air vibration (that is, a capability of responding to the mechanical vibration by the vibration sensor exceeds a capability of responding to the air vibration by the vibration sensor). Since the bone conduction microphone may directly pick a vibration signal of a sound generation part, the bone conduction microphone may reduce impact of ambient noise.


In some exemplary embodiments, the microphone 222 may alternatively be an air conduction microphone that directly captures air vibration signals. The air conduction microphone captures an air vibration signal caused when the user generates sound, and converts the air vibration signal into an electrical signal.


In some exemplary embodiments, the M microphones 222 may be M bone conduction microphones. In some exemplary embodiments, the M microphones 2202 may alternatively be M air conduction microphones. In some exemplary embodiments, the M microphones 222 may include both bone conduction microphone(s) and air conduction microphone(s). Certainly, the microphone 222 may alternatively be another type of microphone, for example, an optical microphone, a microphone receiving a myoelectric signal.


The computing apparatus 240 may be in communication with the microphone array 220. The communication herein may be a communication in any form and capable of directly or indirectly receiving information. In some exemplary embodiments, the computing apparatus 240 and the microphone array 220 may transfer data to each other over a wireless communication connection. In some exemplary embodiments, the computing apparatus 240 and the microphone array 220 may alternatively transfer data to each other over a direct connection by using a wire. In some exemplary embodiments, the computing apparatus 240 may alternatively be connected directly to another circuit by using a wire and hence connected indirectly to the microphone array 220 to implement mutual data transfer. The direct connection between the computing apparatus 240 and the microphone array 220 by using a wire is used as an example for description in this disclosure.


The computing apparatus 240 may be a hardware device having a data information processing function. In some exemplary embodiments, the voice activity detection system may include the computing apparatus 240. In some exemplary embodiments, the voice activity detection system may be applied to the computing apparatus 240. In other words, the voice activity detection system may operate on the computing apparatus 240. The voice activity detection system may include a hardware device having a data information processing function and a program required to drive the hardware device to work. Certainly, the voice activity detection system may also be only a hardware device having a data processing capability or only a program executed by a hardware device.


The voice activity detection system may store data or an instruction(s) for performing a voice activity detection method described in this disclosure, and may execute the data and/or the instruction. When the voice activity detection system operates on the computing apparatus 240, the voice activity detection system may obtain the microphone signals from the microphone array 220 based on the communication, and execute the data or the instruction of the voice activity detection method described in this disclosure, so as to determine whether a target voice signal is present in the microphone signals. The voice activity detection method is described in other parts of this disclosure. For example, the voice activity detection method is described in the descriptions of FIG. 3 to FIG. 8.


As shown in FIG. 1, the computing apparatus 240 may include at least one storage medium 243 and at least one processor 242. In some exemplary embodiments, the electronic device 200 may further include a communications port 245 and an internal communications bus 241.


The internal communications bus 241 may connect different system components, including the storage medium 243, the processor 242, and the communications port 245.


The communications port 245 may be used for data communication between the computing apparatus 240 and the outside world. For example, the computing apparatus 240 may obtain the microphone signals from the microphone array 220 through the communications port 245.


The at least one storage medium 243 may include a data storage apparatus. The data storage apparatus may be a non-transitory storage medium, or may be a transitory storage medium. For example, the data storage apparatus may include one or more of a magnetic disk, a read-only memory (ROM), or a random access memory (RAM). When the voice activity detection system operates on the computing apparatus 240, the storage medium 243 may further include at least one instruction set stored in the data storage apparatus, where the instruction set is used to perform voice activity detection on the microphone signals. The instruction is computer program code. The computer program code may include a program, a routine, an object, a component, a data structure, a process, a module, or the like for performing the voice activity detection method provided in this disclosure.


The at least one processor 242 may be in communication with the at least one storage medium 243 via the internal communications bus 241. The communication connection may be a communication in any form and capable of directly or indirectly receiving information. The at least one processor 242 is configured to execute the at least one instruction set. When the voice activity detection system may run on the computing apparatus 240, the at least one processor 242 reads the at least one instruction set, and implements, based on the at least one instruction set, the voice activity detection method provided in this disclosure. The processor 242 may perform all steps included in the voice activity detection method. The processor 242 may be in a form of one or more processors. In some exemplary embodiments, the processor 242 may include one or more hardware processors, for example, a microcontroller, a microprocessor, a reduced instruction set computer (RISC), an application-specific integrated circuit (ASIC), an application-specific instruction set processor (ASIP), a central processing unit (CPU), a graphics processing unit (GPU), a physical processing unit (PPU), a microcontroller unit, a digital signal processor (DSP), a field programmable gate array (FPGA), an advanced RISC machine (ARM), a programmable logic device (PLD), any circuit or processor that may implement one or more functions, and the like, or any combination thereof. For illustration only, one processor 242 in the computing apparatus 240 is described in this disclosure. However, it should be noted that the computing apparatus 240 in this disclosure may further include a plurality of processors 242. Therefore, operations and/or method steps disclosed in this disclosure may be performed by one processor in this disclosure, or may be performed jointly by a plurality of processors. For example, if the processor 242 of the computing apparatus 240 in this disclosure performs step A and step B, it should be understood that step A and step B may also be performed jointly or separately by two different processors 242 (for example, the first processor performs step A, and the second processor performs step B, or the first processor and the second processor jointly perform step A and step B).



FIG. 2A is a schematic exploded structural diagram of an electronic device 200 according to some exemplary embodiments of this disclosure. As shown in FIG. 2A, the electronic device 200 may include a microphone array 220, a computing apparatus 240, a first housing 260, and a second housing 280.


The first housing 260 may be a mounting base of the microphone array 220. The microphone array 220 may be mounted inside the first housing 260. A shape of the first housing 260 may be adaptively designed based on a distribution shape of the microphone array 220. This is not limited in this disclosure. The second housing 280 may be a mounting base of the computing apparatus 240. The computing apparatus 240 may be mounted in the second housing 280. A shape of the second housing 280 may be adaptively designed based on a shape of the computing apparatus 240. This is not limited in this disclosure. When the electronic device 200 is a head phone, the second housing 280 may be connected to a wearing part. The second housing 280 may be connected to the first housing 260. As described above, the microphone array 220 may be electrically connected to the computing apparatus 240. Specifically, the microphone array 220 may be electrically connected to the computing apparatus 240 through the connection of the first housing 260 and the second housing 280.


In some exemplary embodiments, the first housing 260 may be fixedly connected, for example, integrated, welded, riveted, or bonded, to the second housing 280. In some exemplary embodiments, the first housing 260 may be detachably connected to the second housing 280. The computing apparatus 240 may be in communication with different microphone arrays 220. Specifically, a difference between the different microphone arrays 220 may lie in different quantities of microphones 222 in the microphone arrays 220, different array shapes, different spacings between the microphones 222, different mounting angles of the microphone arrays 220 in the first housing 260, different mounting positions of the microphone arrays 220 in the first housing 260, or the like. Depending on different application scenarios, the user may change corresponding microphone arrays 220, so that the electronic device 200 may be applied to a wider range of scenarios. For example, when the user is closer to the electronic device 200 in an application scenario, the user may replace the microphone array 220 with a microphone array 220 having a smaller microphone spacing. In another example, when the user is closer to the electronic device 200 in an application scenario, the user may replace the microphone array 220 with a microphone array 220 having a larger microphone spacing and a larger microphone quantity.


The detachable connection may be a physical connection in any form, such as a threaded connection, a snap connection, or a magnetic connection. In some exemplary embodiments, there may be a magnetic connection between the first housing 260 and the second housing 280. To be specific, the first housing 260 and the second housing 280 are detachably connected to each other by a magnetic apparatus.



FIG. 2B is a front view of the first housing 260 according to some exemplary embodiments of this disclosure. FIG. 2C is a top view of the first housing 260 according to some exemplary embodiments of this disclosure. As shown in FIG. 2B and FIG. 2C, the first housing 260 may include a first interface 262. In some exemplary embodiments, the first housing 260 may further include contacts 266. In some exemplary embodiments, the first housing 260 may further include an angle sensor (not shown in FIG. 2B and FIG. 2C).


The first interface 262 may be a mounting interface of the first housing 260 and the second housing 280. In some exemplary embodiments, the first interface 262 may be circular. The first interface 262 may be rotatably connected to the second housing 280. When the first housing 260 is mounted on the second housing 280, the first housing 260 may be rotated relative to the second housing 280 to adjust an angle of the first housing 260 relative to the second housing 280, thereby adjusting an angle of the microphone array 220.


A first magnetic apparatus 263 may be disposed on the first interface 262. The first magnetic apparatus 263 may be disposed at a position of the first interface 262 close to the second housing 280. The first magnetic apparatus 263 may generate magnetic adherence to achieve a detachable connection to the second housing 280. When the first housing 260 approaches the second housing 280, the first housing 260 may be quickly connected to the second housing 280 by the adherence. In some exemplary embodiments, after the first housing 260 is connected to the second housing 280, the first housing 260 may also be rotated relative to the second housing 280 to adjust the angle of the microphone array 220. Due to the adherence, the connection between the first housing 260 and the second housing 280 may still be maintained while the first housing 260 is rotated relative to the second housing 280.


In some exemplary embodiments, a first positioning apparatus (not shown in FIG. 2B and FIG. 2C) may also be disposed on the first interface 262. The first positioning apparatus may be an externally protruding positioning step or an internally extending positioning hole. The first positioning apparatus may cooperate with the second housing 280 to implement quick mounting of the first housing 260 and the second housing 280.


As shown in FIG. 2B and FIG. 2C, in some exemplary embodiments, the first housing 260 may further include contacts 266. The contacts 266 may be mounted on the first interface 262. The contacts 266 may protrude externally from the first interface 262. The contacts 266 may be elastically connected to the first interface 262. The contacts 266 may be in communication with the M microphones 222 in the microphone array 220. The contacts 266 may be made of an elastic metal to implement data transmission. When the first housing 260 is connected to the second housing 280, the microphone array 220 may be in communication with the computing apparatus 240 through the contacts 266. In some exemplary embodiments, the contacts 266 may be distributed in a circular shape. When the first housing 260 is rotated relative to the second housing 280 after the first housing 260 is connected to the second housing 280, the contacts 266 may also rotate relative to the second housing 280 and maintain a communication connection to the computing apparatus 240.


In some exemplary embodiments, an angle sensor (not shown in FIG. 2B and FIG. 2C) may be further disposed on the first housing 260. The angle sensor may be in communication with the contacts 266, thereby implementing a communication connection to the computing apparatus 240. The angle sensor may collect angle data of the first housing 260 to determine an angle at which the microphone array 220 is located, to provide reference data for subsequent calculation of a voice presence probability.



FIG. 2D is a front view of the second housing 280 according to some exemplary embodiments of this disclosure. FIG. 2E is a bottom view of the second housing 280 according to some exemplary embodiments of this disclosure. As shown in FIG. 2D and FIG. 2E, the second housing 280 may include a second interface 282. In some exemplary embodiments, the second housing 280 may further include a guide rail 286.


The second interface 282 may be a mounting interface of the second housing 280 and the first housing 260. In some exemplary embodiments, the second interface 282 may be circular. The second interface 282 may be rotatably connected to the first interface 262 of the first housing 260. When the first housing 260 is mounted on the second housing 280, the first housing 260 may be rotated relative to the second housing 280 to adjust the angle of the first housing 260 relative to the second housing 280, thereby adjusting the angle of the microphone array 220.


A second magnetic apparatus 283 may be disposed on the second interface 282. The second magnetic apparatus 283 may be disposed at a position of the second interface 282 close to the first housing 260. The second magnetic apparatus 283 may generate magnetic adherence to achieve a detachable connection to the first interface 262. The second magnetic apparatus 283 may be used in cooperation with the first magnetic apparatus 263. When the first housing 260 approaches the second housing 280, the first housing 260 may be quickly mounted on the second housing 280 by the adherence between the second magnetic apparatus 283 and the first magnetic apparatus 263. When the first housing 260 is mounted on the second housing 280, a position of the second magnetic apparatus 283 is opposite to a position of the first magnetic apparatus 263. In some exemplary embodiments, after the first housing 260 is connected to the second housing 280, the first housing 260 may also be rotated relative to the second housing 280 to adjust the angle of the microphone array 220. Under the adherence, the connection between the first housing 260 and the second housing 280 may still be maintained while the first housing 260 is rotated relative to the second housing 280.


In some exemplary embodiments, a second positioning apparatus (not shown in FIG. 2D and FIG. 2E) may also be disposed on the second interface 282. The second positioning apparatus may be an externally protruding positioning step or an internally extending positioning hole. The second positioning apparatus may cooperate with the first positioning apparatus of the first housing 260 to implement quick mounting of the first housing 260 and the second housing 280. When the first positioning apparatus is the positioning step, the second positioning apparatus may be the positioning hole. When the first positioning apparatus is the positioning hole, the second positioning apparatus may be the positioning step.


As shown in FIG. 2D and FIG. 2E, in some exemplary embodiments, the second housing 280 may further include a guide rail 286. The guide rail 286 may be mounted on the second interface 282. The guide rail 286 may be in communication with the computing apparatus 240. The guide rail 286 may be made of a metal material to implement data transmission. When the first housing 260 is connected to the second housing 280, the contacts 266 may contact the guide rail 286 to form a communication connection, to implement the communication between the microphone array 220 and the computing apparatus 240 and implement data transmission. As described above, the contacts 266 may be elastically connected to the first interface 262. Therefore, after the first housing 260 is connected to the second housing 280, the contacts 266 may contact the guide rail 286 under elastic force of the elastic connection, so that a reliable communication may be implemented. In some exemplary embodiments, the guide rail 286 may be distributed in a circular shape. When the first housing 260 is rotated relative to the second housing 280 after the first housing 260 is connected to the second housing 280, the contacts 266 may also rotate relative to the guide rail 286 and maintain a communication connection to the guide rail 286. FIG. 3 is a flowchart of a voice activity detection method P100 according to some exemplary embodiments of this disclosure. The method P100 may determine whether a target voice signal is present in microphone signals. Specifically, a processor 242 may perform the method P100. As shown in FIG. 3, the method P100 may include the following steps.


S120. Obtain microphone signals output by M microphones 222.


As described above, each microphone 222 may output a corresponding microphone signals. The M microphones 222 correspond to M microphone signals. When determining whether a target voice signal is present in the microphone signals, the method P100 may include calculation based on all of the M microphone signals or calculation based on a part of the microphone signals. Therefore, the microphone signals may include the M microphone signals corresponding to the M microphones 222 or a part of microphone signals. In the subsequent description of this disclosure, an example in which the microphone signals may include the M microphone signals corresponding to the M microphones 222 is used for description.


In some exemplary embodiments, the microphone signal may be a time domain signal. In some exemplary embodiments, in step S120, a computing apparatus 240 may perform frame division and windowing processing on the microphone signal to divide the microphone signal into a plurality of continuous audio signals. In some exemplary embodiments, in step S120, the computing apparatus 240 may further perform a time-frequency transform on the microphone signal to obtain a frequency domain signal of the microphone signal. For ease of description, a microphone signal at any frequency is marked as X. In some exemplary embodiments, the microphone signal X may include K frames of continuous audio signals. K is any positive integer greater than 1. For ease of description, a kth frame of microphone signal is marked as xk. The kth frame of microphone signal xk may be represented by the following formula:






x
k
=[x
1,k
,x
2,k
, . . . ,x
M,k]T  formula (1)


The kth frame of microphone signal xk may be an M-dimensional signal vector formed by M microphone signals. The microphone signal X may be represented by an M×K data matrix. The microphone signal X may be represented by the following formula:









X
=


[


x
1

,

x
2

,


,

x
K


]

=

[




x

1
,
1








x

1
,
K


















x

M
,
1








x

M
,
K





]






formula



(
2
)








where the microphone signal X is an M×K data matrix; an mth row in the data matrix represents a microphone signal received by an mth microphone; and a kth column represents the kth frame of microphone signal.


As described above, the microphone 222 may capture noise in an ambient environment and output a noise signal, and may also capture a voice of a target user and output a target voice signal. When the target user does not speak, the microphone signal includes only the noise signal. When the target user speaks, the microphone signal includes the target voice signal and the noise signal. The kth frame of microphone signal xk may be represented by the following formula:






x
k
=Ps
k
+d
k  formula (3)


where k=1, 2, K; dk is a noise signal in the kth frame of microphone signal xk; sk is an amplitude of the target voice signal; and P is a target steering vector of the target voice signal.


The microphone signal X may be represented by the following formula:






X=[x
1
,x
2
, . . . ,x
K
]=PS+D  formula (4)


where S is the amplitude of the target voice signal; S=[s1,s2, . . . ,sK]; D is the noise signal; and D=[d1,d2, . . . ,dK].


The noise signal dk may be represented by the following formula:






d
k
=[d
1,k
,d
2,k
, . . . ,d
M,k]T  formula (5)


The noise signal dk in the kth frame of microphone signal xk may be an M-dimensional signal vector formed by M microphone signals.


In some exemplary embodiments, the noise signal dk may include at least a colored noise signal ck. In some exemplary embodiments, the noise signal dk may further include at least a white noise signal nk. The noise signal dk may be represented by the following formula:






d
k
=c
k
+n
k  formula (6)


In this case, the noise signal D=C+N. C is the colored noise signal, and C=[c1, c2, . . . , cK]. N is the white noise signal, and N=[n1, n2, . . . , nK].


The computing apparatus 240 may use a unified mapping relationship between a cluster feature of a sound source spatial distribution of the noise signal dk and a parameter of the microphone array 220 to establish a parameterized cluster model, and perform clustering on a sound source of the noise signal dk to divide the noise signal dk into the colored noise signal ck and the white noise signal nk.


In some exemplary embodiments, the noise signal D conforms to a Gaussian distribution. The noise signal dk˜CN (0, M). M is a noise covariance matrix of the noise signal dk. The colored noise signal ck conforms to a zero-mean Gaussian distribution, that is, ck˜CN(0, Mc). The noise covariance matrix Mc corresponding to the colored noise signal ck has a low-rank feature and is a low-rank semi-positive definite matrix. The white noise signal nk also conforms to a zero-mean Gaussian distribution, that is, nk˜CN(0, M e). Power of the white noise signal nk is 66. Mn02In, that is, nk˜CN(0,δ02). The noise covariance matrix M of the noise signal dk may be represented by the following formula:






M=M
c
+M
n
=M
c02In  formula (7)


The noise covariance matrix M of the noise signal dk may be decomposed into a sum of an identity matrix In and the low-rank semi-positive definite matrix Mc.


In some exemplary embodiments, the power δ02 of the white noise signal nk may be prestored in the computing apparatus 240. In some exemplary embodiments, the power δ02 of the white noise signal nk may be estimated in advance by the computing apparatus 240. For example, the computing apparatus 240 may estimate the power δ02 of the white noise signal nk based on minimum tracking, a histogram, or the like. In some exemplary embodiments, the computing apparatus 240 may estimate the power δ02 of the white noise signal nk based on the method P100.


sk is a complex amplitude of the target voice signal. In some exemplary embodiments, a target voice signal source is present around the microphone 222. In some exemplary embodiments, there are L target voice signal sources around the microphone 222. In this case, sk may be an L×1-dimensional vector.


The target steering vector P is an M×L-dimensional matrix. The target steering vector P may be represented by the following formula:









P
=

[



1





1





e


-
j


2

π


f
0




dcos


θ
1


c









e


-
j


2

π


f
0




dcos


θ
L


c



















e


-
j


2

π



f
0

(

M
-
1

)




dcos


θ
1


c









e


-
j


2

π



f
0

(

M
-
1

)




dcos


θ
L


c






]





formula



(
8
)








where f0 is a carrier frequency; d is a spacing between adjacent microphones 222; c is a speed of sound; and θ1, . . . , θN are incident angles between L target voice signal sources and microphones 222 respectively. In some exemplary embodiments, angles of the target voice signal source sk are generally distributed in a group of specific angle ranges. Therefore, θ1, . . . , θN are known. Relative position relationships, such as relative distances, or relative coordinates, of the M microphones 222 are prestored in the computing apparatus 240. In other words, the spacing d between adjacent microphones 222 is prestored in the computing apparatus 240.



FIG. 4 is a schematic diagram of a complete observation signal according to some exemplary embodiments of this disclosure. In some exemplary embodiments, the microphone signal X is a complete observation signal, as shown in FIG. 4. All data in the M×K data matrix in the complete observation signal is complete. As shown in FIG. 4, a horizontal direction is a frame number k of the microphone signal X and a vertical direction is a microphone signal number m in the microphone array 220. The mth row represents the microphone signal received by the mth microphone 222, and the kth column represents the kth frame of microphone signal.



FIG. 5A is a schematic diagram of an incomplete observation signal according to some exemplary embodiments of this disclosure. In some exemplary embodiments, the microphone signal X is an incomplete observation signal, as shown in FIG. 5A. Some data in the M×K data matrix is missing in the incomplete observation signal. The computing apparatus 240 may rearrange the incomplete observation signal. As shown in FIG. 5A, a horizontal direction is a frame number k of the microphone signal X and a vertical direction is a microphone signal channel number m. The mth row represents the microphone signal received by the mth microphone 222, and the kth column represents the kth frame of microphone signal.


When the microphone signal X is the incomplete observation signal, step S120 may further include rearranging the incomplete observation signal. FIG. 5B is a schematic diagram of an incomplete observation signal rearrangement according to some exemplary embodiments of this disclosure. FIG. 5C is a schematic diagram of an incomplete observation signal rearrangement according to some exemplary embodiments of this disclosure. That the computing apparatus 240 rearranges the incomplete observation signal may include: the computing apparatus 240 obtaining the incomplete observation signal; and the computing apparatus 240 performing row-column permutation on the microphone signal X based on a position of missing data in each column in the M×K data matrix, and dividing the microphone signal X into at least one sub microphone signal, where the microphone signal X includes the at least one sub microphone signal.


In the incomplete observation signal, because positions of missing data in the microphone signals xk with different frame numbers may be the same, to reduce a calculation amount and calculation time of the algorithm, the computing apparatus 240 may classify K frames of microphone signals X based on the positions of missing data in the microphone signals xk with different frame numbers, classify microphone signals xk with same positions of missing data into a same sub microphone signal, and perform permutation on row positions in the data matrix of the microphone signal X, so that positions of the microphone signals in the same sub microphone signal are adjacent, as shown in FIG. 5B. The K frames of microphone signal X are classified into at least one sub microphone signal. For ease of description, the quantity of at least one sub microphone signal is defined as G, where G is a positive integer not less than 1. A gth sub microphone signal is defined as Xg, where g=1, 2, . . . , G.


The calculation apparatus 240 may further perform row permutation on the microphone signal X based on the position of the missing data in each sub microphone signal Xg, so that positions of missing data in all sub microphone signals are adjacent, as shown in FIG. 5C.


In summary, in the incomplete observation signal, the sub microphone signal Xg may be represented by the following formula:






X
g
=P
g
S
g
+D
g  formula (9)


where Xg=QgXBgT, Dg=QgDBgT, Pg=QgP, Sg=BgS. Matrices Qg and Bg are matrices formed by elements 0 and 1 and determined by the position of the missing data.


The microphone signal X may be represented by the following formula:






X=[X
1
,X
2
, . . . ,X
G]  formula (10)


For ease of description, in the following description, the microphone signal X is described as an incomplete observation signal.


As described above, the microphone 222 may capture both the noise signal D and the target voice signal. When the target voice signal is absent in the microphone signal X, the microphone signal X satisfies a first model corresponding to the noise signal D. When the target voice signal is present in the microphone signal X, the microphone signal satisfies a second model corresponding to the target voice signal mixed with the noise signal D.


For ease of description, the first model is defined by the following formula:






X=D  formula (11)


When the microphone signal X is a complete observation signal, the first model may be represented by the following formula:






x
k
=d
k  formula (12)


When the microphone signal X is an incomplete observation signal, the first model may be represented by the following formula:






X
g
=D
g  formula (13)


The second model is defined as the following formula:






X=PS+D  formula (14)


When the microphone signal X is a complete observation signal, the second model may be represented by the following formula:






x
k
=Ps
k
+d
k  formula (15)


When the microphone signal X is an incomplete observation signal, the second model may be represented by the following formula:






X
g
=P
g
S
g
+D
g  formula (16)


For ease of presentation, in the following description, an example in which the microphone signal X is an incomplete observation signal is used for description.


As shown in FIG. 3, the method P100 may further include:


S140. Optimize the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determine a first estimate {circumflex over (M)}1 of a noise covariance matrix M1 of the first model and a second estimate {circumflex over (M)}2 of a noise covariance matrix M2 of the second model.


A noise covariance matrix M of an unknown parametric noise signal D is present in the first model. For ease of description, the noise covariance matrix M of the unknown parametric noise signal D in the first model is defined as M1. The noise covariance matrix M of the unknown parametric noise signal D and the amplitude S of the target voice signal are present in the second model. For ease of description, the noise covariance matrix M of the unknown parametric noise signal D in the second model is defined as M2. The computing apparatus 240 may optimize the first model and the second model respectively based on an optimization method, determine the first estimate {circumflex over (M)}1 of the unknown parameter M1, the second estimate {circumflex over (M)}2 of M2, and an estimate S of the amplitude S of the target voice signal.


On the one hand, the computing apparatus 240 may be triggered from a perspective of the likelihood function to optimize the design of the first model and the second model respectively by using maximization of the likelihood function as an optimization objective. On the other hand, as described above, because the noise covariance matrix Mc corresponding to the colored noise signal ck has a low-rank feature and is a low-rank semi-positive definite matrix, the noise covariance matrix M of the noise signal dk may also have a low-rank feature. Especially for the incomplete observation signal, the low-rank feature of the noise covariance matrix M of the noise signal dk needs to be maintained in the process of rearranging the incomplete observation signal. Therefore, the computing apparatus 240 may optimize the design of the first model and the second model respectively based on the low-rank feature of the noise covariance matrix M of the noise signal dk by using rank minimization of the noise covariance matrix M as an optimization objective. Therefore, the computing apparatus 240 may optimize the first model and the second model respectively by using maximization of the likelihood function and rank minimization of the noise covariance matrix as joint optimization objectives, to determine the first estimate custom-character of the unknown parameter M1, the second estimate custom-character of M2, and the estimate S of the amplitude S of the target voice signal.



FIG. 6 is a flowchart of iterative optimization according to some exemplary embodiments of this disclosure. FIG. 6 shows step S140. As shown in FIG. 6, step S140 may include:


S142. Establish a first likelihood function L1(M1) corresponding to the first model by using the microphone signal X as sample data.


The likelihood function includes the first likelihood function L1(M1). Based on the formulas (11) to (13), the first likelihood function L1(M1) may be represented by the following formula:









{






L
1

(

M
1

)

=


f
1

(


x
1

,

x
2

,


,


x
K




)









L
1

(

M
1

)

=


f
1

(


X
1

,

X
2

,


,


X
G




)









formula



(
17
)








The formula (17) represents the first likelihood function L1(M1) for the complete observation signal and the incomplete observation signal respectively. custom-character represents a maximum likelihood estimate of the parameter M1. f1(x1, x2, . . . , xK|custom-character) and f1(X1, X2, . . . , XG|custom-character) represent a probability of occurrence of the microphone signal X after the parameter custom-character is given in the first model.


S144. Optimize the first model by using maximization of the first likelihood function L1(M1) and rank Rank(M1) minimization of the noise covariance matrix M1 of the first model as optimization objectives, and determine the first estimate custom-character of M1.


The maximization of the first likelihood function L1(M1) may be represented by min(−log(L1(M1))). The rank Rank(M1) minimization of the noise covariance matrix M1 of the first model may be represented by min(Rank(M1)). As described above, the known noise covariance matrix δ02In of the white noise signal nk is used as an example for description. It can be learned from the formula (7) that the rank Rank(M1) minimization of the noise covariance matrix M1 of the first model may be represented by minimization min(Rank(Mc)) of the noise covariance matrix Mc of the colored noise signal C. Therefore, an objective function of the optimization objective may be represented by the following formula:





min(−log(L1(M1))+γRank(Mc))  formula (18)


where γ is a regularization coefficient. Because minimization of the matrix rank may be relaxed to a nuclear norm minimization, the formula (18) may be represented by the following formula:





min(−log(L1(M1))+γ∥Mc*)  formula (19)


An iteration constraint of the first model may be represented by the following formula:









s
.
t
.






M
c


0







M
1

=


M
c

+

M
n










formula



(
20
)








where Mc≥0 is a positive definite constraint of the noise covariance matrix Mc of the colored noise signal C. The optimization problem of the first model may be represented by the following formula:










min

(


-

log

(


L
1

(

M
1

)

)


+

γ





M
c



*



)




s
.
t
.






M
c


0







M
1

=


M
c

+

M
n











formula



(
21
)








After determining the objective function and the constraint, the computing apparatus 240 may iteratively optimize the unknown parameter M1 of the first model by using the objective function as an optimization objective to determine the first estimate custom-character of the noise covariance matrix M1 of the first model.


The formula (21) is a semi-positive definite programming problem that can be solved by the computing apparatus 240 by using a variety of algorithms. For example, a gradient projection algorithm may be used. Specifically, in each iteration in the gradient projection algorithm, the formula (19) is first solved by using a gradient method without any constraint, and then a resulting solution is projected onto a semi-positive definite cone, so that the cone satisfies the matrix semi-positive definite constraint formula (20).


As shown in FIG. 6, step S140 may further include:


S146. Establish a second likelihood function L2(S, M2) corresponding to the second model by using the microphone signal X as sample data.


The likelihood function includes the second likelihood function L2(S, M2). Based on the formulas (14) to (16), the second likelihood function L2(S, M2) may be represented by the following formula:









{






L
2

(

S
,

M
2


)

=


f
2

(


x
1

,

x
2

,


,


x
K



S
^


,


M
2

^


)









L
2

(

S
,

M
2


)

=


f
2

(


X
1

,

X
2

,


,


X
G



S
^


,


M
2

^


)









formula



(
22
)








where the formula (22) represents the second likelihood function for the complete observation signal and the incomplete observation signal respectively. Ŝ and custom-character represent maximum likelihood estimates of parameters S and M2. f2(x1, x2, . . . , xK|custom-character) and f2(X1, X2, . . . , XG|Ŝ,custom-character) represent probabilities of occurrence of the microphone signal X after the parameters S and M2 are given.


S148. Optimize the second model by using maximization of the second likelihood function L2(S, M2) and rank Rank(M2) minimization of the noise covariance matrix M2 of the second model as optimization objectives, and determine the second estimate custom-character and the estimate S of the amplitude S of the target voice signal.


The maximization of the second likelihood function L2(S, M2) may be represented by min(−log(L2(S, M2))). The rank Rank(M2) minimization of the noise covariance matrix M2 of the second model may be represented by min(Rank(M2)). As described above, the known noise covariance matrix δ02In of the white noise signal nk is used as an example for description. It can be learned from the formula (7) that the rank Rank(M2) minimization of the noise covariance matrix M2 of the second model may be represented by minimization min(Rank(Mc)) of the noise covariance matrix Mc of the colored noise signal C. Therefore, an objective function of the optimization objective may be represented by the following formula:





min(−log(L2(S,M2))+γRank(Mc))  formula (23)


where γ is a regularization coefficient. Because minimization of the matrix rank may be relaxed to a nuclear norm minimization problem, the formula (23) may be represented by the following formula:





min(−log(L2(S,M2))+γ∥Mc*)  formula (24)


An iteration constraint of the second model may be represented by the following formula:









s
.
t
.






M
c


0







M
2

=


M
c

+

M
n










formula



(
25
)








where Mc≥0 is a positive definite constraint of the noise covariance matrix Mc of the colored noise signal C. The optimization problem of the second model may be represented by the following formula:










min

(


-

log

(


L
2

(

S
,

M
2


)

)


+

γ





M
c



*



)




s
.
t
.










M
c


0







M
2

=


M
c

+

M
n











formula



(
26
)








After determining the objective function and the constraint, the computing apparatus 240 may iteratively optimize the unknown parameter M2 of the second model by using the objective function as an optimization objective to determine the second estimate custom-character the noise covariance matrix M2 of the second model and the estimate S of the amplitude S of the target voice signal.


The formula (26) is a semi-positive definite programming problem that can be solved by the computing apparatus 240 by using a variety of algorithms. For example, a gradient projection algorithm may be used. Specifically, in each iteration of the gradient projection algorithm, the formula (24) is first solved by using the gradient method without any constraint, and then a resulting solution is projected onto a semi-positive definite cone, so that the cone satisfies the matrix semi-positive definite constraint formula (25).


In summary, the method P100 may optimize the first model and the second model respectively by using maximization of the likelihood function and rank minimization of the noise covariance matrix as joint optimization objectives, to determine the first estimate custom-character of the unknown parameter M1 and the second estimate custom-character of M2, so that estimation accuracy of M1 and M2 are higher, to provide a higher-accuracy data model for subsequent statistical hypothesis testing, thereby improving accuracy of voice activity detection and a voice enhancement effect.


As shown in FIG. 3, the method P100 may further include:


S160. Determine, based on statistical hypothesis testing, a target model and a noise covariance matrix M corresponding to the microphone signal X.


The target model includes one of the first model and the second model. The noise covariance matrix M of the microphone signal X is the noise covariance matrix of the target model. When the target model of the microphone signal X is the first model, the noise covariance matrix M of the microphone signal X is equal to custom-character When the target model of the microphone signal X is the second model, the noise covariance matrix M of the microphone signal X is equal to custom-character.


The computing apparatus 240 may determine, based on the statistical hypothesis testing method, whether the microphone signal X satisfies the first model or the second model, and therefore determine whether the target voice signal is present in the microphone signal X.



FIG. 7 is a flowchart for determining the target model according to some exemplary embodiments of this disclosure. The flowchart shown in FIG. 7 is step S160. As shown in FIG. 7, step S160 may include:


S162. Establish a binary hypothesis testing model based on the microphone signal X.


An original hypothesis H0 of the binary hypothesis testing model may be that the target voice signal is absent in the microphone signal X, that is, the microphone signal X satisfies the first model. An alternative hypothesis H1 of the binary hypothesis testing model may be that the target voice signal is present in the microphone signal X, that is, the microphone signal satisfies the second model. The binary hypothesis testing model may be represented by the following formula:









{






H
0

:


x
k


=

d
k






k
=
1

,
2
,


,
K








H
1

:


x
k


=


Ps
k

+

d
k







k
=
1

,
2
,


,
K








formula



(
27
)













{






H
0

:


X
g


=

D
g






g
=
1

,
2
,


,
G








H
1

:


X
g


=



P
g



S
g


+

D
g







g
=
1

,
2
,


,
G








formula



(
28
)








where the microphone signal X in the formula (27) is a complete observation signal; and the microphone signal X in the formula (28) is an incomplete observation signal.


S164. Substitute the first estimate custom-character, the second estimate custom-character, and the estimate Ŝ of the amplitude S into a decision criterion of a detector of the binary hypothesis testing model to obtain a test statistic ψ.


The detector may be any one or more detectors. In some exemplary embodiments, the detector may be one or more of a GLRT detector, a Rao detector, and a Wald detector. In some exemplary embodiments, the detector may alternatively be a u-detector, a t-detector, a χ2 detector (a chi-square detector), an F-detector, a rank-sum detector, and the like. Different detectors have different test statistics tp.


The GLRT detector (Generalized Likelihood Ratio Test, generalized likelihood ratio test detector) is used as an example for description. When the microphone signal X is a complete observation signal, in the GLRT detector, the test statistic m be represented b the following formula:









Ψ
=



f

H
1


(


x
1

,

x
2

,


,


x
K



,

)



f

H
0


(


x
1

,

x
2

,


,


x
K




)






formula



(
29
)








where fH0 (x1, x2, . . . ,xK|custom-character) and fH1 (x1, x2, . . . , xK|custom-character,custom-character) are respectively likelihood functions under the original hypothesis H0 and the alterative hypothesis H1; fH1 (x1, x2, . . . , xK|custom-character,custom-character)=f2(x1,x2, . . . ,xK|custom-character); and fH0(x1, x2, . . . , xK|custom-character)=f1(x1, x2, . . . , xK|custom-character).


When the microphone signal X is an incomplete observation signal, in the GLRT detector, the test statistic ψ may be represented by the following formula:









ψ
=



f

H
1


(


X
1

,

X
2

,


,


X
G



,

)



f

H
0


(


X
1

,

X
2

,


,


X
G




)






formula



(
30
)








where fH0(X1, X2, . . . ,XG|custom-character) and fH1(X1, X2, . . . ,XG|custom-character,custom-character) are respectively likelihood functions under the origin hypothesis H0 and the alternative hypothesis H1; fH1(X1,X2, . . . ,XG|custom-character,custom-character)=f2(X1,X2, . . . ,XG|,Ŝ,custom-character and fH0(X1, X2, . . . , XG|custom-character)=f1(X1, X2, . . . , XG|custom-character).


In the GLRT detector, unknown parameters custom-character, custom-character, and custom-character under the original hypothesis H0 and the alternative hypothesis H1 all need to be estimated, so there are many parameters to be estimated. The Rao detector only needs to custom-characterestimate the unknown parameter custom-character under the original hypothesis H0. When the quantity of frames is K, the Rao detector has the same detection performance as the GLRT detector. However, when the quantity K of frames is limited, the Rao detector cannot achieve the same detection performance as the GLRT detector, but the Rao detector has advantages of simpler calculation and being more suitable for cases in which it is difficult to solve unknown parameters under the alternative hypothesis H1.


Therefore, in view of requirements of an actual system for balancing detection performance and calculation complexity, the Rao detector is proposed for the computing apparatus 240 on a basis of the foregoing GLRT detector. Using the incomplete observation signal as an example, the test statistic ψ of the Rao detector may be represented by the following formula:









ψ
=



[


J

-
1


(
)

]



θ
r

,

θ
r








formula



(
31
)








where f(X1, X2, . . . , XG|θ, M) represents a probability density function under the alternative hypothesis H1; M=M2;


θr=[PSR,1, PSR,2, . . . , PSR,M, PSL,1, PSL,2, . . . , PSL,M]T, where PSR,m is a real part of an amplitude of a target voice signal in an audio signal of the mth microphone 222, PSL,m is an imaginary part of the amplitude of the target voice signal in the audio signal of the mth microphone 222, and m=1,2, . . . , M; θr is a 2M-dimensional vector; and θ=[θrTθST]T, where θs is a real vector containing redundant parameters, including real and imaginary parts of M off-diagonal elements and elements on diagonals. The formula (31) may be simplified to the following formula:





ψ=ΣgGXgHMg−1Pg[PgHMg−1Pg]−1PgHMg−1Xg  formula (32)


where Mg=QgMQgT.


In the formula (32), the test statistic ψ of the Rao test may be obtained as long as the estimate custom-character of the unknown parameter custom-character under the original hypothesis H0 may be obtained.


S166. Determine the target model of the microphone signal X based on the test statistic ψ.


Specifically, step S166 may include:


S166-2. Determine that the test statistic ψ is greater than a preset decision threshold η, determine that the target voice signal is present in the microphone signal, and determine that the target model is the second model and that the noise covariance matrix of the microphone signal is the second estimate custom-character; or


S166-4. Determine that the test statistic ψ is less than a preset decision threshold, determine that the target voice signal is absent in the microphone signal, and determine that the target model is the first model and that the noise covariance matrix of the microphone signal is the first estimate custom-character.


Step S166 may be represented by the following formula:









ψ





H
1





>




<





H
0





η




formula



(
33
)








The decision threshold n is a parameter related to a false alarm probability. The false alarm probability may be obtained by experiment, machine learning, or experience.


As shown in FIG. 3, the method P100 may further include:


S180. Output a target pattern of the microphone signal X and the noise covariance matrix M.


The computing apparatus 240 may output the target pattern of the microphone signal X and the noise covariance matrix M to other calculation modules, such as a voice enhancement module.


In summary, in the voice activity detection system and method P100 provided in this disclosure, the computing apparatus 240 may optimize the first model and the second model respectively by using maximization of the likelihood function and rank minimization of the noise covariance matrix as joint optimization objectives, to determine the first estimate custom-character of the known parameter M1 and the second estimate custom-character of M2, so that estimation accuracy of M1 and M2 are higher, to provide a higher-accuracy data model for subsequent statistical hypothesis testing, thereby improving accuracy of voice activity detection and the voice enhancement effect.


This disclosure further provides a voice enhancement system. The voice enhancement system may also be applied to an electronic device 200. In some exemplary embodiments, the voice enhancement system may include a computing apparatus 240. In some exemplary embodiments, the voice enhancement system may be applied to the computing apparatus 240. In other words, the voice enhancement system may operate on the computing apparatus 240. The voice enhancement system may include a hardware device having a data information processing function and a program required to drive the hardware device to work. Certainly, the voice enhancement system may also be only a hardware device having a data processing capability or only a program running in a hardware device.


The voice enhancement system may store data or an instruction for performing a voice enhancement method described in this disclosure and may execute the data and/or the instruction. When the voice enhancement system operates on the computing apparatus 240, the voice enhancement system may obtain a microphone signal(s) from a microphone array 220 based on a communication and execute the data or the instruction of the voice enhancement method described in this disclosure. The voice enhancement method is described in other parts of this disclosure. For example, the voice enhancement method is described in the description of FIG. 8.


When operating on the computing apparatus 240, the voice enhancement system is in communication with the microphone array 220. A storage medium 243 may further include at least one instruction set stored in a data storage apparatus and used for performing voice enhancement calculation on the microphone signal(s). The instruction may be computer program code. The computer program code may include a program, a routine, an object, a component, a data structure, a process, a module, or the like for performing the voice enhancement method provided in this disclosure. A processor 242 may read the at least one instruction set and perform, based on the at least one instruction set, the voice enhancement method provided in this disclosure. The processor 242 may perform all steps included in the voice enhancement method.



FIG. 8 is a flowchart of a voice enhancement method P200 according to some exemplary embodiments of this disclosure. The method P200 may perform voice enhancement on a microphone signal. Specifically, a processor 242 may perform the method P200. As shown in FIG. 8, the method P200 may include the following steps.


S220. Obtain microphone signals X output by M microphones.


This step is the same as step S120, and will not be described herein again.


S240. Determine target models of the microphone signals X and noise covariance matrices M of the microphone signals X based on the voice activity detection method P100.


The noise covariance matrix M of the microphone signal X is a noise covariance matrix of the target model. When the target model of the microphone signal X is a first model the noise covariance matrix M of the microphone signal X is equal to custom-character. When the target model of the microphone signal X is a second model, the noise covariance matrix M of the microphone signal X is equal to custom-character.


S260. Determine, based on an MVDR method and the noise covariance matrices M of the microphone signals X, filter coefficients W corresponding to the microphone signals.


The filter coefficient ω may be an M×1-dimensional vector. The filter coefficient ω may be represented by the following formula:





ω=[ω12, . . . ,ωM]H  formula (34)


where the filter coefficient corresponding to the mth microphone 222 is om, and m=1,2, . . . , M.


The filter coefficient ω may be represented by the following formula:









ω
=



M

-
1



P



P
H



M

-
1



P






formula



(
35
)








As described above, P is a target steering vector of a target voice signal. In some exemplary embodiments, P is known.


S280. Combine the microphone signals X based on the filter coefficients, and output a target audio signal yk.


The target audio signal Y may be represented by the following formula:






Y=ω
H
X  formula (36)


The computing apparatus 240 may output the target audio signal Y to other electronic devices, such as a remote communications device.


In summary, the voice activity detection system and method P100, and the voice enhancement system and method P200 provided in this disclosure may be applied to the microphone array 220 formed by a plurality of microphones 222. The voice activity detection system and method P100, and the voice enhancement system and method P200 may obtain the microphone signal X captured by the microphone array 220. The microphone signal X may be the first model corresponding to the noise signal or may be the second model corresponding to the target voice signal mixed with the noise signal. The voice activity detection system and method P100, and the voice enhancement system and method P200 may optimize the first model and the second model respectively by using the microphone signal X as a sample and using maximization of the likelihood function and rank minimization of the noise covariance matrix M of the microphone signal X as joint optimization objectives, and determine the first estimate custom-character of the noise covariance matrix M1 of the first model and the second estimate custom-character of the noise covariance matrix M2 of the second model; and determine, by using the statistical hypothesis testing method, whether the microphone signal X satisfies the first model or the second model, thereby determining whether the target voice signal is present in the microphone signal X, determine the noise covariance matrix M of the microphone signal X, and further perform voice enhancement on the microphone signal X based on the MVDR method. The voice activity detection system and method P100, and the voice enhancement system and method P200 may make estimation accuracy of the noise covariance matrix M and accuracy of voice activity detection higher, thereby improving the voice enhancement effect.


Another aspect of this disclosure provides a non-transitory storage medium. The non-transitory storage medium stores at least one set of executable instructions for voice activity detection, and when the executable instructions are executed by a processor, the executable instructions instruct the processor to implement steps of the voice activity detection method P100 described in this disclosure. In some possible implementations, each aspect of this disclosure may be further implemented in a form of a program product, where the program product includes program code. When the program product operates on a computing device (for example, the computing apparatus 240), the program code may be used to enable the computing device to perform steps of voice activity detection described in this disclosure. The program product for implementing the foregoing method may use a portable compact disc read-only memory (CD-ROM) including program code, and may run on the computing device. However, the program product in this disclosure is not limited thereto. In this disclosure, a readable storage medium may be any tangible medium containing or storing a program, and the program may be used by or in connection with an instruction execution system (for example, the processor 242). The program product may use any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. For example, the readable storage medium may be but is not limited to an electronic, magnetic, optical, electromagnetic, infrared, or semi-conductor system, apparatus, or device, or any combination thereof. More specific examples of the readable storage medium include: an electrical connection having one or more conducting wires, a portable diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any appropriate combination thereof. The computer-readable storage medium may include a data signal propagated in a baseband or as part of a carrier, where the data signal carries readable program code. The propagated data signal may be in a plurality of forms, including but not limited to an electromagnetic signal, an optical signal, or any appropriate combination thereof. Alternatively, the readable storage medium may be any readable medium other than the readable storage medium. The readable medium may send, propagate, or transmit a program to be used by or in combination with an instruction execution system, apparatus, or device. The program code contained in the readable storage medium may be transmitted by using any appropriate medium, including but not limited to wireless, wired, optical cable, RF, or the like, or any appropriate combination thereof. The program code for performing operations in this disclosure may be compiled in any combination of one or more programming languages. The programming languages include object-oriented programming languages such as Java and C++, and further include conventional procedural programming languages such as a “C” language or a similar programming language. The program code may be fully executed on the computing device, partially executed on the computing device, executed as an independent software package, partially executed on the computing device and partially executed on a remote computing device, or fully executed on a remote computing device.


Specific exemplary embodiments of this disclosure have been described above. Other embodiments also fall within the scope of the appended claims. In some cases, actions or steps described in the claims may be performed in an order different from orders in the embodiments and still achieve expected results. In addition, the processes depicted in the drawings do not necessarily require a specific order or sequence to achieve the expected results. In some implementations, multitask processing and parallel processing are also possible or may be advantageous.


In summary, after reading this detailed disclosure, a person skilled in the art may understand that the foregoing detailed disclosure is illustrative, rather than restrictive. A person skilled in the art may understand that this disclosure is intended to cover various reasonable changes, improvements, and modifications to the embodiments, although this is not stated herein. These changes, improvements, and modifications are intended to be made in this disclosure and are within the spirit and scope of this disclosure.


In addition, some terms in this disclosure have been used to describe the embodiments of this disclosure. For example, “one embodiment”, “an embodiment”, and/or “some exemplary embodiments” mean/means that a specific feature, structure, or characteristic described with reference to the embodiment(s) may be included in at least one embodiment of this disclosure. Therefore, it can be emphasized and should be understood that in various parts of this disclosure, two or more references to “an embodiment” or “one embodiment” or “an alternative embodiment” do not necessarily all refer to the same embodiment. Further, specific features, structures, or characteristics may be appropriately combined in one or more embodiments of this disclosure.


It should be understood that in the foregoing description of the embodiments of this disclosure, to help understand one feature and for the purpose of simplifying this disclosure, various features in this disclosure are combined in a single embodiment, single drawing, or description thereof. However, this does not mean that the combination of these features is necessary. It is entirely possible for a person skilled in the art to extract some of the features as a separate embodiment for understanding when reading this disclosure. In other words, an embodiment of this disclosure may also be understood as the integration of a plurality of sub-embodiments. It is also true when content of each sub-embodiment is less than all features of a single embodiment disclosed above.


Each patent, patent application, patent application publication, and other materials cited herein, such as articles, books, disclosures, publications, documents, and materials, can be incorporated herein by reference, which are applicable to all content used for all purposes, except for any history of prosecution documents associated therewith, any identical, or any identical prosecution document history, which may be inconsistent or conflicting with this document, or any such subject matter that may have a restrictive effect on the broadest scope of the claims associated with this document now or later. For example, if there is any inconsistency or conflict in descriptions, definitions, and/or use of a term associated with this document and descriptions, definitions, and/or use of the term associated with any material, the term in this document shall prevail.


Finally, it should be understood that the implementation solutions of this application disclosed herein illustrate the principles of the implementation solutions of this disclosure. Other modified embodiments also fall within the scope of this disclosure. Therefore, the embodiments disclosed in this disclosure are merely exemplary and not restrictive. A person skilled in the art may use alternative configurations to implement the application in this disclosure according to the embodiments of this disclosure. Therefore, the embodiments of this disclosure are not limited to those specific embodiments specifically described in this application.

Claims
  • 1. A voice activity detection system, comprising: at least one storage medium storing a set of instructions for voice activity detection; andat least one processor in communication with the at least one storage medium, wherein during a process of voice activity detection for M microphones distributed in a preset array shape, wherein M is an integer greater than 1, the at least one processor executes the set of instructions to: obtain microphone signals output by the M microphones, wherein the microphone signals satisfy a first model corresponding to absence of a target voice signal or a second model corresponding to presence of a target voice signal,optimize the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determine a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model, anddetermine, based on statistical hypothesis testing, a target model and a noise covariance matrix corresponding to the microphone signals, wherein the target model includes one of the first model and the second model, and the noise covariance matrix of the microphone signals is a noise covariance matrix of the target model.
  • 2. The voice activity detection system according to claim 1, wherein the microphone signals include K frames of continuous audio signals, K is a positive integer greater than 1, and the microphone signals include an M×K data matrix.
  • 3. The voice activity detection system according to claim 2, wherein the microphone signals are complete observation signals or incomplete observation signals, all data in the M×K data matrix in the complete observation signals is complete, and a part of data in the M×K data matrix in the incomplete observation signals is missing, and when the microphone signals are the incomplete observation signals, to obtain the microphone signals output by the M microphones, the at least one processor executes the set of instructions to: obtain the incomplete observation signals, andperform row-column permutation on the microphone signals based on a position of missing data in each column in the M×K data matrix, and divide the microphone signals into at least one sub microphone signal, wherein the microphone signals include the at least one sub microphone signal.
  • 4. The voice activity detection system according to claim 1, wherein to optimize the first model and the second model respectively by using maximization of the likelihood function and rank minimization of the noise covariance matrix as the joint optimization objectives, the at least one processor executes the set of instructions to: establish a first likelihood function corresponding to the first model by using the microphone signals as sample data, wherein the likelihood function includes the first likelihood function;optimize the first model by using maximization of the first likelihood function and rank minimization of the noise covariance matrix of the first model as optimization objectives, and determine the first estimate;establish a second likelihood function corresponding to the second model by using the microphone signals as sample data, wherein the likelihood function includes the second likelihood function; andoptimize the second model by using maximization of the second likelihood function and rank minimization of the noise covariance matrix of the second model as optimization objectives, and determine the second estimate and an estimate of an amplitude of the target voice signal.
  • 5. The voice activity detection system according to claim 4, wherein the microphone signals include a noise signal, the noise signal conforms to a Gaussian distribution, and the noise signal includes at least: a colored noise signal conforming to a zero-mean Gaussian distribution, wherein a noise covariance matrix corresponding to the colored noise signal is a low-rank semi-positive definite matrix.
  • 6. The voice activity detection system according to claim 4, wherein to determine, based on statistical hypothesis testing, the target model and the noise covariance matrix corresponding to the microphone signals, the at least one processor executes the set of instructions to: establish a binary hypothesis testing model based on the microphone signals, wherein an original hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the first model, and an alternative hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the second model;substitute the first estimate, the second estimate, and the estimate of the amplitude into a decision criterion of a detector of the binary hypothesis testing model to obtain a test statistic; anddetermine the target model of the microphone signals based on the test statistic.
  • 7. The voice activity detection system according to claim 6, wherein to determine the target model of the microphone signals based on the test statistic, the at least one processor executes the set of instructions to: determine that the test statistic is greater than a preset decision threshold, determine that the target voice signal is present in the microphone signals, and determine that the target model is the second model and that the noise covariance matrix of the microphone signals is the second estimate; ordetermine that the test statistic is less than the preset decision threshold, determine that the target voice signal is absent in the microphone signals, and determine that the target model is the first model and that the noise covariance matrix of the microphone signals is the first estimate.
  • 8. The voice activity detection system according to claim 6, wherein the detector includes at least one of a generalized likelihood ratio test (GLRT) detector, a Rao detector, or a Wald detector.
  • 9. A voice activity detection method, wherein the method is for M microphones distributed in a preset array shape, and M is an integer greater than 1, the voice activity detection method comprising: obtaining microphone signals output by the M microphones, wherein the microphone signals satisfy a first model corresponding to absence of a target voice signal or a second model corresponding to presence of a target voice signal;optimizing the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determining a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model; anddetermining, based on statistical hypothesis testing, a target model and a noise covariance matrix corresponding to the microphone signals, wherein the target model includes one of the first model and the second model, and the noise covariance matrix of the microphone signals is a noise covariance matrix of the target model.
  • 10. The voice activity detection method according to claim 9, wherein the determining, based on the statistical hypothesis testing, of the target model and the noise covariance matrix corresponding to the microphone signals includes: establishing a binary hypothesis testing model based on the microphone signals, wherein an original hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the first model, and an alternative hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the second model;substituting the first estimate, the second estimate, and an amplitude estimate into a decision criterion of a detector of the binary hypothesis testing model to obtain a test statistic; anddetermining the target model of the microphone signals based on the test statistic.
  • 11. A voice enhancement system, comprising: at least one storage medium storing a set of instructions for voice enhancement; andat least one processor in communication with the at least one storage medium, wherein during a process of voice enhancement for M microphones distributed in a preset array shape, wherein M is an integer greater than 1, the at least one processor executes the set of instructions to: obtain microphone signals output by the M microphones,determine target models of the microphone signals and noise covariance matrices of the microphone signals, wherein the noise covariance matrices of the microphone signals are noise covariance matrices of the target models,determine, based on an MVDR method and the noise covariance matrices of the microphone signals, filter coefficients corresponding to the microphone signals, andcombine the microphone signals based on the filter coefficients, and output a target audio signal.
  • 12. The voice enhancement system according to claim 11, wherein to determine the target models of the microphone signals and the noise covariance matrices of the microphone signals, the at least one processor executes the set of instructions to: obtain microphone signals output by the M microphones, wherein the microphone signals satisfy a first model corresponding to absence of a target voice signal or a second model corresponding to presence of a target voice signal,optimize the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determine a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model, anddetermine, based on statistical hypothesis testing, a target model and a noise covariance matrix corresponding to the microphone signals, wherein the target model includes one of the first model and the second model, and the noise covariance matrix of the microphone signals is a noise covariance matrix of the target model; andthe microphone signals include K frames of continuous audio signals, K is a positive integer greater than 1, and the microphone signals include an M×K data matrix.
  • 13. The voice enhancement system according to claim 12, wherein the microphone signals are complete observation signals or incomplete observation signals, all data in the M×K data matrix in the complete observation signals is complete, and a part of data in the M×K data matrix in the incomplete observation signals is missing, and when the microphone signals are the incomplete observation signals, to obtain the microphone signals output by the M microphones, the at least one processor executes the set of instructions to: obtain the incomplete observation signals, andperform row-column permutation on the microphone signals based on a position of missing data in each column in the M×K data matrix, and divide the microphone signals into at least one sub microphone signal, wherein the microphone signals include the at least one sub microphone signal.
  • 14. The voice enhancement system according to claim 11, wherein to optimize the first model and the second model respectively by using maximization of the likelihood function and rank minimization of the noise covariance matrix as the joint optimization objectives, the at least one processor executes the set of instructions to: establish a first likelihood function corresponding to the first model by using the microphone signals as sample data, wherein the likelihood function includes the first likelihood function;optimize the first model by using maximization of the first likelihood function and rank minimization of the noise covariance matrix of the first model as optimization objectives, and determine the first estimate;establish a second likelihood function corresponding to the second model by using the microphone signals as sample data, wherein the likelihood function includes the second likelihood function; andoptimize the second model by using maximization of the second likelihood function and rank minimization of the noise covariance matrix of the second model as optimization objectives, and determine the second estimate and an estimate of an amplitude of the target voice signal.
  • 15. The voice enhancement system according to claim 14, wherein the microphone signals include a noise signal, the noise signal conforms to a Gaussian distribution, and the noise signal includes at least: a colored noise signal conforming to a zero-mean Gaussian distribution, wherein a noise covariance matrix corresponding to the colored noise signal is a low-rank semi-positive definite matrix.
  • 16. The voice enhancement system according to claim 14, wherein to determine determining, based on statistical hypothesis testing, the target model and the noise covariance matrix corresponding to the microphone signals, the at least one processor executes the set of instructions to: establish a binary hypothesis testing model based on the microphone signals, wherein an original hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the first model, and an alternative hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the second model;substitute the first estimate, the second estimate, and the estimate of the amplitude into a decision criterion of a detector of the binary hypothesis testing model to obtain a test statistic; anddetermine the target model of the microphone signals based on the test statistic.
  • 17. The voice enhancement system according to claim 16, wherein to determine the target model of the microphone signals based on the test statistic, the at least one processor executes the set of instructions to: determine that the test statistic is greater than a preset decision threshold, determine that the target voice signal is present in the microphone signals, and determine that the target model is the second model and that the noise covariance matrix of the microphone signals is the second estimate; ordetermine that the test statistic is less than the preset decision threshold, determine that the target voice signal is absent in the microphone signals, and determine that the target model is the first model and that the noise covariance matrix of the microphone signals is the first estimate.
  • 18. A voice enhancement method, wherein the voice enhancement method is for M microphones distributed in a preset array shape, and M is an integer greater than 1, the voice enhancement method comprising: obtaining microphone signals output by the M microphones;determining target models of the microphone signals and noise covariance matrices of the microphone signals, wherein the noise covariance matrices of the microphone signals are noise covariance matrices of the target models;determining, based on an MVDR method and the noise covariance matrices of the microphone signals, filter coefficients corresponding to the microphone signals; andcombining the microphone signals based on the filter coefficients, and outputting a target audio signal.
  • 19. The voice enhancement method according to claim 18, wherein the determining of the target models of the microphone signals and the noise covariance matrices of the microphone signals includes: obtaining microphone signals output by the M microphones, wherein the microphone signals satisfy a first model corresponding to absence of a target voice signal or a second model corresponding to presence of a target voice signal;optimizing the first model and the second model respectively by using maximization of a likelihood function and rank minimization of a noise covariance matrix as joint optimization objectives, and determining a first estimate of a noise covariance matrix of the first model and a second estimate of a noise covariance matrix of the second model; anddetermining, based on statistical hypothesis testing, a target model and a noise covariance matrix corresponding to the microphone signals, wherein the target model includes one of the first model and the second model, and the noise covariance matrix of the microphone signals is a noise covariance matrix of the target model.
  • 20. The voice enhancement method according to claim 18, the determining, based on the statistical hypothesis testing, of the target model and the noise covariance matrix corresponding to the microphone signals includes: establishing a binary hypothesis testing model based on the microphone signals, wherein an original hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the first model, and an alternative hypothesis of the binary hypothesis testing model includes that the microphone signals satisfy the second model;substituting the first estimate, the second estimate, and an amplitude estimate into a decision criterion of a detector of the binary hypothesis testing model to obtain a test statistic; anddetermining the target model of the microphone signals based on the test statistic.
RELATED APPLICATIONS

This application is a continuation application of PCT application No. PCT/CN2021/130035, filed on Nov. 11, 2021, and the content of which is incorporated herein by reference in its entirety.

Continuations (1)
Number Date Country
Parent PCT/CN2021/130035 Nov 2021 US
Child 18370381 US