A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
This disclosure relates to the field of acoustic technology, particularly to a signal processing method and an acoustic system.
Some acoustic systems comprise both a speaker and a sound sensor. In these systems, the ambient sound collected by the sound sensor may comprise sound emitted from the speaker, which is detrimental to the operation of the acoustic system. For example, in a hearing aid system, the sound sensor collects ambient sound during operation, amplifies the gain of the ambient sound, and then plays it through the speaker to compensate for the wearer's hearing loss. When the sound emitted by the speaker is recaptured by the sound sensor, a closed-loop circuit is formed in the acoustic system, causing the sound emitted by the speaker to be continuously amplified in the loop, leading to acoustic feedback, which results in discomfort for the wearer. Additionally, in a telephone system or a conference system, voice signals from a remote user are played through the local speaker and are then collected by the local sound sensor along with the voice from the local user, and transmitted back to the remote end. As a result, the remote user may experience interference from echo.
This disclosure provides a signal processing method and acoustic system that can reduce the signal component from the speaker in the target signal, thereby achieving the effect of suppressing feedback or eliminating echo.
In a first aspect, this disclosure provides an acoustic system, comprising: a speaker, configured to receive an input signal and output a target sound during operation; a sound sensor module, comprising at least a first sound sensor and a second sound sensor, where the first sound sensor collects an ambient sound and generates a first signal during operation, the second sound sensor collects the ambient sound and generates a second signal during operation, and the ambient sound comprises at least the target sound; and a signal processing circuit, connected to the sound sensor module to execute, where during operation, the acoustic system performs: obtaining a first signal, obtaining a second signal, performing a first target operation on the first signal and the second signal to reduce pickup of the target sound by the sound sensor module in a target frequency band and attenuate a signal corresponding to the target sound picked up by the sound sensor module, so as to obtain a target signal, and performing a second target operation on the target signal.
In a second aspect, this disclosure provides a signal processing method, comprising, by a signal processing circuit: obtaining a first signal, where the first signal is obtained by a first sound sensor in a sound sensor module during operation to capture an ambient sound, the ambient sound at least comprises a target sound, and the target sound is a sound output by a speaker during operation; obtaining a second signal, where the second signal is obtained by a second sound sensor in the sound sensor module during operation to capture the ambient sound; performing a first target operation on the first signal and the second signal to reduce pickup of the target sound by the sound sensor module in a target frequency band and attenuate a signal corresponding to the target sound picked up by the sound sensor module, so as to obtain a target signal; and performing a second target operation on the target signal.
From the above technical solution, it can be seen that this disclosure provides a signal processing method and an acoustic system. The signal processing method comprises: obtaining a first signal and a second signal, where the first signal is obtained by a first sound sensor in a sound sensor module collecting ambient sound during operation, and the second signal is obtained by a second sound sensor in the sound sensor module collecting ambient sound during operation. The ambient sound comprises at least a target sound emitted by a speaker during operation. A first target operation is performed on the first signal and the second signal to reduce the pickup of the target sound by the sound sensor module in a target frequency band and attenuate the signal corresponding to the target sound picked up by the sound sensor module, thereby obtaining a target signal. Subsequently, a second target operation is performed on the target signal. In the above solution, the first target operation can, on one hand, reduce the pickup of the target sound by the sound sensor module, and on the other hand, attenuate the signal corresponding to the target sound picked up by the sound sensor module, so that the resulting target signal contains no or minimal signal components from the speaker. As a result, the acoustic system can achieve the effect of suppressing howling or eliminating echo.
Other functions of the acoustic system and signal processing method provided in this disclosure will be partially listed in the following description. The inventive aspects of the acoustic system and signal processing method provided in this disclosure can be fully explained through the practice or use of the methods, devices, and combinations described in the detailed examples below.
To more clearly illustrate the technical solutions in the embodiments of this disclosure, the drawings required for the description of the embodiments will be briefly introduced below. Obviously, the drawings described below are merely some exemplary embodiments of this disclosure. For a person of ordinary skill in the art, other drawings can also be obtained based on these drawings without any creative effort.
The following description provides specific application scenarios and requirements of this disclosure, with the aim of enabling a person skilled in the art to manufacture and use the content of this disclosure. For a person skilled in the art, various local modifications to the disclosed embodiments are apparent, and the general principles defined here can be applied to other embodiments and applications without departing from the spirit and scope of this disclosure. Therefore, this disclosure is not limited to the embodiments shown but is intended to cover the broadest scope consistent with the claims.
The terms used herein are for the purpose of describing specific example embodiments and are not meant to be restrictive. For example, unless otherwise explicitly stated in the context, the singular forms “a,” “an,” and “the” may also include the plural forms. When used in this disclosure, the terms “include,” “comprise,” and/or “contain” mean that the associated integer, step, operation, element, and/or component is present but do not exclude the presence of one or more other features, integers, steps, operations, elements, components, and/or groups, or the possibility of adding other features, integers, steps, operations, elements, components, and/or groups to the system/method.
Given the following description, these features and other features of the disclosure, as well as the operation and functionality of the related elements of the structure, and the combination and manufacturability of the parts, can be significantly improved. The accompanying drawings, which form part of this disclosure, are referenced for illustration. However, it should be clearly understood that the drawings are for illustration and description purposes only and are not intended to limit the scope of this disclosure. It should also be understood that the drawings are not drawn to scale.
The flowcharts used in this disclosure illustrate the operations of the system implementation according to some exemplary embodiments of this disclosure. It should be clearly understood that the operations in the flowcharts may not be implemented in a specific order. Instead, the operations may be performed in reverse order or concurrently. Additionally, one or more other operations may be added to the flowcharts, or one or more operations may be removed from them.
For the convenience of description, the terms appearing in this disclosure are first explained below.
Howling: Howling is a phenomenon that frequently occurs in acoustic systems. The process of howling generation is explained below with reference to
Echo: Echo is also a phenomenon that frequently occurs in acoustic systems. The process of echo generation is explained below with reference to
Continuing with reference to
Background noise: Refers to the surrounding ambient sound other than the sound source being measured. In this disclosure, any sound that is unwanted by the user, undesired by the user, or interferes with the user's hearing can be called noise.
Sound pickup direction pattern: Refers to a pattern used to represent the sensitivity of the sound sensor/sound sensor module to sounds from different directions. In simple terms, the sound pickup direction pattern can represent the ability of the sound sensor/sound sensor module to pick up sounds from different directions. Typically, the sound pickup direction pattern can comprise: omnidirectional, heart-shaped, number 8 shaped, super-cardioid-shaped, etc.
Zero sound pickup direction: Theoretically, if the sound sensor/sound sensor module has a sensitivity to sound from a certain direction of 0 or close to 0, that direction is referred to as the zero sound pickup direction. It should be understood that when the sound source is located at the zero sound pickup direction, the sound sensor/sound sensor module will theoretically not capture any sound emitted by the sound source. In practice, due to manufacturing errors of the sound sensor/sound sensing module and the fact that sound sources in reality are not necessarily ideal point sources, the sound sensor/sound sensing module may still capture a small amount of sound from the zero sound pickup direction. It should be noted that in this disclosure, the zero sound pickup direction can refer to a specific direction or to a range of directions that comprises multiple directions.
Far-field sound source: Refers to a sound source that is relatively far from the sound sensor/sound sensor module. Generally speaking, when the distance between the sound source to be measured and the sound sensor/sound sensor module is greater than N times the physical size of the sound sensor/sound sensor module, the sound source can be approximated as a far-field sound source. It should be noted that in different application scenarios, the value of N can vary. For example, in the case of headphones, the physical size of the sound sensor/sound sensor module may be less than or equal to 0.01 m, and at this time, the value of N can be greater than or equal to 10. This means that a sound source located at a distance greater than or equal to 0.1 m from the sound sensor/sound sensor module can be considered a far-field sound source. Compared to a near-field sound source, the sound waves from the far-field sound source are approximately planar, and the amplitude of the sound waves decreases less as they propagate.
Near-field sound source: Refers to a sound source that is relatively close to the sound sensor/sound sensor module. Generally speaking, when the distance between the sound source to be measured and the sound sensor/sound sensor module is less than 2 to 3 times the physical size of the sound sensor/sound sensor module, the sound source can be approximated as a near-field sound source. For example, in the case of headphones, a sound source at a distance less than 0.1 m can be considered a near-field sound source. Compared to the aforementioned far-field sound source, the sound waves from the near-field sound source are closer to spherical, and the amplitude of the sound waves decreases more significantly as they propagate.
Before describing the specific embodiments of this disclosure, the application scenarios of this disclosure are introduced as follows: The signal processing method and acoustic system provided in this disclosure can be applied to scenarios that require squeal suppression (such as the scenario shown in
It should be noted that the squeal suppression scenario and the echo cancellation scenario are just some of the various application scenarios provided in this disclosure. The signal processing method and acoustic system provided in this disclosure can also be applied to other similar scenarios. A person skilled in the art should understand that the signal processing method and acoustic system provided in this disclosure applied to other usage scenarios are also within the scope of protection of this disclosure.
It should be noted that in the acoustic system 003 shown in
In some exemplary embodiments, the acoustic system 003 is a hearing system, and the acoustic system 003 may also comprise a housing, with the speaker 110, the sound sensor module 120, and the signal processing circuit 150 being disposed inside the housing. The housing serves to protect the internal components and makes it convenient for the user to handle and wear. The acoustic system 003 can be worn on the user's head. For example, the acoustic system 003 can be worn in the ear or over the ear at the user's ear region. When the acoustic system 003 is worn on the user's head, the sound-emitting end of the speaker 110 is oriented towards the user's head, for example, towards the user's ear canal opening or near the ear canal opening. The pickup end of at least one sound sensor in the sound sensor module 120 is located on the side of the housing away from the user's head. In this way, on one hand, it facilitates the pickup of ambient sound, and on the other hand, it can minimize the pickup of sound emitted by the speaker 110.
The first sound sensor 120-1 and the second sound sensor 120-2 can be the same sound sensor or different sound sensors. For ease of description, in this disclosure, the sound sensor in the sound sensor module 120 that is closer to the speaker 110 is referred to as the second sound sensor 120-2, while the sound sensor farther from the speaker 110 is referred to as the first sound sensor 120-1. That is to say, the second sound sensor 120-2 is closer to the speaker 110 compared to the first sound sensor 120-1. In this way, the sound emitted by the speaker 110 will first be captured by the second sound sensor 120-2 and then by the first sound sensor 120-1. In this case, the phase of the signal component corresponding to the target sound in the second signal is earlier than the phase of the signal component corresponding to the target sound in the first signal. When the distances between the two sound sensors and the speaker 110 are equal, either one can be referred to as the first sound sensor 120-1, and the other as the second sound sensor 120-2. In this case, the phase of the signal component corresponding to the target sound in the second signal is equal to the phase of the signal component corresponding to the target sound in the first signal. Additionally, the speaker 110, the first sound sensor 120-1, and the second sound sensor 120-2 can be integrated into the same electronic device or can be independent of each other, and this disclosure does not impose any limitations on this.
The speaker 110 can also be referred to as an electroacoustic transducer, a device used to convert electrical signals into sound signals. For example, the speaker 110 can be a loudspeaker. When operating, the speaker 110 receives an input signal and converts it into audio for playback. Here, the aforementioned input signal refers to an electrical signal carrying sound information, and the aforementioned audio refers to the sound played through the speaker 110. In some exemplary embodiments, the input signal received by the speaker 110 may come from the sound sensor module 120. This situation may correspond to the acoustic scenario shown in
The sound sensor (first sound sensor 120-1 and/or second sound sensor 120-2) can also be referred to as an acoustic-electric transducer or a sound pickup device, which is a device used to capture sound and convert it into an electrical signal. For example, a sound sensor can be a microphone (Microphone, MIC). When operating, the sound sensor captures ambient sound and converts it into an electrical signal carrying sound information. The sound sensor can be an omnidirectional sound sensor, in which case it can capture ambient sound from all directions. The sound sensor can also be a directional sound sensor, in which case it can capture ambient sound from specific directions.
Continuing to refer to
The signal processing circuit 150 is connected to the sound sensor module 120, as shown in
Continuing to refer to
In some exemplary embodiments, when the speaker 110, the sound sensor module 120, and the signal processing circuit 150 are deployed in a first acoustic device, the first acoustic device can be communicatively connected to a second acoustic device. In this case, the second target operation 50 may comprise: sending the target signal to the second acoustic device to reduce echo in the second acoustic device. The above scheme can be applied to the echo cancellation scenario shown in
The signal processing circuit 150 can be configured to execute the signal processing method described in this disclosure. In some exemplary embodiments, the signal processing circuit 150 may comprise multiple hardware circuits with interconnection relationships, each hardware circuit comprising one or more electrical components, which, during operation, implement one or more steps of the signal processing method described in this disclosure. These multiple hardware circuits cooperate with each other during operation to realize the signal processing method described in this disclosure. In some exemplary embodiments, the signal processing circuit 150 may comprise hardware equipment with data information processing capabilities and the necessary programs required to drive the operation of this hardware equipment. The hardware equipment executes these programs to implement the signal processing method described in this disclosure. The signal processing method will be described in detail in the following sections.
Continuing to refer to
The storage medium 210 may comprise a data storage device. The data storage device can be a non-transitory storage medium or a transitory storage medium. For example, the data storage device may comprise one or more of a magnetic disk 2101, a read-only storage medium (ROM) 2102, or a random access storage medium (RAM) 2103. The storage medium 210 also comprises at least one instruction set stored in the data storage device. The instruction set comprises instructions, which are computer program code. The computer program code may comprise programs, routines, objects, components, data structures, procedures, modules, etc., for executing the signal processing method provided in this disclosure.
The at least one processor 220 is configured to execute the aforementioned at least one instruction set. When the acoustic system 003 is running, the at least one processor 220 reads the at least one instruction set and, according to the instructions of the at least one instruction set, executes the signal processing method provided in this disclosure. The processor 220 can perform all or part of the steps included in the aforementioned signal processing method. The processor 220 may be in the form of one or more processors. In some exemplary embodiments, the processor 220 may comprise one or more hardware processors, such as a microcontroller, microprocessor, reduced instruction set computer (RISC), application-specific integrated circuit (ASIC), application-specific instruction set processor (ASIP), central processing unit (CPU), graphics processing unit (GPU), physics processing unit (PPU), microcontroller unit, digital signal processor (DSP), field-programmable gate array (FPGA), advanced RISC machine (ARM), programmable logic device (PLD), any circuit or processor capable of executing one or more functions, or any combination thereof. For illustrative purposes only, the acoustic system 003 shown in
S310: Obtain a first signal, where the first signal is obtained by a first sound sensor in a sound sensor module collecting an ambient sound during operation, the ambient sound comprises at least a target sound, and the target sound is a sound output by a speaker during operation.
S320: Obtain a second signal, where the second signal is obtained by a second sound sensor in the sound sensor module collecting the ambient sound during operation.
The signal processing circuit 150 can obtain the first signal from the first sound sensor 120-1 and the second signal from the second sound sensor 120-2. As mentioned earlier, since the ambient sound comprises both the sound emitted by the target sound source 160 and the target sound emitted by the speaker 110, the first signal and the second signal obtained by the signal processing circuit 150 both contain signal components from the target sound source 160 as well as signal components from the speaker 110.
It should be noted that this disclosure does not impose any restrictions on the execution order of S310 and S320. The order of execution of the two can be interchangeable, or they can also be executed simultaneously.
S330: Perform a first target operation on the first signal and the second signal to reduce pickup of the target sound by the sound sensor module in a target frequency band and attenuate a signal corresponding to the target sound picked up by the sound sensor module, thereby obtaining a target signal.
The signal processing circuit 150 can generate a target signal by performing a first target operation 40 on the first signal and the second signal. The first target operation 40 can serve the following two purposes:
On the one hand, by performing the first target operation 40 on the first signal and the second signal, the signal processing circuit 150 can reduce the pickup of the target sound (i.e., the sound from the speaker 110) by the sound sensor module 120 in a target frequency band. Herein, that “reduce the pickup of the target sound by the sound sensor module in the target frequency band” means that, compared to the pickup of the target sound by the sound sensor module 120 without performing the first target operation 40, the pickup of the target sound by the sound sensor module 120 is reduced when the first target operation 40 is performed.
In some exemplary embodiments, the aforementioned first target operation 40 can direct the zero sound pickup direction of the sound sensor module 120 in the target frequency band toward the speaker. The phrase “zero sound pickup direction directed toward the speaker” should be understood as the zero sound pickup direction generally pointing toward the speaker. For example, the zero sound pickup direction may point to the center point of the speaker. Alternatively, the zero sound pickup direction may point to any point on the sound output surface of the speaker. As another example, the zero sound pickup direction may point to a preset area on the sound output surface of the speaker. For yet another example, assuming the directional angle corresponding to the center point of the speaker is θ, the directional angle corresponding to the zero sound pickup direction may fall within the range θ−Δφ, θ+Δφ. It should be understood that when the zero sound pickup direction of the sound sensor module 120 is directed toward the speaker, the sound emitted by the speaker is either not picked up by the sound sensor module 120 or is picked up to a lesser extent. Therefore, by directing the zero sound pickup direction of the sound sensor module 120 in the target frequency band toward the speaker, the signal processing circuit 150 can reduce the pickup of the target sound by the sound sensor module 120.
It should be noted that the aforementioned target frequency band can be either a full frequency band or a partial frequency band. For example, the full frequency band may refer to the sound pickup frequency band of the sound sensor module, while the partial frequency band may refer to a specific sub-frequency band within the sound pickup frequency band of the sound sensor module. Here, the sound pickup frequency band refers to the range of frequency bands supported by the sound sensor module for picking up sound during operation. It should also be noted that when the target frequency band is a partial frequency band, the target frequency band can correspond to any sub-frequency band within the sound pickup frequency band of the sound sensor module. For instance, it may correspond to a high-frequency sub-band within the sound pickup frequency band, a low-frequency sub-band within the sound pickup frequency band, or a mid-frequency sub-band within the sound pickup frequency band, among others.
On the other hand, the signal processing circuit 150 can also attenuate the signal corresponding to the target sound picked up by the sound sensor module 120 by performing the first target operation 40 on the first signal and the second signal. Here, “signal corresponding to the target sound” refers to the signal component picked up by the sound sensor module 120 from the speaker. In some exemplary embodiments, the signal processing circuit 150 can attenuate the signal corresponding to the target sound picked up by the sound sensor module 120 by filtering the signal picked up by the sound sensor module 120. It is not difficult to understand that by attenuating the signal picked up by the sound sensor module 120, the signal processing circuit 150 can reduce the signal component from the speaker 110 in the target signal.
As can be seen, the first target operation 40 can, on one hand, minimize the pickup of the target sound by the sound sensor module 120 (that is, to make the sound sensor module 120 pick up as little as possible or none of the sound emitted by the speaker), and on the other hand, can attenuate the signal corresponding to the target sound picked up by the sound sensor module 120. By combining the effects of these two aspects, the signal processing circuit 150 can minimize the signal component from the speaker in the target signal.
In some exemplary embodiments, the above sound pickup control operation 43 and/or adaptive filtering operation 44 can be implemented by the processor 220 in the signal processing circuit 150, that is, the processor 220 executes the instruction set and performs the sound pickup control operation 43 and/or adaptive filtering operation 44 according to the instructions of the instruction set. In some exemplary embodiments, the signal processing circuit 150 may comprise a sound pickup control circuit, and the above sound pickup control operation 43 can be implemented through the sound pickup control circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise an adaptive filtering circuit, and the above adaptive filtering operation 44 can be implemented through the adaptive filtering circuit.
In some exemplary embodiments, as shown in
(1) Perform sound pickup control operation 43 on the first signal and the second signal to point the zero sound pickup direction of the sound sensor module 120 in the target frequency band toward the speaker, thereby obtaining a sound pickup control output signal.
By executing the sound pickup control operation 43, the signal processing circuit 150 can adjust the zero sound pickup direction of the sound sensor module 120 in the target frequency band to point toward the speaker 110. In this case, the single-channel signal picked up by the sound sensor module 120 is called the sound pickup control output signal. It should be understood that when the zero sound pickup direction of the sound sensor module 120 in the target frequency band points toward speaker 110, the sound sensor module 120 can reduce the pickup of the target sound in the target frequency band; therefore, the sound pickup control output signal contains less signal component from speaker 110. It should be noted that the specific implementation method of the sound pickup control operation 43 will be described in detail later, and will not be repeated herein.
(2) Obtain the input signal of the speaker, and perform adaptive filtering operation 44 on the sound pickup control output signal based on the input signal to obtain the target signal.
Since the target sound is emitted by speaker 110 based on the input signal, the signal processing circuit 150 can predict how the sound sensor module 120 picks up the target sound based on the input signal, and then perform adaptive filtering on the sound pickup control output signal based on the pickup situation to obtain the target signal. This further reduces the signal component from speaker 110 in the target signal.
In some exemplary embodiments, continuing to refer to
Where the target gain corresponds to the transfer function between speaker 110 and sound sensor module 120; assuming the transfer function between speaker 110 and sound sensor module 120 is F, the signal processing circuit 150 can predict the transfer function F to obtain a predicted transfer function, and determine the predicted transfer function as the target gain F. In this way, providing the target gain to the input signal can simulate how the target sound is picked up by the sound sensor module 120. Furthermore, considering the influence of factors such as environment and user actions, the transfer function F between speaker 110 and sound sensor module 120 is constantly changing. Therefore, in some exemplary embodiments, a preset algorithm can be used to adaptively update the target gain {circumflex over (F)}, making the target gain {circumflex over (F)} continuously track the transfer function F (i.e., the target gain changes with changes in transfer function F). This makes the third signal (u*{circumflex over (F)}) obtained based on the input signal u and the target gain {circumflex over (F)} more accurate, that is, the third signal (u*{circumflex over (F)}) can more accurately characterize how the sound sensor module 120 picks up the target sound. It should be noted that this disclosure does not limit the preset algorithm, for example, it can use algorithms such as Least Mean Square (LMS), Normalized Least Mean Square (NLMS), Recursive Least Square (RLS), etc.
In some exemplary embodiments, as shown in
(1) obtaining the input signal of the speaker, and based on the input signal, performing the adaptive filtering operation 44 on the first signal and the second signal to obtain a first filtered signal corresponding to the first signal and a second filtered signal corresponding to the second signal. Since the target sound is emitted by the speaker 110 based on the input signal, the signal processing circuit 150 can predict the pickup condition of the first sound sensor 120-1 and the second sound sensor 120-2 for the target sound based on the input signal, and then perform adaptive filtering on the first signal and the second signal based on the pickup condition to obtain the first filtered signal and the second filtered signal. That is, the first filtered signal can be regarded as the signal obtained by attenuating the signal component from the speaker 110 in the first signal, and the second filtered signal can be regarded as the signal obtained by attenuating the signal component from the speaker 110 in the second signal. Therefore, the signal components from the speaker 110 contained in the first filtered signal and the second filtered signal are reduced.
In some exemplary embodiments, continuing with reference to ) to the input signal to obtain a fourth signal, and subtracting the fourth signal from the first signal to obtain the first filtered signal, as well as providing a second target gain (
) to the input signal to obtain a fifth signal, and subtracting the fifth signal from the second signal to obtain the second filtered signal. For example, assuming the input signal is u, the first signal is y1, and the second signal is y2, then the first filtered signal s1 and the second filtered signal s2 can be represented by the following formulas:
Where, the first target gain () corresponds to the transfer function between the speaker 110 and the first sound sensor 120-1, and the second target gain (
) corresponds to the transfer function between the speaker 110 and the second sound sensor 120-2. It should be understood that the method of determining the first target gain (
) and the second target gain (
) is similar to the method of determining the target gain ({circumflex over (F)}), which will not be elaborated herein.
(2) Perform the sound pickup control operation 43 on the first filtered signal and the second filtered signal to obtain the target signal.
The signal processing circuit 150, by performing the sound pickup control operation 43 on the first filtered signal and the second filtered signal, can adjust the zero sound pickup direction of the sound sensor module 120 in the target frequency band to point toward the speaker 110. It should be understood that when the zero sound pickup direction of the sound sensor module 120 in the target frequency band points toward the speaker 110, the sound sensor module 120 is able to reduce the pickup of the target signal in the target frequency band, thereby further reducing the signal components from the speaker 110 contained in the target signal.
The specific implementation of the adaptive filtering operation 44 is shown in
The zero differential operation 41a shown in
In some exemplary embodiments, when the signal processing circuit 150 performs the first delay operation 411, it can determine the delay duration T corresponding to the second body signal based on the following formula (4):
Where, d is the distance between the first sound sensor 120-1 and the second sound sensor 120-2, and c is the speed of sound.
After performing the first delay operation 411, since the phases of the signal components from the speaker 110 in the second delayed signal and the first body signal have been aligned, the signal processing circuit 150, by performing the first differential operation 413 (i.e., subtracting the second delayed signal from the first body signal), can cause the signal components from the speaker 110 in the first body signal and the second delayed signal to cancel each other out. This results in the first differential signal exhibiting a zero sound pickup characteristic in the direction of the speaker 110.
Furthermore, the inventors analyzed and experimented with the zero differential operation 41a in
Therefore, in some exemplary embodiments, continuing to refer to
As mentioned earlier, the zero differential operation 41a shown in
Among them, the first delay operation 411 is configured to delay the second individual signal to obtain a second delayed signal. In some exemplary embodiments, the delay duration for the second individual signal can be determined based on the aforementioned formula (4), which will not be elaborated herein. The second delay operation 412 is configured to delay the first individual signal to obtain a first delayed signal. In some exemplary embodiments, the delay duration for the first individual signal can be determined based on the aforementioned formula (4), which will not be elaborated herein. The first differential operation 413 is configured to perform a differential operation on the first individual signal and the second delayed signal (i.e., subtract the second delayed signal from the first individual signal) to obtain a first differential signal. The second differential operation 414 is configured to perform a differential operation on the second individual signal and the first delayed signal (i.e., subtract the first delayed signal from the second individual signal) to obtain a second differential signal. The third differential operation 415 is configured to perform a differential operation on the first differential signal and the second differential signal (i.e., subtract the second differential signal from the first differential signal) to obtain a third differential signal.
It should be understood that the principle of adjusting the sound pickup direction in the zero differential operation 41b shown in
From the aforementioned number 8 shaped sound pickup direction pattern (i.e., pattern III), it can be seen that the sound sensor module 120 exhibits pickup zero characteristics in the 90-degree and 270-degree directions. When the speaker 110 is located at 90 degrees/270 degrees or near these directions, the sound sensor module 120 will not pick up (or will pick up very little of) the sound emitted by the speaker 110. Thus, it is evident that the dual-delay-based zero differential operation 41b shown in
Furthermore, continuing to refer to
When using the zero differential operation 41b shown in
It should be understood that the second differential signal corresponds to pattern II in
It should be noted that the various zero differential operations 41 involved in this disclosure (e.g., the zero differential operation 41a shown in
In some exemplary embodiments, any one or more of the aforementioned first delay operation 411, second delay operation 412, first differential operation 413, second differential operation 414, third differential operation 415, gain compensation operation 416, multiplication operation 417, and target parameter generation operation 418 may be implemented by the processor 220 in the signal processing circuit 150. That is, the processor 220 executes an instruction set and performs one or more of the above operations according to the instructions in the set. In some exemplary embodiments, the signal processing circuit 150 may comprise a first delay circuit, and the aforementioned first delay operation 411 may be implemented by the first delay circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a second delay circuit, and the aforementioned second delay operation 412 may be implemented by the second delay circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a first differential circuit, and the aforementioned first differential operation 413 may be implemented by the first differential circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a second differential circuit, and the aforementioned second differential operation 414 may be implemented by the second differential circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a third differential circuit, and the aforementioned third differential operation 415 may be implemented by the third differential circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a gain compensation circuit, and the aforementioned gain compensation operation 416 may be implemented by the gain compensation circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a multiplication circuit, and the aforementioned multiplication operation 417 may be implemented by the multiplication circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a target parameter generation circuit, and the aforementioned target parameter generation operation 418 may be implemented by the target parameter generation circuit.
The zero differential operation 41 described earlier performs zero differential across the entire frequency band of the first individual signal and the second individual signal (i.e., the sound pickup frequency band of the sound sensor module 120). Therefore, the zero differential operation 41 can direct the zero sound pickup direction of the sound sensor module 120 toward the speaker across the entire frequency band. From this, it can be seen that the sound pickup control operation 43a shown in
The signal synthesis operation 42 is configured to synthesize the differential result signal and the first individual signal to obtain a composite signal, where the components of the first frequency band in the composite signal come from the differential result signal, and the components of the second frequency band come from either the first individual signal or the second individual signal. Since the first individual signal and the second individual signal have not undergone the zero differential operation 41, compared to the differential result signal, they can more accurately reflect certain characteristics of the real ambient sound (e.g., background noise characteristics). Therefore, when the signal in the second frequency band comes from the first individual signal or the second individual signal, the composite signal retains components of the original signal, allowing the composite signal to more accurately reflect the characteristics of the real ambient sound. It can be understood that, in the case where the signal in the first frequency band of the composite signal comes from the differential result signal and the signal in the second frequency band comes from the first individual signal or the second individual signal, the composite signal both preserves components of the original signal and reduces signal components from the speaker 110. As a result, the composite signal can reduce the signal components from the speaker 110 while striving to accurately reflect the real ambient sound, thereby improving the accuracy of the composite signal. For ease of description, the following descriptions will take the example where the signal in the second frequency band comes from the first individual signal. It should be understood that when the signal in the second frequency band comes from the second individual signal, the implementation is similar, and this disclosure will not elaborate further on this.
From the previous description, it is clear that the zero differential operation 41 (e.g., the zero differential operation 41a shown in
In practical applications, considering that background noise in the environment is typically low-frequency noise, in some exemplary embodiments, the frequencies in the first frequency band may be higher than those in the second frequency band. For example, the first frequency band could be a high frequency band, while the second frequency band could be a low frequency band. Alternatively, the first frequency band could be a mid-high frequency band, and the second frequency band could be a low frequency band. Another example could be that the first frequency band is a high frequency band, and the second frequency band is a mid-low frequency band. Yet another example could be that the first frequency band is a mid-frequency band, and the second frequency band is a low frequency band, and so on. In other words, the lower frequency range within the sound pickup frequency band of the sound sensor module 120 is designated as the second frequency band, and the components of the second frequency band in the composite signal are taken from the first signal, thereby avoiding the issue of amplified background noise components.
Herein, the aforementioned low frequency band generally refers to a frequency band below approximately 1 KHz, the mid frequency band refers to a frequency band roughly between 1 KHz and 4 KHz, the high frequency band refers to a frequency band above 4 KHz, the mid-low frequency band refers to a frequency band roughly below 4 KHz, and the mid-high frequency band refers to a frequency band roughly above 1 KHz. A person of ordinary skill in the art should understand that the distinction of these frequency bands is provided as an example with approximate ranges. The definitions of these frequency bands may vary depending on different industries, application scenarios, and classification standards. For instance, in some application scenarios, the low frequency band may refer to a frequency band roughly from 20 Hz to 150 Hz, the mid frequency band may refer to a frequency band roughly from 150 Hz to 5 KHz, the high frequency band may refer to a frequency band roughly from 5 KHz to 20 KHz, the mid-low frequency band may refer to a frequency band roughly from 150 Hz to 500 Hz, and the mid-high frequency band may refer to a frequency band roughly from 500 Hz to 5 KHz. In other application scenarios, the low frequency band may refer to a frequency band roughly from 20 Hz to 80 Hz, the mid-low frequency band may refer to a frequency band roughly between 80 Hz and 160 Hz, the mid frequency band may refer to a frequency band roughly from 160 Hz to 1280 Hz, the mid-high frequency band may refer to a frequency band roughly from 1280 Hz to 2560 Hz, and the high frequency band may refer to a frequency band roughly from 2560 Hz to 20 KHz.
In some exemplary embodiments, the aforementioned zero differential operation 41 and/or signal synthesis operation 42 may be implemented by the processor 220 in the signal processing circuit 150. That is, the processor 220 executes an instruction set and performs the zero differential operation 41 and/or the signal synthesis operation 42 according to the instructions in the set. In some exemplary embodiments, the signal processing circuit 150 may comprise a zero differential circuit, and the aforementioned zero differential operation 41 may be implemented by the zero differential circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a signal synthesis circuit, and the aforementioned signal synthesis operation 42 may be implemented by the signal synthesis circuit.
In some exemplary embodiments, the aforementioned first filtering and second filtering may be complementary filters, meaning that the sum of the transfer function of the first filtering and the transfer function of the second filtering equals 1. For example, assuming the transfer function corresponding to the first filtering is expressed as the following formula (5), the transfer function corresponding to the second filtering can be expressed as the following formula (6).
From formulas (5) and (6), it can be seen that the denominator expressions corresponding to the transfer functions of the two filtering operations are identical, both being A(z), while the numerator expressions corresponding to the transfer functions of the two filtering operations are B(z) and A(z)−B(z), respectively. This design ensures that the sum of the transfer functions of the two filtering operations equals 1, thereby exhibiting all-pass characteristics when the two filtering operations are used in combination. For ease of understanding,
It should be noted that the complementary filtering operations exemplified by the aforementioned formulas (5) and (6) are merely one possible example. A person skilled in the art can understand that any filter bank capable of achieving frequency division and synthesis is feasible. For instance, in some exemplary embodiments, the aforementioned first filtering and second filtering can also be implemented using a first filter and a second filter with the same cutoff frequency. Here, the cutoff frequency of the first filter is w1, and the cutoff frequency of the second filter is w2, with the two filters having identical cutoff frequencies: w1=w2. Additionally, the amplitude responses of the two filters at the cutoff frequency satisfy
Continuing to refer to
In some exemplary embodiments, when there are multiple speakers 110, the sound emission frequency bands corresponding to different speakers may vary. In such cases, different zero differential schemes can be applied to different frequency bands, ensuring that the zero sound pickup direction of the sound sensor module 120 in each frequency band points to the speaker corresponding to that frequency band, thereby minimizing the signal components from each speaker in the composite signal as much as possible. Taking two speakers as an example, for ease of description, the two speakers are referred to as the first speaker and the second speaker, respectively. The first frequency band may comprise a first sub-frequency band and a second sub-frequency band. Accordingly, the signal in the first sub-frequency band of the composite signal comes from the first pickup result signal obtained by the sound sensor module 120 in a first pickup state, where the first pickup state corresponds to the zero sound pickup direction of the sound sensor module 120 pointing to the first speaker. The signal in the second sub-frequency band of the composite signal comes from the second pickup result signal obtained by the sound sensor module 120 in a second pickup state, where the second pickup state corresponds to the zero sound pickup direction of the sound sensor module 120 pointing to the second speaker. The sound pickup direction patterns corresponding to the first pickup state and the second pickup state are different. In other words, when the signal processing circuit 150 executes the sound pickup control operation 43, it can apply different zero differential operations 41 to the first sub-frequency band and the second sub-frequency band, resulting in different sound pickup direction patterns for the first sub-frequency band and the second sub-frequency band. In practical applications, considering that each zero differential scheme may have varying zero attenuation effects (i.e., the sound intensity attenuation in the zero sound pickup direction) across different frequency bands, the zero differential operation 41 with better zero attenuation performance can be selected for the first sub-frequency band and the second sub-frequency band based on the zero attenuation performance of each zero differential operation 41 in different frequency bands, thereby enhancing the overall zero attenuation effect across the entire frequency band.
Continuing to refer to
Based on the comparison of zero attenuation effects shown in
Continuing to refer to
Similar to what is shown in
In some exemplary embodiments, any one or more of the aforementioned first filtering operation 421, second filtering operation 422, third filtering operation 423, and synthesis operation 424 may be implemented by the processor 220 in the signal processing circuit 150. That is, the processor 220 executes an instruction set and performs one or more of the above operations according to the instructions in the set. In some exemplary embodiments, the signal processing circuit 150 may comprise a first filtering circuit, and the aforementioned first filtering operation 421 may be implemented by the first filtering circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a second filtering circuit, and the aforementioned second filtering operation 422 may be implemented by the second filtering circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a third filtering circuit, and the aforementioned third filtering operation 423 may be implemented by the third filtering circuit. In some exemplary embodiments, the signal processing circuit 150 may comprise a synthesis circuit, and the aforementioned synthesis operation 424 may be implemented by the synthesis circuit.
S340: Perform a second target operation on the target signal.
After obtaining the target signal, the signal processing circuit 150 can perform a second target operation 50 on the target signal based on the requirements of the application scenario. In some exemplary embodiments, continuing to refer to
In some exemplary embodiments, the speaker 110, sound sensor module 120, and signal processing circuit 150 are integrated in a first acoustic device, which is communicatively connected to a second acoustic device. In this scenario, the second target operation 50 may comprise: sending the target signal to the second acoustic device to reduce the echo of the second acoustic device. The above solution can be applied to the echo cancellation scenario shown in
In summary, this disclosure provides a signal processing method and an acoustic system, where the signal processing method comprises: obtaining a first signal and a second signal, the first signal being obtained by the first sound sensor in the sound sensor module when collecting ambient sound, the second signal being obtained by the second sound sensor in the sound sensor module when collecting ambient sound, the ambient sound including at least the target sound emitted when the speaker is working, performing a first target operation on the first signal and the second signal to reduce the pickup of the target sound by the sound sensor module in the target frequency band and attenuate the signal corresponding to the target sound picked up by the sound sensor module, thereby obtaining a target signal, and then performing a second target operation on the target signal. In the above solution, since the first target operation can reduce the pickup of the target sound by the sound sensor module on the one hand, and can also attenuate the signal corresponding to the target sound picked up by the sound sensor module on the other hand, the obtained target signal does not contain or contains less signal components from the speaker, so that the acoustic system can achieve the effect of suppressing howling or eliminating echo.
Another aspect of this disclosure provides a non-transitory storage medium storing at least one set of executable instructions for performing image recognition. When the executable instructions are executed by a processor, the executable instructions direct the processor to implement the steps of the image recognition method P100 described in this disclosure. In some possible embodiments, various aspects of this disclosure may also be implemented in the form of a program product, which comprises program code. When the program product runs on a computing device 600, the program code is used to cause the computing device 600 to perform the steps of the image recognition method P100 described in this disclosure. The program product for implementing the above method may adopt a portable compact disc read-only memory (CD-ROM) that comprises program code and can run on the computing device 600. However, the program product of this disclosure is not limited to this. In this disclosure, the readable storage medium may be any tangible medium that contains or stores a program, which can be used by or in combination with an instruction execution system. The program product may adopt any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof. More specific examples of readable storage media comprise: an electrical connection with one or more wires, a portable disk, a hard disk, random access memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disc read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above. The computer-readable storage medium may comprise a data signal propagated in baseband or as part of a carrier wave, which carries readable program code. Such propagated data signals may take various forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above. The readable storage medium may also be any readable medium other than a readable storage medium that can send, propagate, or transmit a program for use by or in combination with an instruction execution system, apparatus, or device. The program code contained on the readable storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the above. Program code for performing the operations of this disclosure may be written in any combination of one or more programming languages, including object-oriented programming languages-such as Java, C++, etc.—as well as conventional procedural programming languages-such as the “C” language or similar programming languages. The program code may be executed entirely on the computing device 600, partially on the computing device 600, as a standalone software package, partially on the computing device 600 and partially on a remote computing device, or entirely on a remote computing device.
The above description pertains to specific embodiments of the present specification. Other embodiments are within the scope of the appended claims. In some cases, the actions or steps described in the claims can be performed in a sequence different from the one in the embodiments and still achieve the desired result. Additionally, the processes depicted in the drawings do not necessarily require a specific order or continuous sequence to achieve the desired outcome. In certain embodiments, multitasking and parallel processing are also possible or may be beneficial.
In summary, after reading this detailed disclosure, a person skilled in the art can understand that the aforementioned detailed disclosure is presented only by way of example and is not intended to be limiting. Although not explicitly stated here, a person skilled in the art will appreciate that the disclosure encompasses various reasonable alterations, improvements, and modifications to the embodiments. These alterations, improvements, and modifications are intended to be within the spirit and scope of the exemplary embodiments presented in this specification.
In addition, certain terms in this specification have been used to describe the embodiments of the specification. For example, the terms “one embodiment,” “embodiment,” and/or “some exemplary embodiments” mean that specific features, structures, or characteristics described in connection with that embodiment may be included in at least one embodiment of the specification. Therefore, it should be emphasized and understood that references to “embodiment,” “one embodiment,” or “alternative embodiment” in various parts of this specification do not necessarily refer to the same embodiment. Additionally, specific features, structures, or characteristics may be appropriately combined in one or more embodiments of the specification.
It should be understood that in the foregoing description of the embodiments of the specification, in order to aid in understanding a feature and simplify the presentation, various features are combined in a single embodiment, drawing, or description. However, this does not mean that the combination of these features is required. A person skilled in the art, upon reading this specification, could very well consider part of the equipment marked as a separate embodiment. In other words, the embodiments in this specification can also be understood as the integration of multiple sub-embodiments. And each sub-embodiment is valid even when it includes fewer features than a single full embodiment disclosed above.
Each patent, patent application, publication of a patent application, and other materials, such as articles, books, specifications, publications, documents, articles, etc., cited herein, except for any historical prosecution documents to which it relates, which may be inconsistent with or any identities that conflict, or any identities that may have a restrictive effect on the broadest scope of the claims, are hereby incorporated by reference for all purposes now or hereafter associated with this document. Furthermore, in the event of any inconsistency or conflict between the description, definition, and/or use of a term associated with any contained material, the term used in this document shall prevail.
Finally, it should be understood that the embodiments of the application disclosed herein are illustrative of the principles of the embodiments of this specification. Other modified embodiments are also within the scope of this specification. Therefore, the embodiments disclosed in this specification are merely examples and not limitations. A person skilled in the art can adopt alternative configurations based on the embodiments in this specification to implement the application in this specification. Thus, the embodiments of this specification are not limited to the embodiments described in the application in precise detail.
This application is a continuation application of PCT application No. PCT/CN2023/094374, filed on May 15, 2023, and the content of which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2023/094374 | May 2023 | WO |
Child | 19075842 | US |