System and method for automatic adjustment of reference gain

Abstract
Systems and methods are provided for automatically adjusting a reference gain of an audio mixer having a reference channel for receiving a far end audio signal from a remote location as a reference signal and a plurality of audio input channels for receiving audio signals captured by a plurality of microphone element. An exemplary method includes determining an echo level in an input audio signal received at a given audio input channel, and automatically determining a gain amount for the reference channel based on the echo level. An exemplary system includes a reference gain adjuster configured to automatically determine a gain amount for the reference channel based on an echo level detected in an input audio signal received at a given audio input channel.
Description
TECHNICAL FIELD

This disclosure generally relates to gain adjustment of a reference signal provided to a mixer for mixing audio signals from a plurality of acoustic sources, such as microphones used in a conferencing system. In particular, the disclosure relates to systems and methods for automatically adjusting the reference gain.


BACKGROUND

Conferencing environments, such as boardrooms, conferencing settings, and the like, can involve the use of microphones for capturing sound from audio sources (also known as the “near end”) and loudspeakers for presenting audio from a remote location (also known as the “far end”). For example, persons in a conference room may be conducting a conference call with persons at a remote location. Typically, speech and sound from the conference room may be captured by microphones and transmitted to the remote location, while speech and sound from the remote location may be received and played on loudspeakers in the conference room. Multiple microphones may be used in order to optimally capture the speech and sound in the conference room.


In some cases, the microphones may pick up the speech and sound from the remote location that is played on the loudspeakers. In such situations, the audio transmitted to the remote location may include not only the speech and sound from the conference room, but also the speech and sound from the remote location, thus creating an undesirable echo for the persons at the remote location hearing their own speech and sound. If there is no correction, the audio transmitted to the remote location may therefore be low quality or unacceptable due to this echo.


Some existing communication systems utilize echo cancellation techniques to minimize, or cancel out, the echo in the mixed audio signal. For example, some systems utilize an acoustic echo canceller for each of the multiple microphones, and a mixer to subsequently mix and process each echo-cancelled microphone signal. However, these types of systems may be computationally intensive and complex. For example, separate and dedicated processing may be needed to perform acoustic echo cancellation on each microphone signal. Other systems place an acoustic echo canceller after the mixer. However, some such systems work poorly due to the need to constantly re-adapt to the mixed signal generated by the mixer, specifically where the mixer is dynamic and therefore, the gains on one or more of the mixer channels changes over time.


Accordingly, there is an opportunity for improved techniques that address these concerns.


SUMMARY

The invention is intended to solve the above-noted problems by providing systems and methods designed to automatically adjust a gain of a reference input channel that is fed into a mixer for generating a mixed audio signal based on near end audio signals, wherein the reference input channel provides a far end audio signal to the mixer for gate-inhibit purposes.


One example embodiment provides a system, comprising a plurality of microphone elements, each microphone element configured to detect sound and convert the sound to a near end audio signal; at least one audio output device configured to output a far end audio signal comprising sound captured at a remote location; an audio mixer having a plurality of audio input channels for receiving audio signals captured by the plurality of microphone elements and a reference channel for receiving the far end audio signal as a reference signal; and a reference gain adjuster configured to automatically determine a gain amount for the reference channel based on an echo level detected in an input audio signal received at a given audio input channel.


Another example embodiment provides a method of automatically adjusting a reference gain of an audio mixer having a reference channel for receiving a far end audio signal from a remote location as a reference signal and a plurality of audio input channels for receiving audio signals captured by a plurality of microphone elements, the method comprising determining an echo level in an input audio signal received at a given audio input channel; and automatically determining a gain amount for the reference channel based on the echo level.


These and other embodiments, and various permutations and aspects, will become apparent and be more fully understood from the following detailed description and accompanying drawings, which set forth illustrative embodiments that are indicative of the various ways in which the principles of the invention may be employed.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a schematic diagram of a communication system including an automatic reference gain adjustment module, in accordance with some embodiments.



FIG. 2 is a flowchart illustrating operations for performing automatic reference gain adjustment using the communication system of FIG. 1, in accordance with some embodiments.



FIG. 3 is a flowchart illustrating operations for performing automatic reference gain adjustment when one or more microphone inputs are boosted, using the communication system of FIG. 1, in accordance with some embodiments.





DETAILED DESCRIPTION

The description that follows describes, illustrates, and exemplifies one or more particular embodiments of the invention in accordance with its principles. This description is not provided to limit the invention to the embodiments described herein, but rather to explain and teach the principles of the invention in such a way to enable one of ordinary skill in the art to understand these principles and, with that understanding, be able to apply them to practice not only the embodiments described herein, but also other embodiments that may come to mind in accordance with these principles. The scope of the invention is intended to cover all such embodiments that may fall within the scope of the appended claims, either literally or under the doctrine of equivalents.


It should be noted that in the description and drawings, like or substantially similar elements may be labeled with the same reference numerals. However, sometimes these elements may be labeled with differing numbers, such as, for example, in cases where such labeling facilitates a more clear description. Additionally, the drawings set forth herein are not necessarily drawn to scale and in some instances, proportions may have been exaggerated to more clearly depict certain features. Such labeling and drawing practices do not necessarily implicate an underlying substantive purpose. As stated above, the specification is intended to be taken as a whole and interpreted in accordance with the principles of the invention as taught herein and understood to one of ordinary skill in the art.



FIG. 1 is a schematic diagram of a communication system 100 comprising a plurality of microphones 102 for capturing sound from one or more audio sources in an environment (also referred to herein as “near end audio”), a loudspeaker 104 for presenting audio received from a remote location (also referred to herein as “far end audio”), and an audio mixer 106 for generating a mix of the near end audio for sending to the remote location. For example, environments such as conference rooms may utilize the communication system 100 to facilitate communication with persons situated at the remote location, which may be another room, building, city, or any other area that is separate from the given environment. The components of the communication system 100 may be in communication with each other using wired and/or wireless connections. In the latter case, the communication system 100 may further include a communications module for enabling wireless communication between various components of the communication system 100 itself, and/or to facilitate wireless communication with the remote location.


The types of microphones 102 and their placement in a particular environment may depend on the locations of the audio sources, physical space requirements, aesthetics, room layout, and/or other considerations. For example, in some environments, the microphones 102 may be placed on a table or lectern near the audio sources. In other environments, the microphones 102 may be mounted overhead to capture the sound from the entire room, for example. The communication system 100 may work in conjunction with any type and any number of microphones 102.


Various components included in the communication system 100 may be implemented using software executable by one or more servers or computers, such as a computing device with a processor and memory, by hardware (e.g., discrete logic circuits, application specific integrated circuits (ASIC), programmable gate arrays (PGA), field programmable gate arrays (FPGA), etc.), or using a combination software and hardware.



FIGS. 2 and 3 illustrate exemplary processes for utilizing the communication system 100, in accordance with embodiments. In particular, FIG. 2 illustrates a process 200 for automatically adjusting a gain of a reference input signal representing far end audio, using the communication system 100, and FIG. 3 illustrates a method 300 for performing automatic reference gain adjustment when one or more of microphone inputs are boosted, or have an extra gain structure, using the communication system 100. One or more processors and/or other processing components within the communication system 100 may perform any, some, or all of the steps of each of the processes 200 and 300. One or more other types of components (e.g., memory, input and/or output devices, transmitters, receivers, buffers, drivers, discrete components, etc.) may also be utilized in conjunction with the processors and/or other processing components to perform any, some, or all of the steps of each of the processes 200 and 300. In general, a computer program product in accordance with the embodiments includes a computer-usable storage medium (e.g., standard random access memory (RAM), an optical disc, a universal serial bus (USB) drive, or the like) having computer-readable program code embodied therein, wherein the computer-readable program code is adapted to be executed by a processor (e.g., working in connection with an operating system) to implement the processes described herein. In this regard, the program code may be implemented in any desired language, and may be implemented as machine code, assembly code, byte code, interpretable source code or the like (e.g., via C, C++, Java, Actionscript, Objective-C, Javascript, CSS, XML, and/or others).


Referring now to FIG. 1, the communication system 100 can include the microphones 102, loudspeaker 104, audio mixer 106, and an automatic reference gain adjustment module 108. Each of the microphones 102 (also referred to herein as a “microphone element”) can be configured to detect sound in the environment and convert the sound to an audio signal. In embodiments, some or all of the audio signals (also referred to herein as “near end audio signals”) from the microphones 102 may be processed by a beamformer 110 to generate one or more beamformed audio signals (or lobes), as is known in the art. In such cases, the beamformed audio signals may be provided to the audio mixer 106 as input audio signals, as shown in FIG. 1. While the following description may refer to audio signals from microphones 102 for the sake of simplicity, it is contemplated that the techniques described herein may be utilized with any type of acoustic source or signal, including the beamformed audio signals (or lobes) generated by beamformer 110. The beamformer 110 may be any suitable beamformer, such as, e.g., a delay and sum beamformer, etc.


As shown in FIG. 1, the audio mixer 106 receives audio signals, or audio input signals, from the microphones 102 and/or the beamformer 110 at respective audio input channels and generates a mixed audio signal using the received audio signals. The mixed audio signal generated by the mixer 106 may conform to a desired audio mix such that the audio signals from certain microphones 102 are emphasized and the audio signals from other microphones 102 are deemphasized or suppressed, as described herein. Exemplary embodiments of audio mixers are disclosed in commonly-assigned patents, U.S. Pat. Nos. 4,658,425 and 5,297,210, each of which is incorporated by reference in its entirety. In some embodiments, the mixed audio signal may be provided to an acoustic echo canceller (not shown) that is either included in, or coupled to, the communication system 100.


The loudspeaker 104 (also referred to herein as “audio output device”) can be configured to output audio signals received from the remote location (e.g., via wired or wireless communication). The received audio signals (also referred to herein as “far end audio signals”) may include speech and/or sound captured by microphones located at the remote location. Sometimes, one or more of the microphones 102 may pick up the far end audio as it is being played on the loudspeaker 104, in addition to the near end audio that is present in the environment. In such cases, if left unmitigated, the mixed audio signal may produce an undesirable echo due to persons at the remote location hearing their own speech and sound. One technique for minimizing this echo is to utilize a “gate-inhibit” technique that prevents the mixer 106 from outputting a mixed audio signal that includes the far end audio being played by the loudspeaker. This technique is achieved, at least in part, by providing the far end audio signal to a reference input channel of the mixer and adjusting a gain of the reference input channel so that the far end signal level is high enough to “win” against any “far end echoes” picked up by the microphones 102 and provided to the audio input channels of the mixer 106.


Existing reference gain adjustment systems require manual configuration by an installer while physically testing the environment for echo issues. Specifically, the installer manually boosts a gain of the reference input channel, for example, using a physical slider or other manual input device, until the gate inhibit feature of the mixer is activated. However, if the installer sets the gain too high, this may result in drowning out or suppressing the near end audio, particularly in double-talk situations, sluggish gating, and other undesirable effects. Thus, the effectiveness of existing systems depends greatly on the skill level of the installer or operator.


The communication system 100 shown in FIG. 1 includes an automatic reference gain adjustment module 108 configured to automatically adjust a gain amount applied to the reference input channel of the mixer 106 for receiving the far end audio signal transmitted by the remote location, based on one or more inputs from the communication system 100. According to embodiments, this gain amount (also referred to herein as “reference gain”) can be controlled using a reference gain 112, or reference gain device, which may be implemented in hardware (e.g., an amplifier or the like), software, or a combination thereof. Thus, the communication system 100 can precisely control the far end reference gain so as to mitigate an echo in the mixed audio signal provided to the remote location.


As shown in FIG. 1, the mixer 106 may have a plurality of audio input channels (e.g., N channels) for receiving audio signals captured by a plurality of microphones 102 (e.g., N or more microphones), a corresponding plurality of reference input channels (e.g., N channels) for receiving the same audio signals (e.g., to help the mixer 106 avoid reverberation and other undesirable effects due to multiple microphones picking up the same audio source), and an additional reference input channel for receiving the audio signal received from the remote location (i.e. the far end reference input channel). Based on the N+1 reference input channels (or “reference channels”), the mixer 106 selects mixer-matrix gains for each of the N channel inputs (i.e. the near end audio signals), such that the channels considered “active” are gated-on (e.g., by applying a gain of 0 dB) and the remaining channels, i.e. those considered “inactive,” or gated-off (e.g., by applying a low gain or attenuation, such as −20 dB). The input gains 114 are applied to all channel inputs, and the mixed output, or mixed audio signal, is obtained based on the selected combination of channel inputs and gains at the mixer 106. As described further below, when a far-end reference signal with sufficient gain is applied to the appropriate reference input channel of the mixer 106, the mixer 106 can gate-off select channel inputs, as needed, to prevent or minimize the presence of an echo in the mixed audio signal output by the mixer 106.


In the illustrated embodiment, each microphone 102 is coupled to, or corresponds to, a respective one of the audio input channels of the mixer 106. In other embodiments, the number of microphones 102 (i.e. “M” microphones) may exceed the number of audio input channels (i.e. M>N) and/or each audio input channel may be configured to receive a respective one of the plurality of beamformed audio signals (e.g., N signals) output by the beamformer 110, each beamformed audio signal including audio signals captured by multiple microphones 102.


As shown, each audio input channel has individually controllable gain and mute settings. The gain settings can be controlled using an input gain 114, or input gain device, that is coupled to each input channel of the mixer 106 and can be implemented using hardware (e.g., an amplifier or the like), software, or a combination thereof. Likewise, the mute settings can be controlled using a mute device 116 that is coupled to each input channel and can be implemented using hardware (e.g., a mute button or the like), software, or a combination thereof. In various embodiments, the gain devices 114 and/or the mute devices 116 may be included in the mixer 106, embedded in or coupled to the respective microphones 102, included in a separate component of the communication system 100, or a combination thereof.


The mute settings provide a mute status (e.g., muted or unmuted) of the audio input channel, or the microphone 102 coupled thereto, and based thereon, determine whether the corresponding audio input channel is active or inactive. Specifically, if a given microphone 102 is unmuted, the corresponding input channel will be deemed “active.” If, on the other hand, a given audio input channel or microphone 102 is muted, the audio input channel will be deemed “inactive.” In embodiments, muted or inactive input channels do not contribute to calculation of a median gain or “outliers” and thus, may not participate in the algorithm decision making described herein, for example, with respect to FIG. 3.


The input gain settings can be used to emphasize or de-emphasize certain active microphone(s) (e.g., the chairperson of a meeting) or audio input channels, while keeping the rest of the active microphones or channels at the same gain level. Since input gain levels affect the overall gain structure of the audio input channels, the input gain settings may be referred to as an input gain structure of the microphones 102. As shown in FIG. 1, the input gain structure is provided as an input to the automatic reference gain adjustment module 108 for calculating echo-to-reference estimates and the final reference gain 112, as described in further detail herein.


In embodiments, the input gain settings enable one or more active channels to be gated-on, or emphasized, more easily by the mixer 106. For example, if the input gain 114 for a first channel is increased or set to a higher level than that of the other channels, and audio at substantially similar sound levels are simultaneously received across multiple channels, including the first channel, the mixer 106 will perceive the first channel as having a higher gain level than the other channels and accordingly, will select, or gate-on, the first channel. As will be appreciated, decreasing the input gain 114 of a given channel can have the opposite effect, thus de-emphasizing the channel having a lower input gain level.


In some embodiments, the input gain for each microphone 102, or audio input channel, is a static setting that is either controlled manually by the installer or operator, or automatically set using an algorithm. For example, in a conference room situation, the input gain settings may be configured to emphasize the microphone 102 located at the head of the room or table. In other embodiments, the input gain is a flexible setting that can be adjusted for each meeting or event. For example, a select input gain 114 may be configured so that the microphone 102 being used by a specific speaker or participant is emphasized (or given preference), irrespective of their location in the room or table seat. In another example, each input gain 114 may be set to an equal gain level for one event, and a different combination of gain levels for another event. As will be appreciated, in the case of flexible input gain settings, the reference gain settings are adjusted accordingly.


As described herein, the mixer 106 can be configured to use a “gate-inhibit” technique, or algorithm, to prevent the audio input channels from “gating on,” or being activated, when they pick up the audio being played by the loudspeaker 104 (i.e. an echo). This involves boosting a gain of the far end reference input channel so that, when the far end signal is playing on the loudspeaker 104, the far end reference channel is the only channel that is gating on, and the remaining audio input channels are “gate-inhibited,” or prevented from gating on.


Typically, gating decisions are made by comparing an energy level (e.g., decibel level) of a given audio input channel to a current maximum energy level, e.g., the highest energy level of all input channels for a preset duration of time (e.g., 400 milliseconds), to determine whether the corresponding microphone 102 is picking up audio from an audio source, as will be appreciated. If the energy level is sufficiently high enough (e.g., equal to 90% or more of the maximum level), the audio input channel becomes active and, for example, a lobe is pointed towards the audio source (e.g., talker). To prevent the microphones 102 from gating on in response to picking up the loudspeaker audio, the audio input channels are additionally compared to the far end reference channel during a gating decision. If the signal level of a given audio input channel does not exceed a signal level of the far end reference channel, i.e. due to the audio input channel picking up the loudspeaker audio instead of the audio sources in the environment, the far end reference channel will “win” the gating decision, and the audio input channel will not be activated. However, since the far end reference channel is not actually a physical output channel, no outputs are provided to the mixer 106 for generating the mixed audio signal. In this manner, the far end audio can be prevented from causing audio input channels connected to the mixer 106 from gating on and creating an echo in the mixed audio signal.


The automatic reference gain adjustment module 108 (also referred to herein as a “reference gain adjuster”) can improve the functionality of this gate inhibit technique by automatically selecting an appropriate gain amount for the reference gain device 112 coupled to the far end reference input channel (or far end reference channel) based on an echo signal detected at one or more audio input channels (or audio channels), so that the far end audio signal “wins” the gating decision each time the microphones 102 pick up the audio playing on the loudspeaker 104 (i.e. an echo). In this manner, the reference gain amount can be configured to remove the echo signal from, or reduce a contribution of the echo signal to, the input audio signals before the audio mixer 106 generates the mixed audio signal for sending to the remote location. To achieve this, the reference gain adjuster 108 utilizes a reference gain calculator 118 and a coherence estimator 120 (also referred to herein as an “echo-to-reference estimator”), which may be included in the adjustment module 108, as shown in FIG. 1. In various embodiments, the reference gain adjuster 108 may be included in the mixer 106, provided as a separate component of the communication system 100, or a combination thereof. In some embodiments, the coherence estimator 120 may be part of an acoustic echo canceller (AEC) coupled to or included in the system 100, or may be configured to receive coherence information from the AEC.


According to embodiments, the reference gain calculator 118 is configured to calculate or determine an appropriate value or level for the reference gain 112 based on an echo-to-reference ratio received from the coherence estimator 120 and, in at least some cases, the input gain structure received from the input gain devices 114 (e.g., as described with reference to FIG. 3). The coherence estimator 120 is configured to determine the echo-to-reference ratio based on an echo level detected in the audio signal received at a given audio input channel. More specifically, the coherence estimator 120 may be configured to compare the reference signal to the near end audio signal to detect the presence of an echo, or echo signal, and determine a level of coherence based on the echo signal and the reference signal, and if there is high coherence, determine the echo-to-reference ratio for the given audio input channel by estimating an energy level of the echo signal (or “echo level”) and an energy level of the reference signal (or “reference level”).


The reference gain adjuster 108 is further configured to determine which active audio input channel has the highest (or worst) echo-to-reference ratio and use that highest ratio to determine the gain amount applied by the reference gain 112 to the far end reference input channel. More specifically, the reference gain calculator 118 is configured to automatically adjust or set the reference gain amount to a preset value that is a predetermined number of decibels (dB) higher than the highest echo-to-reference ratio. For example, in at least one embodiment, the reference gain calculator 118 calculates the reference gain amount by adding 6 dB to the highest echo-to-reference ratio determined for the active audio input channels. In this manner, the calculated reference gain will be high enough to “defeat” any echo signals present in the audio input channels of the mixer 106, but not so high that it would prevent the audio input channels from gating on when there is speech at the near end (e.g., a double-talk situation).


Referring now to FIG. 2, the above-described technique for automatically adjusting reference gain may be implemented using exemplary method or process 200, in accordance with embodiments. For example, process 200 may be performed by one or more components of the communication system 100, including the reference gain adjuster 108. Accordingly, process 200 will be described below with reference to the communication system 100 of FIG. 1.


As shown, the process 200 may begin at block 202 by selecting a first one of the N active audio input channels. As described herein, each audio input channel may have a mute setting that determines whether the channel is identified as “active” or “inactive” (e.g., an unmuted or muted status of the mute device 116, respectively). In embodiments, the reference gain adjuster 108 is configured to calculate echo-to-reference ratios for all active audio input channels, or otherwise analyze only the active audio channels for echo signals. Accordingly, block 202 may also include determining whether a given audio channel is active or inactive based on a mute setting associated with that channel.


At block 204, the coherence estimator 120 estimates or calculates a coherence level for the given active audio input channel, or how much of the audio signal received at the input channel is due to the far end audio signal (i.e. corresponds to echo) versus the near end audio signal. For example, the coherence estimator 120 may determine the coherence level by comparing the reference signal to the input audio signal using cross-correlation techniques or other suitable coherence computation involving the signal envelope and/or Fourier Transform of each signal. The coherence level may be a cross-correlation coefficient or other suitable metric obtained using said techniques.


Once the coherence level is estimated for the first channel, the process 200 continues to block 206 where the coherence estimator 120 determines whether there is high coherence between the input audio signal and the reference signal. For example, high coherence may be found if the coherence level calculated at step 204 is greater than a preset threshold (e.g., 80%). In embodiments, a high level of coherence can indicate that only the far-end audio is present in the audio signal received at the given input channel, i.e. the input audio signal only or mostly contains a far end echo (or echo signal) and/or no near end audio is present in the environment (e.g., because the near end talkers are presumably silent). A low level of coherence can indicate that the input audio signal is most likely due to near-end talkers (i.e. no or very little far-end echo is present in the environment). A medium level of coherence can indicate the presence of “double-talk” in the input audio signal (i.e. both near end audio and far end audio are present in the environment).


If there is high coherence between the two signals at block 206 (i.e. the input audio signal primarily contains a far-end echo), the input audio signal is treated as the “echo signal,” and the process 200 continues to block 208, where the coherence estimator 120 calculates a new estimate for the echo level (“Echo(Ch)”) based on an energy level of the input audio signal (which is the sum of the far-end echo and near-end speech, if any) and a new estimate for the reference level (“Ref”) based on an energy level of the reference signal (i.e. the far end audio). If, on the other hand, high coherence is not found at block 206 (i.e. the coherence level is determined to be medium or low), the process 200 continues to block 209, where the coherence estimator 120 uses or retrieves previous estimates for the echo level and reference level. That is, at block 208, the Echo(Ch) and Ref levels are updated based on current estimates, while at block 209, the Echo(Ch) and Ref levels remain at previously calculated values, i.e. are not adjusted or re-calculated.


In embodiments, the process 200 may be an iterative process that improves over time as the estimates calculated at block 208, and therefore, the final auto reference gain amount, becomes more accurate. In such cases, when the process 200 is initialized or executed for the first time, a set of initial values may be assigned to each of the echo level and the reference level (e.g., −60 dB or other small positive value) and the determination at block 206 may necessarily be “no” (i.e. no high coherence). The initial values may be replaced with newly estimated values for the echo and reference levels after the coherence estimator 120 first identifies the existence of high coherence at block 206. The estimated values may be stored in a memory of the system 100 and continuously updated at block 208 each time high coherence is found at block 206. When high coherence is not found, the last set of estimates may be used at block 209 as the echo level and reference level for the current auto reference gain calculation. In this manner, the reference gain amount can remain high enough to offset future echo situations, even if a current situation does not contain an echo.


In embodiments, the echo level and reference level estimates may be calculated using respective exponentially weighted first order moving average equations, namely:

Echo(Ch)=A*Echo(Ch)+(1−A)*AVG_ECHO_ENERGY
Ref=A*Ref+(1−A)*AVG_REF_ENERGY


In the above equations, the parameter “A” is a smoothing factor, such as, for example, a discrete-time equivalent of a time constant (for a low-pass first order filter). This parameter, along with the first order moving average equations, may be used to smooth out the energy estimates over time and prevent sudden and/or random jumps in the estimates. In other embodiments, other suitable equations or energy estimators may be used to smooth out the energy estimates, such as, for example, a different smoothing method or a higher-order low pass filter.


From either block 208 or 209, the process 200 continues to block 210. At block 210, an echo-to-reference level, or ratio, for the given channel (“Echo2Ref(Ch)”) is calculated converting a ratio of echo level to reference level (Echo(Ch)/Ref) to decibel (dB) units using the equation: 10*log(Echo(Ch)/Ref). As should be appreciated, the exact values used for “Echo(Ch)” and “Ref” at block 210 will vary depending on whether or not these values were updated at block 208.


At block 212, the coherence estimator 120 determines whether there are more channels to analyze. If there are (i.e. Ch<N), the next channel is selected at block 214 and blocks 204 through 210 are repeated for that channel. This continues until an echo-to-reference ratio has been calculated for each remaining active audio channel, or all N active channels have been analyzed (i.e. Ch=N).


Once all active channels have been analyzed, the process 200 further includes, either at block 216 or just prior to block 216, comparing the echo-to-reference ratios calculated for the N active channels using the coherence estimator 120 to determine which audio input channel has the highest echo-to-reference ratio (“Max_Echo2Ref”). In some embodiments, the coherence estimator 120 provides the echo-to-reference ratio estimated for each audio channel to the reference gain calculator 118, and the reference gain calculator 118 determines which of the received ratios is the highest. In other embodiments, the coherence estimator 120 determines which echo-to-reference ratio is the highest and provides only the highest ratio to the reference gain calculator 118.


At block 216, the reference gain calculator 118 calculates a reference gain amount using the highest echo-to-reference ratio. In particular, the reference gain calculator 118 sets or adjusts the reference gain to a preset value (e.g., 6 dB) above the highest echo-to-reference ratio (i.e. AutoRefGain=6 dB+Max_Echo2Ref). The reference gain calculator 118 then uses the final reference gain amount to adjust the gain 112 applied to the far end reference channel.


In process 200, it is assumed that all input gains 114 are set to the same value or level and therefore, it is sufficient to use the worst case scenario, i.e. the highest echo-to-reference ratio among all active audio channels, to determine the reference gain amount. However, in cases where there is an extra gain structure, or boost, for certain input channels (referred to herein as “outliers”), the appropriate reference gain for the far end reference input channel must be adjusted to “defeat” the outliers. In such cases, process 300 shown in FIG. 3 may be used to calculate the reference gain amount.


Referring now to FIG. 3, process 300 is configured for performing automatic reference gain adjustment when one or more of the microphone input gains (e.g., input gains 114 in FIG. 1) are boosted, or has a specific input gain setting that affects the echo-to-reference level estimate for that channel. The process 300 may be performed by one or more components of the communication system 100, such as the reference gain adjuster 108 shown in FIG. 1, in accordance with embodiments. The input gain “outliers” may exist because, for example, the installer manually boosted the input gain 114 of one or more audio input channels (e.g., to boost the microphone 102 located at the head of the room, etc.).


In various embodiments, the process 300 may begin after completing block 216 of the process 200 shown in FIG. 2. Namely, once the echo-to-reference ratios are calculated for all N active channels, and a maximum echo-to-reference ratio is determined and provided to the reference gain calculator 118, the process 300 may begin at block 302 with the reference gain calculator 118 receiving an input gain structure of the audio input channels from the input gains 114 shown in FIG. 1, and calculating a median input gain (“MIG”) for all N active input channels based on these input gain amounts.


At block 304, the reference gain calculator 118 selects a first active audio input channel. At block 306, the reference gain calculator 118 scans or analyzes the selected channel in order to determine whether the channel is an outlier, or a channel whose input gain 114 is set to a level that is higher than the sum of MIG and XdB (where X is typically 8). If the determination at block 306 is “yes,” a counter for the number of outliers (NumOutliers) is incremented by 1 at block 308, and the process 300 goes directly to block 314. If the determination at block 306 is “no,” the process 300 continues to block 310.


At block 310, the reference gain calculator 118 determines whether the echo-to-reference level for the given channel (“Echo2Ref(Ch)”) is greater than a maximum echo-to-reference level estimate over all channels but excluding outlier channels (denoted as “Max_Echo2Ref_withoutOutliers”). If the answer is “yes,” the process 300 continues to block 312, where the Max_Echo2Ref_withoutOutliers estimate is set equal to the echo-to-reference level for the given channel. From block 312, the process 300 continues to block 314. If the answer at block 310 is “no,” the process 300 goes directly to block 314.


At block 314, the reference gain calculator 118 determines whether the echo-to-reference level for the given channel (“Echo2Ref(Ch)”) is greater than a maximum echo-to-reference estimate over all channels, including outlier channels (denoted as “Max_Echo2Ref total”). If the answer is “yes,” the process 300 continues to block 316, where the Max_Echo2Ref total estimate is set equal to the echo-to-reference level for the given channel. From block 316, the process 300 continues to block 318. If the answer at block 314 is “no,” the process 300 goes directly to block 318. In embodiments, the initial values for Max_Echo2Ref_withoutOutliers and Max_Echo2Ref total may be set to −∞ dB.


At block 318, the reference gain calculator 118 determines whether there are any other active audio input channels (i.e. CH<N). If there are, process 300 continues to block 319, where the next active channel is selected and blocks 306 through 316 are repeated for that channel. This loop may continue until there are no more active channels remaining (i.e. CH=N or the determination at block 318 is “no”), at which point the process 300 continues to block 320.


At block 320, if there is only one input channel with a very high input gain (i.e. NumOutliers=1, or only one outlier channel), the process 300 continues to block 322, where the reference gain (“AutoRefGain”) is automatically set to YdB higher than the Max_Echo2Ref_withoutOutliers estimate (where Y is typically 6), thus basically ignoring the outlier channel. This may be preferred in embodiments so as to not allow the outlier channel with the high input gain to set the reference gain to a value that is too high, as that would make it very difficult for near-end talkers to “break in,” or for the near end audio to be picked up by the mixer 106.


If there is more than one outlier channel at block 320 (i.e. NumOutliers>1), the process 300 continues to block 324, where the reference gain (“AutoRefGain”) is calculated using the following equation:

AutoRefGain=YdB+Max_Echo2Ref_withoutOutliers+((NumOutliers/N)*(Max_Echo2Ref total−Max_Echo2Ref_withoutOutliers)),


where Y is typically 6. In other words, the reference gain is linearly increased based on the number of outliers when there is more than one outlier. This allows the channels with high input gains to at least somewhat raise the amount of reference gain that is automatically applied.


The process 300 may end once the reference gain amount calculated at blocks 322 or 324 is applied to the reference gain 112 of the reference input channel. In this manner, the echo signal can be removed from the near end audio signals before a mixed audio signal is generated by the audio mixer 106 based on the near end audio signals.


Referring back to FIG. 1, the communication system 100 may also include a speaker gain 122 for controlling a gain level of the loudspeaker 104. The speaker gain 122 may affect the overall gain structure of the input channels and the “Echo2Ref” ratios estimated in method 200, as well as the final value of the reference gain calculated by either of the methods 200 and 300. According to embodiments, the speaker gain 122, or speaker gain device, may be implemented in hardware (e.g., an amplifier or the like), software, or a combination thereof.


Thus, systems and methods are provided herein that can mitigate an echo in a mixed audio signal provided to a remote location, wherein the mixed audio signal is generated by a mixer using audio signals captured by microphones in an environment that further includes a loudspeaker for playing a far end audio signal received from the remote location. The microphones can be prevented from contributing far end audio to the mixed audio signal by automatically adjusting a gain applied to a far end reference input channel that is used by the mixer to determine whether the audio channels should be activated due to detection of near end audio.


While embodiments shown and described herein include multiple audio input channels, other embodiments may include a communication system with a single audio input channel. Also, though a particular audio communication system is shown and described herein, other embodiments may include any type of system with multiple, or single, input channels and a reference input channel for controlling gating decisions based on a contribution of the reference signal in the environment. For example, the techniques described herein may be used in other situations where setting the reference gain automatically to an appropriate level can help determine the activity/contribution of the reference signal in the near-end signal (mix), and can allow the mixer decisions to be used to temporarily inhibit, disable, enable, or otherwise control other types of actions or mechanisms, such as, for example, voice, noise, and/or disturbance activity detectors in the near end, which attempt to determine if near end sounds in the room represent talkers or noise but should not do so if the near-end source of sound is associated with an echo; automatic focus of existing lobes, which attempts to better focus a beam to the exact position of the near-end talker in the room, but should not do so if the near-end source of sound is associated with an echo; and/or automatic placement of various lobes, which attempts to initially place a beam or lobe towards the position of the near-end talker in the room, but should not do so if that near-end source of sound is associated with an echo. Examples of such other systems are disclosed in commonly-assigned patent application U.S. Ser. No. 16/887,790 and U.S. Ser. No. 16/826,115, each of which is incorporated by reference in its entirety.


Any process descriptions or blocks in the figures, such as, e.g., FIGS. 2 and 3, should be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process, and alternate implementations are included within the scope of the embodiments of the invention in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those having ordinary skill in the art.


This disclosure is intended to explain how to fashion and use various embodiments in accordance with the technology rather than to limit the true, intended, and fair scope and spirit thereof. The foregoing description is not intended to be exhaustive or to be limited to the precise forms disclosed. Modifications or variations are possible in light of the above teachings. The embodiment(s) were chosen and described to provide the best illustration of the principle of the described technology and its practical application, and to enable one of ordinary skill in the art to utilize the technology in various embodiments and with various modifications as are suited to the particular use contemplated. All such modifications and variations are within the scope of the embodiments as determined by the appended claims, as may be amended during the pendency of this application for patent, and all equivalents thereof, when interpreted in accordance with the breadth to which they are fairly, legally and equitably entitled.

Claims
  • 1. A system, comprising: a plurality of microphone elements, each microphone element configured to detect sound and convert the sound to a near end audio signal;at least one audio output device configured to output a far end audio signal comprising sound captured at a remote location;an audio mixer having a plurality of audio input channels for receiving audio signals captured by the plurality of microphone elements and a reference channel for receiving the far end audio signal as a reference signal; anda reference gain adjuster configured to automatically determine a reference gain amount for the reference channel based on an echo level detected in an input audio signal received at a given audio input channel,wherein the reference gain adjuster determines the reference gain amount based on an echo-to-reference ratio calculated using the echo level and a reference level calculated for the reference signal.
  • 2. The system of claim 1, wherein the reference gain adjuster is further configured to determine a coherence level based on the input audio signal and the reference signal, and if high coherence is found, calculating the echo-to-reference ratio based on the echo level.
  • 3. The system of claim 1, wherein the reference gain adjuster is further configured to: calculate an echo-to-reference ratio for each remaining audio input channel,determine which audio input channel has a highest echo-to-reference ratio, andcalculate the reference gain amount using the highest echo-to-reference ratio.
  • 4. The system of claim 3, wherein the reference gain adjuster further determines the reference gain amount by adding a preset value to the highest echo-to-reference ratio.
  • 5. The system of claim 3, wherein each audio input channel has a mute setting that determines if the channel is active or inactive, and the reference gain adjuster is configured to calculate echo-to-reference ratios for all active audio input channels.
  • 6. The system of claim 1, wherein the reference gain adjuster is configured to determine the reference gain amount based further on an input gain structure of the audio input channels.
  • 7. The system of claim 6, wherein each audio input channel has an individually-controllable input gain setting that is included in the input gain structure of the audio input channels.
  • 8. The system of claim 6, wherein the reference gain adjuster is further configured to: calculate an average input gain based on the input gain structure of the audio input channels;identify a number of the audio input channels as having an input gain setting that is higher than the average input gain; andadjust the reference gain amount based on the echo-to-reference ratios calculated for the number of audio input channels.
  • 9. The system of claim 1, wherein the audio mixer is configured to generate a mixed audio signal based on the audio signals captured by the microphones for sending to the remote location.
  • 10. The system of claim 1, wherein the audio mixer is configured to apply the reference gain amount to the reference channel and make gating decisions for the plurality of audio input channels based on the reference channel.
  • 11. A method of automatically adjusting a reference gain of an audio mixer having a reference channel for receiving a far end audio signal from a remote location as a reference signal and a plurality of audio input channels for receiving audio signals captured by a plurality of microphone elements, the method comprising: determining an echo level in an input audio signal received at a given audio input channel; andautomatically determining a reference gain amount for the reference channel based on the echo level by: calculating an echo-to-reference ratio for the given audio input channel using the echo level and a reference level calculated for the reference signal, andcalculating the reference gain amount based on the echo-to-reference ratio.
  • 12. The method of claim 11, further comprising: determining a coherence level based on the input audio signal and the reference signal; andif high coherence is found, calculating the echo-to-reference ratio based on the echo level.
  • 13. The method of claim 11, further comprising: calculating an echo-to-reference ratio for each remaining audio input channel,determining which audio input channel has a highest echo-to-reference ratio, andcalculating the reference gain amount using the highest echo-to-reference ratio.
  • 14. The method of claim 13, wherein calculating the reference gain amount further comprises adding a preset value to the highest echo-to-reference ratio.
  • 15. The method of claim 13, further comprising: determining whether each audio input channel is active or inactive based on a mute setting associated with that channel, andanalyzing all active audio input channels for echo signals.
  • 16. The method of claim 11, further comprising: determining the reference gain amount based further on an input gain structure of the audio input channels.
  • 17. The method of claim 16, wherein each audio input channel has an individually-controllable input gain setting that is included in the input gain structure of the audio input channels.
  • 18. The method of claim 16, further comprising: calculating an average input gain based on the input gain structure of the audio input channels;identifying a number of the audio input channels as having an input gain setting that is higher than the average input gain; andadjusting the reference gain amount based on the echo-to-reference ratios calculated for the number of audio input channels.
  • 19. The method of claim 11, further comprising: generating a mixed audio signal based on said audio signals.
  • 20. The method of claim 11, further comprising: applying the reference gain amount to the reference channel; andmaking gating decisions for the plurality of audio input channels based on the reference channel.
CROSS-REFERENCE

This application claims the benefit of U.S. Provisional Patent Application No. 62/971,850, filed on Feb. 7, 2020, the contents of which is incorporated by reference herein in its entirety.

US Referenced Citations (977)
Number Name Date Kind
1535408 Fricke Apr 1925 A
1540788 McClure Jun 1925 A
1965830 Hammer Jul 1934 A
2075588 Meyers Mar 1937 A
2113219 Olson Apr 1938 A
2164655 Kleerup Jul 1939 A
D122771 Doner Oct 1940 S
2233412 Hill Mar 1941 A
2268529 Stiles Dec 1941 A
2343037 Adelman Feb 1944 A
2377449 Prevette Jun 1945 A
2481250 Schneider Sep 1949 A
2521603 Prew Sep 1950 A
2533565 Eichelman Dec 1950 A
2539671 Olson Jan 1951 A
2777232 Kulicke Jan 1957 A
2828508 Labarre Apr 1958 A
2840181 Wildman Jun 1958 A
2882633 Howell Apr 1959 A
2912605 Tibbetts Nov 1959 A
2938113 Schnell May 1960 A
2950556 Larios Aug 1960 A
3019854 Obryant Feb 1962 A
3132713 Seeler May 1964 A
3143182 Sears Aug 1964 A
3160225 Sechrist Dec 1964 A
3161975 McMillan Dec 1964 A
3205601 Gawne Sep 1965 A
3239973 Hannes Mar 1966 A
3240883 Seeler Mar 1966 A
3310901 Sarkisian Mar 1967 A
3321170 Vye May 1967 A
3509290 Mochida Apr 1970 A
3573399 Schroeder Apr 1971 A
3657490 Scheiber Apr 1972 A
3696885 Grieg Oct 1972 A
3755625 Maston Aug 1973 A
3828508 Moeller Aug 1974 A
3857191 Sadorus Dec 1974 A
3895194 Fraim Jul 1975 A
3906431 Clearwaters Sep 1975 A
D237103 Fisher Oct 1975 S
3936606 Wanke Feb 1976 A
3938617 Forbes Feb 1976 A
3941638 Horky Mar 1976 A
3992584 Dugan Nov 1976 A
4007461 Luedtke Feb 1977 A
4008408 Kodama Feb 1977 A
4029170 Phillips Jun 1977 A
4032725 McGee Jun 1977 A
4070547 Dellar Jan 1978 A
4072821 Bauer Feb 1978 A
4096353 Bauer Jun 1978 A
4127156 Brandt Nov 1978 A
4131760 Christensen Dec 1978 A
4169219 Beard Sep 1979 A
4184048 Alcaide Jan 1980 A
4198705 Massa Apr 1980 A
D255234 Wellward Jun 1980 S
D256015 Doherty Jul 1980 S
4212133 Lufkin Jul 1980 A
4237339 Bunting Dec 1980 A
4244096 Kashichi Jan 1981 A
4244906 Heinemann Jan 1981 A
4254417 Speiser Mar 1981 A
4275694 Nagaishi Jun 1981 A
4296280 Richie Oct 1981 A
4305141 Massa Dec 1981 A
4308425 Momose Dec 1981 A
4311874 Wallace, Jr. Jan 1982 A
4330691 Gordon May 1982 A
4334740 Wray Jun 1982 A
4365449 Liautaud Dec 1982 A
4373191 Fette Feb 1983 A
4393631 Krent Jul 1983 A
4414433 Horie Nov 1983 A
4429850 Weber Feb 1984 A
4436966 Botros Mar 1984 A
4449238 Lee May 1984 A
4466117 Goerike Aug 1984 A
4485484 Flanagan Nov 1984 A
4489442 Anderson Dec 1984 A
4518826 Caudill May 1985 A
4521908 Miyaji Jun 1985 A
4566557 Lemaitre Jan 1986 A
4593404 Bolin Jun 1986 A
4594478 Gumb Jun 1986 A
D285067 Delbuck Aug 1986 S
4625827 Bartlett Dec 1986 A
4653102 Hansen Mar 1987 A
4658425 Julstrom Apr 1987 A
4669108 Deinzer May 1987 A
4675906 Sessler Jun 1987 A
4693174 Anderson Sep 1987 A
4696043 Iwahara Sep 1987 A
4712231 Julstrom Dec 1987 A
4741038 Elko Apr 1988 A
4752961 Kahn Jun 1988 A
4805730 O'Neill Feb 1989 A
4815132 Minami Mar 1989 A
4860366 Fukushi Aug 1989 A
4862507 Woodard Aug 1989 A
4866868 Kass Sep 1989 A
4881135 Heilweil Nov 1989 A
4888807 Reichel Dec 1989 A
4903247 Van Gerwen Feb 1990 A
4923032 Nuernberger May 1990 A
4928312 Hill May 1990 A
4969197 Takaya Nov 1990 A
5000286 Crawford Mar 1991 A
5038935 Wenkman Aug 1991 A
5058170 Kanamori Oct 1991 A
5088574 Kertesz, III Feb 1992 A
D324780 Sebesta Mar 1992 S
5121426 Baumhauer Jun 1992 A
D329239 Hahn Sep 1992 S
5189701 Jain Feb 1993 A
5204907 Staple Apr 1993 A
5214709 Ribic May 1993 A
D340718 Leger Oct 1993 S
5289544 Franklin Feb 1994 A
D345346 Alfonso Mar 1994 S
D345379 Chan Mar 1994 S
5297210 Julstrom Mar 1994 A
5322979 Cassity Jun 1994 A
5323459 Hirano Jun 1994 A
5329593 Lazzeroni Jul 1994 A
5335011 Addeo Aug 1994 A
5353279 Koyama Oct 1994 A
5359374 Schwartz Oct 1994 A
5371789 Hirano Dec 1994 A
5383293 Royal Jan 1995 A
5384843 Masuda Jan 1995 A
5396554 Hirano Mar 1995 A
5400413 Kindel Mar 1995 A
D363045 Phillips Oct 1995 S
5473701 Cezanne Dec 1995 A
5509634 Gebka Apr 1996 A
5513265 Hirano Apr 1996 A
5525765 Freiheit Jun 1996 A
5550924 Helf Aug 1996 A
5550925 Hori Aug 1996 A
5555447 Kotzin Sep 1996 A
5574793 Hirschhorn Nov 1996 A
5602962 Kellermann Feb 1997 A
5633936 Oh May 1997 A
5645257 Ward Jul 1997 A
D382118 Ferrero Aug 1997 S
5657393 Crow Aug 1997 A
5661813 Shimauchi Aug 1997 A
5673327 Julstrom Sep 1997 A
5687229 Sih Nov 1997 A
5706344 Finn Jan 1998 A
5715319 Chu Feb 1998 A
5717171 Miller Feb 1998 A
D392977 Kim Mar 1998 S
D394061 Fink May 1998 S
5761318 Shimauchi Jun 1998 A
5766702 Lin Jun 1998 A
5787183 Chu Jul 1998 A
5796819 Romesburg Aug 1998 A
5848146 Slattery Dec 1998 A
5870482 Loeppert Feb 1999 A
5878147 Killion Mar 1999 A
5888412 Sooriakumar Mar 1999 A
5888439 Miller Mar 1999 A
D416315 Nanjo Nov 1999 S
5978211 Hong Nov 1999 A
5991277 Maeng Nov 1999 A
6035962 Lin Mar 2000 A
6039457 O'Neal Mar 2000 A
6041127 Elko Mar 2000 A
6049607 Marash Apr 2000 A
D424538 Hayashi May 2000 S
6069961 Nakazawa May 2000 A
6125179 Wu Sep 2000 A
D432518 Muto Oct 2000 S
6128395 De Vries Oct 2000 A
6137887 Anderson Oct 2000 A
6144746 Azima Nov 2000 A
6151399 Killion Nov 2000 A
6173059 Huang Jan 2001 B1
6198831 Azima Mar 2001 B1
6205224 Underbrink Mar 2001 B1
6215881 Azima Apr 2001 B1
6266427 Mathur Jul 2001 B1
6285770 Azima Sep 2001 B1
6301357 Romesburg Oct 2001 B1
6329908 Frecska Dec 2001 B1
6332029 Azima Dec 2001 B1
D453016 Nevill Jan 2002 S
6386315 Roy May 2002 B1
6393129 Conrad May 2002 B1
6424635 Song Jul 2002 B1
6442272 Osovets Aug 2002 B1
6449593 Valve Sep 2002 B1
6481173 Roy Nov 2002 B1
6488367 Debesis Dec 2002 B1
D469090 Tsuji Jan 2003 S
6505057 Finn Jan 2003 B1
6507659 Iredale Jan 2003 B1
6510919 Roy Jan 2003 B1
6526147 Rung Feb 2003 B1
6556682 Gilloire Apr 2003 B1
6592237 Pledger Jul 2003 B1
6622030 Romesburg Sep 2003 B1
D480923 Neubourg Oct 2003 S
6633647 Markow Oct 2003 B1
6665971 Lowry Dec 2003 B2
6694028 Matsuo Feb 2004 B1
6704422 Jensen Mar 2004 B1
D489707 Kobayashi May 2004 S
6731334 Maeng May 2004 B1
6741720 Myatt May 2004 B1
6757393 Spitzer Jun 2004 B1
6768795 Feltstrom Jul 2004 B2
6868377 Laroche Mar 2005 B1
6885750 Egelmeers Apr 2005 B2
6885986 Gigi Apr 2005 B1
D504889 Andre May 2005 S
6889183 Gunduzhan May 2005 B1
6895093 Ali May 2005 B1
6931123 Hughes Aug 2005 B1
6944312 Mason Sep 2005 B2
D510729 Chen Oct 2005 S
6968064 Ning Nov 2005 B1
6990193 Beaucoup Jan 2006 B2
6993126 Kyrylenko Jan 2006 B1
6993145 Combest Jan 2006 B2
7003099 Zhang Feb 2006 B1
7013267 Huart Mar 2006 B1
7031269 Lee Apr 2006 B2
7035398 Matsuo Apr 2006 B2
7035415 Belt Apr 2006 B2
7050576 Zhang May 2006 B2
7054451 Janse May 2006 B2
D526643 Ishizaki Aug 2006 S
D527372 Allen Aug 2006 S
7092516 Furuta Aug 2006 B2
7092882 Arrowood Aug 2006 B2
7098865 Christensen Aug 2006 B2
7106876 Santiago Sep 2006 B2
7120269 Lowell Oct 2006 B2
7130309 Boaz Oct 2006 B2
D533177 Andre Dec 2006 S
7149320 Haykin Dec 2006 B2
7161534 Tsai Jan 2007 B2
7187765 Popovic Mar 2007 B2
7203308 Kubota Apr 2007 B2
D542543 Bruce May 2007 S
7212628 Mirjana May 2007 B2
D546318 Yoon Jul 2007 S
D546814 Takita Jul 2007 S
D547748 Tsuge Jul 2007 S
7239714 de Blok Jul 2007 B2
D549673 Niitsu Aug 2007 S
7269263 Dedieu Sep 2007 B2
D552570 Niitsu Oct 2007 S
D559553 Mischel Jan 2008 S
7333476 LeBlanc Feb 2008 B2
D566685 Koller Apr 2008 S
7359504 Reuss Apr 2008 B1
7366310 Stinson Apr 2008 B2
7387151 Payne Jun 2008 B1
7412376 Florencio Aug 2008 B2
7415117 Tashev Aug 2008 B2
D578509 Thomas Oct 2008 S
D581510 Albano Nov 2008 S
D582391 Morimoto Dec 2008 S
D587709 Niitsu Mar 2009 S
D589605 Reedy Mar 2009 S
7503616 Linhard Mar 2009 B2
7515719 Hooley Apr 2009 B2
7536769 Pedersen May 2009 B2
D595402 Miyake Jun 2009 S
D595736 Son Jul 2009 S
7558381 Ali Jul 2009 B1
7565949 Tojo Jul 2009 B2
D601585 Andre Oct 2009 S
7651390 Profeta Jan 2010 B1
7660428 Rodman Feb 2010 B2
7667728 Kenoyer Feb 2010 B2
7672445 Zhang Mar 2010 B1
D613338 Marukos Apr 2010 S
7701110 Fukuda Apr 2010 B2
7702116 Stone Apr 2010 B2
D614871 Tang May 2010 S
7724891 Beaucoup May 2010 B2
D617441 Koury Jun 2010 S
7747001 Kellermann Jun 2010 B2
7756278 Moorer Jul 2010 B2
7783063 Pocino Aug 2010 B2
7787328 Chu Aug 2010 B2
7830862 James Nov 2010 B2
7831035 Stokes Nov 2010 B2
7831036 Beaucoup Nov 2010 B2
7856097 Tokuda Dec 2010 B2
7881486 Killion Feb 2011 B1
7894421 Kwan Feb 2011 B2
D636188 Kim Apr 2011 S
7925006 Hirai Apr 2011 B2
7925007 Stokes Apr 2011 B2
7936886 Kim May 2011 B2
7970123 Beaucoup Jun 2011 B2
7970151 Oxford Jun 2011 B2
D642385 Lee Aug 2011 S
D643015 Kim Aug 2011 S
7991167 Oxford Aug 2011 B2
7995768 Miki Aug 2011 B2
8000481 Nishikawa Aug 2011 B2
8005238 Tashev Aug 2011 B2
8019091 Burnett Sep 2011 B2
8041054 Yeldener Oct 2011 B2
8059843 Hung Nov 2011 B2
8064629 Jiang Nov 2011 B2
8085947 Haulick Dec 2011 B2
8085949 Kim Dec 2011 B2
8095120 Blair Jan 2012 B1
8098842 Florencio Jan 2012 B2
8098844 Elko Jan 2012 B2
8103030 Barthel Jan 2012 B2
8109360 Stewart, Jr. Feb 2012 B2
8112272 Nagahama Feb 2012 B2
8116500 Oxford Feb 2012 B2
8121834 Rosec Feb 2012 B2
D655271 Park Mar 2012 S
D656473 Laube Mar 2012 S
8130969 Buck Mar 2012 B2
8130977 Chu Mar 2012 B2
8135143 Ishibashi Mar 2012 B2
8144886 Ishibashi Mar 2012 B2
D658153 Woo Apr 2012 S
8155331 Nakadai Apr 2012 B2
8170882 Davis May 2012 B2
8175291 Chan May 2012 B2
8175871 Wang May 2012 B2
8184801 Hamalainen May 2012 B1
8189765 Nishikawa May 2012 B2
8189810 Wolff May 2012 B2
8194863 Takumai Jun 2012 B2
8199927 Raftery Jun 2012 B1
8204198 Adeney Jun 2012 B2
8204248 Haulick Jun 2012 B2
8208664 Iwasaki Jun 2012 B2
8213596 Beaucoup Jul 2012 B2
8213634 Daniel Jul 2012 B1
8219387 Cutler Jul 2012 B2
8229134 Duraiswami Jul 2012 B2
8233352 Beaucoup Jul 2012 B2
8243951 Ishibashi Aug 2012 B2
8244536 Arun Aug 2012 B2
8249273 Inoda Aug 2012 B2
8259959 Marton Sep 2012 B2
8275120 Stokes, III Sep 2012 B2
8280728 Chen Oct 2012 B2
8284949 Farhang Oct 2012 B2
8284952 Reining Oct 2012 B2
8286749 Stewart Oct 2012 B2
8290142 Lambert Oct 2012 B1
8291670 Gard Oct 2012 B2
8297402 Stewart Oct 2012 B2
8315380 Liu Nov 2012 B2
8331582 Steele Dec 2012 B2
8345898 Reining Jan 2013 B2
8355521 Larson Jan 2013 B2
8370140 Vitte Feb 2013 B2
8379823 Ratmanski Feb 2013 B2
8385557 Tashev Feb 2013 B2
D678329 Lee Mar 2013 S
8395653 Feng Mar 2013 B2
8403107 Stewart Mar 2013 B2
8406436 Craven Mar 2013 B2
8428661 Chen Apr 2013 B2
8433061 Cutler Apr 2013 B2
D682266 Wu May 2013 S
8437490 Marton May 2013 B2
8443930 Stewart, Jr. May 2013 B2
8447590 Ishibashi May 2013 B2
8472639 Reining Jun 2013 B2
8472640 Marton Jun 2013 B2
D685346 Szymanski Jul 2013 S
D686182 Ashiwa Jul 2013 S
8479871 Stewart Jul 2013 B2
8483398 Fozunbal Jul 2013 B2
8498423 Thaden Jul 2013 B2
D687432 Duan Aug 2013 S
8503653 Ahuja Aug 2013 B2
8515089 Nicholson Aug 2013 B2
8515109 Dittberner Aug 2013 B2
8526633 Ukai Sep 2013 B2
8553904 Said Oct 2013 B2
8559611 Ratmanski Oct 2013 B2
D693328 Goetzen Nov 2013 S
8583481 Viveiros Nov 2013 B2
8599194 Lewis Dec 2013 B2
8600443 Kawaguchi Dec 2013 B2
8605890 Zhang Dec 2013 B2
8620650 Walters Dec 2013 B2
8631897 Stewart Jan 2014 B2
8634569 Lu Jan 2014 B2
8638951 Zurek Jan 2014 B2
D699712 Bourne Feb 2014 S
8644477 Gilbert Feb 2014 B2
8654955 Lambert Feb 2014 B1
8654990 Faller Feb 2014 B2
8660274 Wolff Feb 2014 B2
8660275 Buck Feb 2014 B2
8670581 Harman Mar 2014 B2
8672087 Stewart Mar 2014 B2
8675890 Schmidt Mar 2014 B2
8675899 Jung Mar 2014 B2
8676728 Velusamy Mar 2014 B1
8682675 Togami Mar 2014 B2
8724829 Visser May 2014 B2
8730156 Weising May 2014 B2
8744069 Cutler Jun 2014 B2
8744101 Burns Jun 2014 B1
8755536 Chen Jun 2014 B2
8811601 Mohammad Aug 2014 B2
8818002 Tashev Aug 2014 B2
8824693 Åhgren Sep 2014 B2
8842851 Beaucoup Sep 2014 B2
8855326 Derkx Oct 2014 B2
8855327 Tanaka Oct 2014 B2
8861713 Xu Oct 2014 B2
8861756 Zhu Oct 2014 B2
8873789 Bigeh Oct 2014 B2
D717272 Kim Nov 2014 S
8886343 Ishibashi Nov 2014 B2
8893849 Hudson Nov 2014 B2
8898633 Bryant Nov 2014 B2
D718731 Lee Dec 2014 S
8903106 Meyer Dec 2014 B2
8923529 McCowan Dec 2014 B2
8929564 Kikkeri Jan 2015 B2
8942382 Elko Jan 2015 B2
8965546 Visser Feb 2015 B2
D725059 Kim Mar 2015 S
D725631 McNamara Mar 2015 S
8976977 De Sena et al. Mar 2015 B2
8983089 Chu Mar 2015 B1
8983834 Davis Mar 2015 B2
D726144 Kang Apr 2015 S
D727968 Onoue Apr 2015 S
9002028 Haulick Apr 2015 B2
D729767 Lee May 2015 S
9038301 Zelbacher May 2015 B2
9088336 Mani Jul 2015 B2
9094496 Teutsch Jul 2015 B2
D735717 Lam Aug 2015 S
D737245 Fan Aug 2015 S
9099094 Burnett Aug 2015 B2
9107001 Diethorn Aug 2015 B2
9111543 Åhgren Aug 2015 B2
9113242 Hyun Aug 2015 B2
9113247 Chatlani Aug 2015 B2
9126827 Hsieh Sep 2015 B2
9129223 Velusamy Sep 2015 B1
9140054 Oberbroeckling Sep 2015 B2
D740279 Wu Oct 2015 S
9172345 Kok Oct 2015 B2
D743376 Kim Nov 2015 S
D743939 Seong Nov 2015 S
9196261 Burnett Nov 2015 B2
9197974 Clark Nov 2015 B1
9203494 Tarighat Mehrabani Dec 2015 B2
9215327 Bathurst Dec 2015 B2
9215543 Sun Dec 2015 B2
9226062 Sun Dec 2015 B2
9226070 Hyun Dec 2015 B2
9226088 Pandey Dec 2015 B2
9232185 Graham Jan 2016 B2
9237391 Benesty Jan 2016 B2
9247367 Nobile Jan 2016 B2
9253567 Morcelli Feb 2016 B2
9257132 Gowreesunker Feb 2016 B2
9264553 Pandey Feb 2016 B2
9264805 Buck Feb 2016 B2
9280985 Tawada Mar 2016 B2
9286908 Zhang Mar 2016 B2
9294839 Lambert Mar 2016 B2
9301049 Elko Mar 2016 B2
D754103 Fischer Apr 2016 S
9307326 Elko Apr 2016 B2
9319532 Bao Apr 2016 B2
9319799 Salmon Apr 2016 B2
9326060 Nicholson Apr 2016 B2
D756502 Lee May 2016 S
9330673 Cho May 2016 B2
9338301 Pocino May 2016 B2
9338549 Haulick May 2016 B2
9354310 Visser May 2016 B2
9357080 Beaucoup May 2016 B2
9403670 Schelling Aug 2016 B2
9426598 Walsh Aug 2016 B2
D767748 Nakai Sep 2016 S
9451078 Yang Sep 2016 B2
D769239 Li Oct 2016 S
9462378 Kuech Oct 2016 B2
9473868 Huang Oct 2016 B2
9479627 Rung Oct 2016 B1
9479885 Ivanov Oct 2016 B1
9489948 Chu Nov 2016 B1
9510090 Lissek Nov 2016 B2
9514723 Silfvast Dec 2016 B2
9516412 Shigenaga Dec 2016 B2
9521057 Klingbeil Dec 2016 B2
9549245 Frater Jan 2017 B2
9560446 Chang Jan 2017 B1
9560451 Eichfeld Jan 2017 B2
9565493 Abraham Feb 2017 B2
9578413 Sawa Feb 2017 B2
9578440 Otto Feb 2017 B2
9589556 Gao Mar 2017 B2
9591123 Sorensen Mar 2017 B2
9591404 Chhetri Mar 2017 B1
D784299 Cho Apr 2017 S
9615173 Sako Apr 2017 B2
9628596 Bullough Apr 2017 B1
9635186 Pandey Apr 2017 B2
9635474 Kuster Apr 2017 B2
D787481 Tyss May 2017 S
D788073 Silvera May 2017 S
9640187 Niemisto May 2017 B2
9641688 Pandey May 2017 B2
9641929 Li May 2017 B2
9641935 Ivanov May 2017 B1
9653091 Matsuo May 2017 B2
9653092 Sun May 2017 B2
9655001 Metzger May 2017 B2
9659576 Kotvis May 2017 B1
D789323 Mackiewicz Jun 2017 S
9674604 Deroo Jun 2017 B2
9692882 Mani Jun 2017 B2
9706057 Mani Jul 2017 B2
9716944 Yliaho Jul 2017 B2
9721582 Huang Aug 2017 B1
9734835 Fujieda Aug 2017 B2
9754572 Salazar Sep 2017 B2
9761243 Taenzer Sep 2017 B2
D801285 Timmins Oct 2017 S
9788119 Vilermo Oct 2017 B2
9813806 Graham Nov 2017 B2
9818426 Kotera Nov 2017 B2
9826211 Sawa Nov 2017 B2
9854101 Pandey Dec 2017 B2
9854363 Sladeczek Dec 2017 B2
9860439 Sawa Jan 2018 B2
9866952 Pandey Jan 2018 B2
D811393 Ahn Feb 2018 S
9894434 Rollow, IV Feb 2018 B2
9930448 Chen Mar 2018 B1
9936290 Mohammad Apr 2018 B2
9966059 Ayrapetian May 2018 B1
9973848 Chhetri May 2018 B2
9980042 Benattar May 2018 B1
D819607 Chui Jun 2018 S
D819631 Matsumiya Jun 2018 S
10015589 Ebenezer Jul 2018 B1
10021506 Johnson Jul 2018 B2
10021515 Mallya Jul 2018 B1
10034116 Kadri Jul 2018 B2
10054320 Choi Aug 2018 B2
10153744 Every Dec 2018 B1
10165386 Lehtiniemi Dec 2018 B2
D841589 Böhmer Feb 2019 S
10206030 Matsumoto Feb 2019 B2
10210882 McCowan Feb 2019 B1
10231062 Pedersen Mar 2019 B2
10244121 Mani Mar 2019 B2
10244219 Sawa Mar 2019 B2
10269343 Wingate Apr 2019 B2
10367948 Wells-Rutherford Jul 2019 B2
D857873 Shimada Aug 2019 S
10389861 Mani Aug 2019 B2
10389885 Sun Aug 2019 B2
D860319 Beruto Sep 2019 S
D860997 Sae Sep 2019 S
D864136 Kim Oct 2019 S
10440469 Barnett Oct 2019 B2
D865723 Cho Nov 2019 S
10566008 Thorpe Feb 2020 B2
10602267 Grosche Mar 2020 B2
D883952 Lucas May 2020 S
10650797 Kumar May 2020 B2
D888020 Lyu Jun 2020 S
10728653 Graham Jul 2020 B2
D900070 Lantz Oct 2020 S
D900071 Lantz Oct 2020 S
D900072 Lantz Oct 2020 S
D900073 Lantz Oct 2020 S
D900074 Lantz Oct 2020 S
10827263 Christoph Nov 2020 B2
10863270 O'Neill Dec 2020 B1
10930297 Christoph Feb 2021 B2
10959018 Shi Mar 2021 B1
10979805 Chowdhary Apr 2021 B2
D924189 Park Jul 2021 S
11109133 Lantz Aug 2021 B2
D940116 Cho Jan 2022 S
20010031058 Anderson Oct 2001 A1
20020015500 Belt Feb 2002 A1
20020041679 Beaucoup Apr 2002 A1
20020048377 Vaudrey Apr 2002 A1
20020064158 Yokoyama May 2002 A1
20020064287 Kawamura May 2002 A1
20020069054 Arrowood Jun 2002 A1
20020110255 Killion Aug 2002 A1
20020126861 Colby Sep 2002 A1
20020131580 Smith Sep 2002 A1
20020140633 Rafii Oct 2002 A1
20020146282 Wilkes Oct 2002 A1
20020149070 Sheplak Oct 2002 A1
20020159603 Hirai Oct 2002 A1
20030026437 Janse Feb 2003 A1
20030053639 Beaucoup Mar 2003 A1
20030059061 Tsuji Mar 2003 A1
20030063762 Tajima Apr 2003 A1
20030063768 Cornelius Apr 2003 A1
20030072461 Moorer Apr 2003 A1
20030107478 Hendricks Jun 2003 A1
20030118200 Beaucoup Jun 2003 A1
20030122777 Grover Jul 2003 A1
20030138119 Pocino Jul 2003 A1
20030156725 Boone Aug 2003 A1
20030161485 Smith Aug 2003 A1
20030163326 Maase Aug 2003 A1
20030169888 Subotic Sep 2003 A1
20030185404 Milsap Oct 2003 A1
20030198339 Roy Oct 2003 A1
20030198359 Killion Oct 2003 A1
20030202107 Slattery Oct 2003 A1
20040013038 Kajala Jan 2004 A1
20040013252 Craner Jan 2004 A1
20040076305 Santiago Apr 2004 A1
20040105557 Matsuo Jun 2004 A1
20040125942 Beaucoup Jul 2004 A1
20040175006 Kim Sep 2004 A1
20040202345 Stenberg Oct 2004 A1
20040240664 Freed Dec 2004 A1
20050005494 Way Jan 2005 A1
20050041530 Goudie Feb 2005 A1
20050069156 Haapapuro Mar 2005 A1
20050094580 Kumar May 2005 A1
20050094795 Rambo May 2005 A1
20050149320 Kajala Jul 2005 A1
20050157897 Saltykov Jul 2005 A1
20050175189 Lee Aug 2005 A1
20050175190 Tashev Aug 2005 A1
20050213747 Popovich Sep 2005 A1
20050221867 Zurek Oct 2005 A1
20050238196 Furuno Oct 2005 A1
20050270906 Ramenzoni Dec 2005 A1
20050271221 Cerwin Dec 2005 A1
20050286698 Bathurst Dec 2005 A1
20050286729 Harwood Dec 2005 A1
20060083390 Kaderavek Apr 2006 A1
20060088173 Rodman Apr 2006 A1
20060093128 Oxford May 2006 A1
20060098403 Smith May 2006 A1
20060104458 Kenoyer May 2006 A1
20060109983 Young May 2006 A1
20060151256 Lee Jul 2006 A1
20060159293 Azima Jul 2006 A1
20060161430 Schweng Jul 2006 A1
20060165242 Miki Jul 2006 A1
20060192976 Hall Aug 2006 A1
20060198541 Henry Sep 2006 A1
20060204022 Hooley Sep 2006 A1
20060215866 Francisco Sep 2006 A1
20060222187 Jarrett Oct 2006 A1
20060233353 Beaucoup Oct 2006 A1
20060239471 Mao Oct 2006 A1
20060262942 Oxford Nov 2006 A1
20060269080 Oxford Nov 2006 A1
20060269086 Page Nov 2006 A1
20070006474 Taniguchi Jan 2007 A1
20070009116 Reining Jan 2007 A1
20070019828 Hughes Jan 2007 A1
20070053524 Haulick Mar 2007 A1
20070093714 Beaucoup Apr 2007 A1
20070116255 Derkx May 2007 A1
20070120029 Keung May 2007 A1
20070165871 Roovers Jul 2007 A1
20070230712 Belt Oct 2007 A1
20070253561 Williams Nov 2007 A1
20070269066 Derleth Nov 2007 A1
20080008339 Ryan Jan 2008 A1
20080033723 Jang Feb 2008 A1
20080046235 Chen Feb 2008 A1
20080056517 Algazi Mar 2008 A1
20080101622 Sugiyama May 2008 A1
20080130907 Sudo Jun 2008 A1
20080144848 Buck Jun 2008 A1
20080168283 Penning Jul 2008 A1
20080188965 Bruey Aug 2008 A1
20080212805 Fincham Sep 2008 A1
20080232607 Tashev Sep 2008 A1
20080247567 Kjolerbakken Oct 2008 A1
20080253553 Li Oct 2008 A1
20080253589 Trahms Oct 2008 A1
20080259731 Happonen Oct 2008 A1
20080260175 Elko Oct 2008 A1
20080279400 Knoll Nov 2008 A1
20080285772 Haulick Nov 2008 A1
20090003586 Lai Jan 2009 A1
20090030536 Gur Jan 2009 A1
20090052684 Ishibashi Feb 2009 A1
20090086998 Jeong Apr 2009 A1
20090087000 Ko Apr 2009 A1
20090087001 Jiang Apr 2009 A1
20090094817 Killion Apr 2009 A1
20090129609 Oh May 2009 A1
20090147967 Ishibashi Jun 2009 A1
20090150149 Cutter Jun 2009 A1
20090161880 Hooley Jun 2009 A1
20090169027 Ura Jul 2009 A1
20090173030 Gulbrandsen Jul 2009 A1
20090173570 Levit Jul 2009 A1
20090226004 Moeller Sep 2009 A1
20090233545 Sutskover Sep 2009 A1
20090237561 Kobayashi Sep 2009 A1
20090254340 Sun Oct 2009 A1
20090274318 Ishibashi Nov 2009 A1
20090310794 Ishibashi Dec 2009 A1
20100011644 Kramer Jan 2010 A1
20100034397 Nakadai Feb 2010 A1
20100074433 Zhang Mar 2010 A1
20100111323 Marton May 2010 A1
20100111324 Yeldener May 2010 A1
20100119097 Ohtsuka May 2010 A1
20100123785 Chen May 2010 A1
20100128892 Chen May 2010 A1
20100128901 Herman May 2010 A1
20100131749 Kim May 2010 A1
20100142721 Wada Jun 2010 A1
20100150364 Buck Jun 2010 A1
20100158268 Marton Jun 2010 A1
20100165071 Ishibashi Jul 2010 A1
20100166219 Marton Jul 2010 A1
20100189275 Christoph Jul 2010 A1
20100189299 Grant Jul 2010 A1
20100202628 Meyer Aug 2010 A1
20100208605 Wang Aug 2010 A1
20100215184 Buck Aug 2010 A1
20100215189 Marton Aug 2010 A1
20100217590 Nemer Aug 2010 A1
20100245624 Beaucoup Sep 2010 A1
20100246873 Chen Sep 2010 A1
20100284185 Ngai Nov 2010 A1
20100305728 Aiso Dec 2010 A1
20100314513 Evans Dec 2010 A1
20110002469 Ojala Jan 2011 A1
20110007921 Stewart Jan 2011 A1
20110033063 McGrath Feb 2011 A1
20110038229 Beaucoup Feb 2011 A1
20110096136 Liu Apr 2011 A1
20110096631 Kondo Apr 2011 A1
20110096915 Nemer Apr 2011 A1
20110164761 McCowan Jul 2011 A1
20110194719 Frater Aug 2011 A1
20110211706 Tanaka Sep 2011 A1
20110235821 Okita Sep 2011 A1
20110268287 Ishibashi Nov 2011 A1
20110311064 Teutsch Dec 2011 A1
20110311085 Stewart Dec 2011 A1
20110317862 Hosoe Dec 2011 A1
20120002835 Stewart Jan 2012 A1
20120014049 Ogle Jan 2012 A1
20120027227 Kok Feb 2012 A1
20120076316 Zhu Mar 2012 A1
20120080260 Stewart Apr 2012 A1
20120093344 Sun Apr 2012 A1
20120117474 Miki May 2012 A1
20120128160 Kim May 2012 A1
20120128175 Visser May 2012 A1
20120155688 Wilson Jun 2012 A1
20120155703 Hernandez-Abrego Jun 2012 A1
20120163625 Siotis Jun 2012 A1
20120169826 Jeong Jul 2012 A1
20120177219 Mullen Jul 2012 A1
20120182429 Forutanpour Jul 2012 A1
20120207335 Spaanderman Aug 2012 A1
20120224709 Keddem Sep 2012 A1
20120243698 Elko Sep 2012 A1
20120262536 Chen Oct 2012 A1
20120288079 Burnett Nov 2012 A1
20120288114 Duraiswami Nov 2012 A1
20120294472 Hudson Nov 2012 A1
20120327115 Chhetri Dec 2012 A1
20120328142 Horibe Dec 2012 A1
20130002797 Thapa Jan 2013 A1
20130004013 Stewart Jan 2013 A1
20130015014 Stewart Jan 2013 A1
20130016847 Steiner Jan 2013 A1
20130028451 De Roo Jan 2013 A1
20130029684 Kawaguchi Jan 2013 A1
20130034241 Pandey Feb 2013 A1
20130039504 Pandey Feb 2013 A1
20130083911 Bathurst Apr 2013 A1
20130094689 Tanaka Apr 2013 A1
20130101141 McElveen Apr 2013 A1
20130136274 Ahgren May 2013 A1
20130142343 Matsui Jun 2013 A1
20130147835 Lee Jun 2013 A1
20130156198 Kim Jun 2013 A1
20130182190 McCartney Jul 2013 A1
20130206501 Yu Aug 2013 A1
20130216066 Yerrace Aug 2013 A1
20130226593 Magnusson Aug 2013 A1
20130251181 Stewart Sep 2013 A1
20130264144 Hudson Oct 2013 A1
20130271559 Feng Oct 2013 A1
20130294616 Mulder Nov 2013 A1
20130297302 Pan Nov 2013 A1
20130304476 Kim Nov 2013 A1
20130304479 Teller Nov 2013 A1
20130329908 Lindahl Dec 2013 A1
20130332156 Tackin Dec 2013 A1
20130336516 Stewart Dec 2013 A1
20130343549 Vemireddy Dec 2013 A1
20140003635 Mohammad Jan 2014 A1
20140010383 Mackey Jan 2014 A1
20140016794 Lu Jan 2014 A1
20140029761 Maenpaa Jan 2014 A1
20140037097 Labosco Feb 2014 A1
20140050332 Nielsen Feb 2014 A1
20140072151 Ochs Mar 2014 A1
20140098233 Martin Apr 2014 A1
20140098964 Rosca Apr 2014 A1
20140122060 Kaszczuk May 2014 A1
20140177857 Kuster Jun 2014 A1
20140233777 Tseng Aug 2014 A1
20140233778 Hardiman Aug 2014 A1
20140264654 Salmon Sep 2014 A1
20140265774 Stewart Sep 2014 A1
20140270271 Dehe Sep 2014 A1
20140286518 Stewart Sep 2014 A1
20140295768 Wu Oct 2014 A1
20140301586 Stewart Oct 2014 A1
20140307882 Leblanc Oct 2014 A1
20140314251 Rosca Oct 2014 A1
20140341392 Lambert Nov 2014 A1
20140357177 Stewart Dec 2014 A1
20140363008 Chen Dec 2014 A1
20150003638 Kasai Jan 2015 A1
20150025878 Gowreesunker Jan 2015 A1
20150030172 Gaensler Jan 2015 A1
20150033042 Iwamoto Jan 2015 A1
20150050967 Bao Feb 2015 A1
20150055796 Nugent Feb 2015 A1
20150055797 Nguyen Feb 2015 A1
20150063579 Bao Mar 2015 A1
20150070188 Aramburu Mar 2015 A1
20150078581 Etter Mar 2015 A1
20150078582 Graham Mar 2015 A1
20150097719 Balachandreswaran Apr 2015 A1
20150104023 Bilobrov Apr 2015 A1
20150117672 Christoph Apr 2015 A1
20150118960 Petit Apr 2015 A1
20150126255 Yang May 2015 A1
20150156578 Alexandridis Jun 2015 A1
20150163577 Benesty Jun 2015 A1
20150185825 Mullins Jul 2015 A1
20150189423 Giannuzzi Jul 2015 A1
20150208171 Funakoshi Jul 2015 A1
20150237424 Wilker Aug 2015 A1
20150281832 Kishimoto Oct 2015 A1
20150281833 Shigenaga Oct 2015 A1
20150281834 Takano Oct 2015 A1
20150312662 Kishimoto Oct 2015 A1
20150312691 Virolainen Oct 2015 A1
20150326968 Shigenaga Nov 2015 A1
20150341734 Sherman Nov 2015 A1
20150350621 Sawa Dec 2015 A1
20150358734 Butler Dec 2015 A1
20160011851 Zhang Jan 2016 A1
20160021478 Katagiri Jan 2016 A1
20160029120 Nesta Jan 2016 A1
20160031700 Sparks Feb 2016 A1
20160037277 Matsumoto Feb 2016 A1
20160055859 Finlow-Bates Feb 2016 A1
20160080867 Nugent Mar 2016 A1
20160088392 Huttunen Mar 2016 A1
20160100092 Bohac Apr 2016 A1
20160105473 Klingbeil Apr 2016 A1
20160111109 Tsujikawa Apr 2016 A1
20160127527 Mani May 2016 A1
20160134928 Ogle May 2016 A1
20160142548 Pandey May 2016 A1
20160142814 Deroo May 2016 A1
20160142815 Norris May 2016 A1
20160148057 Oh May 2016 A1
20160150315 Tzirkel-Hancock May 2016 A1
20160150316 Kubota May 2016 A1
20160155455 Ojanperä Jun 2016 A1
20160165340 Benattar Jun 2016 A1
20160173976 Podhradsky Jun 2016 A1
20160173978 Li Jun 2016 A1
20160189727 Wu Jun 2016 A1
20160192068 Ng Jun 2016 A1
20160196836 Yu Jul 2016 A1
20160234593 Matsumoto Aug 2016 A1
20160275961 Yu Sep 2016 A1
20160295279 Srinivasan Oct 2016 A1
20160300584 Pandey Oct 2016 A1
20160302002 Lambert Oct 2016 A1
20160302006 Pandey Oct 2016 A1
20160323667 Shumard Nov 2016 A1
20160323668 Abraham Nov 2016 A1
20160330545 McElveen Nov 2016 A1
20160337523 Pandey Nov 2016 A1
20160353200 Bigeh Dec 2016 A1
20160357508 Moore Dec 2016 A1
20170019744 Matsumoto Jan 2017 A1
20170064451 Park Mar 2017 A1
20170105066 McLaughlin Apr 2017 A1
20170134849 Pandey May 2017 A1
20170134850 Graham May 2017 A1
20170164101 Rollow, IV Jun 2017 A1
20170180861 Chen Jun 2017 A1
20170206064 Breazeal Jul 2017 A1
20170230748 Shumard Aug 2017 A1
20170264999 Fukuda Sep 2017 A1
20170303887 Richmond Oct 2017 A1
20170308352 Kessler Oct 2017 A1
20170374454 Bernardini Dec 2017 A1
20180083848 Siddiqi Mar 2018 A1
20180102136 Ebenezer Apr 2018 A1
20180109873 Xiang Apr 2018 A1
20180115799 Thiele Apr 2018 A1
20180160224 Graham Jun 2018 A1
20180196585 Densham Jul 2018 A1
20180219922 Bryans Aug 2018 A1
20180227666 Barnett Aug 2018 A1
20180292079 Branham Oct 2018 A1
20180310096 Shumard Oct 2018 A1
20180313558 Byers Nov 2018 A1
20180338205 Abraham Nov 2018 A1
20180359565 Kim Dec 2018 A1
20190042187 Truong Feb 2019 A1
20190166424 Harney May 2019 A1
20190215540 Nicol Jul 2019 A1
20190230436 Tsingos Jul 2019 A1
20190259408 Freeman Aug 2019 A1
20190268683 Miyahara Aug 2019 A1
20190295540 Grima Sep 2019 A1
20190295569 Wang Sep 2019 A1
20190319677 Hansen Oct 2019 A1
20190371354 Lester Dec 2019 A1
20190373362 Ansai Dec 2019 A1
20190385629 Moravy Dec 2019 A1
20190387311 Schultz Dec 2019 A1
20200015021 Leppanen Jan 2020 A1
20200021910 Rollow, IV Jan 2020 A1
20200037068 Barnett Jan 2020 A1
20200068297 Rollow, IV Feb 2020 A1
20200100009 Lantz Mar 2020 A1
20200100025 Shumard Mar 2020 A1
20200137485 Yamakawa Apr 2020 A1
20200145753 Rollow, IV May 2020 A1
20200152218 Kikuhara May 2020 A1
20200162618 Enteshari May 2020 A1
20200228663 Wells-Rutherford Jul 2020 A1
20200251119 Yang Aug 2020 A1
20200275204 Labosco Aug 2020 A1
20200278043 Cao Sep 2020 A1
20200288237 Abraham Sep 2020 A1
20210012789 Husain Jan 2021 A1
20210021940 Petersen Jan 2021 A1
20210044881 Lantz Feb 2021 A1
20210051397 Veselinovic Feb 2021 A1
20210098014 Tanaka Apr 2021 A1
20210098015 Pandey Apr 2021 A1
20210120335 Veselinovic Apr 2021 A1
20210200504 Park Jul 2021 A1
20210375298 Zhang Dec 2021 A1
Foreign Referenced Citations (147)
Number Date Country
2359771 Apr 2003 CA
2475283 Jan 2005 CA
2505496 Oct 2006 CA
2838856 Dec 2012 CA
2846323 Sep 2014 CA
1780495 May 2006 CN
101217830 Jul 2008 CN
101833954 Sep 2010 CN
101860776 Oct 2010 CN
101894558 Nov 2010 CN
102646418 Aug 2012 CN
102821336 Dec 2012 CN
102833664 Dec 2012 CN
102860039 Jan 2013 CN
104036784 Sep 2014 CN
104053088 Sep 2014 CN
104080289 Oct 2014 CN
104347076 Feb 2015 CN
104581463 Apr 2015 CN
105355210 Feb 2016 CN
105548998 May 2016 CN
106162427 Nov 2016 CN
106251857 Dec 2016 CN
106851036 Jun 2017 CN
107221336 Sep 2017 CN
107534725 Jan 2018 CN
108172235 Jun 2018 CN
109087664 Dec 2018 CN
208190895 Dec 2018 CN
109727604 May 2019 CN
110010147 Jul 2019 CN
306391029 Mar 2021 CN
2941485 Apr 1981 DE
0077546430001 Mar 2020 EM
0381498 Aug 1990 EP
0594098 Apr 1994 EP
0869697 Oct 1998 EP
1180914 Feb 2002 EP
1184676 Mar 2002 EP
0944228 Jun 2003 EP
1439526 Jul 2004 EP
1651001 Apr 2006 EP
1727344 Nov 2006 EP
1906707 Apr 2008 EP
1952393 Aug 2008 EP
1962547 Aug 2008 EP
2133867 Dec 2009 EP
2159789 Mar 2010 EP
2197219 Jun 2010 EP
2360940 Aug 2011 EP
2721837 Apr 2014 EP
2772910 Sep 2014 EP
2778310 Sep 2014 EP
2942975 Nov 2015 EP
2988527 Feb 2016 EP
3131311 Feb 2017 EP
2393601 Mar 2004 GB
2446620 Aug 2008 GB
S63144699 Jun 1988 JP
H01260967 Oct 1989 JP
H0241099 Feb 1990 JP
H05260589 Oct 1993 JP
H07336790 Dec 1995 JP
3175622 Jun 2001 JP
2003060530 Feb 2003 JP
2003087890 Mar 2003 JP
2004349806 Dec 2004 JP
2004537232 Dec 2004 JP
2005323084 Nov 2005 JP
2006094389 Apr 2006 JP
2006101499 Apr 2006 JP
4120646 Aug 2006 JP
4258472 Aug 2006 JP
4196956 Sep 2006 JP
2006340151 Dec 2006 JP
4760160 Jan 2007 JP
4752403 Mar 2007 JP
2007089058 Apr 2007 JP
4867579 Jun 2007 JP
2007208503 Aug 2007 JP
2007228069 Sep 2007 JP
2007228070 Sep 2007 JP
2007274131 Oct 2007 JP
2007274463 Oct 2007 JP
2007288679 Nov 2007 JP
2008005347 Jan 2008 JP
2008042754 Feb 2008 JP
2008154056 Jul 2008 JP
2008259022 Oct 2008 JP
2008263336 Oct 2008 JP
2008312002 Dec 2008 JP
2009206671 Sep 2009 JP
2010028653 Feb 2010 JP
2010114554 May 2010 JP
2010268129 Nov 2010 JP
2011015018 Jan 2011 JP
4779748 Sep 2011 JP
2012165189 Aug 2012 JP
5028944 Sep 2012 JP
5139111 Feb 2013 JP
5306565 Oct 2013 JP
5685173 Mar 2015 JP
2016051038 Apr 2016 JP
100298300 May 2001 KR
100901464 Jun 2009 KR
100960781 Jun 2010 KR
1020130033723 Apr 2013 KR
300856915 May 2016 KR
201331932 Aug 2013 TW
I484478 May 2015 TW
1997008896 Mar 1997 WO
1998047291 Oct 1998 WO
2000030402 May 2000 WO
2003073786 Sep 2003 WO
2003088429 Oct 2003 WO
2004027754 Apr 2004 WO
2004090865 Oct 2004 WO
2006049260 May 2006 WO
2006071119 Jul 2006 WO
2006114015 Nov 2006 WO
2006121896 Nov 2006 WO
2007045971 Apr 2007 WO
2008074249 Jun 2008 WO
2008125523 Oct 2008 WO
2009039783 Apr 2009 WO
2009109069 Sep 2009 WO
2010001508 Jan 2010 WO
2010091999 Aug 2010 WO
2010140084 Dec 2010 WO
2010144148 Dec 2010 WO
2011104501 Sep 2011 WO
2012122132 Sep 2012 WO
2012140435 Oct 2012 WO
2012160459 Nov 2012 WO
2012174159 Dec 2012 WO
2013016986 Feb 2013 WO
2013182118 Dec 2013 WO
2014156292 Oct 2014 WO
2016176429 Nov 2016 WO
2016179211 Nov 2016 WO
2017208022 Dec 2017 WO
2018140444 Aug 2018 WO
2018140618 Aug 2018 WO
2018211806 Nov 2018 WO
2019231630 Dec 2019 WO
2020168873 Aug 2020 WO
2020191354 Sep 2020 WO
Non-Patent Literature Citations (269)
Entry
“Philips Hue Bulbs and Wireless Connected Lighting System,” Web page https://www.philips-hue.com/en-in, 8 pp, Sep. 23, 2020, retrieved from Internet Archive Wayback Machine, <https://web.archive.org/web/20200923171037/https://www.philips-hue.com/en-in> on Sep. 27, 2021.
“Vsa 2050 II Digitally Steerable Column Speaker,” Web page https://www.rcf.it/en_US/products/product-detail/vsa-2050-ii/972389, 15 pages, Dec. 24, 2018.
Advanced Network Devices, IPSCM Ceiling Tile IP Speaker, Feb. 2011, 2 pgs.
Advanced Network Devices, IPSCM Standard 2′ by 2′ Ceiling Tile Speaker, 2 pgs.
Affes, et al., “A Signal Subspace Tracking Algorithm for Microphone Array Processing of Speech,” IEEE Trans. On Speech and Audio Processing, vol. 5, No. 5, Sep. 1997, pp. 425-437.
Affes, et al., “A Source Subspace Tracking Array of Microphones for Double Talk Situations,” 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings, May 1996, pp. 909-912.
Affes, et al., “An Algorithm for Multisource Beamforming and Multitarget Tracking,” IEEE Trans. On Signal Processing, vol. 44, No. 6, Jun. 1996, pp. 1512-1522.
Affes, et al., “Robust Adaptive Beamforming via LMS-Like Target Tracking,” Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 1994, pp. IV-269-IV-272.
Ahonen, et al, “Directional Analysis of Sound Field with Linear Microphone Array and Applications in Sound Reproduction,” Audio Engineering Socity, Convention Paper 7329, May 2008, 11 pp.
Alarifi, et al., “Ultra Wideband Indoor Positioning Technologies: Analysis and Recent Advances,” Sensors 2016, vol. 16, No. 707, 36 pp.
Amazon webpage for Metalfab MFLCRFG (last visited Apr. 22, 2020) available at <https://www.amazon.com/RETURN-FILTERGRILLE-Drop-Ceiling/dp/B0064Q9A7l/ref=sr 12?dchild=1&keywords=drop+ceiling+return+air+grille&qid=1585862723&s=hi&sr=1-2>, 11 pp.
Armstrong “Walls” Catalog available at <https://www.armstrongceilings.com/content/dam/armstrongceilings/commercial/north-america/catalogs/armstrong-ceilings-wallsspecifiers-reference.pdf>, 2019, 30 pp.
Armstrong Tectum Ceiling & Wall Panels Catalog available at <https://www.armstrongceilings.com/content/dam/armstrongceilings/commercial/north-america/brochures/tectum-brochure.pdf>, 2019, 16 pp.
Armstrong Woodworks Concealed Catalog available at <https://sweets.construction.com/swts_content_files/3824/442581.pdf>, 2014, 6 pp.
Armstrong Woodworks Walls Catalog available at <https://www.armstrongceilings.com/pdbupimagesclg/220600.pdf/download/data-sheet-woodworks-walls.pdf>, 2019, 2 pp.
Armstrong World Industries, Inc., I-Ceilings Sound Systems Speaker Panels, 2002, 4 pgs.
Armstrong, Acoustical Design: Exposed Structure, available at <https://www.armstrongceilings.com/pdbupimagesclg/217142.pdf/download/acoustical-design-exposed-structurespaces-brochure.pdf>, 2018, 19 pp.
Armstrong, Ceiling Systems, Brochure page for Armstrong Softlook, 1995, 2 pp.
Armstrong, Excerpts from Armstrong 2011-2012 Ceiling Wall Systems Catalog, available at <https://web.archive.org/web/20121116034120/http://www.armstrong.com/commceilingsna/en_us/pdf/ceilings_catalog_screen-2011 .pdf>, as early as 2012, 162 pp.
Armstrong, i-Ceilings, Brochure, 2009, 12 pp.
Arnold, et al., “A Directional Acoustic Array Using Silicon Micromachined Piezoresistive Microphones,” Journal of the Acoustical Society of America, 113(1), Jan. 2003, 10 pp.
Atlas Sound, I128SYSM IP Compliant Loudspeaker System with Microphone Data Sheet, 2009, 2 pgs.
Atlas Sound, 1′X2′ IP Speaker with Micophone for Suspended Ceiling Systems, https://www.atlasied.com/i128sysm, retrieved Oct. 25, 2017, 5 pgs.
Audio Technica, ES945 Omnidirectional Condenser Boundary Microphones, https://eu.audio-technica.com/resources/ES945%20Specifications.pdf, 2007, 1 pg.
Audix Microphones, Audix Introduces Innovative Ceiling Mies, http://audixusa.com/docs_12/latest_news/EFpIFkAAkIOtSdolke.shtml, Jun. 2011, 6 pgs.
Audix Microphones, M70 Flush Mount Ceiling Mic, May 2016, 2 pgs.
Automixer Gated, Information Sheet, MIT, Nov. 2019, 9 pp.
Beh, et al., “Combining Acoustic Echo Cancellation and Adaptive Beamforming for Achieving Robust Speech Interface in Mobile Robot,” 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Sep. 2008, pp. 1693-1698.
Benesty, et al., “A New Class of Doubletalk Detectors Based on Cross-Correlation,” IEEE Transactions on Speech and Audio Processing, vol. 8, No. 2, Mar. 2000, pp. 168-172.
Benesty, et al., “Adaptive Algorithms for Mimo Acoustic Echo Cancellation,” AI2 Allen Institute for Artifical Intelligence, 2003.
Benesty, et al., “Differential Beamforming,” Fundamentals of Signal Enhancement and Array Signal Processing, First Edition, 2017, 39 pp.
Benesty, et al., “Frequency-Domain Adaptive Filtering Revisited, Generalization to the Multi-Channel Case, and Application to Acoustic Echo Cancellation,” 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing Proceedings, Jun. 2000, pp. 789-792.
Benesty, et. al., “Microphone Array Signal Processing,” Springer, 2010, 20 pp.
Berkun, et al., “Combined Beamformers for Robust Broadband Regularized Superdirective Beamforming,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 23, No. 5, May 2015, 10 pp.
Beyer Dynamic, Classis BM 32-33-34 DE-EN-FR 2016, 1 pg.
Beyer Dynamic, Classis-BM-33-PZ A1, 2013, 1 pg.
BNO055, Intelligent 9-axis absolute orientation sensor, Data sheet, Bosch, Nov. 2020, 118 pp.
Boyd, et al., Convex Optimization, Mar. 15, 1999, 216 pgs.
Brandstein, et al., “Microphone Arrays: Signal Processing Techniques and Applications,” Digital Signal Processing, Springer-Verlag Berlin Heidelberg, 2001, 401 pgs.
Brooks, et al., “A Quantitative Assessment of Group Delay Methods for Identifying Glottal Closures in Voiced Speech,” IEEE Transaction on Audio, Speech, and Language Processing, vol. 14, No. 2, Mar. 2006, 11 pp.
Bruel & Kjaer, by J.J. Christensen and J. Hald, Technical Review: Beamforming, No. 1, 2004, 54 pgs.
BSS Audio, Soundweb London Application Guides, 2010, 120 pgs.
Buchner, et al., “An Acoustic Human-Machine Interface with Multi-Channel Sound Reproduction,” IEEE Fourth Workshop on Multimedia Signal Processing, Oct. 2001, pp. 359-364.
Buchner, et al., “An Efficient Combination of Multi-Channel Acoustic Echo Cancellation with a Beamforming Microphone Array,” International Workshop on Hands-Free Speech Communication (HSC2001), Apr. 2001, pp. 55-58.
Buchner, et al., “Full-Duplex Communication Systems Using Loudspeaker Arrays and Microphone Arrays,” IEEE International Conference on Multimedia and Expo, Aug. 2002, pp. 509-512.
Buchner, et al., “Generalized Multichannel Frequency-Domain Adaptive Filtering: Efficient Realization and Application to Hands-Free Speech Communication,” Signal Processing 85, 2005, pp. 549-570.
Buchner, et al., “Multichannel Frequency-Domain Adaptive Filtering with Application to Multichannel Acoustic Echo Cancellation,” Adaptive Signal Processing, 2003, pp. 95-128.
Buck, “Aspects of First-Order Differential Microphone Arrays in the Presence of Sensor Imperfections,” Transactions on Emerging Telecommunications Technologies, 13.2, 2002, 8 pp.
Buck, et al., “First Order Differential Microphone Arrays for Automotive Applications,” 7th International Workshop on Acoustic Echo and Noise Control, Darmstadt University of Technology, Sep. 10-13, 2001, 4 pp.
Buck, et al., “Self-Calibrating Microphone Arrays for Speech Signal Acquisition: A Systematic Approach,” Signal Processing, vol. 86, 2006, pp. 1230-1238.
Burton, et al., “A New Structure for Combining Echo Cancellation and Beamforming in Changing Acoustical Environments,” IEEE International Conference on Acoustics, Speech and Signal Processing, 2007, pp. 1-77-1-80.
BZ-3a Installation Instructions, XEDIT Corporation, Available at <chrome-extension://efaidnbmnnnibpcajpcglclefindmkaj/viewer.html?pdfurl=https%3A%2F%2Fwww.servo reelers.com%2Fmt-content%2Fuploads%2F2017%2F05%2Fbz-a-3universal-2017c.pdf&clen=189067&chunk=true>, 1 p.
Cabral, et al., Glottal Spectral Separation for Speech Synthesis, IEEE Journal of Selected Topics in Signal Processing, 2013, 15 pp.
Campbell, “Adaptive Beamforming Using a Microphone Array for Hands-Free Telephony,” Virginia Polytechnic Institute and State University, Feb. 1999, 154 pgs.
Canetto, et al., “Speech Enhancement Systems Based on Microphone Arrays,” VI Conference of the Italian Society for Applied and Industrial Mathematics, May 27, 2002, 9 pp.
Cech, et al., “Active-Speaker Detection and Localization with Microphones and Cameras Embedded into a Robotic Head,” IEEE-RAS International Conference on Humanoid Robots, Oct. 2013, pp. 203-210.
Chan, et al., “Uniform Concentric Circular Arrays with Frequency-Invariant Characteristics-Theory, Design, Adaptive Beamforming and DOA Estimation,” IEEE Transactions on Signal Processing, vol. 55, No. 1, Jan. 2007, pp. 165-177.
Chau, et al., “A Subband Beamformer on an Ultra Low-Power Miniature DSP Platform,” 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, 4 pp.
Chen, et al., “A General Approach to the Design and Implementation of Linear Differential Microphone Arrays,” Signal and Information Processing Association Annual Summit and Conference, 2013 Asia-Pacific, IEEE, 7 pp.
Chen, et al., “Design and Implementation of Small Microphone Arrays,” PowerPoint Presentation, Northwestern Polytechnical University and Institut national de la recherche scientifique, Jan. 1, 2014, 56 pp.
Chen, et al., “Design of Robust Broadband Beamformers with Passband Shaping Characteristics using Tikhonov Regularization,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 17, No. 4, May 2009, pp. 565-681.
Chou, “Frequency-Independent Beamformer with Low Response Error,” 1995 International Conference on Acoustics, Speech, and Signal Processing, pp. 2995-2998, May 9, 1995, 4 pp.
Chu, “Desktop Mic Array for Teleconferencing,” 1995 International Conference on Acoustics, Speech, and Signal Processing, May 1995, pp. 2999-3002.
Circuit Specialists webpage for an aluminum enclosure, available at <https://www.circuitspecialists.com/metal-instrument-enclosure-la7.html?otaid=gpl&gclid=EAlalQobChMI2JTw-Ynm6AIVgbbICh3F4QKuEAkYBiABEgJZMPD_BwE>, 3 pp.
ClearOne Introduces Ceiling Microphone Array With Built-In Dante Interface, Press Release; GlobeNewswire, Jan. 8, 2019, 2 pp.
ClearOne Launches Second Generation of its Groundbreaking Beamforming Microphone Array, Press Release, Acquire Media, Jun. 1, 2016, 2 pp.
ClearOne to Unveil Beamforming Microphone Array with Adaptive Steering and Next Generation Acoustic Echo Cancellation Technology, Press Release, InfoComm, Jun. 4, 2012, 1 p.
ClearOne, Clearly Speaking Blog, “Advanced Beamforming Microphone Array Technology for Corporate Conferencing Systems,” Nov. 11, 2013, 5 pp., http://www.clearone.com/blog/advanced-beamforming-microphone-array-technology-for-corporate-conferencing-systems/.
ClearOne, Beamforming Microphone Array, Mar. 2012, 6 pgs.
ClearOne, Ceiling Microphone Array Installation Manual, Jan. 9, 2012, 20 pgs.
ClearOne, Converge/Converge Pro, Manual, 2008, 51 pp.
ClearOne, Professional Conferencing Microphones, Brochure, Mar. 2015, 3 pp.
Coleman, “Loudspeaker Array Processing for Personal Sound Zone Reproduction,” Centre for Vision, Speech and Signal Processing, 2014, 239 pp.
Cook, et al., An Alternative Approach to Interpolated Array Processing for Uniform Circular Arrays, Asia-Pacific Conference on Circuits and Systems, 2002, pp. 411-414.
Cox, et al., “Robust Adaptive Beamforming,” IEEE Trans. Acoust., Speech, and Signal Processing, vol. ASSP-35, No. 10, Oct. 1987, pp. 1365-1376.
CTG Audio, Ceiling Microphone CTG CM-01, Jun. 5, 2008, 2 pgs.
CTG Audio, CM-01 & CM-02 Ceiling Microphones Specifications, 2 pgs.
CTG Audio, CM-01 & CM-02 Ceiling Microphones, 2017, 4 pgs.
CTG Audio, CTG FS-400 and RS-800 with “Beamforming” Technology, Datasheet, as early as 2009, 2 pp.
CTG Audio, Expand Your IP Teleconferencing to Full Room Audio, Obtained from website htt. )://www ct audio com/ex and-, our-i - teleconforencino-to-ful-room-audio-while-conquennc.1-echo-cancelation-issues Mull, 2014.
CTG Audio, Frequently Asked Questions, As early as 2009, 2 pp.
CTG Audio, Installation Manual, Nov. 21, 2008, 25 pgs.
CTG Audio, Introducing the CTG FS-400 and FS-800 with Beamforming Technology, as early as 2008, 2 pp.
CTG Audio, Meeting the Demand for Ceiling Mics in the Enterprise 5 Best Practices, Brochure, 2012, 9 pp.
Dahl et al., Acoustic Echo Cancelling with Microphone Arrays, Research Report Mar. 1995, Univ. of Karlskrona/Ronneby, Apr. 1995, 64 pgs.
Decawave, Application Note: APR001, Uwb Regulations, A Summary of Worldwide Telecommunications Regulations governing the use of Ultra-Wideband radio, Version 1.2, 2015, 63 pp.
Desiraju, et al., “Efficient Multi-Channel Acoustic Echo Cancellation Using Constrained Sparse Filter Updates in the Subband Domain,” Acoustic Speech Enhancement Research, Sep. 2014, 4 pp.
DiBiase. et al., Robust Localization in Reverberent Rooms, in Brandstein, ed., Microphone Arrays: Techniques and Applications, 2001, Springer-Verlag Berlin Heidelberg, pp. 157-180.
Diethorn, “Audio Signal Processing for Next-Generation Multimedia Communication Systems,” Chapter 4, 2004, 9 pp.
Digikey webpage for Converta box (last visited Apr. 22, 2020) <https://www.digikey.com/product-detail/en/bud-industries/CU-452-A/377-1969-ND/439257?utm_adgroup=Boxes&utm_source=google&utm_medium=cpc&utm_campaign=Shopping_Boxes%2C%20Enclosures%2C%20Racks_NEW&utm_term=&utm_content=Boxes&gclid=EAIalQobChMI2JTw-Ynm6AIVgbblCh3F4QKuEAkYCSABEgKybPD_BwE>, 3 pp.
Digikey webpage for Pomona Box (last visited Apr. 22, 2020) available at <https://www.digikey.com/product-detail/en/pomonaelectronics/3306/501-2054-ND/736489>, 2 pp.
Digital Wireless Conference System, MCW-D 50, Beyerdynamic Inc., 2009, 18 pp.
Do et al., A Real-Time SRP-PHAT Source Location Implementation using Stochastic Region Contraction (SRC) on a Large-Aperture Microphone Array, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing—ICASSP '07, , Apr. 2007, pp. I-121-I-124.
Dominguez, et al., “Towards an Environmental Measurement Cloud: Delivering Pollution Awareness to the Public,” International Journal of Distributed Sensor Networks, vol. 10, Issue 3, Mar. 31, 2014, 17 pp.
Dormehl, “HoloLens concept lets you control your smart home via augmented reality,” digitaltrends, Jul. 26, 2016, 12 pp.
Double Condenser Microphone SM 69, Datasheet, Georg Neumann GmbH, available at <https://ende.neumann.com/product_files/7453/download>, 8 pp.
Eargle, “The Microphone Handbook,” Elar Publ. Co., 1st ed., 1981, 4 pp.
Enright, Notes From Logan, June edition of Scanlines, Jun. 2009, 9 pp.
Fan, et al., “Localization Estimation of Sound Source by Microphones Array,” Procedia Engineering 7, 2010, pp. 312-317.
Firoozabadi, et al., “Combination of Nested Microphone Array and Subband Processing for Multiple Simultaneous Speaker Localization,” 6th International Symposium on Telecommunications, Nov. 2012, pp. 907-912.
Flanagan et al., Autodirective Microphone Systems, Acustica, vol. 73, 1991, pp. 58-71.
Flanagan, et al., “Computer-Steered Microphone Arrays for Sound Transduction in Large Rooms,” J. Acoust. Soc. Am. 78 (5), Nov. 1985, pp. 1508-1518.
Fohhn Audio New Generation of Beam Steering Systems Available Now, audioXpress Staff, May 10, 2017, 8 pp.
Fox, et al., “A Subband Hybrid Beamforming for In-Car Speech Enhancement,” 20th European Signal rocessing Conference, Aug. 2012, 5 pp.
Frost, III, An Algorithm for Linearly Constrained Adaptive Array Processing, Proc. IEEE, vol. 60, No. 8, Aug. 1972, pp. 926-935.
Gannot et al., Signal Enhancement using Beamforming and Nonstationarity with Applications to Speech, IEEE Trans. On Signal Processing, vol. 49, No. 8, Aug. 2001, pp. 1614-1626.
Gansler et al., A Double-Talk Detector Based on Coherence, IEEE Transactions on Communications, vol. 44, No. 11, Nov. 1996, pp. 1421-1427.
Gazor et al., Robust Adaptive Beamforming via Target Tracking, IEEE Transactions on Signal Processing, vol. 44, No. 6, Jun. 1996, pp. 1589-1593.
Gazor et al., Wideband Multi-Source Beamforming with Adaptive Array Location Calibration and Direction Finding, 1995 International Conference on Acoustics, Speech, and Signal Processing, May 1995, pp. 1904-1907.
Gentner Communications Corp., AP400 Audio Perfect 400 Audioconferencing System Installation & Operation Manual, Nov. 1998, 80 pgs.
Gentner Communications Corp., XAP 800 Audio Conferencing System Installation & Operation Manual, Oct. 2001, 152 pgs.
Gil-Cacho et al., Multi-Microphone Acoustic Echo Cancellation Using Multi-Channel Warped Linear Prediction of Common Acoustical Poles, 18th European Signal Processing Conference, Aug. 2010, pp. 2121-2125.
Giuliani, et al., “Use of Different Microphone Array Configurations for Hands-Free Speech Recognition in Noisy and Reverberant Environment,” IRST-Istituto per la Ricerca Scientifica e Tecnologica, Sep. 22, 1997, 4 pp.
Gritton et al., Echo Cancellation Algorithms, IEEE ASSP Magazine, vol. 1, issue 2, Apr. 1984, pp. 30-38.
Hald, et al., “A class of optimal broadband phased array geometries designed for easy construction,” 2002 Int'l Congress & Expo. on Noise Control Engineering, Aug. 2002, 6 pp.
Hamalainen, et al., “Acoustic Echo Cancellation for Dynamically Steered Microphone Array Systems,” 2007 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, Oct. 2007, pp. 58-61.
Hayo, Virtual Controls for Real Life, Web page downloaded from https://hayo.io/ on Sep. 18, 2019, 19 pp.
Herbordt et al., A Real-time Acoustic Human-Machine Front-End for Multimedia Applications Integrating Robust Adaptive Beamforrning and Stereophonic Acoustic Echo Cancellation, 7th International Conference on Spoken Language Processing, Sep. 2002, 4 pgs.
Herbordt et al., GSAEC—Acoustic Echo Cancellation embedded into the Generalized Sidelobe Canceller, 10th European Signal Processing Conference, Sep. 2000, 5 pgs.
Herbordt et al., Multichannel Bin-Wise Robust Frequency-Domain Adaptive Filtering and its Application to Adaptive Beamforming, IEEE Transactions on Audio, Speech, and Language Processing, vol. 15, No. 4, May 2007, pp. 1340-1351.
Herbordt, “Combination of Robust Adaptive Beamforming with Acoustic Echo Cancellation for Acoustic Human/Machine Interfaces,” Friedrich-Alexander University, 2003, 293 pgs.
Herbordt, et al., Joint Optimization of LCMV Beamforming and Acoustic Echo Cancellation for Automatic Speech Recognition, IEEE International Conference on Acoustics, Speech, and Signal Processing, Mar. 2005, pp. III-77-III-80.
Holm, “Optimizing Microphone Arrays for use in Conference Halls,” Norwegian University of Science and Technology, Jun. 2009, 101 pp.
Huang et al., Immersive Audio Schemes: The Evolution of Multiparty Teleconferencing, IEEE Signal Processing Magazine, Jan. 2011, pp. 20-32.
ICONYX Gen5, Product Overview; Renkus-Heinz, Dec. 24, 2018, 2 pp.
International Search Report and Written Opinion for PCT/US2016/022773 dated Jun. 10, 2016.
International Search Report and Written Opinion for PCT/US2016/029751 dated Nov. 28, 2016, 21 pp.
International Search Report and Written Opinion for PCT/US2018/013155 dated Jun. 8, 2018.
International Search Report and Written Opinion for PCT/US2019/031833 dated Jul. 24, 2019, 16 pp.
International Search Report and Written Opinion for PCT/US2019/033470 dated Jul. 31, 2019, 12 pp.
International Search Report and Written Opinion for PCT/US2019/051989 dated Jan. 10, 2020, 15 pp.
International Search Report and Written Opinion for PCT/US2020/024063 dated Aug. 31, 2020, 18 pp.
International Search Report and Written Opinion for PCT/US2020/035185 dated Sep. 15, 2020, 11 pp.
International Search Report and Written Opinion for PCT/US2020/058385 dated Mar. 31, 2021, 20 pp.
International Search Report and Written Opinion for PCT/US2021/070625 dated Sep. 17, 2021, 17 pp.
International Search Report for PCT/US2020/024005 dated Jun. 12, 2020, 12 pp.
Invensense, “Microphone Array Beamforming,” Application Note AN-1140, Dec. 31, 2013, 12 pp.
Invensense, Recommendations for Mounting and Connecting InvenSense MEMS Microphones, Application Note AN-1003, 2013, 11 pp.
Ito et al., Aerodynamic/Aeroacoustic Testing in Anechoic Closed Test Sections of Low-speed Wind Tunnels, 16th AIAA/CEAS Aeroacoustics Conference, 2010, 11 pgs.
Johansson et al., Robust Acoustic Direction of Arrival Estimation using Root-SRP-PHAT, a Realtime Implementation, IEEE International Conference on Acoustics, Speech, and Signal Processing, Mar. 2005, 4 pgs.
Johansson, et al., Speaker Localisation using the Far-Field SRP-PHAT in Conference Telephony, 2002 International Symposium on Intelligent Signal Processing and Communication Systems, 5 pgs.
Johnson, et al., “Array Signal Processing: Concepts and Techniques,” p. 59, Prentice Hall, 1993, 3 pp.
Julstrom et al., Direction-Sensitive Gating: A New Approach to Automatic Mixing, J. Audio Eng. Soc., vol. 32, No. 7/8, Jul./Aug. 1984, pp. 490-506.
Kahrs, Ed., The Past, Present, and Future of Audio Signal Processing, IEEE Signal Processing Magazine, Sep. 1997, pp. 30-57.
Kallinger et al., Multi-Microphone Residual Echo Estimation, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, Apr. 2003, 4 pgs.
Kammeyer, et al., New Aspects of Combining Echo Cancellers with Beamformers, IEEE International Conference on Acoustics, Speech, and Signal Processing, Mar. 2005, pp. III-137-III-140.
Kellermann, A Self-Steering Digital Microphone Array, 1991 International Conference on Acoustics, Speech, and Signal Processing, Apr. 1991, pp. 3581-3584.
Kellermann, Acoustic Echo Cancellation for Beamforming Microphone Arrays, in Brandstein, ed., Microphone Arrays: Techniques and Applications, 2001, Springer-Verlag Berlin Heidelberg, pp. 281-306.
Kellermann, Integrating Acoustic Echo Cancellation with Adaptive Beamforming Microphone Arrays, Forum Acusticum, Berlin, Mar. 1999, pp. 1-4.
Kellermann, Strategies for Combining Acoustic Echo Cancellation and Adaptive Beamforming Microphone Arrays, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing, Apr. 1997, 4 pgs.
Klegon, “Achieve Invisible Audio with the MXA910 Ceiling Array Microphone,” Jun. 27, 2016, 10 pp.
Knapp, et al., The Generalized Correlation Method for Estimation of Time Delay, IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. ASSP-24, No. 4, Aug. 1976, pp. 320-327.
Kobayashi et al., A Hands-Free Unit with Noise Reduction by Using Adaptive Beamformer, IEEE Transactions on Consumer Electronics, vol. 54, No. 1, Feb. 2008, pp. 116-122.
Kobayashi et al., A Microphone Array System with Echo Canceller, Electronics and Communications in Japan, Part 3, vol. 89, No. 10, Feb. 2, 2006, pp. 23-32.
Kolund{hacek over (z)}ija, et al., “Baffled circular loudspeaker array with broadband high directivity,” 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, 2010, pp. 73-76.
Lai, et al., “Design of Robust Steerable Broadband Beamformers with Spiral Arrays and the Farrow Filter Structure,” Proc. Intl. Workshop Acoustic Echo Noise Control, 2010, 4 pp.
Lebret, et al., Antenna Array Pattern Synthesis via Convex Cptimization, IEEE Trans. on Signal Processing, vol. 45, No. 3, Mar. 1997, pp. 526-532.
LecNet2 Sound System Design Guide, Lectrosonics, Jun. 2, 2006.
Lectrosonics, LecNet2 Sound System Design Guide, Jun. 2006, 28 pgs.
Lee et al., Multichannel Teleconferencing System with Multispatial Region Acoustic Echo Cancellation, International Workshop on Acoustic Echo and Noise Control (IWAENC2003), Sep. 2003, pp. 51-54.
Li, “Broadband Beamforming and Direction Finding Using Concentric Ring Array,” Ph.D. Dissertation, University of Missouri-Columbia, Jul. 2005, 163 pp.
Lindstrom et al., An Improvement of the Two-Path Algorithm Transfer Logic for Acoustic Echo Cancellation, IEEE Transactions on Audio, Speech, and Language Processing, vol. 15, No. 4, May 2007, pp. 1320-1326.
Liu et al., Adaptive Beamforming with Sidelobe Control: A Second-Order Cone Programming Approach, IEEE Signal Proc. Letters, vol. 10, No. 11, Nov. 2003, pp. 331-334.
Liu, et al., “Frequency Invariant Beamforming in Subbands,” IEEE Conference on Signals, Systems and Computers, 2004, 5 pp.
Liu, et al., “Wideband Beamforming,” Wiley Series on Wireless Communications and Mobile Computing, pp. 143-198, 2010, 297 pp.
Lobo, et al., Applications of Second-Order Cone Programming, Linear Algebra and its Applications 284, 1998, pp. 193-228.
Luo et al., Wideband Beamforming with Broad Nulls of Nested Array, Third Int'l Conf. on Info. Science and Tech., Mar. 23-25, 2013, pp. 1645-1648.
Marquardt et al., A Natural Acoustic Front-End for Interactive TV in the EU-Project DICIT, IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, Aug. 2009, pp. 894-899.
Martin, Small Microphone Arrays with Postfilters for Noise and Acoustic Echo Reduction, in Brandstein, ed., Microphone Arrays: Techniques and Applications, 2001, Springer-Verlag Berlin Heidelberg, pp. 255-279.
Maruo et al., On the Optimal Solutions of Beamformer Assisted Acoustic Echo Cancellers, IEEE Statistical Signal Processing Workshop, 2011, pp. 641-644.
McCowan, Microphone Arrays: A Tutorial, Apr. 2001, 36 pgs.
MFLCRFG Datasheet, Metal_Fab Inc., Sep. 7, 2007, 1 p.
Microphone Array Primer, Shure Question and Answer Page, <https://service.shure.eom/s/article/microphone-array-primer?language=en_US>, Jan. 2019, 5 pp.
Milanovic, et al., “Design and Realization of FPGA Platform for Real Time Acoustic Signal Acquisition and Data Processing” 22nd Telecommunications Forum TELFOR, 2014, 6 pp.
Mohammed, A New Adaptive Beamformer for Optimal Acoustic Echo and Noise Cancellation with Less Computational Load, Canadian Conference on Electrical and Computer Engineering, May 2008, pp. 000123-000128.
Mohammed, A New Robust Adaptive Beamformer for Enhancing Speech Corrupted with Colored Noise, AICCSA, Apr. 2008, pp. 508-515.
Mohammed, Real-time Implementation of an efficient RLS Algorithm based on IIR Filter for Acoustic Echo Cancellation, AICCSA, Apr. 2008, pp. 489-494.
Mohan, et al., “Localization of multiple acoustic sources with small arrays using a coherence test,” Journal Acoustic Soc Am., 123(4), Apr. 2008, 12 pp.
Moulines, et al., “Pitch-Synchronous Waveform Processing Techniques for Text-to-Speech Synthesis Using Diphones,” Speech Communication 9, 1990, 15 pp.
Multichannel Acoustic Echo Cancellation, Obtained from website http://www.buchner-net.com/mcaec.html, Jun. 2011.
Myllyla et al., Adaptive Beamforming Methods for Dynamically Steered Microphone Array Systems, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, Mar.-Apr. 2008, pp. 305-308.
New Shure Microflex Advance MXA910 Microphone With Intellimix Audio Processing Provides Greater Simplicity, Flexibility, Clarity, Press Release, Jun. 12, 2019, 4 pp.
Nguyen-Ky, et al., “An Improved Error Estimation Algorithm for Stereophonic Acoustic Echo Cancellation Systems,” 1st International Conference on Signal Processing and Communication Systems, Dec. 17-19, 2007, 5 pp.
Office Action for Taiwan Patent Application No. 105109900 dated May 5, 2017.
Office Action issued for Japanese Patent Application No. 2015-023781 dated Jun. 20, 2016, 4 pp.
Oh, et al., “Hands-Free Voice Communication in an Automobile With a Microphone Array,” 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing, Mar. 1992, pp. I-281-I-284.
Olszewski, et al., “Steerable Highly Directional Audio Beam Loudspeaker,” Interspeech 2005, 4 pp.
Omologo, Multi-Microphone Signal Processing for Distant-Speech Interaction, Human Activity and Vision Summer School (HAVSS), INRIA Sophia Antipolis, Oct. 3, 2012, 79 pgs.
Order, Conduct of the Proceeding, Clearone, Inc. v. Shure Acquisition Holdings, Inc., Nov. 2, 2020, 10 pp.
Pados et al., An Iterative Algorithm for the Computation of the MVDR Filter, IEEE Trans. On Signal Processing, vol. 49, No. 2, Feb. 2001, pp. 290-300.
Palladino, “This App Lets You Control Your Smarthome Lights via Augmented Reality,” Next Reality Mobile AR News, Jul. 2, 2018, 5 pp.
Parikh, et al., “Methods for Mitigating IP Network Packet Loss in Real Time Audio Streaming Applications,” GatesAir, 2014, 6 pp.
Pasha, et al., “Clustered Multi-channel Dereverberation for Ad-hoc Microphone Arrays,” Proceedings of APSIPA Annual Summit and Conference, Dec. 2015, pp. 274-278.
Petitioner's Motion for Sanctions, Clearone, Inc. v. Shure Acquisition Holdings, Inc., Aug. 24, 2020, 20 pp.
Pettersen, “Broadcast Applications for Voice-Activated Microphones,” db, Jul./Aug. 1985, 6 pgs.
Pfeifenberger, et al., “Nonlinear Residual Echo Suppression using a Recurrent Neural Network,” Interspeech 2020, 5 pp.
Phoenix Audio Technologies, “Beamforming and Microphone Arrays—Common Myths”, Apr. 2016, http://info.phnxaudio.com/blog/microphone-arrays-beamforming-myths-1, 19 pp.
Plascore, PCGA-XR1 3003 Aluminum Honeycomb Data Sheet, 2008, 2 pgs.
Polycom Inc., Vortex EF2211/EF2210 Reference Manual, 2003, 66 pgs.
Polycom, Inc., Polycom Soundstructure C16, C12, C8, and SR12 Design Guide, Nov. 2013, 743 pgs.
Polycom, Inc., Setting Up the Polycom HDX Ceiling Microphone Array Series, https://support.polycom.com/content/dam/polycom-support/products/Telepresence-and-Video/HDX%20Series/setup-maintenance/en/hdx_ceiling_microphone_array_setting_up.pdf, 2010, 16 pgs.
Polycom, Inc., Vortex EF2241 Reference Manual, 2002, 68 pgs.
Polycom, Inc., Vortex EF2280 Reference Manual, 2001, 60 pp.
Pomona, Model 3306, Datasheet, Jun. 9, 1999, 1 p.
Powers, et al., “Proving Adaptive Directional Technology Works: A Review of Studies,” The Hearing Review, Apr. 6, 2004, 5 pp.
Prime, et al., “Beamforming Array Optimisation Averaged Sound Source Mapping on a Model Wind Turbine,” ResearchGate, Nov. 2014, 10 pp.
Rabinkin et al., Estimation of Wavefront Arrival Delay Using the Cross-Power Spectrum Phase Technique, 132nd Meeting of the Acoustical Society of America, Dec. 1996, pp. 1-10.
Rane Corp., Halogen Acoustic Echo Cancellation Guide, AEC Guide Version 2, Nov. 2013, 16 pgs.
Rao, et al., “Fast LMS/Newton Algorithms for Stereophonic Acoustic Echo Cancelation,” IEEE Transactions on Signal Processing, vol. 57, No. 8, Aug. 2009.
Reuven et al., Joint Acoustic Echo Cancellation and Transfer Function GSC in the Frequency Domain, 23rd IEEE Convention of Electrical and Electronics Engineers in Israel, Sep. 2004, pp. 412-415.
Reuven et al., Joint Noise Reduction and Acoustic Echo Cancellation Using the Transfer-Function Generalized Sidelobe Canceller, Speech Communication, vol. 49, 2007, pp. 623-635.
Reuven, et al., “Multichannel Acoustic Echo Cancellation and Noise Reduction in Reverberant Environments Using the Transfer-Function GSC,” 2007 IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 2007, 4 pp.
Ristimaki, Distributed Microphone Array System for Two-Way Audio Communication, Helsinki Univ. of Technology, Master's Thesis, Jun. 15, 2009, 73 pgs.
Rombouts et al., An Integrated Approach to Acoustic Noise and Echo Cancellation, Signal Processing 85, 2005, pp. 849-871.
Sällberg, “Faster Subband Signal Processing,” IEEE Signal Processing Magazine, vol. 30, No. 5, Sep. 2013, 6 pp.
Sasaki et al., A Predefined Command Recognition System Using a Ceiling Microphone Array in Noisy Housing Environments, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Sep. 2008, pp. 2178-2184.
Sennheiser, New microphone solutions for ceiling and desk installation, https://en-us.sennheiser.com/news-new-microphone-solutions-for-ceiling-and-desk-installation, Feb. 2011, 2 pgs.
Sennheiser, TeamConnect Ceiling, https://en-us.sennheiser.com/conference-meeting-rooms-teamconnect-ceiling, 7 pgs.
Serdes, Wikipedia article, last edited on Jun. 25, 2018; retrieved on Jun. 27, 2018, 3 pp., https://en.wikipedia.org/wiki/SerDes.
Sessler, et al., “Directional Transducers,” IEEE Transactions on Audio and Electroacoustics, vol. AU-19, No. 1, Mar. 1971, pp. 19-23.
Sessler, et al., “Toroidal Microphones,” Journal of Acoustical Society of America, vol. 46, No. 1, 1969, 10 pp.
Shure AMS Update, vol. 1, No. 1, 1983, 2 pgs.
Shure AMS Update, vol. 1, No. 2, 1983, 2 pgs.
Shure AMS Update, vol. 4, No. 4, 1997, 8 pgs.
Shure Debuts Microflex Advance Ceiling and Table Array Microphones, Press Release, Feb. 9, 2016, 4 pp.
Shure Inc., A910-HCM Hard Ceiling Mount, retrieved from website <http://www.shure.com/en-US/products/accessories/a910hcm> on Jan. 16, 2020, 3 pp.
Shure Inc., Microflex Advance, http://www.shure.com/americas/microflex-advance, 12 pgs.
Shure Inc., MX395 Low Profile Boundary Microphones, 2007, 2 pgs.
Shure Inc., MXA910 Ceiling Array Microphone, http://www.shure.com/americas/products/microphones/microflex-advance/mxa910-ceiling-array-microphone, 7 pgs.
Shure, New MXA910 Variant Now Available, Press Release, Dec. 13, 2019, 5 pp.
Shure, Q&A in Response to Recent Us Court Ruling on Shure MXA910, Available at <https://www.shure.com/en-US/meta/legal/q-and-a-inresponse-to-recent-US-court-ruling-on-shure-mxa910-response>, As early as 2020, 5 pp.
Shure, RK244G Replacement Screen and Grille, Datasheet, 2013, 1 p.
Shure, The Microflex Advance MXA310 Table Array Microphone, Available at <https://www.shure.com/en-US/products/microphones/mxa310>, As early as 2020, 12 pp.
Signal Processor MRX7-D Product Specifications, Yamaha Corporation, 2016.
Silverman et al., Performance of Real-Time Source-Location Estimators for a Large-Aperture Microphone Array, IEEE Transactions on Speech and Audio Processing, vol. 13, No. 4, Jul. 2005, pp. 593-606.
Sinha, Ch. 9: Noise and Echo Cancellation, in Speech Processing in Embedded Systems, Springer, 2010, pp. 127-142.
SM 69 Stereo Microphone, Datasheet, Georg Neumann GmbH, Available at <https://ende.neumann.com/product_files/6552/download>, 1 p.
Soda et al., Introducing Multiple Microphone Arrays for Enhancing Smart Home Voice Control, The Institute of Electronics, Information and Communication Engineers, Technical Report of IEICE, Jan. 2013, 6 pgs.
Soundweb London Application Guides, BSS Audio, 2010.
Symetrix, Inc., SymNet Network Audio Solutions Brochure, 2008, 32 pgs.
SymNet Network Audio Solutions Brochure, Symetrix, Inc., 2008.
Tan, et al., “Pitch Detection Algorithm: Autocorrelation Method and AMDF,” Department of Computer Engineering, Prince of Songkhla University, Jan. 2003, 6 pp.
Tandon, et al., “An Efficient, Low-Complexity, Normalized LMS Algorithm for Echo Cancellation,” 2nd Annual IEEE Northeast Workshop on Circuits and Systems, Jun. 2004, pp. 161-164.
Tetelbaum et al., Design and Implementation of a Conference Phone Based on Microphone Array Technology, Proc. Global Signal Processing Conference and Expo (GSPx), Sep. 2004, 6 pgs.
Tiete et al., SoundCompass: A Distributed MEMS Microphone Array-Based Sensor for Sound Source Localization, Sensors, Jan. 23, 2014, pp. 1918-1949.
TOA Corp., Ceiling Mount Microphone AN-9001 Operating Instructions, http://www.toaelectronics.com/media/an9001_mt1e.pdf, 1 pg.
Togami, et al., “Subband Beamformer Combined with Time-Frequency ICA for Extraction of Target Source Under Reverberant Environments,” 17th European Signal Processing Conference, Aug. 2009, 5 pp.
U.S. Appl. No. 16/598,918, filed Oct. 10, 2019, 50 pp.
Van Compernolle, Switching Adaptive Filters for Enhancing Noisy and Reverberant Speech from Microphone Array Recordings, Proc. IEEE Inf. Conf. on Acoustics, Speech, and Signal Processing, Apr. 1990, pp. 833-836.
Van Trees, Optimum Array Processing: Part IV of Detection, Estimation, and Modulation Theory, 2002, 54 pgs., pp. i-xxv, 90-95, 201-230.
Van Veen et al., Beamforming: A Versatile Approach to Spatial Filtering, IEEE ASSP Magazine, vol. 5, issue 2, Apr. 1988, pp. 4-24.
Vicente, “Adaptive Array Signal Processing Using the Concentric Ring Array and the Spherical Array,” Ph.D. Dissertation, University of Missouri, May 2009, 226 pp.
Wang et al., Combining Superdirective Beamforming and Frequency-Domain Blind Source Separation for Highly Reverberant Signals, EURASIP Journal on Audio, Speech, and Music Processing, vol. 2010, pp. 1-13.
Warsitz, et al., “Blind Acoustic Beamforming Based on Generalized Eigenvalue Decomposition,” IEEE Transactions on Audio, Speech and Language Processing, vol. 15, No. 5, 2007, 11 pp.
Weinstein, et al., “LOUD: A 1020-Node Microphone Array and Acoustic Beamformer,” 14th International Congress on Sound & Vibration, Jul. 2007, 8 pgs.
Weinstein, et al., “LOUD: A 1020-Node Modular Microphone Array and Beamformer for Intelligent Computing Spaces,” MIT Computer Science and Artifical Intelligence Laboratory, 2004, 18 pp.
Wung, “A System Approach to Multi-Channel Acoustic Echo Cancellation and Residual Echo Suppression for Robust Hands-Free Teleconferencing,” Georgia Institute of Technology, May 2015, 167 pp.
XAP Audio Conferencing Brochure, ClearOne Communications, Inc., 2002.
Yamaha Corp., MRX7-D Signal Processor Product Specifications, 2016, 12 pgs.
Yamaha Corp., PJP-100H IP Audio Conference System Owner's Manual, Sep. 2006, 59 pgs.
Yamaha Corp., PJP-EC200 Conference Echo Canceller Brochure, Oct. 2009, 2 pgs.
Yan et al., Convex Optimization Based Time-Domain Broadband Beamforming with Sidelobe Control, Journal of the Acoustical Society of America, vol. 121, No. 1, Jan. 2007, pp. 46-49.
Yensen et al., Synthetic Stereo Acoustic Echo Cancellation Structure with Microphone Array Beamforming for VOIP Conferences, 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing, Jun. 2000, pp. 817-820.
Yermeche, et al., “Real-Time DSP Implementation of a Subband Beamforming Algorithm for Dual Microphone Speech Enhancement,” 2007 IEEE International Symposium on Circuits and Systems, 4 pp.
Zavarehei, et al., “Interpolation of Lost Speech Segments Using LP-HNM Model with Codebook Post-Processing,” IEEE Transactions on Multimedia, vol. 10, No. 3, Apr. 2008, 10 pp.
Zhang, et al., “F-T-LSTM based Complex Network for Joint Acoustic Echo Cancellation and Speech Enhancement,” Audio, Speech and Language Processing Group, Jun. 2021, 5 pp.
Zhang, et al., “Multichannel Acoustic Echo Cancelation in Multiparty Spatial Audio Conferencing with Constrained Kalman Filtering,” 11th International Workshop on Acoustic Echo and Noise Control, Sep. 14, 2008, 4 pp.
Zhang, et al., “Selective Freguency Invariant Uniform Circular Broadband Beamformer,” EURASIP Journal on Advances in Signal Processing, vol. 2010, pp. 1-11.
Zheng, et al., “Experimental Evaluation of a Nested Microphone Array With Adaptive Noise Cancellers,” IEEE Transactions on Instrumentation and Measurement, vol. 53, No. 3, Jun. 2004, 10 pp.
Related Publications (1)
Number Date Country
20210250009 A1 Aug 2021 US
Provisional Applications (1)
Number Date Country
62971850 Feb 2020 US