The present disclosure relates generally to audio capture optimization and, more particularly, to the optimization of a multi-microphone system for an endpoint device.
The share of the workforce that is working from home has increased dramatically in recent times. Collaboration between remote employees remains necessary, however. As a result, tele- and video-conferences have become a common and valuable resource for many businesses.
Some endpoint devices which are designed specifically with conferencing in mind integrate several individual components-such as a microphone, loudspeaker, camera, and/or video display-into a single unit. These all-in-one endpoints, while certainly convenient, encounter challenges with regard to an acoustical design that ensures high-quality speech acquisition. For instance, the close proximity of the loudspeaker and the microphone frequently results in an increased echo to near-end speech ratio. This phenomenon makes echo control with satisfactory double-talk performance difficult to achieve.
Furthermore, users are often unaware of where the endpoint's microphones and/or loudspeakers are located, or unaware that placing objects too close to the endpoint may degrade speech signal pickup. Placing an object, such as a laptop computer, in front of the microphone, for example, impairs sound quality by removing high-frequency content. It can also increase the acoustic coupling between the loudspeaker and microphone, causing detrimental echo and distortion artifacts at the far-end. Even the sound of the laptop's cooling fan can impair the audio signal through added noise when placed close to the microphone. Although the endpoint device can be elevated from a table surface to reduce the chance of shadowing effects at the microphone, this solution, too, is non-optimal as sound reflection from the table can result in comb-filtering that harms the sound quality all the same.
The embodiments herein may be better understood by referring to the following description in conjunction with the accompanying drawings in which like reference numerals indicate identical or functionally similar units, of which:
Overview
According to one or more embodiments of the disclosure, input signals are acquired for a remote conference between an endpoint device and at least one other endpoint device. The input signals are received via a multi-microphone system including at least a top microphone unit disposed at a top area of the endpoint device and a bottom microphone unit disposed at a bottom area of the endpoint device. A signal degradation event that causes degradation of signals received by the top microphone unit or the bottom microphone unit is detected. Then, based on information regarding the signal degradation event, it is determined whether the signal degradation event affects one or both of the top microphone unit and the bottom microphone unit. In response to determining that the signal degradation event affects one or both of the top microphone unit and the bottom microphone unit, an output signal is generated for transmission to the at least one other endpoint device, and the output signal uses a portion of the input signals that excludes signals received by the top microphone unit and/or the bottom microphone unit determined to be affected by the signal degradation event.
A computer network is a geographically distributed collection of nodes interconnected by communication links and segments for transporting data between end nodes, such as personal computers and workstations, or other devices, such as sensors, etc. Many types of networks are available, with the types ranging from local area networks (LANs) to wide area networks (WANs). LANs typically connect the nodes over dedicated private communications links located in the same general physical location, such as a building or campus. WANs, on the other hand, typically connect geographically dispersed nodes over long-distance communications links, such as common carrier telephone lines, optical lightpaths, synchronous optical networks (SONET), or synchronous digital hierarchy (SDH) links, or Powerline Communications (PLC) such as IEEE 61334, IEEE P1901.2, and others. The Internet is an example of a WAN that connects disparate networks throughout the world, providing global communication between nodes on various networks. The nodes typically communicate over the network by exchanging discrete frames or packets of data according to predefined protocols, such as the Transmission Control Protocol/Internet Protocol (TCP/IP). In this context, a protocol consists of a set of rules defining how the nodes interact with each other. Computer networks may be further interconnected by an intermediate network node, such as a router, to extend the effective “size” of each network.
Smart object networks, such as sensor networks, in particular, are a specific type of network having spatially distributed autonomous devices such as sensors, actuators, etc., that cooperatively monitor physical or environmental conditions at different locations, such as, e.g., energy/power consumption, resource consumption (e.g., water/gas/etc. for advanced metering infrastructure or “AMI” applications) temperature, pressure, vibration, sound, radiation, motion, pollutants, etc. Other types of smart objects include actuators, e.g., responsible for turning on/off an engine or perform any other actions. Sensor networks, a type of smart object network, are typically shared-media networks, such as wireless or PLC networks. That is, in addition to one or more sensors, each sensor device (node) in a sensor network may generally be equipped with a radio transceiver or other communication port such as PLC, a microcontroller, and an energy source, such as a battery. Often, smart object networks are considered field area networks (FANs), neighborhood area networks (NANs), personal area networks (PANs), etc. Generally, size and cost constraints on smart object nodes (e.g., sensors) result in corresponding constraints on resources such as energy, memory, computational speed and bandwidth.
In some implementations, a router or a set of routers may be connected to a private network (e.g., dedicated leased lines, an optical network, etc.) or a virtual private network (VPN), such as an MPLS VPN thanks to a carrier network, via one or more links exhibiting very different network and service level agreement characteristics. For the sake of illustration, a given customer site may fall under any of the following categories:
1.) Site Type A: a site connected to the network (e.g., via a private or VPN link) using a single CE router and a single link, with potentially a backup link (e.g., a 3G/4G/LTE backup connection). For example, a particular CE router 110 shown in network 100 may support a given customer site, potentially also with a backup link, such as a wireless connection.
2.) Site Type B: a site connected to the network using two MPLS VPN links (e.g., from different Service Providers), with potentially a backup link (e.g., a 3G/4G/LTE connection). A site of type B may itself be of different types:
2a.) Site Type B1: a site connected to the network using two MPLS VPN links (e.g., from different Service Providers), with potentially a backup link (e.g., a 3G/4G/LTE connection).
2b.) Site Type B2: a site connected to the network using one MPLS VPN link and one link connected to the public Internet, with potentially a backup link (e.g., a 3G/4G/LTE connection). For example, a particular customer site may be connected to network 100 via PE-3 and via a separate Internet connection, potentially also with a wireless backup link.
2c.) Site Type B3: a site connected to the network using two links connected to the public Internet, with potentially a backup link (e.g., a 3G/4G/LTE connection).
Notably, MPLS VPN links are usually tied to a committed service level agreement, whereas Internet links may either have no service level agreement at all or a loose service level agreement (e.g., a “Gold Package” Internet service connection that guarantees a certain level of performance to a customer site).
3.) Site Type C: a site of type B (e.g., types B1, B2 or B3) but with more than one CE router (e.g., a first CE router connected to one link while a second CE router is connected to the other link), and potentially a backup link (e.g., a wireless 3G/4G/LTE backup link). For example, a particular customer site may include a first CE router 110 connected to PE-2 and a second CE router 110 connected to PE-3.
Servers 152-154 may include, in various embodiments, a network management server (NMS), a dynamic host configuration protocol (DHCP) server, a constrained application protocol (CoAP) server, an outage management system (OMS), an application policy infrastructure controller (APIC), an application server, etc. As would be appreciated, network 100 may include any number of local networks, data centers, cloud environments, devices/nodes, servers, etc.
In some embodiments, the techniques herein may be applied to other network topologies and configurations. For example, the techniques herein may be applied to peering points with high-speed links, data centers, etc.
In various embodiments, network 100 may include one or more mesh networks, such as an Internet of Things network. Loosely, the term “Internet of Things” or “IoT” refers to uniquely identifiable objects (things) and their virtual representations in a network-based architecture. In particular, the next frontier in the evolution of the Internet is the ability to connect more than just computers and communications devices, but rather the ability to connect “objects” in general, such as lights, appliances, vehicles, heating, ventilating, and air-conditioning (HVAC), windows and window shades and blinds, doors, locks, etc. The “Internet of Things” thus generally refers to the interconnection of objects (e.g., smart objects), such as sensors and actuators, over a computer network (e.g., via IP), which may be the public Internet or a private network.
Notably, shared-media mesh networks, such as wireless or PLC networks, etc., are often on what is referred to as Low-Power and Lossy Networks (LLNs), which are a class of network in which both the routers and their interconnect are constrained: LLN routers typically operate with constraints, e.g., processing power, memory, and/or energy (battery), and their interconnects are characterized by, illustratively, high loss rates, low data rates, and/or instability. LLNs are comprised of anything from a few dozen to thousands or even millions of LLN routers, and support point-to-point traffic (between devices inside the LLN), point-to-multipoint traffic (from a central control point such at the root node to a portion of devices inside the LLN), and multipoint-to-point traffic (from devices inside the LLN towards a central control point). Often, an IoT network is implemented with an LLN-like architecture. For example, as shown, local network 160 may be an LLN in which CE-2 operates as a root node for nodes/devices 10-16 in the local mesh, in some embodiments.
In contrast to traditional networks, LLNs face a number of communication challenges. First, LLNs communicate over a physical medium that is strongly affected by environmental conditions that change over time. Some examples include temporal changes in interference (e.g., other wireless networks or electrical appliances), physical obstructions (e.g., doors opening/closing, seasonal changes such as the foliage density of trees, etc.), and propagation characteristics of the physical media (e.g., temperature or humidity changes, etc.). The time scales of such temporal changes can range between milliseconds (e.g., transmissions from other transceivers) to months (e.g., seasonal changes of an outdoor environment). In addition, LLN devices typically use low-cost and low-power designs that limit the capabilities of their transceivers. In particular, LLN transceivers typically provide low throughput. Furthermore, LLN transceivers typically support limited link margin, making the effects of interference and environmental changes visible to link and network protocols. The high number of nodes in LLNs in comparison to traditional networks also makes routing, quality of service (QoS), security, network management, and traffic engineering extremely challenging, to mention a few.
The network interfaces 210 include the mechanical, electrical, and signaling circuitry for communicating data over physical links coupled to the network 100. The network interfaces may be configured to transmit and/or receive data using a variety of different communication protocols. Notably, a physical network interface 210 may also be used to implement one or more virtual network interfaces, such as for virtual private network (VPN) access, known to those skilled in the art.
The memory 240 comprises a plurality of storage locations that are addressable by the processor(s) 220 and the network interfaces 210 for storing software programs and data structures associated with the embodiments described herein. The processor 220 may comprise necessary units or logic adapted to execute the software programs and manipulate the data structures 245. An operating system 242 (e.g., the Internetworking Operating System, or IOS®, of Cisco Systems, Inc., another operating system, etc.), portions of which are typically resident in memory 240 and executed by the processor(s), functionally organizes the node by, inter alia, invoking network operations in support of software processors and/or services executing on the device. These software processors and/or services may comprise a multi-microphone optimization process 248, as described herein.
It will be apparent to those skilled in the art that other processor and memory types, including various computer-readable media, may be used to store and execute program instructions pertaining to the techniques described herein. Also, while the description illustrates various processes, it is expressly contemplated that various processes may be embodied as modules configured to operate in accordance with the techniques herein (e.g., according to the functionality of a similar process). Further, while processes may be shown and/or described separately, those skilled in the art will appreciate that processes may be routines or modules within other processes.
As noted above, remote collaboration has become increasing commonplace in recent times. Video conferencing, for instance, allows employees in disparate locations to view and communicate with each other as if present in the same room. It can increase productivity and worker efficiency, while simultaneously saving costs for the employer. At the same time, in order for video conferencing tools to create an enjoyable user experience without disturbances, it is important to ensure that input speech signals can be captured with high quality, i.e., naturally. Some newer endpoint devices which integrate several individual conferencing components—e.g., a microphone, loudspeaker, camera, video display, etc.-enhance user convenience but encounter challenges with regard to an acoustical design that ensures high-quality speech acquisition. For instance, the close proximity of the loudspeaker and the microphone frequently results in an increased echo-to-near-end speech ratio. This phenomenon makes echo control with satisfactory double-talk performance difficult to achieve.
Furthermore, users are often unaware of where the endpoint's microphones and/or loudspeakers are located, or unaware that placing objects too close to the endpoint may degrade speech signal pickup. Placing an object, such as a laptop computer, in front of the microphone, for example, impairs sound quality by removing high-frequency content. It can also increase the acoustic coupling between the loudspeaker and microphone, causing detrimental echo and distortion artifacts at the far-end. Even the sound of the laptop's cooling fan can add noise to and thus impair the audio signal when placed close to the microphone. Although the endpoint device can be elevated from a table surface to reduce the chance of shadowing effects at the microphone, this solution, too, is non-optimal as sound reflection from the table can result in comb-filtering that harms the sound quality all the same.
Optimization Of Multi-Microphone System For Endpoint Device
The techniques herein introduce techniques for enhancing the quality of speech signal acquisition by tele- or video-conferencing endpoint devices using a multi-microphone system, including at least top and bottom microphones, and intelligent switching between microphones, and combinations of microphones, based on detected signal degradation events. In some aspects, techniques are described for detecting events during an ongoing communication session that would potentially degrade input signal quality, such as physical obstructions, noise, table reflection effects, echo levels, double-talk performance, camera shutter operation, and so on. These events can be used as the basis for deciding which microphone, or combination of microphones, should be used at any given time for generating an optimized output signal. In further aspects, the multi-microphone system may comprise one or more dual-microphone arrays positioned at the top and/or the bottom of the endpoint to receive sound from predefined sectors and suppress noise based on spatial separation.
Specifically, according to one or more embodiments of the disclosure as described in detail below, input signals are acquired for a remote conference between an endpoint device and at least one other endpoint device. The input signals are received via a multi-microphone system including at least a top microphone unit disposed at a top area of the endpoint device and a bottom microphone unit disposed at a bottom area of the endpoint device. A signal degradation event that causes degradation of signals received by the top microphone unit or the bottom microphone unit is detected. Then, based on information regarding the signal degradation event, it is determined whether the signal degradation event affects one or both of the top microphone unit and the bottom microphone unit. In response to determining that the signal degradation event affects one or both of the top microphone unit and the bottom microphone unit, an output signal is generated for transmission to the at least one other endpoint device, and the output signal uses a portion of the input signals that excludes signals received by the top microphone unit and/or the bottom microphone unit determined to be affected by the signal degradation event.
Illustratively, the techniques described herein may be performed by hardware, software, and/or firmware, such as in accordance with the multi-microphone optimization process 248, which may include computer executable instructions executed by the processor 220, to perform functions relating to the techniques described herein.
Operationally, an example video conferencing endpoint device 300 is illustrated in
Endpoint device 300, as shown in
With respect to the microphone elements, in particular, the endpoint device 300 may include a multi-microphone system comprising a plurality of microphone units. In some cases, the microphone units may be omni-directional microphone units. The microphone units may be variously positioned on the endpoint device 300. According to some embodiments, the multi-microphone system may include, at least, a bottom microphone unit 310 disposed at a bottom area of the endpoint device 300 and a top microphone unit 320 disposed at a top area of the endpoint device 300. More precisely, the microphone units may be positioned such that the bottom microphone unit 310 is physically separated from the top microphone unit 320, meaning that the “bottom area” and the “top area” of the endpoint device 300 are similarly physically separated from each other. There may be any number of the bottom and top microphone units, respectively. For instance, as shown in
With respect to the bottom microphone units 310, in particular, the endpoint device 300 may include a bottom-left microphone unit 310a disposed at a bottom-left area of the endpoint device 300 and a bottom-right microphone unit 310b disposed at a bottom-right area of the endpoint device 300. According to some embodiments, and as shown in
However, when these bottom microphones are utilized on its own (i.e., without supplemental microphones), certain challenges arise:
As an example, in the event that a laptop computer resting on the table surface is placed in front of a base (bottom) microphone, the affected microphone may likely experience shadowing which leads to a low-pass effect, as low-frequency sound waves diffract around the obstacle and are picked up by the microphone, while high-frequency sound waves are blocked or severely attenuated. The present application counteracts this problem of shadowing by utilizing multiple, well-separated microphones at the base of the endpoint device 300—e.g., bottom-left microphone unit 310a and bottom-right microphone unit 310b. At any given time, the input signal received by the bottom microphone unit 310 that is least obstructed may be selected, as described in greater detail below.
While using multiple, separated bottom microphone units 310 is beneficial, particularly in the case of a physical obstruction being placed in front of one, but not both, microphone units, the concern of high echo levels, shadowing effects, and poor double-talk performance still exists when the only microphone units are located on the bottom of endpoint device 300. Therefore, the present application seeks to resolve this issue by also utilizing a top microphone unit 320 that is spaced apart from the bottom microphone units 310, as well as the integrated loudspeaker unit 330. According to some embodiments, the top microphone unit 320 may be front-facing and disposed at a top area of the endpoint device 300, as shown in
Nevertheless, even with the benefits that accompany the top and bottom microphone arrangement illustrated in
In detail,
Microphone units, including bottom-left microphone unit 310a, bottom-right microphone unit 310b, and top microphone unit 320, are shown on the left end of architecture 400. As explained above, the microphone units may receive input signals, e.g., speech signals, for a remote conference. For instance, a user of the endpoint device 300 may be engaged in a call with co-workers across one or more other endpoints. The input signals received by the various microphone units may comprise, notably, the user's speech. The input signals may also comprise other artifacts, however, that potentially compromise the signal quality of the user's speech, such as outside noise due to sound emitted from the loudspeaker 330, a fan of a nearby laptop computer, a camera shutter, echo, and so forth. The raw input signals as received from the microphone units may be provided both to the detector module 410 and the signal processing module 420 for further analysis and processing, as will be described below.
According to further embodiments, the top microphone unit 320 may comprise dual microphone units separated from each other (e.g., 17 mm). The dual top microphones may be used to estimate the direction of arrival of the input signals. For instance, assume an audio signal is acquired from one of the two microphones, from a third microphone, or from a microphone array. This audio signal is then processed in such a way that audio originating from undesired directions is attenuated. This enables the attenuation of audio originating from undesired directions. As such, the dual top microphone units may be used specifically for estimating the direction or arrival, whereas the remaining microphone units receive signals specifically for processing, as described herein. In yet further embodiments, face detection based on the video image from the camera 350 may be utilized so any audio originating from a direction where no face is detected can be attenuated.
Meanwhile, the architecture 400 may collect information used for detecting whether a signal degradation event is present. A signal degradation event, for the purposes of the present application, may refer to an event that causes signal degradation of signals received by any one or more of microphone units of the endpoint device 300. In response, a portion of the input signals that is negatively affected by the signal degradation event may be excluded to generate an optimized output signal for transmission to another endpoint device, as will be described in detail later.
Numerous possible signal degradation events are envisioned herein, as are the responses to each signal degradation event, and also the techniques for detecting the signal degradation events for a remote conference. Possible signal degradation events may include, but are not limited to:
Each of the signal degradation events may degrade the quality of speech signals from a user of the endpoint device 300, diminishing the overall experience of the tele- or video-conference. Given the placement of the various microphone units, as illustrated in
Detector module 410 may execute stored program instructions (e.g., multi-microphone optimization process 248) to detect the presence of a signal degradation event. For example, detector module 410 may assess the input signals received by top and bottom microphone units 310/320 to determine whether there is evidence of signal degradation. Additionally, sensor data obtained by one or more additional sensors 430 may be provided to the detector module 410 for processing in conjunction with the input signals received by the microphone units 310/320. The data provided from sensors 430 may be utilized by the detector module 410 to measure noise levels, echo levels, and so on, each of which may represent indicators of a signal degradation event. According to some embodiments, the sensors 430 may include, for example, an ultrasonic sensor, camera, additional microphone units, or any other known sensor (e.g., buttons, touch screen, etc.) or combination of sensors. The sensors 430 may be external to the endpoint device 300 in some cases, or integrated in the endpoint device 300 in other cases.
Detector module 410 may employ a variety of possible techniques based on available information (e.g., from sensors 430, microphone units 310/320, system bus 250, etc.) to detect a signal degradation event. For illustration, provided below is a list of several examples in which detector module 410 detects a signal degradation event that affects the endpoint device 300:
Upon detecting a signal degradation event, the detector module 410 may use information regarding the signal degradation event to determine whether the signal degradation event affects one or both of the top microphone unit 320 and the bottom microphone unit(s) 310. Furthermore, given the microphone unit(s) determined to be affected, detector module 410 may identify the optimal microphone unit whose received input signals will be used for generating an output signal to be transmitted to another endpoint device in communication with endpoint device 300. In many cases, the optimal microphone unit for generating the output signal may be the microphone unit that is not affected, or least affected, by the detected signal degradation event. Input signals received by the affected microphone unit, by contrast, may be excluded from the output signal so as to ensure a high level of quality in the outgoing signal.
To demonstrate,
Another important consideration in the above scenario is preserving full-duplex communication across near- and far-end participants of the conference. In the event that the AEC of signal processing module 420 is unable to remove a residual echo signal, it is commonly removed by non-linear processing (NLP), which also affects the near-end signal and thus compromises full-duplex communication. A larger distance between the active loudspeaker 330 and the active microphone units reduces the level of the significant echo caused by the direct sound from loudspeaker 330 to the bottom microphone units 310. This reduces the level of the echo the AEC seeks to remove and, in cases where it does not satisfactorily remove it, does not require the NLP to behave as aggressively. This in turn may preserve full-duplex communication. In general, the simplest way by which detector module 410 may select the microphone unit most likely to ensure the best full-duplex performance is to identify the microphone unit furthest away from the active loudspeaker 330 while the loudspeaker 330 is playing audio. The AEC's metrics for how much of the AEC reference, the signal received from far-end, is picked up by a microphone unit may be used in cases where a microphone unit further from the active loudspeaker 330 receives higher levels of sound because of reflections, e.g., from the table surface 630. In other embodiments, microphone choice may also be motivated by comparing echo canceller metrics, such as the amount of non-linear processing (NLP) applied. High non-linear processing attenuation for one or more microphone units suggests poor echo canceller performance and may lead processing module 420 to use the signal, or portions of a signal, from a microphone unit for which the AEC performs better.
With these example scenarios in mind, it should be understood that the conditions during any conference may frequently change. To provide the best signal quality as the conference progresses, the various techniques described herein may be employed and combined as needed based on the current circumstances, considering both local conditions and far-end signal characteristics.
Referring briefly again to
After processing the input signals according to the signal processing instructions provided by detector module 410, the signal processing module 420 may create a final signal mix. To this end, a plurality of virtual microphone signals may be created based on the signals from one or more of the microphone units (e.g., bottom-left microphone unit, bottom-right microphone unit, top microphone unit, bottom-left microphone unit and top microphone unit, bottom-right microphone unit and top microphone unit, and so on), and the detector module 410 may select the best signal among the plurality of virtual microphone signals, according to some embodiments. The signal mix may be provided to the encoder module 440, which then encodes the signal mix for transmission. The encoder module 440 may then send the encoded signal to the network interface (e.g., network interface 210) for transmission to at least one other endpoint device. It should be appreciated that this process may repeat throughout the duration of the conference.
At step 715, as detailed above, a signal degradation event that causes degradation of signals received by the top microphone unit or the bottom microphone unit may be detected. Numerous possible signal degradation events are described hereinabove for illustration. According to some embodiments, the signal degradation event may include a physical obstruction affecting any one or more of the bottom microphone units. According to other embodiments, the signal degradation event may include noise at any of the top and bottom microphone units. In such case, noise levels experienced at each microphone unit may be compared to determine which microphone unit is most and least affected by the noise. In further embodiments, spectral analysis or a machine learning-based model may be applied to the input signals to discern between noise and speech. In yet further embodiments, the signal degradation event may include a comb-filtering effect caused by a reflection of signals received by the top microphone unit. In even further embodiments, the signal degradation event may include audio being emitted from a loudspeaker of the endpoint device that interferes with the input signals received at any of the microphone units.
At step 720, as detailed above, information regarding the signal degradation event may be used to determine whether the signal degradation event affects one or both of the top microphone unit and the bottom microphone unit. Knowledge of the affected microphone unit(s) may inform the subsequent processing of the input signals in order to generate an optimized output signal to be transmitted to at least one other endpoint device. In some embodiments, noise levels at individual microphone units may be compared to a predefined threshold to determine whether noise is sufficiently disruptive to exclude a portion of the input signals from the output signal that is generated.
At step 725, as detailed above, the optimized output signal may be generated in response to determining that the signal degradation event affects one or both of the top microphone unit and the bottom microphone unit. The output signal may be generated using a portion of the input signals that excludes signals received by the top microphone unit and/or the bottom microphone unit determined to be affected by the signal degradation event. In other words, based on the signal degradation event and the specific microphone unit(s) impacted by the signal degradation event, the portion of received input signals used for generating the output signal may include input signals received by the unaffected, or less affected, microphone unit, while excluding input signals received by the more affected microphone unit, thereby preventing the inclusion of distortion, echo, and other harmful artifacts that would hinder the audio quality of the output signal and diminish the user experience of the conference.
It should be noted that while certain steps within procedure 700 may be optional as described above, the steps shown in
The techniques described herein, therefore, allow for high-quality speech pickup by a conferencing endpoint device. Particularly, the described techniques enable broadband audio pickup and smooth frequency response, while avoiding shadowing effects from physical objects. Aspects of the present application also enhance the quality of double-talk (full-duplex communication) between two users, thus facilitating a more natural flow of conversation. Further aspects of the present application attenuate sound sources outside a specific sector, for instance, the camera field of view or the table surface.
While there have been shown and described illustrative embodiments that provide for optimization of a multi-microphone system of an endpoint device, it is to be understood that various other adaptations and modifications may be made within the spirit and scope of the embodiments herein. For example, while certain embodiments are described herein specifically with respect to top and bottom microphone units, other microphone units may also be used as desired. In addition, while a certain endpoint device is shown in the accompanying figures, the design in no way limits the scope of the present application, as other suitable endpoint designs may be used, accordingly.
The foregoing description has been directed to specific embodiments. It will be apparent, however, that other variations and modifications may be made to the described embodiments, with the attainment of some or all of their advantages. For instance, it is expressly contemplated that the components and/or units described herein can be implemented as software being stored on a tangible (non-transitory) computer-readable medium (e.g., disks/CDs/RAM/EEPROM/etc.) having program instructions executing on a computer, hardware, firmware, or a combination thereof. Accordingly, this description is to be taken only by way of example and not to otherwise limit the scope of the embodiments herein. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the embodiments herein.
Number | Name | Date | Kind |
---|---|---|---|
9113243 | Nielsen et al. | Aug 2015 | B2 |
9210499 | Sun et al. | Dec 2015 | B2 |
9226062 | Sun et al. | Dec 2015 | B2 |
9571925 | Adva Fish | Feb 2017 | B1 |
9723401 | Chen et al. | Aug 2017 | B2 |
10051396 | Virolainen et al. | Aug 2018 | B2 |
10389885 | Sun et al. | Aug 2019 | B2 |
10863035 | Robison et al. | Dec 2020 | B2 |
10992905 | Therkelsen et al. | Apr 2021 | B1 |
11011182 | Shanmugam et al. | May 2021 | B2 |
11076251 | Burenius | Jul 2021 | B2 |
20130216050 | Chen | Aug 2013 | A1 |
20150271593 | Sun | Sep 2015 | A1 |
20150312691 | Virolanen et al. | Oct 2015 | A1 |
20160078879 | Lu | Mar 2016 | A1 |
20170272878 | Partio | Sep 2017 | A1 |
20180035222 | Anderson | Feb 2018 | A1 |
20190166257 | Robison et al. | May 2019 | A1 |
20190272842 | Bryan | Sep 2019 | A1 |
20200312342 | Shanmugam et al. | Oct 2020 | A1 |
20200321021 | Sereshki et al. | Oct 2020 | A1 |
Number | Date | Country |
---|---|---|
3793212 | Mar 2021 | EP |
Number | Date | Country | |
---|---|---|---|
20230063260 A1 | Mar 2023 | US |