Multi-sourced noise suppression

Information

  • Patent Grant
  • 9799330
  • Patent Number
    9,799,330
  • Date Filed
    Thursday, August 27, 2015
    9 years ago
  • Date Issued
    Tuesday, October 24, 2017
    7 years ago
Abstract
Systems and methods for multi-sourced noise suppression are provided. An example system may receive streams of audio data including a voice signal and noise, the voice signal including a spoken word. The streams of audio data are provided by distributed audio devices. The system can assign weights to the audio streams based at least partially on quality of the audio streams. The weights of audio streams can be determined based on signal-to-noise ratios (SNRs). The system may further process, based on the weights, the audio stream to generate cleaned speech. Each audio device comprises microphone(s) and can be associated with the Internet of Things (IoT), such that the audio devices are Internet of Things devices. The processing can include noise suppression and reduction and echo cancellation. The cleaned speech can be provided to a remote device for further processing which may include Automatic Speech Recognition (ASR).
Description
FIELD

The present application relates generally to audio processing and, more specifically, to systems and methods for providing multi-sourced noise suppression.


BACKGROUND

Automatic Speech Recognition (ASR) and voice user interfaces (VUI) are widely used to control different type of devices, such as TV sets, game consoles, and the like. Usually, a user utters a voice command to control a device when the user is located in near proximity to the device, for example, in the same room as the device. However, such location may not be convenient if the user needs to provide a voice command for a device located in a different room, a garage, a different house, or another remote location. Moreover, the voice command can be unclear due to a noisy environment in which the device operates. Therefore, the device may not recognize the issued command. Accordingly, more robust systems and methods for delivering spoken commands to a device with a VUI interface may be desired.


SUMMARY

This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.


Systems and methods for multi-sourced audio processing are described. An exemplary method for multi-sourced noise suppression comprises: assigning weights to audio streams, the audio streams being provided substantially synchronously by a plurality of audio devices, the weights depending on quality of the audio streams; processing, based on the weights, the audio streams to generate a cleaned voice signal; and providing the cleaned voice signal to at least one remote device for further processing. In some embodiments, each of the audio devices includes at least one microphone and is associated with the Internet of Things, also referred to herein as Internet of Things devices.


Other example embodiments of the disclosure and aspects will become apparent from the following description taken in conjunction with the following drawings.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments are illustrated by way of example and not limitation in the figures of the accompanying drawings, in which like references indicate similar elements.



FIG. 1 is a block diagram illustrating an exemplary environment in which a method for multi-sourced noise suppression can be practiced.



FIG. 2 is a block diagram illustrating an audio device, according to an exemplary embodiment.



FIG. 3 is a block diagram illustrating a system for multi-sourced noise suppression, according to an exemplary embodiment.



FIG. 4 is a flow chart illustrating a method for multi-sourced noise suppression, according to an exemplary embodiment.



FIG. 5 is a block diagram illustrating an exemplary computing system in which embodiments of the disclosed technology are implemented.





DETAILED DESCRIPTION

The technology disclosed herein is directed to systems and methods for multi-sourced noise suppression, also referred to herein as crowd-based noise suppression. Various embodiments of the present technology may be practiced with a plurality of audio devices configured at least to capture acoustic signals. The audio device can include cellular phones, smartphones, wearables, tablets, phablets, video cameras, phone handsets, headsets, conferencing systems, and other devices having one or more microphones and the functionality to capture sounds. In some embodiments, the audio devices are devices that are connected or part of the Internet of Things (IoT), e.g., a dynamic network of globally connected devices, which may include devices not ordinarily considered audio devices, such as smart thermostats, smart appliances and the like.


In various embodiments, the audio devices further includes radio frequency (RF) receivers, transmitters and transceivers, wired and/or wireless telecommunications and/or networking devices, amplifiers, audio and/or video players, encoders, decoders, speakers, inputs, outputs, storage devices, and user input devices. The audio devices may also include input devices such as buttons, switches, keys, keyboards, trackballs, sliders, touch screens, one or more microphones, gyroscopes, accelerometers, global positioning system (GPS) receivers, and the like. The audio devices may also include outputs, such as LED indicators, video displays, touchscreens, speakers, and the like.


In various embodiments, the audio devices are operated in stationary and portable environments. Stationary environments include residential and commercial buildings or structures, and the like. For example, the stationary embodiments include living rooms, bedrooms, home theaters, conference rooms, auditoriums, business premises, and the like. Portable environments include moving vehicles, moving persons, transportation means, and the like.


The present technology may be used for providing remote commands to a device, such as a device located in a different part of the house, in a vehicle, or in another house. Additionally, the present technology may be used to enable live-talk communications (i.e., real-time communications with a second user located in a different part of the house or even in a different house). In some embodiments, the data is relayed to another device through a local wired or local wireless network (see e.g., network 140) or through a computing cloud 160.



FIG. 1 is a block diagram illustrating an environment 100 in which a method for multi-sourced noise suppression can be practiced, according to an exemplary embodiment. The example environment 100 includes one or more audio devices 110. The audio devices 110 may be located at different places inside a residence and/or office. Each of the audio devices 110 may be configured to receive acoustic signals, process the acoustic signal to generate an audio stream and send the audio stream to a remote device. In some embodiments, each of the audio devices 110 may include at least one microphone for capturing the acoustic sound. In various embodiments the acoustic signal may include a voice from a user 120 contaminated by one or more noise sources 130. Noise sources 130 may include street noise, ambient noise, and speech from entities other than an intended speaker 120. For example, noise sources 130 include working air conditioners, ventilation fans, street noise, TV sets, mobile phones, stereo audio systems, and the like.


In various embodiments, the audio devices 110 are interconnected via a network 140. In some embodiments, the network 140 includes a local network, for example a Wi-Fi network, a Bluetooth network, and the like. In addition or alternatively, the audio devices 110 may be interconnected via wired or mesh network. In some embodiments, the audio devices 110 may include a controller/coordinator 150, also referred to as “controller 150” herein. In certain embodiments, the audio devices 110 is synchronized to a common time source, provided either by an external device or the controller 150. The controller/coordinator 150 may be a router, a chip, one of the audio devices 110 (such as the TV set), and so forth. For example, if the audio devices 110 are interconnected via a wireless network, the router may act as the controller/coordinator 150.


In further embodiments, one or more of the audio devices 110 are connected to a cloud-based computing resource(s) 160, also referred to as “computing cloud 160”, and “cloud-based computing resource services 160” herein. In some embodiments, the cloud-based computing resource includes one or more server farms/clusters including a collection of computer servers which may be co-located with network switches and/or routers. The cloud-based computing resource 160 may include an application that interconnects the audio devices 110 for data exchange between the audio devices 110, and applications for processing data received from the audio devices 110, controller 150, and other services.


In various embodiments, audio devices 110 constantly or periodically listening for voice and buffer audio data. The exemplary audio devices 110 communicate with each other via the network 140. In various embodiments, the audio devices are devices that are connected to or part of the Internet of Things. The exemplary audio devices 100 have one or more microphones for capturing sounds and may be connected to a network, e.g., the Internet. Such exemplary audio devices are also referred to herein as “Internet of Things devices” or “IoT devices”. By way of example and not limitation, first and second audio devices 110 may be located at different distances from the speaker 120, also referred to herein as a the talker or user 120. The audio data captured by the first and second audio devices 110 may be provided to controller/coordinator 150 and treated as data coming from a primary microphone and a secondary microphone. With this information, the controller 150 may perform echo and noise suppression. For example, as the user 120 walks around the house, alternate audio devices 110 and microphones positioned throughout the house may become optimal for picking up speech from the user 120. When the user 120 speaks (for example, providing a voice command to an audio device 110), all listening audio devices 110 and microphones send their time-stamped data to the controller/coordinator 150 for further processing.



FIG. 2 is a block diagram illustrating an exemplary audio device 110 suitable for implementing methods for multi-sourced noise suppression in various embodiments. The example audio device 110 may include a transceiver 210, a processor 220, a microphone 230, an audio processing system 240, and an output device 250. The audio device 110 may include more or other components to provide a particular operation or functionality. Similarly, the audio device 110 may comprise fewer components to perform functions similar or equivalent to those depicted in FIG. 2.


In the example in FIG. 2, the transceiver 210 is configured to communicate with a network such as the Internet, Wide Area Network (WAN), Local Area Network (LAN), cellular network, and so forth, to receive and/or transmit audio data stream. The received audio data stream may be forwarded to the audio processing system 240 and the output device 250.


The processor 220 may include hardware, firmware, and software that implement the processing of audio data and various other operations depending on a type of the audio device 110 (e.g., communications device and computer). A memory (e.g., non-transitory computer readable storage medium) may store, at least in part, instructions and data for execution by processor 220.


The audio processing system 240 may include hardware, firmware, and software that implement the encoding of acoustic signals. For example, the audio processing system 240 is further configured to receive acoustic signals from an acoustic source via microphone 230 (which may be one or more microphones or acoustic sensors) and process the acoustic signals. After reception by the microphone 230, the acoustic signals may be converted into electric signals by an analog-to-digital converter.


An exemplary output device 250 includes any device which can provide an audio output to a listener (e.g., the acoustic source). For example, the exemplary output device 250 comprises a speaker, a class-D output, an earpiece of a headset, or a handset on the audio device 110.



FIG. 3 is a block diagram illustrating a system 300 suitable for implementing a method for multi-sourced noise suppression, according to an exemplary embodiment. The example system 300 may be incorporated in the controller 150 (shown in FIG. 1) and operable to receive audio streams from one or more audio devices 110 (shown in FIG. 1) via network 140 (also shown in FIG. 1). The system 300 may include weighting module 310, noise suppression and reduction module 320, echo cancellation module 330, and ASR module 340. In some embodiments, the modules 310-340 of system 300 are implemented as instructions stored in a memory and executed by a processor of the controller/coordinator 150. In other embodiments, the system 300 may be implemented as hardware, a chip, or firmware incorporated in controller/coordinator 150. The system 300 may operate in an environment using a protocol suitable for communication with and among Internet of Things devices.


In further embodiments, some or all of the modules 310-340 of system 300 may be implemented as instructions stored and executed on a remote server or by cloud-based computing resource services 160 (also shown in FIG. 1). The controller 150 may communicate to the computing cloud 160, via network 140, a command to send audio stream and other data for processing, and may receive the results of computations.


In various embodiments, the controller 150 may be operable to perform diversity pooling. That is, the controller 150 may receive N streams of audio data from N audio devices 110. Each audio stream may include a voice signal and noise. The weighting module 310 may execute an algorithm that assigns a weight to each of the received audio data streams based on the quality of the audio data, determined by a quality metric. In certain embodiments, the weight associated with an audio stream is calculated based on signal-to-noise ratio as a quality metric. The quality of the audio data may depend on a particular environment in which the corresponding audio device 110 operates. In certain embodiments, therefore, the weight assigned to a stream of audio data depends on an audio device's 110 environmental conditions. For example, if a user 120 is watching TV, a microphone located directly above the user 120 may be optimal for picking up the user's speech. However, if the microphone is located near a heating, ventilation, or air condition (HVAC) system, the microphone may not be optimal due to the lowered signal-to-noise ratio when, for example, the air conditioner (AC) is in operation. Thus, the weight assigned to the audio data from the microphone may depend on whether a noise source, such as the AC in this example, is active or not.


In some embodiments, quality of audio data and weight assigned to the audio data may depend on particular characteristics of components of the corresponding audio device 110 (for example, a type of a microphone, a type of an audio processing system, and so forth).


The exemplary system 300 performs distributed noise suppression and reduction to separate noise from audio data and distill cleaned speech using multiple audio stream data and weights assigned to the audio stream data, in some embodiments. For example, in audio devices 110 with multiple microphones, an inter-microphone level difference (ILD) between energies of the primary and secondary acoustic signals may be used for acoustic signal enhancement. Methods and systems for acoustic signal enhancement are described, for example, in U.S. patent application Ser. No. 11/343,524 (patented as U.S. Pat. No. 8,345,890), entitled “System and Method for Utilizing Inter-Microphone Level Differences for Speech Enhancement”, the disclosure of which is incorporated herein by reference for the above-identified purposes.


In addition, in some embodiments, by using multiple audio stream data and weights assigned to the audio stream data, the system 300 may perform various other processing such as echo cancellation and gain control, to name a few. Further details regarding applying weighting to modify acoustic signals is found in commonly assigned U.S. patent application Ser. No. 12/893,208 entitled “Systems and Methods for Producing an Acoustic Field Having a Target Spatial Pattern” (patented as U.S. Pat. No. 8,615,392) and incorporated by reference herein. As the user 120 walks around the house, for example, and as environmental conditions change, the weight assigned to each audio stream from each audio device 110 is dynamically adjusted, and signal processing (gain control, echo cancellation, noise suppression, etc.) is performed to ensure optimal audio quality and speech recognition at all times.


The above described embodiments of the method may operate in the IoT environment. Further details regarding the method for operating in an IoT environment according to various embodiments are now described.


In some embodiments, each of the audio devices 110 includes at least one microphone and is associated with the Internet of Things, also referred to herein as Internet of Things devices or IoT devices.


In some embodiments, the method, and in particular the weighting, includes generating acoustic activity maps by locating, identifying, and mapping target sound(s) (e.g., speech) and noise source(s) in a single or multi-room Internet of Things environment by combining multiple audio streams from microphones on multiple Internet of Things devices (e.g., audio devices 110) to create a multidimensional acoustic view of the environment.


Acoustic signatures may be continually updated between the IoT devices using sound sources in the vicinity of the IoT devices.


Auditory scene analysis and scene classifiers may be used to identify noise and target sound types. Further details regarding exemplary scene analysis and scene classifiers may be found in U.S. patent application Ser. No. 14/335,850 entitled “Speech Signal Separation and Synthesis Based on Auditory Scene Analysis and Speech Modeling” and U.S. patent application Ser. No. 12/860,043 (patented as U.S. Pat. No. 8,447,596) entitled “Monaural Noise Suppression Based on Computational Auditory Scene Analysis”, both of which are incorporated by reference herein. In some embodiments, signaling mechanisms, including transmitters and receivers, between the IoT devices are used to identify locations between the IoT devices relative to each other.


In various embodiments, the method includes, based on the acoustic activity maps, identifying the optimal audio device that provides good signal-to-noise ratio (SNR) for the talker (e.g., user 120) along with identification of the optimal audio devices (among the IoT devices) for measuring noise in the talker's environment and surrounding environment. The identification may be used for assigning weights to the audio stream associated with the audio device. In various embodiments, a combination of audio streams from the audio devices is utilized to enhance audio processing (e.g., noise cancellation, noise suppression, etc.) of the target signal. As a result, various embodiments provide for a seamless, hands-free voice communication experience as the talker (e.g., user 120) moves around in a single room or across different rooms. In a further result, various embodiments provide for a graceful, smooth handoff of whichever IoT device has the optimal SNR along with a graceful, smooth handoff of whichever IoT device has optimal noise measurement.


Further, in some embodiments, the method provides for a fluid human-computer voice interface, which can result in high-performing ASR across the IoT devices in the Internet of Things environment.


In addition, the method in certain embodiments provides for having IoT devices communicate with the user 120 (e.g., using a loudspeaker or other communication functionality of the IoT devices) at the optimal place, at the optimal time, and at the optimal volume. Certain embodiments would thus provide for a seamless handoff between and among the IoT devices that are listening to and communicating with the user 120.


In some embodiments, the resulting cleaned voice signal may be provided to an ASR module 340, for example, to distill a spoken command. In some embodiments, the ASR module 340 may associate a remote device 360 with the spoken command (e.g., a television, streaming device, or the like, depending on the command context) and provide the spoken command to the associated remote device 360 for further processing. In other embodiments, the cleaned voice is used for various voice interfaces and other services.


Example 1. Remote Command

By way of example and not limitation, in some embodiments, a user 120 provides a voice command to one device from the audio device 110 (shown in FIG. 1) and the device may relay the command to a different device at a different location via the network 140. The voice command can be picked up by microphones located on, or connected to, various audio devices 110 around the premises (e.g. a house) and sent to the controller/coordinator 150. The controller/coordinator 150 may include a router or a device, such as a TV. Once the controller/coordinator 150 receives the command, it may request that all other devices send a time-stamped audio command (as well as a small portion of the preceding audio for context). Operations such as weighting audio streams, noise suppression, echo cancellation, gain control, and execution of an ASR algorithm may be performed using the multi-microphone data to clean up the voice command. The data processing can be carried out locally, on the controller 150, or on the computing cloud 160. Thus, as the user 120 walks around the premises and utters voice commands in this example, the commands are picked up, processed, and sent to the ASR module 340.


In some embodiments, the user 120 may send remote commands to devices located in other areas of the premises, for example, a garage area of a house. In other embodiments, the user 120 may send remote commands to a vehicle or receive notifications from the vehicle if someone tries to start the vehicle (for example, if the user's teenage son is trying to take the vehicle for a ride).


In further embodiments, the user 120 may send remote commands to a device located in other premises, such as a second house owned by the user's elderly parents, for example, in which case, the command may be relayed through the computing cloud.


Example 2. Live-Talk Communication

The technology described herein may also allow for real-time communications between two or more users 120 located in different parts of the premises or between users in separate premises, (e.g. different houses).


By way of example and not limitation, user #1 utters a voice command, such as “connect with my dad”, and this command may be picked up by various audio devices 110 located near user #1. In various embodiments, different audio streams containing the command are processed to distill cleaned speech and recognize the command, as described in example 1, above. Once the command is understood by one or more controlling devices in this example, communication between audio devices 110 is established with one or more devices located near user #2 (e.g. dad). User #1 and user #2 talk through the established communications link between audio devices 110 located near each user 120. The speech from user #1 is received by one or more audio devices 110 in the vicinity of user #1, processed to distill cleaned speech, as described herein, and transmitted to one or more audio devices 110 in the vicinity of user #2 (e.g. the user's dad). Speech from user #2 (e.g. user's dad) can similarly be processed and received by user #1.


In some embodiments, if user #2 is located in the same house, the data may be transferred through, for example, a local network, using wireless (e.g. WiFi), or wired (e.g. Ethernet) connections. In other embodiments, if user #2 is located in a different house, the data is sent through a WAN, or other infrastructure including a computing cloud environment. A placement of sufficient networked audio devices 110, using the technology described herein, may enable a user 120 to connect to and speak with another person while the user 120 moves throughout the premises (e.g. house).



FIG. 4 is a flow chart illustrating a method 400 for multi-sourced noise suppression, according to an exemplary embodiment. The example method 400 may commence at operation 402 by assigning weights to audio streams. The audio streams can be provided by distributed audio devices 110. The audio streams may contain voice and noise. In various embodiments, the weights applied to an audio stream are determined based on the quality of the audio stream, using a signal-to-noise ratio, for example. Continued processing at operation 404, based on the weights assigned to the audio streams, can generate cleaned speech. Processing may include gain control, noise suppression, noise reduction, echo cancellation, and the like. At operation 406, the exemplary method includes providing cleaned speech to a remote device, (e.g., remote device 360), for further processing such as ASR.



FIG. 5 illustrates an exemplary computer system 500 that may be used to implement various elements (e.g., audio devices, controller, etc.) of various embodiments of the present technology. The computer system 500 of FIG. 5 may be implemented in the context of computing systems, networks, servers, or combinations thereof. The computer system 500 of FIG. 5 includes one or more processor units 510 and main memory 520. Main memory 520 stores, in part, instructions and data for execution by processor units 510. In various embodiments, main memory 520 stores the executable code when in operation. The computer system 500 of FIG. 5 further includes one or more mass data storage device(s) 530, one or more portable storage device 540, output devices 550, user input devices 560, a graphics display system 570, and peripheral devices 580.


The components shown in FIG. 5 are depicted as being connected via a single bus 590. The components may be connected through one or more data transport means. Processor units 510 and main memory 520 are connected via a local microprocessor bus, and the mass data storage device(s) 530, peripheral device(s) 580, portable storage device 540, and graphics display system 570 are connected via one or more input/output (I/O) buses.


Mass data storage device(s) 530, which can be implemented with a magnetic disk drive, solid state drive, or an optical disk drive, is a non-volatile storage device for storing data and instructions for use by processor units 510. Mass data storage device(s) 530 stores the system software for implementing embodiments of the present disclosure, and all or part of the software may be loaded into main memory 520 during program execution.


Portable storage device 540 operates in conjunction with a portable non-volatile storage medium, such as a flash drive, floppy disk, compact disk, digital video disc, or Universal Serial Bus (USB) storage device, to input and output data and software code to and from the computer system 500 of FIG. 5. System software for implementing embodiments of the present disclosure may be stored on portable medium and input into the computer system 500 via the portable storage device 540.


User input devices 560 can provide a portion of a user interface. User input devices 560 may include one or more microphones, an alphanumeric keypad, such as a keyboard, a pointing device, such as a mouse, a trackball, a trackpad, a stylus, or cursor direction keys, for entering and manipulating alphanumeric and other information User input devices 560 may also include a touchscreen. Additionally, the computer system 500 as shown in FIG. 5 includes output devices 550. Suitable output devices 550 include speakers, printers, network interfaces, and monitors.


Graphics display system 570 includes a liquid crystal display (LCD) or other suitable display device. Graphics display system 570 is configurable to receive textual and graphical information and processes the information for output to the display device.


Peripheral devices 580 may include any type of computer support device to add additional functionality to the computer system 500.


The components provided in the computer system 500 of FIG. 5 are those typically found in computer systems that may be suitable for use with embodiments of the present disclosure and are intended to represent a broad category of such computer components that are well known in the art. Thus, the computer system 500 of FIG. 5 can be a personal computer (PC), hand held computer system, telephone, mobile computer system, workstation, tablet, phablet, mobile phone, server, minicomputer, mainframe computer, wearable, embedded device, or any other computer system. The computer may also include different bus configurations, networked platforms, multi-processor platforms, and the like. Various operating systems may be used including UNIX, LINUX, WINDOWS, MAC OS, PALM OS, QNX ANDROID, IOS, CHROME, TIZEN, and other suitable operating systems.


The processing for various embodiments may be implemented in software that is cloud-based. In some embodiments, the computer system 500 is implemented as a cloud-based computing environment, such as a virtual machine operating within a computing cloud. In other embodiments, the computer system 500 may itself include a cloud-based computing environment, where the functionalities of the computer system 500 are executed in a distributed fashion. Thus, the computer system 500, when configured as a computing cloud, may include pluralities of computing devices in various forms, as will be described in greater detail below.


In general, a cloud-based computing environment is a resource that typically combines the computational power of a large grouping of processors (such as within web servers) and/or that combines the storage capacity of a large grouping of computer memories or storage devices. Systems that provide cloud-based resources may be utilized exclusively by their owners or the systems may be accessible to other users who deploy applications within the computing infrastructure to obtain the benefit of large computational or storage resources.


The cloud may be formed, for example, by a network of web servers that comprise a plurality of computing devices, similar in configuration to the computer system 500, with each server, or at least a plurality thereof, providing processor and/or storage resources. These servers may manage workloads provided by multiple users (e.g., cloud resource customers or other users). Typically, each user places workload demands upon cloud resources that vary in real-time. The nature and extent of these variations may depend, for example, on the type of business served by the resources.


The present technology is described above with reference to example embodiments. The illustrative discussions above are not intended to be exhaustive or to limit embodiments of the disclosed subject matter to the forms disclosed. Modifications and variations are possible in view of the above teachings, to enable others skilled in the art to utilize those embodiments as may be suitable to a particular use.

Claims
  • 1. A method for multi-sourced noise suppression, the method comprising: assigning weights to audio streams, the audio streams being provided substantially synchronously by a plurality of audio devices, the weights depending on quality of the audio streams, wherein the assigning weights includes generating an acoustic activity map by locating, identifying and mapping target sounds and noise sources in at least one of a single room and multi-room environment, so as to create a multidimensional acoustic view of the environment;based on the weights, performing noise suppression processing on the audio streams to generate a cleaned voice signal;providing the cleaned voice signal from the noise suppression processing to at least one remote device for further processing; andbased on the acoustic activity map, selecting an optimal one of the plurality of audio devices to communicate with the user.
  • 2. The method of claim 1, wherein each of the audio devices includes at least one microphone and the audio devices are connected in a dynamic network of connected devices, such that the audio devices are connected as part of an Internet of Things environment.
  • 3. The method of claim 1, wherein the weights are proportional to at least one quality metric for the audio stream, the quality metric comprising at least one signal to noise ratio (SNR).
  • 4. The method of claim 1, wherein the performing of the noise suppression processing is provided in combination with at least one of: performing noise reduction; andperforming echo cancellation.
  • 5. The method of claim 1, further including continually updating acoustic signatures between the audio devices based on one or more sound sources located in the vicinity of the audio devices.
  • 6. The method of claim 1, wherein auditory scene analysis and scene classifiers are used for the identifying of target sounds and noise sources.
  • 7. The method of claim 1, wherein the audio streams include time stamps, the method further comprising, based on the time stamps, synchronizing the audio devices to a common time source.
  • 8. The method of claim 7, further comprising, based on the acoustic activity map, assigning weights to the audio streams based on the SNR quality metric.
  • 9. The method of claim 8, further comprising, based on the acoustic activity map, assigning weights to the audio streams further based on the degree to which the associated audio device, that provides a respective one of the audio streams, measures noise.
  • 10. The method of claim 1, wherein the communication with the user is via a loudspeaker of the optimal audio device.
  • 11. The method of claim 7, wherein the performing of the noise suppression processing is provided in combination with at least one of: performing noise reduction; andperforming echo cancellation.
  • 12. The method of claim 1, wherein the audio streams comprise at least one voice command to perform at least one of activating the remote device and communicating with another user.
  • 13. The method of claim 1, wherein the further processing comprises automatic speech recognition (ASR) processing of the cleaned voice signal.
  • 14. The method of claim 13, wherein, based on the ASR processing, a context of a command to connect to another user is recognized and the cleaned voice signal is communicated to at least one of the audio devices located in proximity to the other user for establishing two way communication therewith.
  • 15. A system for multi-sourced audio processing, the system comprising: a processor; anda memory communicatively coupled with the processor, the memory storing instructions, which, when executed by the processor, perform a method comprising: assigning weights to audio streams, the audio streams being provided substantially synchronously by a plurality of audio devices, the weights depending on quality of the audio streams;based on the weights, performing noise suppression processing on the audio streams to generate a cleaned voice signal, andproviding the cleaned voice signal from the noise suppression processing to a remote device for further processing,wherein each of the audio devices includes at least one microphone and wherein the plurality of audio devices are physically separate from each other but connected in a dynamic network of connected devices, such that the audio devices are connected as part of an Internet of Things environment.
  • 16. The system of claim 15, wherein the assigning weights includes generating an acoustic activity map by locating, identifying and mapping target sounds and noise sources in at least one of a single room and multi-room environment, so as to create a multidimensional acoustic view of the environment.
  • 17. A non-transitory computer-readable storage medium having embodied thereon instructions, which, when executed by at least one processor, perform steps of a method, the method comprising: assigning weights to audio streams, the audio streams being provided substantially synchronously by a plurality of audio devices, the weights depending on quality of the audio streams, wherein the assigning weights includes generating an acoustic activity map by locating, identifying and mapping target sounds and noise sources in at least one of a single room and multi-room environment, so as to create a multidimensional acoustic view of the environment;based on the weights, performing noise suppression processing on the audio streams to generate a cleaned voice signal;providing the cleaned voice signal from the noise suppression processing to at least one remote device for further processing; andbased on the acoustic activity map, selecting an optimal one of the plurality of audio devices to communicate with the user.
  • 18. The method of claim 1, wherein one or more of the plurality of audio devices is incorporated in an Internet of Things device.
  • 19. The non-transitory computer-readable medium of claim 17, wherein one or more of the plurality of audio devices is incorporated in an Internet of Things device.
  • 20. The system of claim 15, further comprising: a controller for receiving the audio streams from the plurality of audio devices via the dynamic network and for performing the noise suppression processing on the received audio streams.
CROSS-REFERENCE TO RELATED APPLICATION

The present application claims the benefit of U.S. Provisional Application No. 62/043,344, filed on Aug. 28, 2014. The subject matter of the aforementioned application is incorporated herein by reference for all purposes.

US Referenced Citations (550)
Number Name Date Kind
3976863 Engel Aug 1976 A
3978287 Fletcher et al. Aug 1976 A
4137510 Iwahara Jan 1979 A
4433604 Ott Feb 1984 A
4516259 Yato et al. May 1985 A
4535473 Sakata Aug 1985 A
4536844 Lyon Aug 1985 A
4581758 Coker et al. Apr 1986 A
4628529 Borth et al. Dec 1986 A
4630304 Borth et al. Dec 1986 A
4649505 Zinser, Jr. et al. Mar 1987 A
4658426 Chabries et al. Apr 1987 A
4674125 Carlson et al. Jun 1987 A
4718104 Anderson Jan 1988 A
4811404 Vilmur et al. Mar 1989 A
4812996 Stubbs Mar 1989 A
4864620 Bialick Sep 1989 A
4920508 Yassaie et al. Apr 1990 A
4991166 Julstrom Feb 1991 A
5011853 Olney Apr 1991 A
5027410 Williamson et al. Jun 1991 A
5054085 Meisel et al. Oct 1991 A
5058419 Nordstrom et al. Oct 1991 A
5099738 Hotz Mar 1992 A
5115404 Lo et al. May 1992 A
5119711 Bell et al. Jun 1992 A
5142961 Paroutaud Sep 1992 A
5150413 Nakatani et al. Sep 1992 A
5175769 Hejna, Jr. et al. Dec 1992 A
5177482 Cideciyan et al. Jan 1993 A
5187776 Yanker Feb 1993 A
5208864 Kaneda May 1993 A
5210366 Sykes, Jr. May 1993 A
5216423 Mukherjee Jun 1993 A
5222251 Roney, IV et al. Jun 1993 A
5224170 Waite, Jr. Jun 1993 A
5230022 Sakata Jul 1993 A
5319736 Hunt Jun 1994 A
5323459 Hirano Jun 1994 A
5341432 Suzuki et al. Aug 1994 A
5381473 Andrea et al. Jan 1995 A
5381512 Holton et al. Jan 1995 A
5400409 Linhard Mar 1995 A
5402493 Goldstein Mar 1995 A
5402496 Soli et al. Mar 1995 A
5406635 Jarvinen Apr 1995 A
5416847 Boze May 1995 A
5471195 Rickman Nov 1995 A
5473759 Slaney et al. Dec 1995 A
5479564 Vogten et al. Dec 1995 A
5502663 Lyon Mar 1996 A
5544250 Urbanski Aug 1996 A
5546458 Iwami Aug 1996 A
5550924 Helf et al. Aug 1996 A
5574824 Slyh et al. Nov 1996 A
5590241 Park et al. Dec 1996 A
5602962 Kellermann Feb 1997 A
5625697 Bowen et al. Apr 1997 A
5633631 Teckman May 1997 A
5675778 Jones Oct 1997 A
5694474 Ngo et al. Dec 1997 A
5706395 Arslan et al. Jan 1998 A
5717829 Takagi Feb 1998 A
5729612 Abel et al. Mar 1998 A
5732189 Johnston et al. Mar 1998 A
5749064 Pawate et al. May 1998 A
5754665 Hosoi May 1998 A
5757937 Itoh et al. May 1998 A
5774837 Yeldener et al. Jun 1998 A
5777658 Kerr et al. Jul 1998 A
5792971 Timis et al. Aug 1998 A
5796819 Romesburg Aug 1998 A
5806025 Vis et al. Sep 1998 A
5809463 Gupta et al. Sep 1998 A
5819215 Dobson et al. Oct 1998 A
5839101 Vahatalo et al. Nov 1998 A
5845243 Smart et al. Dec 1998 A
5887032 Cioffi Mar 1999 A
5917921 Sasaki et al. Jun 1999 A
5920840 Satyamurti et al. Jul 1999 A
5933495 Oh Aug 1999 A
5943429 Handel Aug 1999 A
5978824 Ikeda Nov 1999 A
5983139 Zierhofer Nov 1999 A
5990405 Auten et al. Nov 1999 A
6002776 Bhadkamkar et al. Dec 1999 A
6061456 Andrea et al. May 2000 A
6072881 Linder Jun 2000 A
6084916 Ott Jul 2000 A
6092126 Rossum Jul 2000 A
6097820 Turner Aug 2000 A
6098038 Hermansky et al. Aug 2000 A
6108626 Cellario et al. Aug 2000 A
6122384 Mauro Sep 2000 A
6122610 Isabelle Sep 2000 A
6125175 Goldberg et al. Sep 2000 A
6134524 Peters et al. Oct 2000 A
6137349 Menkhoff et al. Oct 2000 A
6140809 Doi Oct 2000 A
6144937 Ali Nov 2000 A
6173255 Wilson et al. Jan 2001 B1
6188797 Moledina et al. Feb 2001 B1
6205421 Morii Mar 2001 B1
6205422 Gu et al. Mar 2001 B1
6208671 Paulos et al. Mar 2001 B1
6216103 Wu et al. Apr 2001 B1
6222927 Feng et al. Apr 2001 B1
6223090 Brungart Apr 2001 B1
6263307 Arslan et al. Jul 2001 B1
6266633 Higgins et al. Jul 2001 B1
6317501 Matsuo Nov 2001 B1
6321193 Nystrom et al. Nov 2001 B1
6324235 Savell et al. Nov 2001 B1
6327370 Killion et al. Dec 2001 B1
6339706 Tillgren et al. Jan 2002 B1
6339758 Kanazawa et al. Jan 2002 B1
6343267 Kuhn et al. Jan 2002 B1
6355869 Mitton Mar 2002 B1
6363345 Marash et al. Mar 2002 B1
6381469 Wojick Apr 2002 B1
6381570 Li et al. Apr 2002 B2
6389142 Hagen et al. May 2002 B1
6411930 Burges Jun 2002 B1
6424938 Johansson et al. Jul 2002 B1
6430295 Handel et al. Aug 2002 B1
6434417 Lovett Aug 2002 B1
6449586 Hoshuyama Sep 2002 B1
6453284 Paschall Sep 2002 B1
6453289 Ertem et al. Sep 2002 B1
6456209 Savari Sep 2002 B1
6469732 Chang et al. Oct 2002 B1
6477489 Lockwood et al. Nov 2002 B1
6480610 Fang et al. Nov 2002 B1
6487257 Gustafsson et al. Nov 2002 B1
6496795 Malvar Dec 2002 B1
6513004 Rigazio et al. Jan 2003 B1
6516066 Hayashi Feb 2003 B2
6516136 Lee Feb 2003 B1
6526140 Marchok et al. Feb 2003 B1
6529606 Jackson, Jr. II et al. Mar 2003 B1
6531970 McLaughlin et al. Mar 2003 B2
6549630 Bobisuthi Apr 2003 B1
6584203 Elko et al. Jun 2003 B2
6615170 Liu et al. Sep 2003 B1
6647067 Hjelm et al. Nov 2003 B1
6683938 Henderson Jan 2004 B1
6717991 Gustafsson et al. Apr 2004 B1
6718309 Selly Apr 2004 B1
6738482 Jaber May 2004 B1
6745155 Andringa et al. Jun 2004 B1
6760450 Matsuo Jul 2004 B2
6768979 Menendez-Pidal et al. Jul 2004 B1
6778954 Kim et al. Aug 2004 B1
6782363 Lee et al. Aug 2004 B2
6785381 Gartner et al. Aug 2004 B2
6792118 Watts Sep 2004 B2
6795558 Matsuo Sep 2004 B2
6798886 Smith et al. Sep 2004 B1
6804203 Benyassine et al. Oct 2004 B1
6804651 Juric et al. Oct 2004 B2
6810273 Mattila et al. Oct 2004 B1
6859508 Koyama et al. Feb 2005 B1
6882736 Dickel et al. Apr 2005 B2
6915257 Heikkinen et al. Jul 2005 B2
6915264 Baumgarte Jul 2005 B2
6917688 Yu et al. Jul 2005 B2
6934387 Kim Aug 2005 B1
6978159 Feng et al. Dec 2005 B2
6982377 Sakurai et al. Jan 2006 B2
6990196 Zeng et al. Jan 2006 B2
7010134 Jensen Mar 2006 B2
7016507 Brennan Mar 2006 B1
7020605 Gao Mar 2006 B2
RE39080 Johnston Apr 2006 E
7031478 Belt et al. Apr 2006 B2
7035666 Silberfenig et al. Apr 2006 B2
7042934 Zamir May 2006 B2
7050388 Kim et al. May 2006 B2
7054452 Ukita May 2006 B2
7054808 Yoshida May 2006 B2
7058572 Nemer Jun 2006 B1
7065485 Chong-White et al. Jun 2006 B1
7065486 Thyssen Jun 2006 B1
7072834 Zhou Jul 2006 B2
7076315 Watts Jul 2006 B1
7092529 Yu et al. Aug 2006 B2
7092882 Arrowood et al. Aug 2006 B2
7099821 Visser et al. Aug 2006 B2
7110554 Brennan et al. Sep 2006 B2
7127072 Rademacher et al. Oct 2006 B2
7142677 Gonopolskiy et al. Nov 2006 B2
7146013 Saito et al. Dec 2006 B1
7146316 Alves Dec 2006 B2
7155019 Hou Dec 2006 B2
7165026 Acero et al. Jan 2007 B2
7171008 Elko Jan 2007 B2
7171246 Mattila et al. Jan 2007 B2
7174022 Zhang et al. Feb 2007 B1
7190665 Warke et al. Mar 2007 B2
7190775 Rambo Mar 2007 B2
7206418 Yang et al. Apr 2007 B2
7209567 Kozel et al. Apr 2007 B1
7221622 Matsuo et al. May 2007 B2
7225001 Eriksson et al. May 2007 B1
7242762 He et al. Jul 2007 B2
7245767 Moreno et al. Jul 2007 B2
7246058 Burnett Jul 2007 B2
7254242 Ise et al. Aug 2007 B2
7254535 Kushner et al. Aug 2007 B2
7289554 Alloin Oct 2007 B2
7289955 Deng et al. Oct 2007 B2
7327985 Morfitt, III et al. Feb 2008 B2
7330138 Mallinson et al. Feb 2008 B2
7339503 Elenes Mar 2008 B1
7359520 Brennan et al. Apr 2008 B2
7376558 Gemello et al. May 2008 B2
7383179 Alves et al. Jun 2008 B2
7395298 Debes et al. Jul 2008 B2
7412379 Taori et al. Aug 2008 B2
7433907 Nagai et al. Oct 2008 B2
7436333 Forman et al. Oct 2008 B2
7469208 Kincaid Dec 2008 B1
7516067 Seltzer et al. Apr 2009 B2
7555434 Nomura et al. Jun 2009 B2
7561627 Chow et al. Jul 2009 B2
7562140 Clemm et al. Jul 2009 B2
7574352 Quatieri, Jr. Aug 2009 B2
7577084 Tang et al. Aug 2009 B2
7617099 Yang et al. Nov 2009 B2
7617282 Han Nov 2009 B2
7657038 Doclo et al. Feb 2010 B2
7664640 Webber Feb 2010 B2
7725314 Wu et al. May 2010 B2
7764752 Langberg et al. Jul 2010 B2
7777658 Nguyen et al. Aug 2010 B2
7783032 Abutalebi et al. Aug 2010 B2
7783481 Endo et al. Aug 2010 B2
7791508 Wegener Sep 2010 B2
7895036 Hetherington et al. Feb 2011 B2
7912567 Chhatwal et al. Mar 2011 B2
7925502 Droppo et al. Apr 2011 B2
7949522 Hetherington et al. May 2011 B2
7953596 Pinto May 2011 B2
8010355 Rahbar Aug 2011 B2
8032364 Watts Oct 2011 B1
8046219 Zurek et al. Oct 2011 B2
8081878 Zhang et al. Dec 2011 B1
8098812 Fadili et al. Jan 2012 B2
8103011 Mohammad et al. Jan 2012 B2
8107656 Dreβler et al. Jan 2012 B2
8126159 Goose et al. Feb 2012 B2
8140331 Lou Mar 2012 B2
8143620 Malinowski et al. Mar 2012 B1
8150065 Solbach et al. Apr 2012 B2
8155953 Park et al. Apr 2012 B2
8175291 Chan et al. May 2012 B2
8180064 Avendano et al. May 2012 B1
8184818 Ishiguro May 2012 B2
8189429 Chen et al. May 2012 B2
8194880 Avendano Jun 2012 B2
8194882 Every et al. Jun 2012 B2
8204252 Avendano Jun 2012 B1
8204253 Solbach Jun 2012 B1
8223988 Wang et al. Jul 2012 B2
8280731 Yu Oct 2012 B2
8345890 Avendano et al. Jan 2013 B2
8359195 Li Jan 2013 B2
8363850 Amada Jan 2013 B2
8369973 Risbo Feb 2013 B2
8378871 Bapat Feb 2013 B1
8447596 Avendano et al. May 2013 B2
8467891 Huang et al. Jun 2013 B2
8473285 Every et al. Jun 2013 B2
8488805 Santos et al. Jul 2013 B1
8494193 Zhang et al. Jul 2013 B2
8521530 Every et al. Aug 2013 B1
8538035 Every et al. Sep 2013 B2
8606249 Goodwin Dec 2013 B1
8639516 Lindahl et al. Jan 2014 B2
8682006 Laroche et al. Mar 2014 B1
8705759 Wolff et al. Apr 2014 B2
8718290 Murgia et al. May 2014 B2
8737188 Murgia et al. May 2014 B1
8737532 Green et al. May 2014 B2
8744844 Klein Jun 2014 B2
8750526 Santos et al. Jun 2014 B1
8762144 Cho et al. Jun 2014 B2
8774423 Solbach Jul 2014 B1
8781137 Goodwin Jul 2014 B1
8804865 Elenes et al. Aug 2014 B2
8867759 Avendano et al. Oct 2014 B2
8880396 Laroche et al. Nov 2014 B1
8886525 Klein Nov 2014 B2
8949120 Every et al. Feb 2015 B1
8949266 Phillips Feb 2015 B2
8965942 Rossum et al. Feb 2015 B1
9008329 Mandel et al. Apr 2015 B1
9049282 Murgia et al. Jun 2015 B1
9076456 Avendano et al. Jul 2015 B1
9143857 Every et al. Sep 2015 B2
9185487 Solbach et al. Nov 2015 B2
9197974 Clark et al. Nov 2015 B1
9236874 Rossum Jan 2016 B1
9253560 Goldstein Feb 2016 B2
9343056 Goodwin May 2016 B1
9544687 Srinivasan Jan 2017 B2
20010016020 Gustafsson et al. Aug 2001 A1
20010031053 Feng et al. Oct 2001 A1
20010044719 Casey Nov 2001 A1
20010053228 Jones Dec 2001 A1
20020002455 Accardi et al. Jan 2002 A1
20020009203 Erten Jan 2002 A1
20020041693 Matsuo Apr 2002 A1
20020080980 Matsuo Jun 2002 A1
20020106092 Matsuo Aug 2002 A1
20020116187 Erten Aug 2002 A1
20020133334 Coorman et al. Sep 2002 A1
20020138263 Deligne et al. Sep 2002 A1
20020147595 Baumgarte Oct 2002 A1
20020156624 Gigi Oct 2002 A1
20020160751 Sun et al. Oct 2002 A1
20020176589 Buck et al. Nov 2002 A1
20020177995 Walker Nov 2002 A1
20020194159 Kamath et al. Dec 2002 A1
20030014248 Vetter Jan 2003 A1
20030026437 Janse et al. Feb 2003 A1
20030033140 Taori et al. Feb 2003 A1
20030038736 Becker et al. Feb 2003 A1
20030039369 Bullen Feb 2003 A1
20030040908 Yang et al. Feb 2003 A1
20030056220 Thornton et al. Mar 2003 A1
20030061032 Gonopolskiy Mar 2003 A1
20030063759 Brennan et al. Apr 2003 A1
20030072382 Raleigh et al. Apr 2003 A1
20030072460 Gonopolskiy et al. Apr 2003 A1
20030095667 Watts May 2003 A1
20030099345 Gartner et al. May 2003 A1
20030099370 Moore May 2003 A1
20030101048 Liu May 2003 A1
20030103632 Goubran et al. Jun 2003 A1
20030118200 Beaucoup et al. Jun 2003 A1
20030128851 Furuta Jul 2003 A1
20030138116 Jones et al. Jul 2003 A1
20030147538 Elko Aug 2003 A1
20030169891 Ryan et al. Sep 2003 A1
20030177006 Ichikawa et al. Sep 2003 A1
20030191641 Acero et al. Oct 2003 A1
20030228023 Burnett et al. Dec 2003 A1
20040001450 He et al. Jan 2004 A1
20040013276 Ellis et al. Jan 2004 A1
20040015348 McArthur et al. Jan 2004 A1
20040042616 Matsuo Mar 2004 A1
20040047464 Yu et al. Mar 2004 A1
20040078199 Kremer et al. Apr 2004 A1
20040102967 Furuta et al. May 2004 A1
20040125965 Alberth, Jr. et al. Jul 2004 A1
20040131178 Shahaf et al. Jul 2004 A1
20040133421 Burnett et al. Jul 2004 A1
20040148166 Zheng Jul 2004 A1
20040165736 Hetherington et al. Aug 2004 A1
20040185804 Kanamori et al. Sep 2004 A1
20040196989 Friedman et al. Oct 2004 A1
20040263636 Cutler et al. Dec 2004 A1
20050008179 Quinn Jan 2005 A1
20050025263 Wu Feb 2005 A1
20050027520 Mattila et al. Feb 2005 A1
20050049857 Seltzer et al. Mar 2005 A1
20050049864 Kaltenmeier et al. Mar 2005 A1
20050060142 Visser et al. Mar 2005 A1
20050066279 LeBarton et al. Mar 2005 A1
20050069162 Haykin et al. Mar 2005 A1
20050075866 Widrow Apr 2005 A1
20050114123 Lukac et al. May 2005 A1
20050114128 Hetherington et al. May 2005 A1
20050152559 Gierl et al. Jul 2005 A1
20050152563 Amada et al. Jul 2005 A1
20050185813 Sinclair et al. Aug 2005 A1
20050203735 Ichikawa Sep 2005 A1
20050213778 Buck et al. Sep 2005 A1
20050216259 Watts Sep 2005 A1
20050228518 Watts Oct 2005 A1
20050238238 Xu et al. Oct 2005 A1
20050240399 Makinen Oct 2005 A1
20050261894 Balan et al. Nov 2005 A1
20050276423 Aubauer et al. Dec 2005 A1
20050288923 Kok Dec 2005 A1
20060053007 Niemisto Mar 2006 A1
20060058998 Yamamoto et al. Mar 2006 A1
20060072768 Schwartz et al. Apr 2006 A1
20060074646 Alves et al. Apr 2006 A1
20060098809 Nongpiur et al. May 2006 A1
20060120537 Burnett et al. Jun 2006 A1
20060122832 Takiguchi et al. Jun 2006 A1
20060133621 Chen et al. Jun 2006 A1
20060136201 Landron Jun 2006 A1
20060149535 Choi et al. Jul 2006 A1
20060153391 Hooley et al. Jul 2006 A1
20060160581 Beaugeant et al. Jul 2006 A1
20060165202 Thomas et al. Jul 2006 A1
20060184363 McCree et al. Aug 2006 A1
20060206320 Li Sep 2006 A1
20060222184 Buck et al. Oct 2006 A1
20060224382 Taneda Oct 2006 A1
20070021958 Visser et al. Jan 2007 A1
20070027685 Arakawa et al. Feb 2007 A1
20070033020 (Kelleher) Francois et al. Feb 2007 A1
20070033032 Schubert Feb 2007 A1
20070041589 Patel et al. Feb 2007 A1
20070055508 Zhao et al. Mar 2007 A1
20070071206 Gainsboro et al. Mar 2007 A1
20070078649 Hetherington et al. Apr 2007 A1
20070094031 Chen Apr 2007 A1
20070110263 Brox May 2007 A1
20070116300 Chen May 2007 A1
20070127668 Ahya et al. Jun 2007 A1
20070136059 Gadbois Jun 2007 A1
20070150268 Acero et al. Jun 2007 A1
20070154031 Avendano et al. Jul 2007 A1
20070165879 Deng et al. Jul 2007 A1
20070195968 Jaber Aug 2007 A1
20070230712 Belt et al. Oct 2007 A1
20070230913 Ichimura Oct 2007 A1
20070237339 Konchitsky Oct 2007 A1
20070276656 Solbach et al. Nov 2007 A1
20070294263 Punj et al. Dec 2007 A1
20080019548 Avendano Jan 2008 A1
20080033723 Jang et al. Feb 2008 A1
20080059163 Ding et al. Mar 2008 A1
20080071540 Nakano et al. Mar 2008 A1
20080140391 Yen et al. Jun 2008 A1
20080152157 Lin et al. Jun 2008 A1
20080159507 Virolainen et al. Jul 2008 A1
20080160977 Ahmaniemi et al. Jul 2008 A1
20080170703 Zivney Jul 2008 A1
20080192955 Merks Aug 2008 A1
20080201138 Visser et al. Aug 2008 A1
20080228474 Huang et al. Sep 2008 A1
20080228478 Hetherington et al. Sep 2008 A1
20080233934 Diethorn Sep 2008 A1
20080259731 Happonen Oct 2008 A1
20080260175 Elko Oct 2008 A1
20080273476 Cohen et al. Nov 2008 A1
20080298571 Kurtz et al. Dec 2008 A1
20080304677 Abolfathi et al. Dec 2008 A1
20080317259 Zhang et al. Dec 2008 A1
20080317261 Yoshida et al. Dec 2008 A1
20090012783 Klein Jan 2009 A1
20090012786 Zhang et al. Jan 2009 A1
20090034755 Short et al. Feb 2009 A1
20090063142 Sukkar Mar 2009 A1
20090089054 Wang et al. Apr 2009 A1
20090116652 Kirkeby et al. May 2009 A1
20090129610 Kim et al. May 2009 A1
20090141908 Jeong et al. Jun 2009 A1
20090144053 Tamura et al. Jun 2009 A1
20090147942 Culter Jun 2009 A1
20090150149 Culter et al. Jun 2009 A1
20090154717 Hoshuyama Jun 2009 A1
20090164905 Ko Jun 2009 A1
20090177464 Gao et al. Jul 2009 A1
20090220107 Every et al. Sep 2009 A1
20090228272 Herbig et al. Sep 2009 A1
20090240497 Usher et al. Sep 2009 A1
20090245335 Fang Oct 2009 A1
20090245444 Fang Oct 2009 A1
20090253418 Makinen Oct 2009 A1
20090264114 Virolainen et al. Oct 2009 A1
20090271187 Yen et al. Oct 2009 A1
20090292536 Hetherington et al. Nov 2009 A1
20090323925 Sweeney et al. Dec 2009 A1
20090323981 Cutler Dec 2009 A1
20090323982 Solbach et al. Dec 2009 A1
20100017205 Visser et al. Jan 2010 A1
20100027799 Romesburg et al. Feb 2010 A1
20100036659 Haulick et al. Feb 2010 A1
20100082339 Konchitsky et al. Apr 2010 A1
20100092007 Sun Apr 2010 A1
20100094622 Cardillo et al. Apr 2010 A1
20100103776 Chan Apr 2010 A1
20100105447 Sibbald et al. Apr 2010 A1
20100128123 DiPoala May 2010 A1
20100130198 Kannappan et al. May 2010 A1
20100138220 Matsumoto et al. Jun 2010 A1
20100166199 Seydoux Jul 2010 A1
20100177916 Gerkmann et al. Jul 2010 A1
20100215184 Buck et al. Aug 2010 A1
20100278352 Petit et al. Nov 2010 A1
20100282045 Chen et al. Nov 2010 A1
20100290615 Takahashi Nov 2010 A1
20100303298 Marks et al. Dec 2010 A1
20100309774 Astrom Dec 2010 A1
20100315482 Rosenfeld et al. Dec 2010 A1
20110019833 Kuech et al. Jan 2011 A1
20110026734 Hetherington et al. Feb 2011 A1
20110035213 Malenovsky et al. Feb 2011 A1
20110060587 Phillips Mar 2011 A1
20110081026 Ramakrishnan et al. Apr 2011 A1
20110091047 Konchitsky Apr 2011 A1
20110101654 Cech May 2011 A1
20110123019 Gowreesunker et al. May 2011 A1
20110178800 Watts Jul 2011 A1
20110182436 Murgia et al. Jul 2011 A1
20110261150 Goyal et al. Oct 2011 A1
20110286605 Furuta et al. Nov 2011 A1
20110300806 Lindahl et al. Dec 2011 A1
20110305345 Bouchard et al. Dec 2011 A1
20120010881 Avendano et al. Jan 2012 A1
20120027217 Jun et al. Feb 2012 A1
20120027218 Every et al. Feb 2012 A1
20120050582 Seshadri et al. Mar 2012 A1
20120062729 Hart et al. Mar 2012 A1
20120063609 Triki et al. Mar 2012 A1
20120087514 Williams et al. Apr 2012 A1
20120093341 Kim et al. Apr 2012 A1
20120116758 Murgia et al. May 2012 A1
20120121096 Chen et al. May 2012 A1
20120133728 Lee May 2012 A1
20120140917 Nicholson et al. Jun 2012 A1
20120143363 Liu et al. Jun 2012 A1
20120179461 Every et al. Jul 2012 A1
20120179462 Klein Jul 2012 A1
20120182429 Forutanpour et al. Jul 2012 A1
20120197898 Pandey et al. Aug 2012 A1
20120220347 Davidson Aug 2012 A1
20120237037 Ninan et al. Sep 2012 A1
20120249785 Sudo et al. Oct 2012 A1
20120250871 Lu et al. Oct 2012 A1
20130011111 Abraham et al. Jan 2013 A1
20130024190 Fairey Jan 2013 A1
20130034243 Yermeche et al. Feb 2013 A1
20130051543 McDysan et al. Feb 2013 A1
20130096914 Avendano et al. Apr 2013 A1
20130182857 Namba et al. Jul 2013 A1
20130196715 Hansson Aug 2013 A1
20130231925 Avendano et al. Sep 2013 A1
20130251170 Every et al. Sep 2013 A1
20130268280 Del Galdo Oct 2013 A1
20130332156 Tackin et al. Dec 2013 A1
20140039888 Taubman Feb 2014 A1
20140098964 Rosca et al. Apr 2014 A1
20140108020 Sharma Apr 2014 A1
20140112496 Murgia et al. Apr 2014 A1
20140142958 Sharma May 2014 A1
20140241702 Solbach et al. Aug 2014 A1
20140337016 Herbig Nov 2014 A1
20150025881 Carlos et al. Jan 2015 A1
20150030163 Sokolov Jan 2015 A1
20150100311 Kar Apr 2015 A1
20150195647 Srinivasan Jul 2015 A1
20160066089 Klein Mar 2016 A1
Foreign Referenced Citations (77)
Number Date Country
0756437 Jan 1997 EP
1232496 Aug 2002 EP
1474755 Nov 2004 EP
20080428 Jul 2008 FI
20100431 Dec 2010 FI
20125812 Oct 2012 FI
20135038 Apr 2013 FI
124716 Dec 2014 FI
62110349 May 1987 JP
4184400 Jul 1992 JP
5053587 Mar 1993 JP
6269083 Sep 1994 JP
H07248793 Sep 1995 JP
H10-313497 Nov 1998 JP
H11-249693 Sep 1999 JP
2001159899 Jun 2001 JP
2002366200 Dec 2002 JP
2002542689 Dec 2002 JP
2003514473 Apr 2003 JP
2003271191 Sep 2003 JP
2004187283 Jul 2004 JP
2005110127 Apr 2005 JP
2005518118 Jun 2005 JP
2005195955 Jul 2005 JP
2006094522 Apr 2006 JP
2006337415 Dec 2006 JP
2007006525 Jan 2007 JP
2008015443 Jan 2008 JP
2008135933 Jun 2008 JP
2009522942 Jun 2009 JP
2010532879 Oct 2010 JP
2011527025 Oct 2011 JP
5007442 Jun 2012 JP
2013517531 May 2013 JP
2013534651 Sep 2013 JP
5762956 Jun 2015 JP
1020080092404 Oct 2008 KR
1020100041741 Apr 2010 KR
1020110038024 Apr 2011 KR
1020120116442 Oct 2012 KR
101210313 Dec 2012 KR
1020130117750 Oct 2013 KR
101461141 Nov 2014 KR
101610656 Apr 2016 KR
526468 Apr 2003 TW
200305854 Nov 2003 TW
200629240 Aug 2006 TW
I279776 Apr 2007 TW
200910793 Mar 2009 TW
201009817 Mar 2010 TW
201214418 Apr 2012 TW
I463817 Dec 2014 TW
I465121 Dec 2014 TW
201513099 Apr 2015 TW
I488179 Jun 2015 TW
WO0137265 May 2001 WO
WO0141504 Jun 2001 WO
WO0156328 Aug 2001 WO
WO0174118 Oct 2001 WO
WO03043374 May 2003 WO
WO03069499 Aug 2003 WO
WO2006027707 Mar 2006 WO
WO2007001068 Jan 2007 WO
WO2007049644 May 2007 WO
WO2007081916 Jul 2007 WO
WO2008045476 Apr 2008 WO
WO2008101198 Aug 2008 WO
WO2009008998 Jan 2009 WO
WO2010005493 Jan 2010 WO
WO2011091068 Jul 2011 WO
WO2011129725 Oct 2011 WO
WO2012009047 Jan 2012 WO
WO2012097016 Jul 2012 WO
WO2014063099 Apr 2014 WO
WO2014131054 Aug 2014 WO
WO2015010129 Jan 2015 WO
WO2016033364 Mar 2016 WO
Non-Patent Literature Citations (228)
Entry
International Search Report & Written Opinion dated Nov. 27, 2015 in Patent Cooperation Treaty Application No. PCT/US2015/047263, filed Aug. 27, 2015.
Allen, Jont B. “Short Term Spectral Analysis, Synthesis, and Modification by Discrete Fourier Transform”, IEEE Transactions on Acoustics, Speech, and Signal Processing. vol. ASSP-25, No. 3, Jun. 1977. pp. 235-238.
Allen, Jont B. et al., “A Unified Approach to Short-Time Fourier Analysis and Synthesis”, Proceedings of the IEEE. vol. 65, No. 11, Nov. 1977. pp. 1558-1564.
Avendano, Carlos, “Frequency-Domain Source Identification and Manipulation in Stereo Mixes for Enhancement, Suppression and Re-Panning Applications,” 2003 IEEE Workshop on Application of Signal Processing to Audio and Acoustics, Oct. 19-22, pp. 55-58, New Paltz, New York, USA.
Boll, Steven F. “Suppression of Acoustic Noise in Speech using Spectral Subtraction”, IEEE Transactions on Acoustics, Speech and Signal Processing, vol. ASSP-27, No. 2, Apr. 1979, pp. 113-120.
Boll, Steven F. et al., “Suppression of Acoustic Noise in Speech Using Two Microphone Adaptive Noise Cancellation”, IEEE Transactions on Acoustic, Speech, and Signal Processing, vol. ASSP-28, No. 6, Dec. 1980, pp. 752-753.
Boll, Steven F. “Suppression of Acoustic Noise in Speech Using Spectral Subtraction”, Dept. of Computer Science, University of Utah Salt Lake City, Utah, Apr. 1979, pp. 18-19.
Chen, Jingdong et al., “New Insights into the Noise Reduction Wiener Filter”, IEEE Transactions on Audio, Speech, and Language Processing. vol. 14, No. 4, Jul. 2006, pp. 1218-1234.
Cohen, Israel et al., “Microphone Array Post-Filtering for Non-Stationary Noise Suppression”, IEEE International Conference on Acoustics, Speech, and Signal Processing, May 2002, pp. 1-4.
Cohen, Israel, “Multichannel Post-Filtering in Nonstationary Noise Environments”, IEEE Transactions on Signal Processing, vol. 52, No. 5, May 2004, pp. 1149-1160.
Dahl, Mattias et al., “Simultaneous Echo Cancellation and Car Noise Suppression Employing a Microphone Array”, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing, Apr. 21-24, pp. 239-242.
Elko, Gary W., “Chapter 2: Differential Microphone Arrays”, “Audio Signal Processing for Next-Generation Multimedia Communication Systems”, 2004, pp. 12-65, Kluwer Academic Publishers, Norwell, Massachusetts, USA.
“ENT 172.” Instructional Module. Prince George's Community College Department of Engineering Technology. Accessed: Oct. 15, 2011. Subsection: “Polar and Rectangular Notation”. <http://academic.ppgcc.edu/ent/ent172—instr—mod.html>.
Fuchs, Martin et al., “Noise Suppression for Automotive Applications Based on Directional Information”, 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, May 17-21, pp. 237-240.
Fulghum, D. P. et al., “LPC Voice Digitizer with Background Noise Suppression”, 1979 IEEE International Conference on Acoustics, Speech, and Signal Processing, pp. 220-223.
Goubran, R.A. et al., “Acoustic Noise Suppression Using Regressive Adaptive Filtering”, 1990 IEEE 40th Vehicular Technology Conference, May 6-9, pp. 48-53.
Graupe, Daniel et al., “Blind Adaptive Filtering of Speech from Noise of Unknown Spectrum Using a Virtual Feedback Configuration”, IEEE Transactions on Speech and Audio Processing, Mar. 2000, vol. 8, No. 2, pp. 146-158.
Haykin, Simon et al., “Appendix A.2 Complex Numbers.” Signals and Systems. 2nd Ed. 2003. p. 764.
Hermansky, Hynek “Should Recognizers Have Ears?”, In Proc. ESCA Tutorial and Research Workshop on Robust Speech Recognition for Unknown Communication Channels, pp. 1-10, France 1997.
Hohmann, V. “Frequency Analysis and Synthesis Using a Gammatone Filterbank”, ACTA Acustica United with Acustica, 2002, vol. 88, pp. 433-442.
Jeffress, Lloyd A. et al., “A Place Theory of Sound Localization,” Journal of Comparative and Physiological Psychology, 1948, vol. 41, p. 35-39.
Jeong, Hyuk et al., “Implementation of a New Algorithm Using the STFT with Variable Frequency Resolution for the Time-Frequency Auditory Model”, J. Audio Eng. Soc., Apr. 1999, vol. 47, No. 4., pp. 240-251.
Kates, James M. “A Time-Domain Digital Cochlear Model”, IEEE Transactions on Signal Processing, Dec. 1991, vol. 39, No. 12, pp. 2573-2592.
Kato et al., “Noise Suppression with High Speech Quality Based on Weighted Noise Estimation and MMSE STSA” Proc. IWAENC [Online] 2001, pp. 183-186.
Lazzaro, John et al., “A Silicon Model of Auditory Localization,” Neural Computation Spring 1989, vol. 1, pp. 47-57, Massachusetts Institute of Technology.
Lippmann, Richard P. “Speech Recognition by Machines and Humans”, Speech Communication, Jul. 1997, vol. 22, No. 1, pp. 1-15.
Liu, Chen et al., “A Two-Microphone Dual Delay-Line Approach for Extraction of a Speech Sound in the Presence of Multiple Interferers”, Journal of the Acoustical Society of America, vol. 110, No. 6, Dec. 2001, pp. 3218-3231.
Martin, Rainer et al., “Combined Acoustic Echo Cancellation, Dereverberation and Noise Reduction: A two Microphone Approach”, Annales des Telecommunications/Annals of Telecommunications. vol. 49, No. 7-8, Jul.-Aug. 1994, pp. 429-438.
Martin, Rainer “Spectral Subtraction Based on Minimum Statistics”, in Proceedings Europe. Signal Processing Conf., 1994, pp. 1182-1185.
Mitra, Sanjit K. Digital Signal Processing: a Computer-based Approach. 2nd Ed. 2001. pp. 131-133.
Mizumachi, Mitsunori et al., “Noise Reduction by Paired-Microphones Using Spectral Subtraction”, 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, May 12-15. pp. 1001-1004.
Moonen, Marc et al., “Multi-Microphone Signal Enhancement Techniques for Noise Suppression and Dereverbration,” http://www.esat.kuleuven.ac.be/sista/yearreport97//node37.html, accessed on Apr. 21, 1998.
Watts, Lloyd Narrative of Prior Disclosure of Audio Display on Feb. 15, 2000 and May 31, 2000.
Cosi, Piero et al., (1996), “Lyon's Auditory Model Inversion: a Tool for Sound Separation and Speech Enhancement,” Proceedings of ESCA Workshop on ‘The Auditory Basis of Speech Perception,’ Keele University, Keele (UK), Jul. 15-19, 1996, pp. 194-197.
Parra, Lucas et al., “Convolutive Blind Separation of Non-Stationary Sources”, IEEE Transactions on Speech and Audio Processing. vol. 8, No. 3, May 2008, pp. 320-327.
Rabiner, Lawrence R. et al., “Digital Processing of Speech Signals”, (Prentice-Hall Series in Signal Processing). Upper Saddle River, NJ: Prentice Hall, 1978.
Weiss, Ron et al., “Estimating Single-Channel Source Separation Masks: Revelance Vector Machine Classifiers vs. Pitch-Based Masking”, Workshop on Statistical and Perceptual Audio Processing, 2006.
Schimmel, Steven et al., “Coherent Envelope Detection for Modulation Filtering of Speech,” 2005 IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1, No. 7, pp. 221-224.
Slaney, Malcom, “Lyon's Cochlear Model”, Advanced Technology Group, Apple Technical Report #13, Apple computer, Inc., 1988, pp. 1-79.
Slaney, Malcom, et al., “Auditory Model Inversion for Sound Separation,” 1994 IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 19-22, vol. 2, pp. 77-80.
Slaney, Malcom. “An Introduction to Auditory Model Inversion”, Interval Technical Report IRC 1994-014, http://coweb.ecn.purdue.edu/˜maclom/interval/1994-014/, Sep. 1994, accessed on Jul. 6, 2010.
Solbach, Ludger “An Architecture for Robust Partial Tracking and Onset Localization in Single Channel Audio Signal Mixes”, Technical University Hamburg-Harburg, 1998.
Soon et al., “Low Distortion Speech Enhancement” Proc. Inst. Elect. Eng. [Online] 2000, vol. 147, pp. 247-253.
Stahl, V. et al., “Quantile Based Noise Estimation for Spectral Subtraction and Wiener Filtering,” 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing, Jun. 5-9, vol. 3, pp. 1875-1878.
Syntrillium Software Corporation, “Cool Edit User's Manual”, 1996, pp. 1-74.
Tashev, Ivan et al., “Microphone Array for Headset with Spatial Noise Suppressor”, http://research.microsoft.com/users/ivantash/Documents/Tashev—MAforHeadset—HSCMA—05.pdf. (4 pages).
Tchorz, Jurgen et al., “SNR Estimation Based on Amplitude Modulation Analysis with Applications to Noise Suppression”, IEEE Transactions on Speech and Audio Processing, vol. 11, No. 3, May 2003, pp. 184-192.
Valin, Jean-Marc et al., “Enhanced Robot Audition Based on Microphone Array Source Separation with Post-Filter”, Proceedings of 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems, Sep. 28-Oct. 2, 2004, Sendai, Japan. pp. 2123-2128.
Watts, Lloyd, “Robust Hearing Systems for Intelligent Machines,” Applied Neurosystems Corporation, 2001, pp. 1-5.
Widrow, B. et al., “Adaptive Antenna Systems,” Proceedings of the IEEE, vol. 55, No. 12, pp. 2143-2159, Dec. 1967.
Yoo, Heejong et al., “Continuous-Time Audio Noise Suppression and Real-Time Implementation”, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, May 13-17, pp. IV3980-IV3983.
International Search Report dated Jun. 8, 2001 in Patent Cooperation Treaty Application No. PCT/US2001/008372.
International Search Report dated Apr. 3, 2003 in Patent Cooperation Treaty Application No. PCT/US2002/036946.
International Search Report dated May 29, 2003 in Patent Cooperation Treaty Application No. PCT/US2003/004124.
International Search Report and Written Opinion dated Oct. 19, 2007 in Patent Cooperation Treaty Application No. PCT/US2007/000463.
International Search Report and Written Opinion dated Apr. 9, 2008 in Patent Cooperation Treaty Application No. PCT/US2007/021654.
International Search Report and Written Opinion dated Sep. 16, 2008 in Patent Cooperation Treaty Application No. PCT/US2007/012628.
International Search Report and Written Opinion dated Oct. 1, 2008 in Patent Cooperation Treaty Application No. PCT/US2008/008249.
International Search Report and Written Opinion dated Aug. 27, 2009 in Patent Cooperation Treaty Application No. PCT/US2009/003813.
Dahl, Mattias et al., “Acoustic Echo and Noise Cancelling Using Microphone Arrays”, International Symposium on Signal Processing and its Applications, ISSPA, Gold coast, Australia, Aug. 25-30, 1996, pp. 379-382.
Demol, M. et al., “Efficient Non-Uniform Time-Scaling of Speech With WSOLA for CALL Applications”, Proceedings of InSTIL/ICALL2004—NLP and Speech Technologies in Advanced Language Learning Systems—Venice Jun. 17-19, 2004.
Laroche, Jean. “Time and Pitch Scale Modification of Audio Signals”, in “Applications of Digital Signal Processing to Audio and Acoustics”, The Kluwer International Series in Engineering and Computer Science, vol. 437, pp. 279-309, 2002.
Moulines, Eric et al., “Non-Parametric Techniques for Pitch-Scale and Time-Scale Modification of Speech”, Speech Communication, vol. 16, pp. 175-205, 1995.
Verhelst, Werner, “Overlap-Add Methods for Time-Scaling of Speech”, Speech Communication vol. 30, pp. 207-221, 2000.
Bach et al., Learning Spectral Clustering with application to spech separation, Journal of machine learning research, 2006.
Mokbel et al., 1995, IEEE Transactions of Speech and Audio Processing, vol. 3, No. 5, Sep. 1995, pp. 346-356.
Office Action mailed Oct. 14, 2013 in Taiwanese Patent Application 097125481, filed Jul. 4, 2008.
Office Action mailed Oct. 29, 2013 in Japanese Patent Application 2011-516313, filed Jun. 26, 2009.
Office Action mailed Dec. 20, 2013 in Taiwanese Patent Application 096146144, filed Dec. 4, 2007.
Office Action mailed Dec. 9, 2013 in Finnish Patent Application 20100431, filed Jun. 26, 2009.
Office Action mailed Jan. 20, 2014 in Finnish Patent Application 20100001, filed Jul. 3, 2008.
Office Action mailed Mar. 10, 2014 in Taiwanese Patent Application 097125481, filed Jul. 4, 2008.
Bai et al., “Upmixing and Downmixing Two-channel Stereo Audio for Consumer Electronics”. IEEE Transactions on Consumer Electronics [Online] 2007, vol. 53, Issue 3, pp. 1011-1019.
Jo et al., “Crosstalk cancellation for spatial sound reproduction in portable devices with stereo loudspeakers”. Communications in Computer and Information Science [Online] 2011, vol. 266, pp. 114-123.
Nongpuir et al., “NEXT cancellation system with improved convergence rate and tracking performance”. IEEE Proceedings—Communications [Online] 2005, vol. 152, Issue 3, pp. 378-384.
Ahmed et al., “Blind Crosstalk Cancellation for DMT Systems” IEEE—Emergent Technologies Technical Committee. Sep. 2001. pp. 1-5.
Allowance mailed May 21, 2014 in Finnish Patent Application 20100001, filed Jan. 4, 2010.
Office Action mailed May 2, 2014 in Taiwanese Patent Application 098121933, filed Jun. 29, 2009.
Office Action mailed Apr. 15, 2014 in Japanese Patent Application 2010-514871, filed Jul. 3, 2008.
Elhilali et al.,“A cocktail party with a cortical twist: How cortical mechanisms contribute to sound segregation.” J Acoust Soc Am. Dec. 2008; 124(6): 3751-3771).
Jin et al., “HMM-Based Multipitch Tracking for Noisy and Reverberant Speech.”
Kawahara, W, et al., “Tandem-Straight: A temporally stable power spectral representation for periodic signals and applications to interference-free spectrum, F0, and aperiodicity estimation.” IEEE ICASSP 2008.
Office Action mailed Jun. 27, 2014 in Korean Patent Application No. 10-2010-7000194, filed Jan. 6, 2010.
Office Action mailed Jun. 18, 2014 in Finnish Patent Application No. 20080428, filed Jul. 4, 2008.
International Search Report & Written Opinion dated Jul. 15, 2014 in Patent Cooperation Treaty Application No. PCT/US2014/018443, filed Feb. 25, 2014.
Notice of Allowance dated Aug. 26, 2014 in Taiwanese Application No. 096146144, filed Dec. 4, 2007.
Notice of Allowance dated Sep. 16, 2014 in Korean Application No. 10-2010-7000194, filed Jul. 3, 2008.
Notice of Allowance dated Sep. 29, 2014 in Taiwanese Application No. 097125481, filed Jul. 4, 2008.
Notice of Allowance dated Oct. 10, 2014 in Finnish Application No. 20100001, filed Jul. 3, 2008.
International Search Report & Written Opinion dated Nov. 12, 2014 in Patent Cooperation Treaty Application No. PCT/US2014/047458, filed Jul. 21, 2014.
Office Action mailed Oct. 28, 2014 in Japanese Patent Application No. 2011-516313, filed Dec. 27, 2012.
Heiko Pumhagen, “Low Complexity Parametric Stereo Coding in MPEG-4,” Proc. of the 7th Int. Conference on Digital Audio Effects (DAFx'04), Naples, Italy, Oct. 5-8, 2004.
Chun-Ming Chang et al., “Voltage-Mode Multifunction Filter with Single Input and Three Outputs Using Two Compound Current Conveyors” IEEE Transactions on Circuits and Systems-I: Fundamental Theory and Applications, vol. 46, No. 11, Nov. 1999.
Notice of Allowance mailed Feb. 10, 2015 in Taiwanese Patent Application No. 098121933, filed Jun. 29, 2009.
Office Action mailed Jan. 30, 2015 in Finnish Patent Application No. 20080623, filed May 24, 2007.
Office Action mailed Mar. 24, 2015 in Japanese Patent Application No. 2011-516313, filed Jun. 26, 2009.
Office Action mailed Apr. 16, 2015 in Korean Patent Application No. 10-2011-7000440, filed Jun. 26, 2009.
Notice of Allowance mailed Jun. 2, 2015 in Japanese Patent Application 2011-516313, filed Jun. 26, 2009.
Office Action mailed Jun. 4, 2015 in Finnish Patent Application 20080428, filed Jan. 5, 2007.
Office Action mailed Jun. 9, 2015 in Japanese Patent Application 2014-165477 filed Jul. 3, 2008.
Notice of Allowance mailed Aug. 13, 2015 in Finnish Patent Application 20080623, filed May 24, 2007.
Non-Final Office Action, Oct. 27, 2003, U.S. Appl. No. 09/534,682, filed Mar. 24, 2000.
Non-Final Office Action, Feb. 10, 2004, U.S. Appl. No. 09/534,682, filed Mar. 24, 2000.
Final Office Action, Dec. 17, 2004, U.S. Appl. No. 09/534,682, filed Mar. 24, 2000.
Non-Final Office Action, Apr. 20, 2005, U.S. Appl. No. 09/534,682, filed Mar. 24, 2000.
Notice of Allowance, Oct. 26, 2005, U.S. Appl. No. 09/534,682, filed Mar. 24, 2000.
Non-Final Office Action, May 3, 2005, U.S. Appl. No. 09/993,442, filed Nov. 13, 2001.
Final Office Action, Oct. 19, 2005, U.S. Appl. No. 09/993,442, filed Nov. 13, 2001.
Advisory Action, Jan. 20, 2006, U.S. Appl. No. 09/993,442, filed Nov. 13, 2001.
Non-Final Office Action, May 17, 2006, U.S. Appl. No. 09/993,442, filed Nov. 13, 2001.
Non-Final Office Action, Nov. 16, 2006, U.S. Appl. No. 09/993,442, filed Nov. 13, 2001.
Final Office Action, Jun. 15, 2007, U.S. Appl. No. 09/993,442, filed Nov. 13, 2001.
Non-Final Office Action, Oct. 8, 2003, U.S. Appl. No. 10/004,141, filed Nov. 14, 2001.
Notice of Allowance, Feb. 24, 2004, U.S. Appl. No. 10/004,141, filed Nov. 14, 2001.
Non-Final Office Action, May 9, 2003, U.S. Appl. No. 10/074,991, filed Feb. 13, 2002.
Notice of Allowance, Jun. 4, 2003, U.S. Appl. No. 10/074,991, filed Feb. 13, 2002.
Non-Final Office Action, Jun. 26, 2006, U.S. Appl. No. 10/074,991, filed Feb. 13, 2002.
Final Office Action, Feb. 23, 2007, U.S. Appl. No. 10/074,991, filed Feb. 13, 2002.
Non-Final Office Action, Oct. 6, 2005, U.S. Appl. No. 10/177,049, filed Jun. 21, 2002.
Final Office Action, Mar. 28, 2006, U.S. Appl. No. 10/177,049, filed Jun. 21, 2002.
Advisory Action, Jun. 19, 2006, U.S. Appl. No. 10/177,049, filed Jun. 21, 2002.
Non-Final Office Action, Dec. 13, 2006, U.S. Appl. No. 10/613,224, filed Jul. 3, 2003.
Non-Final Office Action, Jun. 13, 2007, U.S. Appl. No. 10/613,224, filed Jul. 3, 2003.
Non-Final Office Action, Jun. 13, 2006, U.S. Appl. No. 10/840,201, filed May 5, 2004.
Non-Final Office Action, Mar. 30, 2010, U.S. Appl. No. 11/343,524, filed Jan. 30, 2006.
Non-Final Office Action, Sep. 13, 2010, U.S. Appl. No. 11/343,524, filed Jan. 30, 2006.
Final Office Action, Mar. 30, 2011, U.S. Appl. No. 11/343,524, filed Jan. 30, 2006.
Final Office Action, May 21, 2012, U.S. Appl. No. 11/343,524, filed Jan. 30, 2006.
Notice of Allowance, Oct. 9, 2012, U.S. Appl. No. 11/343,524, filed Jan. 30, 2006.
Non-Final Office Action, Aug. 5, 2008, U.S. Appl. No. 11/441,675, filed May 25, 2006.
Non-Final Office Action, Jan. 21, 2009, U.S. Appl. No. 11/441,675, filed May 25, 2006.
Final Office Action, Sep. 3, 2009, U.S. Appl. No. 11/441,675, filed May 25, 2006.
Non-Final Office Action, May 10, 2011, U.S. Appl. No. 11/441,675, filed May 25, 2006.
Final Office Action, Oct. 24, 2011, U.S. Appl. No. 11/441,675, filed May 25, 2006.
Notice of Allowance, Feb. 13, 2012, U.S. Appl. No. 11/441,675, filed May 25, 2006.
Non-Final Office Action, Apr. 7, 2011, U.S. Appl. No. 11/699,732, filed Jan. 29, 2007.
Final Office Action, Dec. 6, 2011, U.S. Appl. No. 11/699,732, filed Jan. 29, 2007.
Advisory Action, Feb. 14, 2012, U.S. Appl. No. 11/699,732, filed Jan. 29, 2007.
Notice of Allowance, Mar. 15, 2012, U.S. Appl. No. 11/699,732, filed Jan. 29, 2007.
Non-Final Office Action, Aug. 18, 2010, U.S. Appl. No. 11/825,563, filed Jul. 6, 2007.
Final Office Action, Apr. 28, 2011, U.S. Appl. No. 11/825,563, filed Jul. 6, 2007.
Non-Final Office Action, Apr. 24, 2013, U.S. Appl. No. 11/825,563, filed Jul. 6, 2007.
Final Office Action, Dec. 30, 2013, U.S. Appl. No. 11/825,563, filed Jul. 6, 2007.
Notice of Allowance, Mar. 25, 2014, U.S. Appl. No. 11/825,563, filed Jul. 6, 2007.
Non-Final Office Action, Oct. 3, 2011, U.S. Appl. No. 12/004,788, filed Dec. 21, 2007.
Notice of Allowance, Feb. 23, 2012, U.S. Appl. No. 12/004,788, filed Dec. 21, 2007.
Non-Final Office Action, Sep. 14, 2011, U.S. Appl. No. 12/004,897, filed Dec. 21, 2007.
Notice of Allowance, Jan. 27, 2012, U.S. Appl. No. 12/004,897, filed Dec. 21, 2007.
Non-Final Office Action, Jul. 28, 2011, U.S. Appl. No. 12/072,931, filed Feb. 29, 2008.
Notice of Allowance, Mar. 1, 2012, U.S. Appl. No. 12/072,931, filed Feb. 29, 2008.
Notice of Allowance, Mar. 1, 2012, U.S. Appl. No. 12/080,115, filed Mar. 31, 2008.
Non-Final Office Action, Nov. 14, 2011, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Final Office Action, Apr. 24, 2012, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Advisory Action, Jul. 3, 2012, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Non-Final Office Action, Mar. 11, 2014, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Final Office Action, Jul. 11, 2014, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Non-Final Office Action, Dec. 8, 2014, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Notice of Allowance, Jul. 7, 2015, U.S. Appl. No. 12/215,980, filed Jun. 30, 2008.
Non-Final Office Action, Jul. 13, 2011, U.S. Appl. No. 12/217,076, filed Jun. 30, 2008.
Final Office Action, Nov. 16, 2011, U.S. Appl. No. 12/217,076, filed Jun. 30, 2008.
Non-Final Office Action, Mar. 14, 2012, U.S. Appl. No. 12/217,076, filed Jun. 30, 2008.
Final Office Action, Sep. 19, 2012, U.S. Appl. No. 12/217,076, filed Jun. 30, 2008.
Notice of Allowance, Apr. 15, 2013, U.S. Appl. No. 12/217,076, filed Jun. 30, 2008.
Non-Final Office Action, Sep. 1, 2011, U.S. Appl. No. 12/286,909, filed Oct. 2, 2008.
Notice of Allowance, Feb. 28, 2012, U.S. Appl. No. 12/286,909, filed Oct. 2, 2008.
Non-Final Office Action, Nov. 15, 2011, U.S. Appl. No. 12/286,995, filed Oct. 2, 2008.
Final Office Action, Apr. 10, 2012, U.S. Appl. No. 12/286,995, filed Oct. 2, 2008.
Notice of Allowance, Mar. 13, 2014, U.S. Appl. No. 12/286,995, filed Oct. 2, 2008.
Non-Final Office Action, Dec. 28, 2011, U.S. Appl. No. 12/288,228, filed Oct. 16, 2008.
Non-Final Office Action, Dec. 30, 2011, U.S. Appl. No. 12/422,917, filed Apr. 13, 2009.
Final Office Action, May 14, 2012, U.S. Appl. No. 12/422,917, filed Apr. 13, 2009.
Advisory Action, Jul. 27, 2012, U.S. Appl. No. 12/422,917, filed Apr. 13, 2009.
Notice of Allowance, Sep. 11, 2014, U.S. Appl. No. 12/422,917, filed Apr. 13, 2009.
Non-Final Office Action, Jun. 20, 2012, U.S. Appl. No. 12/649,121, filed Dec. 29, 2009.
Final Office Action, Nov. 28, 2012, U.S. Appl. No. 12/649,121, filed Dec. 29, 2009.
Advisory Action, Feb. 19, 2013, U.S. Appl. No. 12/649,121, filed Dec. 29, 2009.
Notice of Allowance, Mar. 19, 2013, U.S. Appl. No. 12/649,121, filed Dec. 29, 2009.
Non-Final Office Action, Feb. 19, 2013, U.S. Appl. No. 12/944,659, filed Nov. 11, 2010.
Notice of Allowance, May 25, 2011, U.S. Appl. No. 13/016,916, filed Jan. 28, 2011.
Notice of Allowance, Aug. 4, 2011, U.S. Appl. No. 13/016,916, filed Jan. 28, 2011.
Non-Final Office Action, Nov. 22, 2013, U.S. Appl. No. 13/363,362, filed Jan. 31, 2012.
Final Office Action, Sep. 12, 2014, U.S. Appl. No. 13/363,362, filed Jan. 31, 2012.
Non-Final Office Action, Oct. 28, 2015, U.S. Appl. No. 13/363,362, filed Jan. 31, 2012.
Non-Final Office Action, Dec. 4, 2013, U.S. Appl. No. 13/396,568, filed Feb. 14, 2012.
Final Office Action, Sep. 23, 2014, U.S. Appl. No. 13/396,568, filed Feb. 14, 2012.
Non-Final Office Action, Nov. 5, 2015, U.S. Appl. No. 13/396,568, filed Feb. 14, 2012.
Non-Final Office Action, Sep. 17, 2013, U.S. Appl. No. 13/397,597, filed Feb. 15, 2012.
Final Office Action, Apr. 1, 2014, U.S. Appl. No. 13/397,597, filed Feb. 15, 2012.
Non-Final Office Action, Nov. 21, 2014, U.S. Appl. No. 13/397,597, filed Feb. 15, 2012.
Non-Final Office Action, Jun. 7, 2012, U.S. Appl. No. 13/426,436, filed Mar. 21, 2012.
Final Office Action, Dec. 31, 2012, U.S. Appl. No. 13/426,436, filed Mar. 21, 2012.
Non-Final Office Action, Sep. 12, 2013, U.S. Appl. No. 13/426,436, filed Mar. 21, 2012.
Notice of Allowance, Jul. 16, 2014, U.S. Appl. No. 13/426/,436, filed Mar. 21, 2012.
Non-Final Office Action, Jul. 15, 2014, U.S. Appl. No. 13/432,490, filed Mar. 28, 2012.
Notice of Allowance, Apr. 3, 2015, U.S. Appl. No. 13/432,490, filed Mar. 28, 2012.
Notice of Allowance, Oct. 17, 2012, U.S. Appl. No. 13/565,751, filed Aug. 2, 2012.
Non-Final Office Action, Jan. 9, 2012, U.S. Appl. No. 13/664,299, filed Oct. 30, 2012.
Non-Final Office Action, Dec. 28, 2012, U.S. Appl. No. 13/664,299, filed Oct. 30, 2012.
Non-Final Office Action, Mar. 7, 2013, U.S. Appl. No. 13/664,299, filed Oct. 30, 2012.
Final Office Action, Apr. 29, 2013, U.S. Appl. No. 13/664,299, filed Oct. 30, 2012.
Non-Final Office Action, Nov. 27, 2013, U.S. Appl. No. 13/664,299, filed Oct. 30, 2012.
Notice of Allowance, Jan. 30, 2014, U.S. Appl. No. 13/664,299, filed Oct. 30, 2012.
Non-Final Office Action, Jun. 4, 2013, U.S. Appl. No. 13/705,132, filed Dec. 4, 2012.
Final Office Action, Dec. 19, 2013, U.S. Appl. No. 13/705,132, filed Dec. 4, 2012.
Notice of Allowance, Jun. 19, 2014, U.S. Appl. No. 13/705,132, filed Dec. 4, 2012.
Non-Final Office Action, Jul. 14, 2015, U.S. Appl. No. 14/046,551, filed Oct. 4, 2013.
Non-Final Office Action, May 21, 2015, U.S. Appl. No. 14/189,817, filed Feb. 25, 2014.
Final Office Action, Dec. 15, 2015, U.S. Appl. No. 14/189,817, filed Feb. 25, 2014.
Notice of Allowance, Oct. 7, 2014, U.S. Appl. No. 14/207,096, filed Mar. 12, 2014.
Non-Final Office Action, Oct. 28, 2015, U.S. Appl. No. 14/216,567, filed Mar. 17, 2014.
Non-Final Office Action, Jul. 10, 2014, U.S. Appl. No. 14/279,092, filed May 15, 2014.
Notice of Allowance, Jan. 29, 2015, U.S. Appl. No. 14/279,092, filed May 15, 2014.
Non-Final Office Action, Feb. 27, 2015, U.S. Appl. No. 14/336,934, filed Jul. 21, 2014.
Notice of Allowance, Aug. 28, 2015, U.S. Appl. No. 14/336,934, filed Jul. 21, 2014.
International Search Report and Written Opinion dated Sep. 1, 2011 in Patent Cooperation Treaty Application No. PCT/US11/37250.
Fazel et al., “An overview of statistical pattern recognition techniques for speaker verification,” IEEE, May 2011.
Sundaram et al., “Discriminating Two Types of Noise Sources Using Cortical Representation and Dimension Reduction Technique,” IEEE, 2007.
Tognieri et al., “A Comparison of the LBG, LVQ, MLP, SOM and GMM Algorithms for Vector Quantisation and Clustering Analysis,” University of Western Australia, 1992.
Klautau et al., “Discriminative Gaussian Mixture Models a Comparison with Kernel Classifiers,” ICML, 2003.
International Search Report & Written Opinion dated Mar. 18, 2014 in Patent Cooperation Treaty Application No. PCT/US2013/065752, filed Oct. 18, 2013.
Kim et al., “Improving Speech Intelligibility in Noise Using Environment-Optimized Algorithms,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 18, No. 8, Nov. 2010, pp. 2080-2090.
Sharma et al., “Rotational Linear Discriminant Analysis Technique for Dimensionality Reduction,” IEEE Transactions on Knowledge and Data Engineering, vol. 20, No. 10, Oct. 2008, pp. 1336-1347.
Temko et al., “Classifiation of Acoustic Events Using SVM-Based Clustering Schemes,” Pattern Recognition 39, No. 4, 2006, pp. 682-694.
Office Action mailed Jun. 17, 2015 in Japan Patent Application 2013-519682 filed May 19, 2011.
Notice of Allowance dated Feb. 24, 2016 in Korean Application No. 10-2011-7000440, filed Jun. 26, 2009.
Hu et al., “Robust Speaker's Location Detection in a Vehicle Environment Using GMM Models,” IEEE Transactions on Systems, Man, and Cybernetics—Part B: Cybernetics, vol. 36, No. 2, Apr. 2006, pp. 403-412.
Laroche, Jean et al., “Noise Suppression Assisted Automatic Speech Recognition”, U.S. Appl. No. 12/962,519, filed Dec. 7, 2010.
Goodwin, Michael M. et al., “Key Click Suppression”, U.S. Appl. No. 14/745,176, filed Jun. 19, 2015.
Related Publications (1)
Number Date Country
20160063997 A1 Mar 2016 US
Provisional Applications (1)
Number Date Country
62043344 Aug 2014 US