Occlusion reduction and active noise reduction based on seal quality

Information

  • Patent Grant
  • 9779716
  • Patent Number
    9,779,716
  • Date Filed
    Wednesday, December 30, 2015
    8 years ago
  • Date Issued
    Tuesday, October 3, 2017
    6 years ago
Abstract
Systems and methods for active noise reduction and occlusion reduction based on seal quality of an in-the-ear (ITE) module inserted into a user's ear canal are provided. An example method includes receiving one or more acoustic signals. Each of the acoustic signals represents at least one captured sound having at least one of a voice component and an unwanted noise. The voice component may include the user's own voice. A quality of a seal of an ear canal is determined based at least partially on the acoustic signals. If the quality of the seal exceeds a predetermined threshold value, an occlusion reduction is performed on the acoustic signals to improve the voice component. If the quality of the seal is below a predetermined threshold value, active noise reduction is performed on the acoustic signals to reduce the unwanted noise.
Description
FIELD

The present application relates generally to audio processing and, more specifically, to systems and methods for occlusion reduction and active noise cancellation based on seal quality.


BACKGROUND

An active noise reduction (ANR) system in an earpiece-based audio device can be used to reduce background noise. The ANR system can form a compensation signal adapted to cancel background noise at a listening position inside the earpiece. The compensation signal is provided to an audio transducer (e.g., a loudspeaker), which generates an “anti-noise” acoustic wave. The anti-noise acoustic wave is intended to attenuate or eliminate the background noise at the listening position via a destructive interference, so that only the desired audio remains. Consequently, a combination of the anti-noise acoustic wave and the background noise at the listening position results in cancellation of both and, hence, a reduction in noise.


An occlusion effect occurs when earpieces of a headset seal a person's (user's) ear canals. The person may hear uncomfortable sounds from their own voice caused by bone-conducted sound reverberating off the earpiece blocking the ear canal. The occlusion effect is more pronounced if the seal is very good. The occlusion effect can boost low frequency (usually below 500 Hz) sound pressure in the ear canal by 20 dB or more.


SUMMARY

This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.


Methods and systems for occlusion reduction and ANR based on a determination of a quality of a seal are provided. The method may provide for more uniform performance of a headset across different seal qualities. An example method includes receiving acoustic signals. Each of the acoustic signals may represent at least one captured sound having at least one of a voice component and an unwanted noise, the voice component including the voice of a user. The example method further includes determining, based at least partially on the acoustic signals, a quality of a seal, provided by an in-the-ear module of a headset, of the user's ear canal. The example method switches between operational modes depending on seal quality. For example, if the quality of the seal is above a predetermined threshold value, the method may proceed with performing an occlusion reduction on the acoustic signals to improve the voice component. If the quality of the seal is below the predetermined threshold value, the method may proceed with performing an active noise reduction (ANR) on the acoustic signals to reduce the unwanted noise.


According to another example embodiment of the present disclosure, the steps of the method for occlusion reduction and the ANR based on a quality of a seal are stored on a non-transitory machine-readable medium comprising instructions, which, when implemented by one or more processors, perform the recited steps.


Other example embodiments of the disclosure and aspects will become apparent from the following description taken in conjunction with the following drawings.





BRIEF DESCRIPTION OF DRAWINGS

Embodiments are illustrated by way of example and not limitation in the figures of the accompanying drawings, in which like references indicate similar elements and in which:



FIG. 1 is a block diagram of a system and an environment in which the system is used, according to an example embodiment.



FIG. 2 is a block diagram of a headset suitable for implementing the present technology, according to an example embodiment.



FIG. 3 is a block diagram illustrating a system for performing occlusion reduction and active noise reduction based on a determination of seal quality, according to an example embodiment.



FIG. 4 is a flow chart showing steps of a method for performing either occlusion reduction or active noise reduction based on a determination of seal quality, according to an example embodiment.



FIG. 5 illustrates an example of a computer system that may be used to implement embodiments of the disclosed technology.





DETAILED DESCRIPTION

The present technology provides systems and methods for occlusion reduction and ANR based on a determination of a quality of a seal, which can overcome or substantially alleviate problems associated with uncomfortable sounds in an ear canal. Embodiments of the present technology may be practiced on any earpiece-based audio device that is configured to receive and/or provide audio such as, but not limited to, cellular phones, MP3 players, phone handsets, hearing aids, and headsets. While some embodiments of the present technology are described in reference to operation of a cellular phone, the present technology may be practiced on any audio device.


According to an example embodiment, the method for occlusion reduction and ANR based on a determination of a quality of a seal includes receiving acoustic signals. The method may provide for more uniform performance of a headset across different seal qualities. For the example method, each of the acoustic signals represents at least one captured sound. The captured sound may include at least one of a voice component and an unwanted noise. The voice component may include the voice of a user.


The method further includes determining, based at least partially on the acoustic signals, at least the quality of a seal of an ear canal. If the quality of the seal is above a predetermined threshold value, the example method proceeds with performing an occlusion reduction on the acoustic signals in order to improve the voice component. Alternatively, if the quality of the seal is below the predetermined threshold value, the example method proceeds with performing an ANR on the acoustic signals to reduce the unwanted noise.


Referring now to FIG. 1, a block diagram of an example system 100 suitable for performing occlusion reduction and ANR and an environment thereof are shown. The example system 100 includes at least an internal microphone 106, an external microphone 108, a digital signal processor (DSP) 112, and a wireless or wired interface 114. The internal microphone 106 is located inside a user's ear canal 104 and is relatively shielded from the outside acoustic environment 102. The external microphone 108 is located outside of the user's ear canal 104 and is exposed to the outside acoustic environment 102. In some embodiments, the example system 100 includes an accelerometer 120. The accelerometer 120 is located inside user's ear canal 104.


In various embodiments, the microphones 106 and 108 are either analog or digital. In either case, the outputs from the microphones are converted into synchronized pulse code modulation (PCM) format at a suitable sampling frequency and connected to the input port of the DSP 112. The signals xin and xex denote signals representing sounds captured by internal microphone 106 and external microphone 108, respectively.


The DSP 112 performs appropriate signal processing tasks to improve the quality of microphone signals xin and xex, according to some embodiments. The output of DSP 112, referred to as the send-out signal (sout), is transmitted to the desired destination, for example, to a network or host device 116 (see signal identified as sout uplink), through a radio or wired interface 114.


In certain embodiments, a signal is received by the network or host device 116 from a suitable source (e.g., via the wireless radio or wired interface 114). This is referred to as the receive-in signal (rin) (identified as rin downlink at the network or host device 116). The receive-in signal can be coupled via the radio or wired interface 114 to the DSP 112 for processing. The resulting signal, referred to as the receive-out signal (rout), is converted into an analog signal through a digital-to-analog convertor (DAC) 110 and then connected to a loudspeaker 118 in order to be presented to the user. In some embodiments, the loudspeaker 118 is located in the same ear canal 104 as the internal microphone 106. In other embodiments, the loudspeaker 118 is located in the ear canal opposite the ear canal 104. In the example of FIG. 1, the loudspeaker 118 is found in the same ear canal 104 as the internal microphone 106; therefore, an acoustic echo canceller (AEC) may be needed to prevent the feedback of the received signal to the other end. Optionally, if no further processing of the received signal is necessary, the receive-in signal (rin) can be coupled to the loudspeaker 118 without going through the DSP 112. In some embodiments, the receive-in signal rin includes an audio content (for example, music) presented to the user.



FIG. 2 shows an example headset 200 suitable for implementing methods of the present disclosure. The headset 200 includes example in-the-ear (ITE) module(s) 202 and behind-the-ear (BTE) modules 204 and 206 for each ear of a user. The ITE module(s) 202 are configured to be inserted into the user's ear canals. The BTE modules 204 and 206 are configured to be placed behind (or otherwise near) the user's ears. In some embodiments, the headset 200 communicates with host devices through a wireless radio link. The wireless radio link may conform to a Bluetooth Low Energy (BLE), other Bluetooth, 802.11, or other suitable wireless standard and may be variously encrypted for privacy. The example headset 200 is a nonlimiting example, other variations having just an in-the-ear “earpiece” may be used to practice the present technology.


In various embodiments, ITE module(s) 202 include internal microphone 106 and the loudspeaker(s) 118 (shown in FIG. 1), all facing inward with respect to the ear canals. The ITE module(s) 202 can provide acoustic isolation between the ear canal(s) 104 and the outside acoustic environment 102. In some embodiments, ITE module(s) 202 includes at least one accelerometer 120 (also shown in FIG. 1).


In some embodiments, each of the BTE modules 204 and 206 includes at least one external microphone 108 (also shown in FIG. 1). The BTE module 204 may include a DSP 112 (as shown in FIG. 1), control button(s), and Bluetooth radio link to host devices. In certain embodiments, the BTE module 206 includes a suitable battery with charging circuitry.


The system and headset in FIGS. 1 and 2 is discussed in more detail in U.S. patent application Ser. No. 14/853,947, entitled “Microphone Signal Fusion,” filed on Sep. 14, 2015, the disclosure of which is incorporated herein by reference for all purposes.


In certain embodiments, the seal of the ITE module(s) 202 is good enough to isolate acoustic waves coming from the outside acoustic environment 102. However, when speaking or singing, a user can hear the user's own voice reflected by ITE module(s) 202 back into the corresponding ear canal. The sound of the voice of the user is distorted since, while traveling through the user's skull, the high frequencies of the voice are substantially attenuated and thus has a much narrower effective bandwidth compared to voice conducted through air. As a result, the user can hear mostly the low frequencies of the voice.



FIG. 3 is a block diagram showing a system 300 for performing occlusion reduction and ANR based on a determination of a seal quality, according to an example embodiment. The example system 300 includes seal quality determination module 310, an active noise reduction (ANR) module 320, and an occlusion reduction module 330. The modules of system 300 can be implemented as instructions stored in a memory and executed by at least one processor, for example, DSP 112. In certain embodiments, at least some of the instructions performing the functionalities of the modules 310-330 are stored in a memory and executed by at least one processor of the network or host device 116.


In some embodiments, the occlusion reduction module 330 is operable to receive at least internal microphone signal xin and perform active occlusion reduction. The active occlusion reduction may be used to cancel some components of the distorted voice to restore a natural voice sound inside ear canal 104. The distorted voice is captured by the internal microphone inside the ear cancel. The active occlusion reduction generates, based on the internal microphone signal xin, a first signal. When played by loudspeaker 118, the first signal cancels out some low frequencies (e.g., where the distortion due to the skull is found) of the distorted voice and by doing so improves voice quality distorted by travelling through the skull.


In other embodiments, the ANR module 320 is used to reduce outside unwanted noise (also referred to as background noise) captured by external microphone 108 from outside acoustic environment 102. ANR module 320 receives signal xex captured by external microphone 108. ANR module 320 generates, based on the signal xex, a second signal. When played by the loudspeaker 118, the second signal cancels the outside unwanted noise within the ear canal 104.


In various embodiments, the occlusion reduction can be carried via use of a limited bandwidth noise cancellation since, while traveling through human tissue, the high frequencies of the user's voice are substantially attenuated and thus has a much narrower effective bandwidth compared to voice conducted through air. Thus, the bandwidth of noise cancellation for occlusion reduction may be limited to between 100 Hz and 1 KHz, for example.


In various embodiments, switching between the first operational mode for the occlusion reduction (e.g., using occlusion reduction module 330) and the second operational mode for the ANR (e.g., using the ANR module 320) is based on the determination of the quality of the seal of the ear canal. In various embodiments, the seal quality determination module 310 is operable to determine the quality of the seal by comparing signal xex captured by the external microphone 108 and signal xin captured by internal microphone 106. If signal xin includes noise components similar to the noise components of signal xex, it indicates that outside noise is heard inside the earbud, reflective of a bad seal quality, according to various embodiments. The quality of the ear seal might be determined by any of a variety of suitable methods/including comparing the internal and external mic, but is not limited to that method. An example system suitable for determining seal quality is discussed in more detail in U.S. patent application Ser. No. 14/985,187, entitled “Audio Monitoring and Adaptation Using Headset Microphones Inside of User's Ear Canal,” filed on Dec. 30, 2015, the disclosure of which is incorporated herein by reference for all purposes.


In various embodiments, when the ANR is performed in response to the determination that the seal of the ear canal is poor, accelerometer data from accelerometer 120 located inside the ITE module(s) 202 can be used to discriminate between the voice of the user and background noise in the external microphone signal xex. For example, the accelerometer may be used to detect signals (e.g., motion of the user's head) that are indicative of the user speaking. In various embodiments, if it is determined that the user is speaking then the ANR module 320 reduces noise in a way that reduces or cancels the background noise without suppressing the voice components of the user's voice in a way that would distort it. That is, the background noise in the received acoustic signal is suppressed, in various embodiments, in a way that does not result in also causing distortion of the part of acoustic signal that represents the users's voice. An example audio processing system suitable for performing this balance between noise cancellation and voice quality is discussed in more detail in U.S. patent application Ser. No. 12/832,901 (now U.S. Pat. No. 8,473,287), entitled “Method for Jointly Optimizing Noise Reduction and Voice Quality in a Mono or Multi-Microphone System,” filed on Jul. 8, 2010, the disclosure of which is incorporated herein by reference for all purposes.


Although separate modules are shown in FIG. 3 for ANR and occlusion reduction, the ANR module 320 may be configured to perform ANR and the noise cancellation for the occlusion reduction.


In certain embodiments, the ITE module(s) 202 may include a mechanical vent. The mechanical vent may include an electroactive polymer. The mechanical vent may be configured to be closed to make a better seal. In response to the determination that a seal of the ear is good (e.g., the quality of the seal is above a predetermined threshold) and the voice of the user sounds distorted inside the ear canal, the mechanical vent may be opened to let the user's voice that is inside the ear canal 104 travel outside the ITE module(s) 202. When the mechanical vent is open, the distorted user's voice may bounce back less to the ear canal so as to reduce the uncomfortable sound presented to the user. At the same time, opening of the mechanical vent would let in the outside acoustic signals which may not only let in the undistorted user's voice from outside, but also let in background noise inside the ear canal. Active noise cancellation may be performed to cancel just this background noise so that the opening of the mechanical vent does not cause additional outside background noise to be heard by the user. By way of example and not limitation, the mechanical vent may be activated when the user starts a phone call. In certain embodiments, the mechanical vent is activated when the seal quality is above a threshold and speech (for example, from speakers other than the user) is detected, while an external noise is present and the user is listening to music without talking or singing along. The mechanical vent may also actively relieve air pressure in the ear to provide greater comfort for the user.


An example audio processing system suitable for performing noise cancellation and/or noise reduction is discussed in more detail in U.S. patent application Ser. No. 12/832,901 (now U.S. Pat. No. 8,473,287), entitled “Method for Jointly Optimizing Noise Reduction and Voice Quality in a Mono or Multi-Microphone System,” filed on Jul. 8, 2010, the disclosure of which is incorporated herein by reference for all purposes. By way of example and not limitation, noise reduction methods are described in U.S. patent application Ser. No. 12/215,980 (now U.S. Pat. No. 9,185,487), entitled “System and Method for Providing Noise Suppression Utilizing Null Processing Noise Subtraction,” filed Jun. 30, 2008, and in U.S. patent application Ser. No. 11/699,732 (now U.S. Pat. No. 8,194,880), entitled “System and Method for Utilizing Omni-Directional Microphones for Speech Enhancement,” filed Jan. 29, 2007, which are incorporated herein by reference in their entireties.



FIG. 4 is a flow chart showing steps of method 400 for performing either occlusion reduction or ANR based on a determination of a seal quality, according to various example embodiments. The example method 400 can commence with determining a quality of the seal of a user's ear canal that is provided by an in-the-ear (ITE) module inserted therein, in block 402. In some embodiments, the quality of the seal can be determined based on a difference of signal xex captured by the external microphone 108 and signal xin captured by the internal microphone 106. If signal xin includes components similar to components of signal xex, it indicates that outside noise is captured by the internal microphone (e.g., in the ITE module) inside the ear canal.


In decision block 404, a decision is made based on the quality of the seal of the ear canal. If the quality of the seal is above a predetermined threshold value, method 400, in this example, proceeds with performing occlusion reduction in block 406. Alternatively, if the quality of the seal is below a predetermined threshold value, then method 400, in this example, performs ANR in block 408. The predetermined threshold value may be determined based on, for example, the difference in signal between the signal xex captured by the external microphone 108 and signal xin captured by internal microphone 106 being over a certain threshold, indicating the seal is such that outside noise that the external microphone 108 captures is not being captured by the internal microphone 106 because of the seal. In some embodiments, the predetermined threshold value may be a table of values or other relationship, such that there is continually varying, e.g., including a mix of occlusion reduction and ANR for certain values, rather than just switching between occlusion reduction and ANR.



FIG. 5 illustrates an exemplary computer system 500 that may be used to implement some embodiments of the present invention. The computer system 500 of FIG. 5 may be implemented in the contexts of the likes of computing systems, networks, servers, or combinations thereof. The computer system 500 of FIG. 5 includes one or instructions and data for execution by processor unit(s) 510. Main memory 520 stores the executable code when in operation, in this example. The computer system 500 of FIG. 5 further includes a mass data storage 530, portable storage device 540, output devices 550, user input devices 560, a graphics display system 570, and peripheral devices 580.


The components shown in FIG. 5 are depicted as being connected via a single bus 590. The components may be connected through one or more data transport means. Processor unit(s) 510 and main memory 520 are connected via a local microprocessor bus, and the mass data storage 530, peripheral device(s) 580, portable storage device 540, and graphics display system 570 are connected via one or more input/output (I/O) buses.


Mass data storage 530, which can be implemented with a magnetic disk drive, solid state drive, or an optical disk drive, is a non-volatile storage device for storing data and instructions for use by processor unit(s) 510. Mass data storage 530 stores the system software for implementing embodiments of the present disclosure for purposes of loading that software into main memory 520.


Portable storage device 540 operates in conjunction with a portable non-volatile storage medium, such as a flash drive, floppy disk, compact disk, digital video disc, or Universal Serial Bus (USB) storage device, to input and output data and code to and from the computer system 500 of FIG. 5. The system software for implementing embodiments of the present disclosure is stored on such a portable medium and input to the computer system 500 via the portable storage device 540.


User input devices 560 can provide a portion of a user interface. User input devices 560 may include one or more microphones, an alphanumeric keypad, such as a keyboard, for inputting alphanumeric and other information, or a pointing device, such as a mouse, a trackball, stylus, or cursor direction keys. User input devices 560 can also include a touchscreen. Additionally, the computer system 500 as shown in FIG. 5 includes output devices 550. Suitable output devices 550 include speakers, printers, network interfaces, and monitors.


Graphics display system 570 includes a liquid crystal display (LCD) or other suitable display device. Graphics display system 570 is configurable to receive textual and graphical information and processes the information for output to the display device.


Peripheral devices 580 may include any type of computer support device to add additional functionality to the computer system.


The components provided in the computer system 500 of FIG. 5 are those typically found in computer systems that may be suitable for use with embodiments of the present disclosure and are intended to represent a broad category of such computer components that are well known in the art. Thus, the computer system 500 of FIG. 5 can be a personal computer (PC), hand held computer system, telephone, mobile computer system, workstation, tablet, phablet, mobile phone, server, minicomputer, mainframe computer, wearable, or any other computer system. The computer may also include different bus configurations, networked platforms, multi-processor platforms, and the like. Various operating systems may be used including UNIX, LINUX, WINDOWS, MAC OS, PALM OS, QNX ANDROID, IOS, CHROME, TIZEN, and other suitable operating systems.


The processing for various embodiments may be implemented in software that is cloud-based. In some embodiments, the computer system 500 is implemented as a cloud-based computing environment, such as a virtual machine operating within a computing cloud. In other embodiments, the computer system 500 may itself include a cloud-based computing environment, where the functionalities of the computer system 500 are executed in a distributed fashion. Thus, the computer system 500, when configured as a computing cloud, may include pluralities of computing devices in various forms, as will be described in greater detail below.


In general, a cloud-based computing environment is a resource that typically combines the computational power of a large grouping of processors (such as within web servers) and/or that combines the storage capacity of a large grouping of computer memories or storage devices. Systems that provide cloud-based resources may be utilized exclusively by their owners or such systems may be accessible to outside users who deploy applications within the computing infrastructure to obtain the benefit of large computational or storage resources.


The cloud may be formed, for example, by a network of web servers that comprise a plurality of computing devices, such as the computer system 500, with each server (or at least a plurality thereof) providing processor and/or storage resources. These servers may manage workloads provided by multiple users (e.g., cloud resource customers or other users). Typically, each user places workload demands upon the cloud that vary in real-time, sometimes dramatically. The nature and extent of these variations typically depends on the type of business associated with the user.


The present technology is described above with reference to example embodiments. Therefore, other variations upon the example embodiments are intended to be covered by the present disclosure.

Claims
  • 1. A method for audio processing, the method comprising: receiving acoustic signals, each of the acoustic signals representing at least one captured sound having a voice component and an unwanted noise;determining, based at least partially on the acoustic signals, a quality of a seal, provided by an in-the-ear module of a headset, of an ear canal of a user;checking the determined quality of the seal against a predetermined threshold value, and based on the checking: if the quality of the seal is above the predetermined threshold value, performing an occlusion reduction on the acoustic signals to improve the voice component; andif the quality of the seal is below the predetermined threshold value, performing an active noise reduction (ANR) on the acoustic signals to reduce the unwanted noise.
  • 2. The method of claim 1, wherein the voice component includes the voice of the user.
  • 3. The method of claim 1, wherein: the acoustic signals include a first acoustic signal captured outside the ear canal and a second acoustic signal captured inside the ear canal; andthe determination of the quality of the seal includes comparing the first acoustic signal and the second acoustic signal.
  • 4. The method of claim 1, wherein the occlusion reduction includes performing active noise cancellation for a limited bandwidth of the acoustic signals.
  • 5. The method of claim 4, wherein the limited bandwidth is within a frequency range between 100 Hz and 1 kHz.
  • 6. The method of claim 1, wherein the predetermined threshold value is a table of values such that occlusion reduction and the ANR are performed on a continually varying basis as a function of the predetermine threshold value.
  • 7. The method of claim 6, further comprising: determining whether the voice component has qualities indicative of the quality of the seal being above the predetermined threshold value,wherein the in-the-ear module operates in a first mode in response to the determining indicating that the voice component has qualities indicative of the quality of the seal being above the predetermined threshold value.
  • 8. The method of claim 1, wherein the ANR includes: discriminating between the voice component and the unwanted noise; andcancelling, based on results of the discrimination, the unwanted noise in the acoustic signals.
  • 9. The method of claim 8, wherein the discrimination is based on data from an accelerometer located inside the ear canal, the accelerometer providing one or more signals indicative of the user speaking.
  • 10. The method of claim 9, wherein, while detecting that the user is speaking, the ANR is configured to limit distortion of the voice components that represents the user's voice while performing the ANR on the acoustic signals.
  • 11. The method of claim 1, wherein the occlusion reduction includes: activating a mechanical vent to allow sound waves from outside of the ear canal to penetrate inside the ear canal, the mechanical vent being activated in response to the checking indicating that the quality of the seal is above the predetermined threshold value; andcancelling noise in the sound waves.
  • 12. A system for audio processing, the system comprising: at least one processor to receive acoustic signals, each acoustic signal representing at least one captured sound having a voice component and an unwanted noise;at least one processor to determine, based at least partially on the acoustic signals, a quality of a seal, provided by an in-the-ear module of a headset, of an ear canal of a user;at least one processor to check the determined quality of the seal against a predetermined threshold value, and based on the checking: if the quality of the seal is above the predetermined threshold value, at least one processor being configured to perform an occlusion reduction on the acoustic signals to improve the voice component; andif the quality of the seal is below the predetermined threshold value, at least one processor being configured to perform an active noise reduction (ANR) on the acoustic signals to reduce the unwanted noise.
  • 13. The system of claim 12, wherein the voice component includes the voice of the user.
  • 14. The system of claim 12, wherein: the acoustic signals include a first acoustic signal captured outside the ear canal and a second acoustic signal captured inside the ear canal; andthe quality of the seal is determined by comparing the first acoustic signal and the second acoustic signal.
  • 15. The system of claim 12, wherein the occlusion reduction includes performing an active noise cancellation for a limited bandwidth of the acoustic signals, the limited bandwidth being within a frequency range between 100 Hz and 1 kHz.
  • 16. The system of claim 12, wherein the occlusion reduction and the ANR are performed by a module configured to operate, based on the determination of the quality of the seal, in a first mode for performing the occlusion reduction and a second mode for performing the ANR.
  • 17. The system of claim 16, further comprising: at least one processor configured to determine whether the voice component has distortion indicative of the quality of the seal being above the predetermined threshold,wherein the module operates in the first mode in response to the at least one processor configured to determine whether the voice component has distortion indicative of the quality of the seal being above the predetermined threshold indicates that the voice component has distortion indicative of the quality of the seal being above the predetermined threshold.
  • 18. The system of claim 12, wherein the ANR includes: discriminating between the voice component and the unwanted noise; andcancelling, based on results of the discriminating, the unwanted noise in the acoustic signals.
  • 19. The system of claim 18, wherein the discriminating is based on data from an accelerometer located inside the ear canal, the accelerometer detecting at least motion indicative of the user speaking.
  • 20. The system of claim 12, wherein the occlusion reduction includes: activating a mechanical vent to allow sound waves from outside of the ear canal to penetrate inside the ear canal, the mechanical vent being activated in response to the checking indicating that the quality of the seal is above the predetermined threshold; andcancelling noise in the sound waves.
  • 21. A non-transitory computer-readable storage medium having embodied thereon instructions, which, when executed by at least one processor, cause the at least one processor to perform steps of a method, the method comprising: receiving acoustic signals, each of the acoustic signals representing at least one captured sound having a voice component and an unwanted noise;determining, based at least partially on the acoustic signals, a quality of a seal, provided by an in-the-ear module of a headset, of a user's ear canal;checking the determined quality of the seal against a predetermined threshold value, and based on the checking: if the quality of the seal is above the predetermined threshold value, performing an occlusion reduction on the acoustic signals to improve the voice component; andif the quality of the seal is below the predetermined threshold value, performing an active noise reduction (ANR) on the acoustic signals to reduce the unwanted noise.
US Referenced Citations (275)
Number Name Date Kind
2535063 Halstead Dec 1950 A
3995113 Tani Nov 1976 A
4150262 Ono Apr 1979 A
4455675 Bose et al. Jun 1984 A
4516428 Konomi May 1985 A
4520238 Ikeda May 1985 A
4588867 Konomi May 1986 A
4596903 Yoshizawa Jun 1986 A
4644581 Sapiejewski Feb 1987 A
4652702 Yoshii Mar 1987 A
4696045 Rosenthal Sep 1987 A
4975967 Rasmussen Dec 1990 A
5208867 Stites, III May 1993 A
5222050 Marren et al. Jun 1993 A
5251263 Andrea et al. Oct 1993 A
5282253 Konomi Jan 1994 A
5289273 Lang Feb 1994 A
5295193 Ono Mar 1994 A
5305387 Sapiejewski Apr 1994 A
5319717 Holesha Jun 1994 A
5327506 Stites, III Jul 1994 A
D360691 Mostardo Jul 1995 S
D360948 Mostardo Aug 1995 S
D360949 Mostardo Aug 1995 S
5490220 Loeppert Feb 1996 A
5734621 Ito Mar 1998 A
5870482 Loeppert et al. Feb 1999 A
D414493 Jiann-Yeong Sep 1999 S
5960093 Miller Sep 1999 A
5983073 Ditzik Nov 1999 A
6044279 Hokao et al. Mar 2000 A
6061456 Andrea et al. May 2000 A
6094492 Boesen Jul 2000 A
6118878 Jones Sep 2000 A
6122388 Feldman Sep 2000 A
6130953 Wilton et al. Oct 2000 A
6184652 Yang Feb 2001 B1
6211649 Matsuda Apr 2001 B1
6219408 Kurth Apr 2001 B1
6255800 Bork Jul 2001 B1
D451089 Hohl et al. Nov 2001 S
6362610 Yang Mar 2002 B1
6373942 Braund Apr 2002 B1
6408081 Boesen Jun 2002 B1
6462668 Foseide Oct 2002 B1
6535460 Loeppert et al. Mar 2003 B2
6567524 Svean et al. May 2003 B1
6661901 Svean et al. Dec 2003 B1
6683965 Sapiejewski Jan 2004 B1
6694180 Boesen Feb 2004 B1
6717537 Fang et al. Apr 2004 B1
6738485 Boesen May 2004 B1
6748095 Goss Jun 2004 B1
6751326 Nepomuceno Jun 2004 B2
6754358 Boesen et al. Jun 2004 B1
6754359 Svean et al. Jun 2004 B1
6757395 Fang et al. Jun 2004 B1
6801632 Olson Oct 2004 B2
6847090 Loeppert Jan 2005 B2
6879698 Boesen Apr 2005 B2
6920229 Boesen Jul 2005 B2
6931292 Brumitt et al. Aug 2005 B1
6937738 Armstrong et al. Aug 2005 B2
6987859 Loeppert et al. Jan 2006 B2
7023066 Lee et al. Apr 2006 B2
7024010 Saunders et al. Apr 2006 B2
7039195 Svean et al. May 2006 B1
7103188 Jones Sep 2006 B1
7132307 Wang et al. Nov 2006 B2
7136500 Collins Nov 2006 B2
7203331 Boesen Apr 2007 B2
7209569 Boesen Apr 2007 B2
7215790 Boesen et al. May 2007 B2
7289636 Saunders et al. Oct 2007 B2
7302074 Wagner et al. Nov 2007 B2
D573588 Warren et al. Jul 2008 S
7406179 Ryan Jul 2008 B2
7433481 Armstrong et al. Oct 2008 B2
7477754 Rasmussen et al. Jan 2009 B2
7477756 Wickstrom et al. Jan 2009 B2
7502484 Ngia et al. Mar 2009 B2
7590254 Olsen Sep 2009 B2
7680292 Warren et al. Mar 2010 B2
7747032 Zei et al. Jun 2010 B2
7773759 Alves et al. Aug 2010 B2
7869610 Jayanth et al. Jan 2011 B2
7889881 Ostrowski Feb 2011 B2
7899194 Boesen Mar 2011 B2
7965834 Alves et al. Jun 2011 B2
7983433 Nemirovski Jul 2011 B2
8005249 Wirola et al. Aug 2011 B2
8019107 Ngia et al. Sep 2011 B2
8027481 Beard Sep 2011 B2
8045724 Sibbald Oct 2011 B2
8072010 Lutz Dec 2011 B2
8077873 Shridhar et al. Dec 2011 B2
8081780 Goldstein et al. Dec 2011 B2
8103029 Ngia et al. Jan 2012 B2
8111853 Isvan Feb 2012 B2
8116489 Mejia et al. Feb 2012 B2
8116502 Saggio, Jr. et al. Feb 2012 B2
8135140 Shridhar et al. Mar 2012 B2
8180067 Soulodre May 2012 B2
8189799 Shridhar et al. May 2012 B2
8194880 Avendano Jun 2012 B2
8199924 Wertz et al. Jun 2012 B2
8213643 Hemer Jul 2012 B2
8213645 Rye et al. Jul 2012 B2
8229125 Short Jul 2012 B2
8229740 Nordholm et al. Jul 2012 B2
8238567 Burge et al. Aug 2012 B2
8249287 Silvestri et al. Aug 2012 B2
8254591 Goldstein et al. Aug 2012 B2
8270626 Shridhar et al. Sep 2012 B2
8285344 Kahn et al. Oct 2012 B2
8295503 Sung et al. Oct 2012 B2
8311253 Silvestri et al. Nov 2012 B2
8315404 Shridhar et al. Nov 2012 B2
8325963 Kimura Dec 2012 B2
8331604 Saito et al. Dec 2012 B2
8363823 Santos Jan 2013 B1
8376967 Mersky Feb 2013 B2
8385560 Solbeck et al. Feb 2013 B2
8401200 Tiscareno et al. Mar 2013 B2
8401215 Warren et al. Mar 2013 B2
8416979 Takai Apr 2013 B2
8462956 Goldstein et al. Jun 2013 B2
8473287 Every et al. Jun 2013 B2
8483418 Platz et al. Jul 2013 B2
8488831 Saggio, Jr. et al. Jul 2013 B2
8494201 Anderson Jul 2013 B2
8498428 Schreuder et al. Jul 2013 B2
8503689 Schreuder et al. Aug 2013 B2
8503704 Francart et al. Aug 2013 B2
8509465 Theverapperuma Aug 2013 B2
8526646 Boesen Sep 2013 B2
8532323 Wickstrom et al. Sep 2013 B2
8553899 Salvetti et al. Oct 2013 B2
8553923 Tiscareno et al. Oct 2013 B2
8571227 Donaldson et al. Oct 2013 B2
8594353 Anderson Nov 2013 B2
8620650 Walters et al. Dec 2013 B2
8634576 Salvetti et al. Jan 2014 B2
8655003 Duisters et al. Feb 2014 B2
8666102 Bruckhoff et al. Mar 2014 B2
8681999 Theverapperuma et al. Mar 2014 B2
8682001 Annunziato et al. Mar 2014 B2
8705787 Larsen et al. Apr 2014 B2
8837746 Burnett Sep 2014 B2
8942976 Li et al. Jan 2015 B2
8983083 Tiscareno et al. Mar 2015 B2
9014382 Van De Par et al. Apr 2015 B2
9025415 Derkx May 2015 B2
9042588 Aase May 2015 B2
9047855 Bakalos Jun 2015 B2
9078064 Wickstrom et al. Jul 2015 B2
9100756 Dusan et al. Aug 2015 B2
9107008 Leitner Aug 2015 B2
9123320 Carreras et al. Sep 2015 B2
9154868 Narayan et al. Oct 2015 B2
9167337 Shin Oct 2015 B2
9185487 Solbach et al. Nov 2015 B2
9208769 Azmi Dec 2015 B2
9226068 Hendrix et al. Dec 2015 B2
9264823 Bajic et al. Feb 2016 B2
20010011026 Nishijima Aug 2001 A1
20010021659 Okamura Sep 2001 A1
20010049262 Lehtonen Dec 2001 A1
20020016188 Kashiwamura Feb 2002 A1
20020021800 Bodley et al. Feb 2002 A1
20020038394 Liang et al. Mar 2002 A1
20020054684 Menzl May 2002 A1
20020056114 Fillebrown et al. May 2002 A1
20020067825 Baranowski et al. Jun 2002 A1
20020098877 Glezerman Jul 2002 A1
20020136420 Topholm Sep 2002 A1
20020159023 Swab Oct 2002 A1
20020176330 Ramonowski et al. Nov 2002 A1
20020183089 Heller et al. Dec 2002 A1
20030002704 Pronk Jan 2003 A1
20030013411 Uchiyama Jan 2003 A1
20030017805 Yeung et al. Jan 2003 A1
20030058808 Eaton et al. Mar 2003 A1
20030085070 Wickstrom May 2003 A1
20030207703 Liou et al. Nov 2003 A1
20030223592 Deruginsky et al. Dec 2003 A1
20050027522 Yamamoto et al. Feb 2005 A1
20050058313 Victorian Mar 2005 A1
20060029234 Sargaison Feb 2006 A1
20060034472 Bazarjani et al. Feb 2006 A1
20060153155 Jacobsen et al. Jul 2006 A1
20060227990 Kirchhoefer Oct 2006 A1
20060239472 Oda Oct 2006 A1
20070104340 Miller et al. May 2007 A1
20070147635 Dijkstra et al. Jun 2007 A1
20080019548 Avendano Jan 2008 A1
20080063228 Mejia et al. Mar 2008 A1
20080101640 Ballad et al. May 2008 A1
20080107287 Beard May 2008 A1
20080181419 Goldstein et al. Jul 2008 A1
20080232621 Burns Sep 2008 A1
20090041269 Hemer Feb 2009 A1
20090080670 Solbeck et al. Mar 2009 A1
20090182913 Rosenblatt et al. Jul 2009 A1
20090207703 Matsumoto et al. Aug 2009 A1
20090214068 Wickstrom Aug 2009 A1
20090323982 Solbach et al. Dec 2009 A1
20100022280 Schrage Jan 2010 A1
20100081487 Chen et al. Apr 2010 A1
20100183167 Phelps et al. Jul 2010 A1
20100233996 Herz et al. Sep 2010 A1
20100270631 Renner Oct 2010 A1
20110116643 Tiscareno et al. May 2011 A1
20110257967 Every et al. Oct 2011 A1
20120008808 Saltykov Jan 2012 A1
20120056282 Van Lippen et al. Mar 2012 A1
20120099753 van der Avoort et al. Apr 2012 A1
20120197638 Li et al. Aug 2012 A1
20120321103 Smailagic et al. Dec 2012 A1
20130024194 Zhao et al. Jan 2013 A1
20130051580 Miller Feb 2013 A1
20130058495 Furst et al. Mar 2013 A1
20130070935 Hui et al. Mar 2013 A1
20130142358 Schultz et al. Jun 2013 A1
20130272564 Miller Oct 2013 A1
20130287219 Hendrix et al. Oct 2013 A1
20130315415 Shin Nov 2013 A1
20130322642 Streitenberger et al. Dec 2013 A1
20130343580 Lautenschlager et al. Dec 2013 A1
20130345842 Karakaya et al. Dec 2013 A1
20140010378 Voix et al. Jan 2014 A1
20140044275 Goldstein et al. Feb 2014 A1
20140086425 Jensen et al. Mar 2014 A1
20140169579 Azmi Jun 2014 A1
20140233741 Gustavsson Aug 2014 A1
20140247948 Goldstein Sep 2014 A1
20140270231 Dusan et al. Sep 2014 A1
20140273851 Donaldson et al. Sep 2014 A1
20140348346 Fukuda Nov 2014 A1
20140355787 Jiles et al. Dec 2014 A1
20150025881 Carlos et al. Jan 2015 A1
20150043741 Shin Feb 2015 A1
20150055810 Shin Feb 2015 A1
20150078574 Shin Mar 2015 A1
20150110280 Wardle Apr 2015 A1
20150161981 Kwatra Jun 2015 A1
20150172814 Usher et al. Jun 2015 A1
20150237448 Loeppert Aug 2015 A1
20150243271 Goldstein Aug 2015 A1
20150245129 Dusan et al. Aug 2015 A1
20150264472 Aase Sep 2015 A1
20150296305 Shao et al. Oct 2015 A1
20150296306 Shao et al. Oct 2015 A1
20150304770 Watson et al. Oct 2015 A1
20150310846 Andersen et al. Oct 2015 A1
20150325229 Carreras et al. Nov 2015 A1
20150325251 Dusan et al. Nov 2015 A1
20150365770 Lautenschlager Dec 2015 A1
20150382094 Grinker et al. Dec 2015 A1
20160007119 Harrington Jan 2016 A1
20160021480 Johnson et al. Jan 2016 A1
20160029345 Sebeni et al. Jan 2016 A1
20160037261 Harrington Feb 2016 A1
20160037263 Pal et al. Feb 2016 A1
20160042666 Hughes Feb 2016 A1
20160044151 Shoemaker et al. Feb 2016 A1
20160044398 Siahaan et al. Feb 2016 A1
20160044424 Dave et al. Feb 2016 A1
20160060101 Loeppert Mar 2016 A1
20160105748 Pal et al. Apr 2016 A1
20160127829 Ring May 2016 A1
20160150335 Qutub et al. May 2016 A1
20160165334 Grossman Jun 2016 A1
20160165361 Miller et al. Jun 2016 A1
20160255433 Grinker Sep 2016 A1
Foreign Referenced Citations (54)
Number Date Country
204119490 Jan 2015 CN
204145685 Feb 2015 CN
204168483 Feb 2015 CN
204669605 Sep 2015 CN
204681587 Sep 2015 CN
204681593 Sep 2015 CN
ZL2015203769650 Sep 2015 CN
ZL2015204747042 Sep 2015 CN
ZL2015204903074 Sep 2015 CN
915826 Jul 1954 DE
3723275 Mar 1988 DE
102009051713 May 2011 DE
102011003470 Aug 2012 DE
0124870 Nov 1984 EP
0500985 Sep 1992 EP
0684750 Nov 1995 EP
0806909 Nov 1997 EP
1299988 Apr 2003 EP
1509065 Feb 2005 EP
1310136 Mar 2006 EP
1469701 Apr 2008 EP
2434780 Mar 2012 EP
S5888996 May 1983 JP
S60103798 Jun 1985 JP
2007150743 Jun 2007 JP
2012169828 Sep 2012 JP
5049312 Oct 2012 JP
20110058769 Jun 2011 KR
101194904 Oct 2012 KR
1020140026722 Mar 2014 KR
WO8303733 Oct 1983 WO
WO9407342 Mar 1994 WO
WO9623443 Aug 1996 WO
WO0025551 May 2000 WO
WO0217835 Mar 2002 WO
WO0217836 Mar 2002 WO
WO0217837 Mar 2002 WO
WO0217838 Mar 2002 WO
WO0217839 Mar 2002 WO
WO03073790 Sep 2003 WO
WO2006114767 Nov 2006 WO
WO2007073818 Jul 2007 WO
WO2007082579 Jul 2007 WO
WO2007147416 Dec 2007 WO
WO2008128173 Oct 2008 WO
WO2009012491 Jan 2009 WO
WO2009023784 Feb 2009 WO
WO2011051469 May 2011 WO
WO2011061483 May 2011 WO
WO-2012093343 Jul 2012 WO
WO2013033001 Mar 2013 WO
WO2016085814 Jun 2016 WO
WO2016089671 Jun 2016 WO
WO2016089745 Jun 2016 WO
Non-Patent Literature Citations (21)
Entry
Hegde, Nagaraj, “Seamlessly Interfacing MEMS Microphones with BlackfinTM Processors”, EE350 Analog Devices, Rev. 1, Aug. 2010, pp. 1-10.
Korean Office Action regarding Application No. 10-2014-7008553, dated May 21, 2015.
Written Opinion of the International Searching Authority and International Search Report mailed Jan. 21, 2013 in Patent Cooperation Treaty Application No. PCT/US2012/052478, filed Aug. 27, 2012.
Langberg, Mike, “Bluelooth Sharpens its Connections,” Chicago Tribune, Apr. 29, 2002, Business Section, p. 3, accessed Mar. 11, 2016 at URL: <http://articles.chicagotribune.com/2002-04-29/business/0204290116—1—bluetooth-enabled-bluetooth-headset-bluetooth-devices>.
Duplan Corporaton vs. Deering Milliken decision, 197 USPQ 342.
Combined Bluetooth Headset and USB Dongle, Advance Information, RTX Telecom A/S, vol. 1, Apr. 6, 2002.
Ephraim, Y. et al., “Speech enhancement using a minimum mean-square error short-time spectral amplitude estimator,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. ASSP-32, No. 6, Dec. 1984, pp. 1109-1121.
Sun et al., “Robust Noise Estimation Using Minimum Correction with Harmonicity Control.” Conference: INTERSPEECH 2010, 11th Annual Conference of the International Speech Communication Association, Makuhari, Chiba, Japan, Sep. 26-30, 2010. p. 1085-1088.
Lomas, “Apple Patents Earbuds With Noise-Canceling Sensor Smarts,” Aug. 27, 2015. [retrieved on Sep. 16, 2015]. TechCrunch. Retrieved from the Internet: <URL: http://techcrunch.com/2015/08/27/apple-wireless-earbuds-at-last/>. 2 pages.
Smith, Gina, “New Apple Patent Applications: The Sound of Hearables to Come,” aNewDomain, Feb. 12, 2016, accessed Mar. 2, 2016 at URL: <http://anewdomain.net/2016/02/12/new-apple-patent-applications-glimpse-hearables-come/>.
Qutub, Sarmad et al., “Acoustic Apparatus with Dual MEMS Devices,” U.S. Appl. No. 14/872,887, filed Oct. 1, 2015.
Office Action dated Feb. 4, 2016 in U.S. Appl. No. 14/318,436, filed Jun. 27, 2014.
Office Action dated Jan. 22, 2016 in U.S. Appl. No. 14/774,666, filed Sep. 10, 2015.
International Search Report and Written Opinion for Patent Cooperation Treaty Application No. PCT/US2015/062940 dated Mar. 28, 2016 (10 pages).
International Search Report and Written Opinion for Patent Cooperation Treaty Application No. PCT/US2015/062393 dated Apr. 8, 2016 (9 pages).
International Search Report and Written Opinion for Patent Cooperation Treaty Application No. PCT/US2015/061871 dated Mar. 29, 2016 (9 pages).
Yen, Kuan-Chieh et al., “Microphone Signal Fusion”, U.S. Appl. No. 14/853,947, filed Sep. 14, 2015.
Yen, Kuan-Chieh et al., “Audio Monitoring and Adaptation Using Headset Microphones Inside User's Ear Canal”, U.S. Appl. No. 14/985,187, filed Dec. 30, 2015.
Miller, Thomas E. et al., “Voice-Enhanced Awareness Mode”, U.S. Appl. No. 14/985,112, filed Dec. 30, 2015.
Verma, Tony, “Context Aware False Acceptance Rate Reduction”, U.S. Appl. No. 14/749,425, filed Jun. 24, 2015.
International Search Report and Written Opinion, PCT/US2016/069020, Knowles Electronics, LLC, 10 pages (May 2, 2017).
Related Publications (1)
Number Date Country
20170193974 A1 Jul 2017 US