This disclosure relates to assisting conversation, and in particular, to allowing two or more headset users near each other in a noisy environment to speak with ease and hear each other with ease.
Carrying on a conversation in a noisy environment, such as a factory floor, aircraft, or crowded restaurant can be very difficult. In particular, the person speaking has trouble hearing their own voice, and must raise it above what may be a comfortable level just to hear themselves, let alone for the other person to hear them. The speaker may also have difficulty gauging how loudly to speak to allow the other person to hear them. Likewise, the person listening must strain to hear the person speaking, and to pick out what was said. Even with raised voices, intelligibility and listening ease suffer. Additionally, speaking loudly can disturb others nearby, and reduce privacy.
Various solutions have been attempted to reduce these problems. Hearing aids intended for those with hearing loss may attempt to amplify the voice of a person speaking to the user while rejecting unwanted noise, but they suffer from poor signal-to-noise ratio due to limitations of the microphone being located at the ear of the listener. Also, hearing aids provide only a listening benefit, and do not address the discomfort of straining to speak loudly. Other communication systems, such as noise-canceling, intercom-connected headsets for use by pilots, may be quite effective for their application, but are tethered to the dashboard intercom, and are not suitable for use by typical consumers in social or mobile environments or, even in an aircraft environment, i.e., by commercial passengers.
In general, in one aspect, a portable system for enhancing communication between at least two users in proximity to each other includes first and second noise-reducing headsets, each headset including an electroacoustic transducer for providing sound to a respective user's ear and a voice microphone for detecting sound of the respective user's voice and providing a microphone input signal. A first electronic device integral to the first headset and in communication with the second headset generates a first side-tone signal based on the microphone input signal from the first headset, generates a first voice output signal based on the microphone input signal from the first headset, combines the first side-tone signal with a first far-end voice signal associated with the second headset to generate a first combined output signal, and provides the first combined output signal to the first headset for output by the first headset's electroacoustic transducer.
Implementations may include one or more of the following, in any combination. The first electronic device may be coupled directly to the second headset, and the electronic device may generate a second side-tone signal based on the microphone input signal from the second headset, generate the first far-end voice signal based on the microphone input signal from the second headset, combine the second side-tone signal with the first voice output signal to generate a second combined output signal, and provide the second combined output signal to the second headset for output by the second headset's electroacoustic transducer. A second electronic device may be integral to the second headset, the first electronic device may be in communication with the second headset through the second electronic device, and the second electronic device may generate a second side-tone signal based on the microphone input signal from the second headset, generate a second voice output signal based on the microphone input signal from the second headset, provide the second voice output signal to the first electronic device as the first far-end voice signal, receive the first voice output signal from the first electronic device as a second far-end voice signal, combine the second side-tone signal with the second far-end voice signal to generate a second combined output signal, and provide the second combined output signal to the second headset for output by the second headset's electroacoustic transducer. A second electronic device may be integral to the second headset, the first electronic device may be in communication with the second headset through the second electronic device, the second electronic device may transmit the microphone input signal from the second headset to the first electronic device, while the first electronic device generates a second side-tone signal based on the microphone input signal from the second headset, generates a second voice output signal for use as the first far-end voice signal based on the microphone input signal from the second headset, combines the second side-tone signal with the first voice output signal as a second far-end voice signal to generate a second combined output signal, and transmits the second combined output signal to the second electronic device, and the second electronic device may be configured to receive the second combined output signal and provide it to the second headset for output by the second headset's electroacoustic transducer.
The voice microphone of the first headset and the first electronic device may be configured to generate the first microphone input signal by rejecting surrounding noise while detecting the respective user's voice. The first and second headsets may each include a noise cancellation circuit including a noise cancellation microphone for providing anti-noise signals to the respective electroacoustic transducer based on the noise cancellation microphone's output, and the first electronic device may be configured to provide the first combined output signal to the first headset for output by the first headset's electroacoustic transducer in combination with the anti-noise signals provided by the first headsets's noise cancellation circuit. The first and second headsets may each include passive noise reducing structures. Generating the first side-tone signal may include applying a frequency-dependent gain to the microphone input signal from the first headset. Generating the first side-tone signal may include filtering the microphone input signal from the first headset and applying a gain to the filtered signal. The first electronic device may control gains applied to the first side-tone signal and the first voice output signal. The first electronic device may control gains applied to the first side-tone signal and the first far-end voice signal when generating the first combined output signal. The first electronic device may control the gains applied to the signals under the direction of a user of the first headset. The first electronic device may control the gains applied to the signals automatically. The first electronic device may control gains applied to the first side-tone signal and the first voice output signal, and control a further gain applied to the first far-end voice signal.
A third noise-reducing headset may be involved, the third headset including an electroacoustic transducer for providing sound to a respective user's ear, and a voice microphone for detecting sound of the respective user's voice and providing a microphone input signal. A second electronic device may be integral to the second headset, and a third electronic device integral to the third headset, with the first electronic device in communication with the second and third headsets through the respective second and third electronic devices, and the far-end voice signal received by the first electronic device may includes voice output signals from both the second and third headsets. The first far-end voice signal received by the first electronic device may include the first voice output signal, and the first device may remove the first voice output signal from the first far-end voice signal before combining the first far-end voice signal with the first side-tone signal to generate the first combined output signal.
The first electronic device may be in communication with the third headset through the third electronic device, and the third electronic device may generate a third side-tone signal based on the microphone input signal from the third headset, generate a third voice output signal based on the microphone input signal from the third headset, transmit the third voice output signal to the first and second electronic devices for use as the first and second far-end voice signals, receive the first voice output signal from the first electronic device and the second voice output signal from the second electronic device, combine the third side-tone signal with the first and second voice output signals as far-end voice signals to generate a third combined output signal, and provide the third combined output signal to the third headset for output by the third headset's electroacoustic transducer. The second electronic device may be in communication with the third headset through the third electronic device. The second electronic device may be in communication with the third headset through the third electronic device by way of the first electronic device.
In general, in one aspect, a noise-reducing headset for use in a portable system for enhancing communication between at least two users in proximity to each other includes an electroacoustic transducer for providing sound to a user's ear, a voice microphone for detecting sound of the user's voice and providing a microphone input signal, and an electronic circuit integral to the headset and including an interface for communication with a second headset. The electronic device generates a first side-tone signal based on the microphone input signal, generate a first voice output signal based on the microphone input signal, combine the first side-tone signal with a first far-end voice signal associated with the second headset to generate a first combined output signal, and provide the first combined output signal to the transducer for output.
Implementations may include one or more of the following, in any combination. The electronic circuit may apply gains to the first side-tone signal and the first voice output signal. The electronic circuit may apply gains to the first side-tone signal and the first far-end voice signal when generating the first combined output signal.
Advantages include allowing users to engage in conversation in a noisy environment, including hearing their own voice, being heard by their conversation partners, and hearing their partners' voices, all without straining to hear or to speak, and without disturbing others.
All examples and features mentioned above can be combined in any technically possible way. Other features and advantages will be apparent from the description and the claims.
A system for allowing two or more headset users near each other in a noisy environment to speak with ease and hear each other with ease includes two headsets and at least one electronic device in communication with both headsets, as shown in
In some examples, as shown in
Although the headsets are shown as connected to the electronic devices by wires, the connection could be wireless, using any suitable wireless communication method, such as Bluetooth®, WiFi, or a proprietary wireless interface. In addition to the headsets, the electronic devices may be in communication with each other using wired or wireless connections. The wireless connections used for communication between the electronic devices may be different than that used with the headsets. For example, the headsets may use Bluetooth to communicate with their respective electronic devices, while the electronic devices use WiFi to communicate with each other. The electronic devices may also use more than one method simultaneously to communicate with each other. Throughout this application, we refer to various acoustic and electronic signals flowing within and between headsets and electronics. The names of the signals and their references in the figures are listed in
In the electronic device or devices, the voice microphone signals from each headset are handled in two different ways, as shown in
Each system includes a voice microphone 206 receiving a voice audio input V1 or V2, a first equalization stage 207, a first gain stage 208, a second equalization stage 209, a second gain stage 210, an attenuation block 212, and an output summation node 214 providing an audio output Out1 or Out2. The voice audio inputs V1 and V2 represent the actual voice of the user, and the audio outputs Out1 and Out2 are the output acoustic signals heard by the users. The microphones 206 also detect ambient noise N1 and N2 and pass that on to the gain stages, filtered according to the microphone's noise rejection capabilities. The microphones are more sensitive to the voice input than to ambient noise, by a noise rejection ratio M. The combined signals 211 from the microphones, V1+N1/M and V2+N2/M, may be referred to as microphone input signals. Within those signals, N1/M and N2/M represent unwanted background noise. Different ambient noise signals N1 and N2 are shown entering the two systems, but depending on the distance between the users and the acoustic environment, the noises may be effectively the same. Ambient noises N3 and N4 at the users ears, which may also be the same as N1 or N2, are attenuated by the attenuation block 212 in each system, which represents the combined passive and active noise reduction capability, if any, of the headsets. The resulting residual noise is shown entering the output summation node, though in actual implementation, the electronic signals are first summed and output by the output transducer, and the output of the transducer is acoustically combined with the residual noise within the user's ear canal. That is, the output summation node 214 represents the output transducer in combination with its acoustic environment, as shown in more detail in
The two circuits 202 and 204 apply the same processing to the two microphone input signals. First, each microphone input signal is filtered by the first equalization stage 207, which applies a filter Ks, and amplified by the first gain stage 208, which applies a gain G. The filter Ks and gain Gs change the shape and level of the voice signal to optimize it for use as a side-tone signal. When a person cannot hear his own voice, such as in loud noise, he will tend to speak more loudly. This has the effect of straining the speaker's voice. On the other hand, if a person in a noisy environment is wearing noise isolating or noise canceling headphones, he will tend to speak at a comfortable, quieter level, but also will suffer from the occlusion effect, which inhibits natural, comfortable speaking. The occlusion effect is the change in how a person's voice sounds to themselves when the ear is covered or blocked. For example, occlusion may produce low-frequency amplification, and cause a person's voice to sound unnatural to themselves. A side-tone signal is a signal played back to the ear of the speaker, so that he can hear his own voice. If the side-tone signal is appropriately scaled, the speaker will intuitively control the level of his voice to a comfortable level, and be able to speak naturally. The side-tone filter Ks shapes the voice signal to compensate for the way the occlusion effect changes the sound of a speaker's voice when his ear is plugged, so that in addition to being at the appropriate level, the side-tone signal sounds, to the user, like his actual voice sounds when not wearing a headset.
The microphone input signal 211 is also equalized and scaled by the second filter 209 and gain stage 210, applying a voice output filter Ko and a voice output gain G0. The voice output filter and gain are selected to make the voice signal from one headset's microphone audible and intelligible to the user of the second headset, when played back in the second headset. The filtered and scaled voice output signals 213 are each delivered to the other headset, where they are combined with the filtered and scaled side-tone signals 215 within each headset to produce a combined audio output Out1 or Out2. When discussing one headset, we may refer to the voice output signal 213 from the other headset, played back by the headset under consideration, as the far-end voice signal. As mentioned above, the microphones 206 pick up ambient noise N1 and N2, and deliver that to the filter and gain stages along with voice signals V1 and V2. Ambient noise N3 and N4 are attenuated by noise reduction features of the headsets, whether active or only passive, shown as attenuation blocks A, such that an attenuated noise signal A•N3 or A•N4 is heard in each headset, along with the combined side-tone signal 215 and far-end voice signal 213 (i.e., the voice output signal from the other headset), the side-tone signal and far-end voice signal each including the unwanted background noise N1/M and N2/M from their respective microphones.
The gain Gs is selected, taking into consideration the noise rejection capabilities of the voice microphones and the noise attenuation capabilities of the headsets, to provide the side-tone signal at a level that will allow the user to hear his own voice over the residual noise and naturally speak at a comfortable level. At the same time, the gain Go is selected, taking the same factors into account, to provide the voice output signals to each headset at a level that will allow each user to hear the other user's voice at a comfortable and intelligible level. In some examples, the gain Gs is set to balance the user's own comfort, by providing an appropriate side-tone level, with making sure the user speaks loudly enough for the voice microphone to detect the speaker's voice with enough signal-to-noise (SNR) ratio to provide a useful voice signal. The circuits shown in
The examples of
In some examples, as shown in
As can be seen in
The electronic signals to be output, which include the side-tone signal Gs(V1+N1/M), far-end voice signal (voice output signal Vo2 from the other headset), and anti-noise signal Aa•N3, are summed electronically to produce a combined output signal 511 at the input 214a of the output electroacoustic transducer 510. The acoustic output of the transducer is then summed acoustically with the residual noise Ap•N3 penetrating the headphone, represented as an acoustic sum 214b, to form the audio output Out1 referred to in earlier figures. The combined acoustic signals of the audio output are detected by both the feed-back microphone 506 and the eardrum 512.
Embodiments of the systems and methods described above comprise computer components and computer-implemented steps that will be apparent to those skilled in the art. For example, it should be understood by one of skill in the art that the computer-implemented steps may be stored as computer-executable instructions on a computer-readable medium such as, for example, Flash ROMS, nonvolatile ROM, and RAM. Furthermore, it should be understood by one of skill in the art that the computer-executable instructions may be executed on a variety of processors such as, for example, microprocessors, digital signal processors, gate arrays, etc. For ease of exposition, not every step or element of the systems and methods described above is described herein as part of a computer system, but those skilled in the art will recognize that each step or element may have a corresponding computer system or software component. Such computer system and/or software components are therefore enabled by describing their corresponding steps or elements (that is, their functionality), and are within the scope of the disclosure.
A number of implementations have been described. Nevertheless, it will be understood that additional modifications may be made without departing from the scope of the inventive concepts described herein, and, accordingly, other embodiments are within the scope of the following claims.
Number | Name | Date | Kind |
---|---|---|---|
3889059 | Thompson et al. | Jun 1975 | A |
3992584 | Dugan | Nov 1976 | A |
3999015 | Snyder et al. | Dec 1976 | A |
4600208 | Morishima | Jul 1986 | A |
4941187 | Slater | Jul 1990 | A |
5243659 | Stafford et al. | Sep 1993 | A |
5640450 | Watanabe | Jun 1997 | A |
5983183 | Tabet et al. | Nov 1999 | A |
6493450 | Scheuer et al. | Dec 2002 | B1 |
7065198 | Brown et al. | Jun 2006 | B2 |
7260231 | Wedge | Aug 2007 | B1 |
7620419 | Gandolfo | Nov 2009 | B1 |
7627352 | Gauger, Jr. et al. | Dec 2009 | B2 |
8139744 | Carlson et al. | Mar 2012 | B2 |
8208650 | Joho et al. | Jun 2012 | B2 |
8363820 | Graham | Jan 2013 | B1 |
20020059064 | Tabata et al. | May 2002 | A1 |
20030118197 | Nagayasu et al. | Jun 2003 | A1 |
20050282592 | Frerking et al. | Dec 2005 | A1 |
20060293092 | Yard et al. | Dec 2006 | A1 |
20080037749 | Metzger et al. | Feb 2008 | A1 |
20080201138 | Visser et al. | Aug 2008 | A1 |
20090023417 | Davis et al. | Jan 2009 | A1 |
20100119077 | Platz et al. | May 2010 | A1 |
20100150383 | Sampat | Jun 2010 | A1 |
20110033064 | Johnson et al. | Feb 2011 | A1 |
20110044474 | Grover et al. | Feb 2011 | A1 |
20110135086 | Sun et al. | Jun 2011 | A1 |
20110288860 | Schevciw et al. | Nov 2011 | A1 |
20110293103 | Park et al. | Dec 2011 | A1 |
20120140941 | Kuhtz et al. | Jun 2012 | A1 |
20150063601 | Briggs et al. | Mar 2015 | A1 |
Number | Date | Country |
---|---|---|
2555189 | Feb 2013 | EP |
S57124960 | Aug 1982 | JP |
9911045 | Mar 1999 | WO |
9911047 | Mar 1999 | WO |
2009097009 | Aug 2009 | WO |
Entry |
---|
International Search Report and Written Opinion dated Oct. 2, 2014 for International application No. PCT/US2014/049741. |
International Search Report and Written Opinion dated Oct. 27, 2014 for International application No. PCT/US2014/049750. |
Number | Date | Country | |
---|---|---|---|
20150063584 A1 | Mar 2015 | US |