Earbuds transmit and receive sound signals, convert sound signals to electromagnetic signals, and transmit and receive electromagnetic signals. A challenge is to reduce the size and weight of the earbud while enhancing the transmission and reception characteristics of the sound and electromagnetic signals.
In general, in one aspect, one or more embodiments relate to a method that uses an earloop microphone. A first audio signal is received from a first microphone acoustically coupled to a first opening in a headset. A second audio signal is received from a second microphone acoustically coupled to a second opening in the headset. The second opening and the first opening are separated by a first spacing. The first spacing creates first phase and amplitude differences between the second audio signal and the first audio signal. A third audio signal is received from a third microphone acoustically coupled to a third opening in an earloop of the headset. The third opening and the second opening are separated by a second spacing. The second spacing creates second phase and amplitude differences between the third audio signal and the first audio signal. A source signal is identified using the first phase and amplitude differences and the second phase and amplitude differences. A gain is applied to amplify the source signal.
In general, in one aspect, one or more embodiments relate to an apparatus that includes an earloop, a processor, a memory connected to the processor, a first microphone acoustically coupled to a first opening, a second microphone acoustically coupled to a second opening, a third microphone acoustically coupled to a third opening in the earloop, and program code stored on the memory that is executed by the processor. A first audio signal is received from a first microphone acoustically coupled to a first opening in a headset. A second audio signal is received from a second microphone acoustically coupled to a second opening in the headset. The second opening and the first opening are separated by a first spacing. The first spacing creates first phase and amplitude differences between the second audio signal and the first audio signal. A third audio signal is received from a third microphone acoustically coupled to a third opening in an earloop of the headset. The third opening and the second opening are separated by a second spacing. The second spacing creates second phase and amplitude differences between the third audio signal and the first audio signal. A source signal is identified using the first phase and amplitude differences and the second phase and amplitude differences. A gain is applied to amplify the source signal.
In general, in one aspect, one or more embodiments relate to a headset that implements an earloop microphone and includes a housing. An earloop of the headset secures the headset to an ear of a user. A first microphone is acoustically coupled to a first opening in the housing. A second microphone is acoustically coupled to a second opening in the housing. A third microphone is acoustically coupled to a third opening in the earloop.
Other aspects of the invention will be apparent from the following description and the appended claims.
Specific embodiments of the invention will now be described in detail with reference to the accompanying figures. Like elements in the various figures are denoted by like reference numerals for consistency.
In the following detailed description of embodiments of the invention, numerous specific details are set forth in order to provide a more thorough understanding of the invention. However, it will be apparent to one of ordinary skill in the art that the invention may be practiced without these specific details. In other instances, well-known features have not been described in detail to avoid unnecessarily complicating the description.
Throughout the application, ordinal numbers (e.g., first, second, third, etc.) may be used as an adjective for an element (i.e., any noun in the application). The use of ordinal numbers is not to imply or create any particular ordering of the elements nor to limit any element to being only a single element unless expressly disclosed, such as by the use of the terms “before”, “after”, “single”, and other such terminology. Rather, the use of ordinal numbers is to distinguish between the elements. By way of an example, a first element is distinct from a second element, and the first element may encompass more than one element and succeed (or precede) the second element in an ordering of elements.
In general, one or more embodiments of the disclosure reduce the size and weight of the earbuds while enhancing the transmission and reception characteristics of the sound and electromagnetic signals with an earloop microphone. A microphone, of a microphone array, is placed in the earloop of an earbud to increase the spacing between the microphones.
Increased spacing between the microphones increases the phase and amplitude differences between sound signals from the same sound source. The phase and amplitude differences may be used by sound source identification algorithms and beamforming algorithms to amplify (apply a gain) to sound signals from a particular source, e.g., the user of the earbuds.
The phase difference between two audio signals, generated from a sound signal of a sound source, is the difference between a reference point that occurs in both of the audio signals. The phase difference between the two audio signals identifies how much the sound signal captured in one audio signal is shifted in time with respect to the sound signal captured in the other audio signal. The phase difference may be measured in radians or degrees. The amplitude difference between two audio signals is the difference between the extreme values (e.g., peak values) of the audio signals.
Embodiments of the disclosure may also locate an antenna in the earloop of the earbud. The antenna may be colocated with or connected to the structures of the microphone in the earloop. The antenna may be part of a set of antennas used by the earbud to communicate with a media device for interactive voice communication with the user of the earbud.
Turning to
The earloop (106) is a part of the housing (104) that extends from the base (120) and wraps behind the cartilage of the ear of the user. The earloop (106) may wrap behind the helix of the ear of the user. The earloop (106) fits between the head of the user and the ear and secures the headset A (102) to the user. The earloop (106) includes the antenna (108) and the opening C (110). In one embodiment, the earloop (106) is formed as part of, and is an extension to, the base (120). The cross-sectional thickness of the earloop (106), in the dimension perpendicular to the skull of the user, may be about 1.5 millimeters. In additional embodiments, the cross-sectional thickness may range from about 1 millimeter to about 8 millimeters.
The antenna (108) is located in the earloop (106). The antenna (108) connects to the circuitry (121) in the headset A (102), e.g., the data interface adapter (176) (of
The opening C (110) is located on the earloop (106). In general, an opening is one or more holes in the housing that allow for the passage of sound signals. The opening C (110) allows sound signals (acoustic waves) to reach the microphone C (114). The opening C (110) is formed with the direction C (112), which points in a direction perpendicular to a plane formed by the opening C (110). In one embodiment, the other directions A (124) and B (130) of the openings A (122) and B (128) may be different from the direction C (112) of the opening C (110).
The microphone C (114) is acoustically coupled to the opening C (110). The microphone C (114) may be located in the earloop (106). In one embodiment, the microphone C (114) may be located in the base (120) and acoustically coupled to the opening C (110) through an acoustic waveguide (e.g., a cavity) extending from the base (120) into the earloop (106) to the opening C (110).
The base (120) is part of the housing (104) that includes the openings A (122) and B (128) and contains other components of the headset A (102), including the circuitry (121). The circuitry (121) includes the electronic components of the headset A (102), which includes, from
The openings A (122) and B (128) are located at different positions on the base (120). In one embodiment the openings A (122) and B (128) are at least about 20 millimeters apart. The openings A (122) and B (128) are respectively formed with the directions A (124) and B (130), which point in directions perpendicular to planes formed by the openings A (122) and B (128). In one embodiment, the directions A (124) and B (130) may be different from each other without affecting the phase and amplitude differences in the signals captured by the microphones A (126) and B (132). In one embodiment, the microphone pair axis that passes through the centers of the openings A (122) and B (128) may point towards the mouth of the user.
The microphones A (126) and B (132) are acoustically coupled to the openings A (122) and B (128). In one embodiment, the microphones A (126) and B (132) may be colocated with the openings A (122) and B (128) in the base (120). One or both of the microphones A (126) and B (132) may also be acoustically coupled to the openings A (122) and B (128) with acoustic waveguides to separate the microphones A (126) and B (132) away from the location of the openings A (122) and B (128).
Turning to
The headset A (102) includes several components to send and receive sound signals, data signals, electromagnetic signals, etc. The headset A (102) may be an embedded device as described below with reference to the computing system (800) of
The housing (104) of the headset A (102) covers the components of the headset A (102). In one embodiment, the earloop (106) is integrally formed as a part of the housing (104). The housing (104) may be shaped to fit a left ear or a right ear of the user.
The earloop (106) secures the headset A (102) to the user by looping around the cartilage of the ear of the user. In one embodiment, the earloop (106) includes the opening C (110), the microphone C (114), and the antenna (108).
The openings (152) include the openings A (122) (of
The microphones (154) include the microphones A (126) (of
The antennas (156) include the antenna (108). The antennas (156) convert between free space electromagnetic signals and electrical signals in the headset (102). Electromagnetic signals propagate through the space around the headset A (102) and the electrical signals (also referred to as data signals) propagate between the processor (170) and the antennas (156) using the data interface adapter (176). The signal reception and transmission allows data communications to be sent to and received from the headset A (102).
The speakers (158) include the speaker (159). The speakers (158) generate the sound signals that are transmitted to the ear of the user from the audio signals generated by the processor (170).
The processor (170) is a set of one or more processors that receives, processes, and transmits data using electrical signals between the components of the headset A (102). The processor (170) may include one or more embedded processors, digital signal processors (DSPs), systems on chip (SoCs), etc. The processor (170) reads instructions from the memory (172) to process the signals received from the microphones (154) and antennas (156) and generate signals transmitted by the speakers (158) and the antennas (156). In one embodiment, the processor (170) executes instructions from the memory to receive audio signals from the microphones (154), identify a source signal from the audio signals using phase and amplitude differences between the audio signals, and applies a gain to amplify the source signal.
The memory (172) is a set of one or more memories that stores data and instructions captured and used by the headset A (102), including the program code (174). The program code (174) includes the instructions for converting the sound signals from the microphones (154) to audio signals, converting electromagnetic signals from and to the antennas (156) to data signals, and converting data signals to audio signals sent to the speakers (158).
In one embodiment, the program code (174) includes programs for locating sound signal sources (e.g., the user of the headset A (102)) and amplifying selected sound signals from selected sources. For example, with execution of the program code (174) by the processor (170), the headset A (102) may amplify the speech of the user of the headset A (102) by about 20 decibels (dB). The amplification is generated by processing the data signals converted from the sound signals received from the microphones (154) through the openings (152). The spacing between the openings (152) (and the microphones (154)) sense phase and amplitude differences in the sound signals for the sources of the sounds in the sound signals. The phase and amplitude differences are used to identify the source of the sounds and selectively amplify the sound of the speech of the user of the system (100).
The data interface adapter (176) includes components and protocols that transmit and receive data signals to and from the headset A (102). In one embodiment, the data interface adapter (176) includes the antenna (108) and uses a protocol for a personal area network to send and receive data between the headset A (102), the headset B (180), and the media device (182). Through the data interface adapter (176), the headset A (102) may receive data signals from the headset B (180) that correspond to sound signals from the microphones of the headset B (180). The sound signals from the headset B (180) may be used in conjunction with the sound signals from the headset A (102) by the program code (174) to identify and amplify the speech of the user.
The battery (178) is a source of energy. The battery (178) provides electrical power to the components of the headset A (102).
The headset B (180) is complimentary to the headset A (102) and may be configured for the other ear of the user of the system (100). For example, the headset A (102) may be configured for the left ear of the user and the headset B (180) may be configured for the right ear of the user. The hardware and software components and structure may be similar to that of the headset A (102).
The media device (182) includes a computing system, as described in
Turning to
In Block 204, a second audio signal is received from a second microphone acoustically coupled to a second opening in the headset. The second opening and the first opening are separated by a first spacing. The first spacing causes a first amplitude and phase difference between the second audio signal and the first audio signal for the source signal. The two microphones sample the source signal (also referred to as a sound signal) at different points along the wavelength of the source signal as governed by the frequency of the sound and the speech of sound in the source signal. Amplitude of the source signal is governed by the inverse square law equating amplitude to distance from the source. Both of these properties, phase and amplitude, may be used to identify the source signal. In one embodiment, the first spacing between the first opening and the second opening is in the range of about 10 millimeters to about 30 millimeters.
In Block 206, a third audio signal is received from a third microphone acoustically coupled to a third opening in an earloop of the headset. The third opening and the second opening are separated by a second spacing. The second spacing creates a second phase and amplitude differences between the third audio signal and the first audio signal for the source signal. The earloop is configured to secure the headset to an ear of a user. The first opening and the third opening may be separated by a third spacing. The third spacing may be about 30 millimeters or more. In one embodiment, the third spacing may be about 40 millimeters.
The openings may each face different directions without affecting the differences in phase and amplitude. The openings sample the sound wave at different points in space resulting in different amplitudes and phases for the source signal. The differences in amplitude may be used by the headset to identify the location of the source (e.g., the mouth of the user of the headset) in combination with the phase differences created by spacings of the openings.
In one embodiment, a fourth audio signal from a fourth microphone acoustically coupled to a fourth opening may be received. The fourth audio signal includes additional phase and amplitude differences for the source signal with respect to the other audio signals and is used to increase the accuracy of the source signal amplification.
In one embodiment, one or more audio signals may be received from a second headset coupled to a second ear of a user. The audio signals from the second headset may be transmitted wirelessly from the second headset to the first headset. The first headset may process the one or more audio signals having additional phase and amplitude differences to increase the accuracy of the source signal amplification.
In Block 208, a source signal is identified using the first phase and amplitude differences and the second phase and amplitude differences. Identification of the source signal may be performed by the processor of the headset with a signal source identification algorithm. The signal source identification algorithm may identify multiple sources of sound signals in the combined audio signals and identify the locations of the sources relative to the location of the headset. The sound source located at the appropriate direction and distance to the headset may be identified as the source signal.
The voice or source signal is identified and separated from the background noise using the multiple microphones and the time difference of arrival. With the different time differences of arrival and the known spacing between the openings, this sound signal may be identified as speech. If sound or noise is captured by each of the microphones all at roughly the same time, this sound or noise may be identified as background noise and not speech from the direction of the mouth of the user. By utilizing three or more microphones, speech of the user (i.e., the desired signal) is more accurately identified by triangulating on the direction of sound. Microphone spacings of between about 10 millimeters and about 30 millimeters may be used to generate sufficient time differences of arrival and phase differences in the signals received by the headset.
In one embodiment, the headset further uses third phase and amplitude differences between the third audio signal and the second audio signal to identify the source signal. In one embodiment, the source signal is identified by further using a fourth audio signal from a fourth microphone of the headset.
In one embodiment, the source signal is identified using three or more audio signals from a second headset. Instead of merely identifying the closest signal to the headset, the headset may identify the closest signal that is between the two headsets.
In Block 210, a gain is applied to amplify the source signal. The gain increases the amplitude of the source signal with respect to the background noise. In one embodiment, the gain is about 20 decibels or more.
In one embodiment, the headset converts the source signal to an electromagnetic signal. The headset may transmit, using an antenna proximate to the earloop, the electromagnetic signal as part of an interactive voice communication.
Turning to
The opening A (306) and the opening B (308) are aligned to form a line that points to the mouth location (322) of a user. The mouth location (322) is the location of the source signal in the sound signals and audio signals received and generated by the headset (300).
The openings A (306) and B (308) are separated by a spacing that may be about 20 millimeters. The openings A (306) and C (310) are separated by a spacing that is greater than the spacing between the openings A (306) and B (308), which may be about 40 millimeters.
The spacings between the openings A (306), B (308), and C (310) create phase and amplitude differences in the sound signals received by the headset (300). The phase and amplitude differences may be identified by the headset and used to determine the location of source signals from the audio signals captured by the headset (300).
Turning to
The openings A (406), B (408), and C (410) respectively face the directions A (416), B (418), and C (420). The sound signal from the users mouth may have a higher amplitude for the opening A (406) than for the opening C (410) due to the different distances from the mouth of the user to the openings A (406) and C (410). The differences in amplitude may be proportional to the differences in the distances from the mouth of the user to the openings A (406), B (408), and C (410).
The headset uses the amplitude differences and the phase differences to identify the source signal in the audio signals captured from the sound signals by the headset (400). Once the source signal for the user is identified, the source signal for the user is preferentially amplified above the background noise.
Turning to
The openings A (506) and B (508) are aligned with the mouth location (522) of the user. The spacing A (532) between the openings A (506) and B (508) is about the same as the spacing B (534) between the openings B (508) and C (510).
The spacings between the openings A (506), B (508), and C (510) create phase and amplitude differences between the audio signals captured by the headset (500). The phase and amplitude differences are used to identify and amplify the source signal of the speech of the user in the audio signals captured by the headset (500).
Turning to
The openings A (606) and B (608) are aligned with the mouth location (622) of the user. The spacings between the openings A (606), B (608), and C (610) create phase and amplitude differences between the audio signals captured by the headset (600).
Turning to
The openings A (706) and B (708) are aligned in a linear vertical arrangement. face. The spacings between the openings A (706), B (708), and C (710) create phase and amplitude differences between the audio signals captured by the headset (700). In one embodiment, the openings A (706), B (708), and C (710) may each face substantially the same direction.
Embodiments of the invention may be implemented on a computing system. Any combination of a mobile, a desktop, a server, a router, a switch, an embedded device, or other types of hardware may be used. For example, as shown in
The computer processor(s) (802) may be an integrated circuit for processing instructions. For example, the computer processor(s) (802) may be one or more cores or micro-cores of a processor. The computing system (800) may also include one or more input device(s) (810), such as a touchscreen, a keyboard, a mouse, a microphone, a touchpad, an electronic pen, or any other type of input device.
The communication interface (812) may include an integrated circuit for connecting the computing system (800) to a network (not shown) (e.g., a local area network (LAN), a wide area network (WAN) such as the Internet, a mobile network, or any other type of network) and/or to another device, such as another computing device.
Further, the computing system (800) may include one or more output device(s) (808), such as a screen (e.g., a liquid crystal display (LCD), a plasma display, a touchscreen, a cathode ray tube (CRT) monitor, a projector, or other display device), a printer, an external storage, or any other output device. One or more of the output device(s) (808) may be the same or different from the input device(s) (810). The input and output device(s) (810 and 808) may be locally or remotely connected to the computer processor(s) (802), non-persistent storage (804), and persistent storage (806). Many different types of computing systems exist, and the aforementioned input and output device(s) (810 and 808) may take other forms.
Software instructions in the form of computer readable program code to perform embodiments of the invention may be stored, in whole or in part, temporarily or permanently, on a non-transitory computer readable medium such as a CD, a DVD, a storage device, a diskette, a tape, flash memory, physical memory, or any other computer readable storage medium. Specifically, the software instructions may correspond to computer readable program code that, when executed by a processor(s), is configured to perform one or more embodiments of the invention.
The computing system (800) of
For example, a GUI may first obtain a notification from a software application requesting that a particular data object be presented within the GUI. Next, the GUI may determine a data object type associated with the particular data object, e.g., by obtaining data from a data attribute within the data object that identifies the data object type. Then, the GUI may determine any rules designated for displaying that data object type, e.g., rules specified by a software framework for a data object class or according to any local parameters defined by the GUI for presenting that data object type. Finally, the GUI may obtain data values from the particular data object and render a visual representation of the data values within a display device according to the designated rules for that data object type.
Data may also be presented through various audio methods. In particular, data may be rendered into an audio format and presented as sound through one or more speakers operably connected to a computing device.
Data may also be presented to a user through haptic methods. For example, haptic methods may include vibrations or other physical signals generated by the computing system. For example, data may be presented to a user using a vibration generated by a handheld computer device with a predefined duration and intensity of the vibration to communicate the data.
The above description of functions presents only a few examples of functions performed by the computing system (800) of
While the invention has been described with respect to a limited number of embodiments, those skilled in the art, having benefit of this disclosure, will appreciate that other embodiments can be devised which do not depart from the scope of the invention as disclosed herein. Accordingly, the scope of the invention should be limited only by the attached claims.