Typically, when a binaural sound recording is produced great care is exercised to accurately capture a location or direction of each of the instruments and vocals in the binaural recording. In some cases, the recording may also be captured in a manner to compensate for disturbances or modification to the sound caused by a listener's body and head. However, when an earbud is placed within an ear or a headset is placed over the ear, the earbud or headset causes a modification to or compression of the ear canal which is not compensated for when the audio is recorded.
The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different figures indicates similar or identical components or features.
This disclosure includes techniques and implementation to measure and map an individual's ear canal and the acoustic response of the ear canal when excited by a personal speaker system. In some cases, the measurements may be used to allow a personal speaker system (such as an earbud or headset device) to modify an output audio signal to compensate for changes imposed on the audio as it travels from the personal speaker system to the individual's drum compared to if the individual listened to the same audio occurring in free space without the personal speaker system. For example, the personal speaker system may change the frequency characteristics of the audio and, thereby, make audio output different from the characteristics of the recording. In another example, the tip of the earbud is typically inserted into the ear canal, thereby, shorting the distance the sound travels within the ear prior to impacting the eardrum which changes the frequency characteristics of the sound. The changes to the audio characteristics caused by use of the personal speaker system result in a human detectable change to the sound that the user may perceive as unnatural.
Today, when a sound recording is produced great care is exercised to accurately capture a location or direction of each of the instruments and vocals to generate the most natural hearing experience possible for a listener. For instance, a binaural audio recording may be produced using two microphones that are separated by the approximate distance between a human's ears. Thus, the recording from the microphone on the right is played into the listener's right ear and the recording from the microphone on the left is played into the listeners left ear with the intention that the listener hears the audio as if he or she were present during the recording. In some cases, the audio recording may be recorded in a manner to also compensate for modifications caused by interface by a listener's body and head. For example, the right and left microphones used to record the audio signal may be mounted on a model that is representative of an average user. Therefore, the audio captured by the right and left microphones has undergone interface by at least an average human body and head.
Further, with the advancements in virtual visual environments, the desires to reproduce a ‘live’ experience as accurately as possible has been enhanced. For instance, the user may be immersed in a three-dimensional (3D) virtual scene representative of the environment in which the visual and audio data is captured. Thus, the user may experience both the visual and audio experience as if the user was present at the time and location the scene was captured. In these cases, great care may also be taken to capture audio in a manner in which the user is able to move through the virtual scene with both visual and audio stimulus changing as if the user moved through a real-life environment. For instance, the audio may be recorded from a plurality of positions within an environment, each of which compensates for the head and body interface. However, failure to compensate for the modification to an ear canal due to user of the personal speaker system may undermine the great care taken to reproduce the visual and audio scene and, thus, ruin the virtual experience being enjoyed by the user.
The system and methods discussed herein, allow for the personal speaker system (e.g., the earbud or headset) to measure, model, and map the listener's ear canal in a manner that allows the personal speaker system to modify the audio being output as sound to compensate for the modifications being introduced by the user of such personal speaker system. In this way, the sound captured by the listener's eardrum is received as if the user listening in a more natural setting devoid of the earbud or headset.
In some example, the system and method discussed herein may include a personal speaker system (such as an earbud or headset) configured to provided binaural audio output to a user. For example, the personal speaker system may be configured with speaker and an in-ear-microphone. The in-ear-microphone may be located within the ear canal or at a location closer to the eardrum than the speaker. The system may first determine a first thevinin equivalent pressure (Pth) and a first thevinin equivalent acoustic impedance (Zth) of the speaker system using the data captured by the in-ear-microphone.
The system may then model the distance between the position of the in-ear-microphone and the inner tip of the earbud as a first acoustic transmission line. For instance, the system may determine a first impedance (Zo1), a first attenuation constant (α1), and a first phase constant (β1) associated with the first acoustic transmission line (e.g., the distance between the in-ear-microphone and the inner tip of the earbud). Next the system may model the abrupt diameter change from the outer tip of the earbud and the diameter of the ear canal as a lumped inductance (Ldis) dependent on the ratio of the bud radius and the ear canal radius.
In some examples, the first thevinin equivalent pressure and the first thevinin equivalent acoustic impedance of the speaker system may be determined prior to assertion into the ear of the listener or known at the time of insertion.
The system may then model the length of the ear canal (e.g., the distance between the outer tip of the earbud and the eardrum) as a second acoustic transmission line. In this case, the system may determine a second impedance (Zo2), a second attenuation constant (α2), and a second phase constant (β2) associated with the second acoustic transmission line. Next the eardrum may be modeled as a complex lumped acoustic impedance (Zdrum).
When the thevinin equivalent pressure and thevinin equivalent acoustic impedance of the speaker system are known, the first impedance, the first attenuation constant, the first phase constant, the lumped inductance, the second impedance, the second attenuation constant, the second phase constant, and the complex lumped acoustic impedance may be determined by the system by outputting known sound into the ear canal and capturing audio signals using an in-ear microphone. The captured audio may then be analyzed to estimate the values above. In some cases, the known audio may be a wideband signal including wideband noise or a chirp signal. In some cases, the audio recording itself me be used as the known audio.
Once the system is able to model the ear canal using the thevinin equivalent pressure and thevinin equivalent acoustic impedance of the speaker, the characteristics of the first acoustic line (e.g., the distance between the speaker and the inner tip of the earbud), the lumped inductance associated with the transition between the outer tip of the earbud and the diameter of the ear canal, the characteristics of the second acoustic line (e.g., the distance between the outer tip of the are bud and the eardrum), and the complex lumped acoustic impedance, the system may utilize the model to provide noise cancellation or modify the audio signals to sound more natural at a various locations within the ear canal including at the ear drum. The system may also utilize the modeled ear canal values to monitor, determine, or compensate for acoustic leakage between the ear and the earbud.
For instance, in one example, the system may determine the acoustic sound pressure at a particular location, such as the eardrum. Then, the system, may determine acoustic sound pressure, the thevinin equivalent pressure and the thevinin equivalent acoustic impedance of the personal speaker system and the determined characteristics of the ear canal to perform noise cancelation at the ear drum and/or to modify an output audio signal to sound more natural (e.g., more like sound heard naturally without the listener engaged with a headsets or earbuds). Since the listener's ear canal and ear drum, along with their responses to the personal speaker are determined per listener the audio adjustments applied by the system are specific for each individual listener.
In one example, the system may determine the thevinin equivalent pressure at the eardrum (Pdrum) and the thevinin equivalent acoustic impedance of the eardrum (Zdrum) in order to perform, for instance, noise cancellation at the eardrum. In this example, the system may first measure a pressure (Pmic) associated with the point of the in-ear microphone. The system may then determine an equivalent acoustic impedance (Zin) looking into the ear from the point associated with the in-ear microphone. For example, the system may determine the equivalent acoustic impedance (Zin) at the in-ear microphone using the measured equivalent pressure (Pmic), the first thevinin equivalent pressure (Pth), and the first thevinin equivalent acoustic impedance (Zth) using the following equation:
Once the equivalent acoustic impedance (Zin) is determined the system may, the equivalent acoustic impedance associated with the inner tip of the earbud (Zinside). For example, the system may determine Zinside as follows:
where L is the distance from the in-ear microphone to the inner tip of the earbud and γ=α1+jβ1, and j denotes the square root of negative one and Zo1 represents the characteristic acoustic impedance of the acoustic line. The system also determines the equivalent pressure at the inner tip (Pinside) as follows:
Once the equivalent pressure at the inner tip (Pinside) and the equivalent acoustic impedance associated with the inner tip of the earbud (Zinside) are determined, the system may determine the equivalent pressure at the outer tip (Poutside) and the equivalent acoustic impedance associated with the outer tip of the earbud (Zoutside) as follows:
where ω=2πf and f is the frequency of the audio signal and j denotes the square root of negative one and Ldis is the discontinuity inductance associated with the diameter change between the ear bud radius and the ear canal radius.
Next the system determines based on the impedance looking into the ear from the outer tip the magnitude representing the half wavelength point of the ear canal. For example, the system may identify the first maximum of the impedance magnitude in the 4 kHz-10 kHz range and estimate the length of the ear canal, e.g. the distance between the outer tip and the eardrum, (Lcanal) as
where fmax is the frequency at the maximum point and cAir is the speed of sound in air. In some cases, the radius of the ear canal may also be desirable. In these cases, the system may determine the radius of the ear canal as follows:
where cAir is the speed of sound in air, rhoAir is the density of air and ZVector is the impedance of the ear canal at the frequency (f). In some examples, the radius may be determined using a plurality of frequencies and then the results may be average to generate an average radius of the listener's ear canal. In some instances, the radius determined may be greater than a maximum threshold (e.g., greater than a radius that is physically possible given human anatomy) or less than a minimum threshold (e.g., less than a radius that is physically possible given human anatomy). In these instances, the radius may be discarded or excluded from the radii used to generate the average radius.
In some specific examples, the radius of the ear canal and Lcanal are dependent on each other and, thus, the system may solve for each using an iterative process. For instance, the system may initialize the radius and Lcanal to standard or predetermined values. The system may then iterate solving for the radius and the Lcanal. After each iteration, the system may determine error values for the new radius and/or the Lcanal and adjust the value of the radius and the Lcanal based on the error values determined. The system may continue to update the radius and the Lcanal values until the error values are below an error threshold.
Once the length of the canal (Lcanal) is determines, the system may determine the equivalent pressure at the eardrum (Pdrum) and the equivalent acoustic impedance of the eardrum (Zdrum) as follows:
In some examples, in addition to the in-ear microphone, the system may include an external microphone (or a microphone on the exterior of the personal speaker system exposed to the environment). In one implementation, the system may utilize the external microphone to measure noise in the environment. The measured noise may be used to determine noise leakage into the ear canal from the exterior environment and/or to assist with noise cancelation. For example, the noise captured by the external microphone may be compared with noise captured by the internal microphone to determine leakage (e.g., noise present in the audio captured by both microphones). In some cases, the system may then add anti-noise to the output audio or signals that cause the leakage noise to be canceled. In the system discussed herein, the anti-noise may be configured to cancel the leakage noise at the eardrum rather than at the point of the in-ear microphone. Alternatively, the system discussed herein may cancel the leakage noise at the inner tip of the earbud or at the outer tip of the earbud.
In the various alternatives, the system may measure the noise energy at the external microphone (NEEM), measure the noise energy at the in-ear microphone (NEIM), estimate the noise energy at the desired point using the model of the listener's ear canal. The system may then use the estimated energy as an input to a noise cancelation algorithm or technique. In one specific example, the system may estimate the noise energy at the inner tip (NEIT) of the earbud. In this example, the system may determine the parallel combination (ZP) of the internal microphone impedance (ZMici) and the thevinin impedance (Zth) as follows:
Next, the system may determine the refection coefficient at the inner tip (RCIT) using the following equation:
The noise energy at the inner tip (NEIT) may then be determined as follows:
where Lbud is the length of the in-ear microphone to the inner tip of the earbud.
In another specific example, the system may estimate the noise energy at the outer tip (NEOT) of the earbud. In this example, the system may determine the noise energy at the inner tip (NEIT) as discussed above. Then the system may determine the impedance at the inner tip (Zbudin) as follows:
Next, the system may determine the noise energy at the outer tip (NEOT) of the earbud using the following equation:
where j denotes the square root of negative one, as discussed above.
In yet another specific example, the system may estimate the noise energy at the eardrum (NED) of the earbud. In this example, the system may determine the noise energy at the inner tip (NEIT) and the noise energy at the outer tip (NEOT) as discussed above. Then the system may determine the noise energy at the earbud (NED) using the following equation:
In the above examples, the noise leakage estimations are used to cancel the noise at a point in the system. However, the system discussed herein may also determine the equivalent impedance of the noise leakage. For instance, the system may first determine the noise at the external microphone and the in-ear microphone. Next the system may determine the noise energy at the outer tip (NEOT) as discussed above. The system may then determine the impedance looking back towards the earbud from the outer tip of the earbud (ZCI) and the impedance looking toward the eardrum from the outer tip of the earbud (ZCD). Once, the impedances of ZCI and ZCD are estimated, the system may determine the parallel combination of ZCI and ZCD represented as Zin. In this example, the impedance of the leak (Zleak) may representative of the impedance leakage noise between the external environmental and the eardrum and determines as follows:
Where NE is representative of the noise energy at the exterior microphone.
In some cases, the noise leakage value determined may be used to quantify the quality of the fit of the earbud to the listener, suggest adjustments to the type of bud used by the listener or currently inserted into the ear, and adjusting parameters (such as the noise cancelation settings) to compensate for the leakage at the eardrum.
In some cases, the acoustic model may include a first thevinin equivalent pressure (Pth) and a first thevinin equivalent acoustic impedance (Zth) of the speaker system 102 and a first acoustic transmission line representing the distance between the position of the in-ear-microphone 112 and the inner tip 114 of the earbud 116 of the personal speaker system 102 via a first characteristic impedance (Zo1), a first attenuation constant (α1), and a first phase constant (β1). The model 108 may also represent the abrupt diameter change from the outer tip 118 of the earbud 116 and the diameter 120 of the ear canal 104 as a lumped inductance (Ldis) dependent on the ratio of the bud radius and the ear canal radius. In the illustrated example, the diameter 120 of the ear canal 104 may be modeled as a constant. The model 108 may also represent the length 122 of the ear canal 104 (e.g., the distance between the outer tip 118 and the eardrum 106) as a second acoustic transmission line having a second characteristic impedance (Zo2), a second attenuation constant (α2), and a second phase constant (β2). Finally, the model 108 may represent the eardrum 106 as a complex lumped acoustic impedance (Zdrum).
The system may then determine a equivalent acoustic impedance (Zin) looking into the ear from the position of the in-ear-microphone 112. For example, the system may determine the equivalent acoustic impedance (Zin) at the in-ear microphone 102 using the first equivalent pressure (Pth) 204 and the first equivalent acoustic impedance (Zth) 202 using the following equation:
where Pmic is the measured equivalent pressure at the in-ear-microphone 112.
In some cases, the first characteristic impedance (Zo1), the first attenuation constant (α1), and the first phase constant (β1) of the first transmission line 302 may be determined by knowing the physical characteristics of the earbud. Also in some cases, the acoustic pressure and acoustic impedance at the inner tip 114 may be determined by outputting known sound into the ear canal and capturing audio signals using the in-ear microphone 112. The captured audio may then be analyzed to estimate the values.
The order in which the operations are described should not be construed as a limitation. Any number of the described blocks can be combined in any order and/or in parallel to implement the process, or alternative processes, and not all of the blocks need be executed. For discussion purposes, the processes herein are described with reference to the frameworks, architectures and environments described in the examples herein, although the processes may be implemented in a wide variety of other frameworks, architectures or environments.
The process 700 may being with 702, following the listener inserting the personal speaker system onto or into the listener's ear. At 702, the system may estimate an acoustic pressure and an acoustic input impedance of the personal speaker system at an in-ear-microphone. In some cases, a thevinin equivalent pressure (Pth) and a thevinin equivalent acoustic impedance (Zth) may be estimated in a lab and used in the estimation of the acoustic pressure and the acoustic input impedance.
At 704, the system may estimate the acoustic impedance and acoustic pressure at the end of a first acoustic line between the in-ear microphone and an inner tip of an earbud of the personal speaker system. For example, the acoustic impedance and acoustic pressure may be estimated by outputting by the speaker of the personal speaker system a known sound (such as wideband noise) and capturing reverberations. The characteristics of a first acoustic line between the in-ear microphone and an inner tip of an earbud of the personal speaker system. For example, the acoustical transmission line may be represented using a first impedance (Zo1), a first attenuation constant (α1), and a first phase constant (β1) each of which may be estimated by using information known about the physical dimensions of the earbud.
At 706, the system may estimate a diameter change between a radius of an inner tip of the earbud and a radius of the ear canal and, at 708, the system may estimate an acoustic impedance at the outer tip of the earbud. In some examples, the abrupt diameter change may be modeled using a lumped inductance (Ldis) dependent on the ratio of the earbud radius and the ear canal radius. Again, the lumped inductance (Ldis) may be determined by outputting known sound into the ear canal, capturing audio signals using the in-ear microphone, and analyzing the captured audio to estimate the value. In some examples, an estimate of the acoustic pressure and acoustic impedance at the outer tip may then be made.
At 710, the system may estimate characteristics of a second acoustic line outer tip of the earbud and the listener's eardrum. For example, the acoustical transmission line may be represented using a second impedance (Zo2), a second attenuation constant (α2), and a second phase constant (β2) each of which may be estimated by outputting by the speaker of the personal speaker system a known sound (such as wideband noise) and capturing reverberations.
At 712, the system may estimate a complex lumped acoustic impedance (Zdrum) to represent the eardrum and the acoustic pressure at the eardrum within the acoustic model and the resulting acoustic pressure at the eardrum. In some cases, the complex lumped acoustic impedance (Zdrum) and the acoustic pressure at the eardrum may be estimated by outputting known sound into the ear canal, capturing audio signals using the in-ear microphone, and analyzing the captured audio to estimate the value.
It should be understood, that the process 800 is performed using the acoustic model generated as discussed above with respect to
At 802, the system may determine an impedance at a point associated with the in-ear microphone. In some cases, the system may first measure a pressure (Pmic) associated with the point of the in-ear microphone. The system may then determine an acoustic impedance (Zin) looking into the ear from the point associated with the in-ear microphone. For example, the system may determine the acoustic impedance (Zin) at the in-ear microphone using the measured pressure (Pmic), the first pressure (Pth), and the first thevinin equivalent acoustic impedance (Zth) associated with the acoustic model. In these cases, the acoustic impedance (Zin) may be calculated as follows:
At 804, the system may determine an impedance at an inner tip of the earbud of the personal speaker system (Zinside) based at least in part on characteristics of a first acoustic line. For instance, as discussed above the acoustic model may include a first impedance (Zo1), a first attenuation constant (α1), and a first phase constant (β1) associated with the first acoustic line (e.g., the distance between the in-ear microphone and the inner tip of the earbud).
Using the acoustic model and the acoustic impedance (Zin), the system may determine the acoustic impedance (Zinside) as follows:
In the equations above, L is the distance from the in-ear microphone to the inner tip of the earbud and γ=α1+jβ1, and j denotes the square root of negative one.
At 806, the system may determine a reflection coefficient (R1) at the inner tip of the earbud. For example, reflection coefficient (R1) may be determined using:
At 808, the system may determine a pressure at the inner tip (Pinside). For example, the pressure at the inner tip (Pinside) may be determined as follows:
At 810, the system may determine an acoustic impedance associated with the outer tip of the earbud (Zoutside). For example, the system may determine the acoustic impedance (Zoutside) as follows:
Zoutside=j*ω*Ldis
where ω=2πf, j denotes the square root of negative one, and Ldis is the lumped inductance of the abrupt change in diameter between the earbud and the ear canal from the acoustic model.
At 812, the system may determine a pressure at the outer tip (Poutside). For example, the system may determine the equivalent pressure at the outer tip (Poutside) as follows:
where ω=2ρf, f is the frequency of the audio signal, and j again denotes the square root of negative one.
At 814, the system may determine a magnitude of a half wavelength point of the ear canal. For example, the system may identify the first maximum of the impedance magnitude in the 4 kHz-10 kHz range and compute the length of the ear canal, e.g. the distance between the outer tip and the eardrum, (Lcanal). In some cases, the distance may be determined as follows:
where fmax is the frequency at the maximum point and cAir is the speed of sound in air. In some cases, the radius of the ear canal may also be desirable. In these cases, the system may determine the radius of the ear canal as follows:
where rhoAir is the density of air and ZVector is the impedance of the ear canal at the frequency (f). In some examples, the radius may be determined using a plurality of frequencies and then the results may be average to generate an average radius of the listener's ear canal. In some instances, the radius determined may be greater than a maximum threshold (e.g., greater than a radius that is physically possible given human anatomy) or less than a minimum threshold (e.g., less than a radius that is physically possible given human anatomy). In these instances, the radius may be discarded or excluded from the radii used to generate the average radius.
In some specific examples, the radius of the ear canal and Lcanal are dependent on each other and, thus, the system may solve for each using an iterative process. For instance, the system may initialize the radius and Lcanal to standard or predetermined values. The system may then iterate solving for the radius and the Lcanal. After each iteration, the system may determine error values for the new radius and/or the Lcanal and adjust the value of the radius and the Lcanal based on the error values determined. The system may continue to update the radius and the Lcanal values until the error values are below an error threshold.
At 816, the system may determine an acoustic impedance of the eardrum (Zdrum) as follows:
where γ=α1+jβ1, and j denotes the square root of negative one. Thus, in this case, the system utilizes the characteristics of the second acoustic line representative of the ear canal of the user and the length of the ear canal.
At 818, the system may determine a reflection coefficient (R2) at the eardrum. For example, the system may determine the reflection coefficient (R2) using the following equation:
At 820, the system may determine a pressure at the eardrum (Pdrum). In some cases, the pressure at the eardrum (Pdrum) may be determined using the following:
For instance, the pressure may be used to provided noise cancelation at the eardrum opposed to at the location of the in-ear microphone.
At 902, the system may measure the noise energy at the external microphone (NEEM) and, at 904, the system may also measure the noise energy at the in-ear microphone (NEIM).
At 906, the system may estimate the noise energy at the desired reference point using the model of the listener's ear canal. For instance, if the reference point is the eardrum, the system may determine the pressure at the eardrum (Pdrum) and the acoustic impedance of the eardrum (Zdrum) as discussed above with respect to
At 908, the system may use the estimated noise energy (e.g., the pressure and the acoustic impedance at the reference point) to set at least one parameter of a noise cancelation process. For instance, the system may first estimate the noise energy at the inner tip (NEIT) of the earbud. In this example, the system may determine the parallel combination (ZP) of the in-ear microphone impedance (ZMici) and the thevinin impedance (Zth) as follows:
Next, the system may determine the refection coefficient at the inner tip (RCIT) using the following equation:
The noise energy at the inner tip (NEIT) may then be determined as follows:
where Lbud is the length of the in-ear microphone to the inner tip of the earbud.
In another specific example, the system may estimate the noise energy at the outer tip (NEOT) of the earbud. In this example, the system may determine the noise energy at the inner tip (NEIT) as discussed above. Then the system may determine the impedance at the inner tip (Zbudin) as follows:
Next, the system may determine the estimated noise energy at the outer tip (NEOT) of the earbud using the following equation:
where j denotes the square root of negative one, as discussed above.
Next, the system may estimate the noise energy at the eardrum (NED) of the earbud. In this example, the system may determine the noise energy at the inner tip (NEIT) and the noise energy at the outer tip (NEOT) as discussed above. Then the system may determine the noise energy at the earbud (NED) using the following equation:
In the example process 900, the noise is canceled at the eardrum but in other implementations the noise may be canceled at other points, such as inside the earbud (e.g., at the inner tip (NEIT) or at the outer tip (NEOT)).
At 1002, the system may first measure noise at the external microphone and, at 1004, the system may measure noise at an in-ear microphone.
At 1006, the system may determine the noise energy at the outer tip (NEOT). For example, as discussed above, the system may determine the estimated noise energy at the outer tip (NEOT) of the earbud using the following equation:
where j denotes the square root of negative one, as discussed above.
At 1008, the system may then determine a first impedance looking back towards the earbud from the outer tip of the earbud (ZCI). For example, the first thevinin impedance (ZCI) may be based on the thevinin acoustic pressure of the speaker driver, the thevinin acoustic impedance of the speaker driver, the characteristics of the acoustic line between the internal microphone and the inner tip and the radius discontinuity between the inner tip of the ear bud and the ear canal.
At 1010, the system may determine a second impedance looking toward the eardrum from the outer tip of the earbud (ZCD). Again, the second impedance (ZCD) may be based on characteristics of the acoustic line between the outer tip and the ear drum and on the ear drum acoustic impedance.
At 1012, the system may determine the parallel combination of ZCI and ZCD represented as Zin:
Once the parallel combination (Zin) is determined, the process 1000 proceeds to 1014 and the system determines a value representative of the leakage of the earbud (Zleak). For instance, the impedance of the leak (Zleak) may be determines as follows:
where NE is representative of the noise energy difference between the exterior and in-ear microphones.
In some cases, the noise leakage value determined may be used to quantify the quality of the fit of the earbud to the listener, suggest adjustments to the type of bud used by the listener or currently inserted into the ear, and adjusting parameters (such as the noise cancelation settings) to compensate for the leakage at the eardrum.
The personal speaker system 1100 may include one or more external microphones 1102 to capture audio of an environment surrounding the user (e.g., external noise) and one or more in-ear (or internal) microphones 1104 to capture audio within the ear canal. The microphones 1102 and/or 1104 may be implemented as a single omni-directional microphone, a calibrated microphone group, more than one calibrated microphone group, or one or more microphone arrays.
The personal speaker system 1100 also includes one or more speakers 1106 to output audio signals as sounds. For example, the speakers 1106 may output audio files received from various electronic devices, such as smart phones, tablets, computers, or other audio enabled devices.
The personal speaker system 1100 includes one or more communication interfaces 1108 to facilitate a communication with other devices, such as cloud based devices, an audio source, or other electronic devices via one or more networks. The communication interfaces 1108 may support both wired and wireless connection to various networks, such as cellular networks, radio, WiFi networks, short-range or near-field networks (e.g., Bluetooth®), infrared signals, local area networks, wide area networks, the Internet, and so forth. For example, the communication interfaces 1108 may cause the acoustic model of the listener's ear canal to be transmitted to another device, for instance, to use as part of machine learning or research.
The personal speaker system 1100 includes or accesses components such as at least one or more control logic circuits, central processing units, or processors 1110, and one or more computer-readable media 1112 to perform the function of the personal speaker system 1100. Additionally, each of the processors 1110 may itself comprise one or more processors or processing cores.
Depending on the configuration of the personal speaker system 1100, the computer-readable media 1112 may be an example of tangible non-transitory computer storage media and may include volatile and nonvolatile memory and/or removable and non-removable media implemented in any type of technology for storage of information such as computer-readable instructions or modules, data structures, program modules or other data. Such computer-readable media may include, but is not limited to, RAM, ROM, EEPROM, flash memory or other computer-readable media technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, solid state storage, magnetic disk storage, RAID storage systems, storage arrays, network attached storage, storage area networks, cloud storage, or any other medium that can be used to store information and which can be accessed by the processors 1110.
Various instruction, information, data stores, and so forth may be stored within the computer-readable media 1112 and configured to execute on the processors 1110. For instance, the computer-readable media 1112 may store acoustic model estimation instructions 1114, noise cancelation instructions 1116, noise leakage determining instructions 1118, audio modification instructions 1120, as well as other modules. The computer-readable media 1112 may also store a data, such as acoustic model data 1122, known audio data 1124, and/or noise cancelation parameters 1126.
The acoustic model estimation instructions 1114 may be configured to generate acoustic model data 1122 using the in-ear microphones 1104 and the speaker 1106 of the personal speaker system 1100. In some cases, the acoustic model data 1122 may represent the state of the ear canal and ear drum with the personal speaker system 1100 in place on the listener's head. In some examples, the acoustic model estimation instructions 1114 may first determine a first thevinin equivalent pressure (Pth) and a first thevinin equivalent acoustic impedance (Zth) of the speaker 1106 using captured audio data related to known audio outputs.
The acoustic model estimation instructions 1114 may then model the distance between the position of the in-ear-microphone 1104 and the inner tip of the earbud as a first acoustic transmission line. For instance, the system may determine a first impedance (Zo1), a first attenuation constant (α1), and a first phase constant (β1) associated with the first acoustic transmission line. Next the acoustic model estimation instructions 1114 may model the abrupt diameter change from the outer tip of the earbud and the diameter of the ear canal as a lumped inductance (Ldis) dependent on the ratio of the bud radius and the ear canal radius.
The acoustic model estimation instructions 1114 may then model the length of the ear canal as a second acoustic transmission line. In this case, the acoustic model estimation instructions 1114 may determine a second characteristic impedance (Zo2), a second attenuation constant (α2), and a second phase constant (β2) and a length associated with the second acoustic transmission line. The second characteristic impedance, second attenuation constant and second phase constant may be determined by the estimated radius of the second acoustic transmission line and with the known speed of sound of air, density of air and thermoviscous properties of sound propagation in a cylinder. Next, the acoustic model estimation instructions 1114 may model the eardrum as a complex lumped acoustic impedance (Zdrum).
The noise cancelation instructions 1116 may be configure to cancel noise at various reference points within the ear canal, such as the inner tip, outer tip, or ear drum. For instance, the noise cancelation instructions 1116 may determine a pressure and impedance at the reference point and utilize the pressure and impedance at the reference point as noise cancelation parameters 1126.
The noise leakage determining instructions 1118 may be configured to determine noise leakage at the various reference points within the ear canal and to use the leakage values to determine proper placement of the personal speaker system 1100 on the head of the listener and/or proper fit of the personal speaker system 1100 or earbuds of the system 1100. For instance, the noise leakage determining instructions 1118 may utilize the acoustic model data 1122 to estimate noise energy at the reference points and utilize a comparison of the noise energy at the reference point and the noise energy external (e.g., noise energy captured by the external microphone 1102) to determine leakage.
The audio modification instructions 1120 may be configured to utilize the acoustic model data 1122 to determine audio adjustments for the listener. For example, the audio modification instructions 1120 may determine the difference in acoustic pressure at the eardrum that a listener would experience if the listener was listening to the content in free space without an earbud (the free space ear drum pressure) and the actual acoustic pressure at the eardrum using the personal speaker system (the actual pressure). The audio modification instructions may then adjust the audio so that the actual pressure at the eardrum is adjusted to be equivalent to the free space ear drum pressure, providing the user with a more natural listening experience.
Although the subject matter has been described in language specific to structural features, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features described. Rather, the specific features are disclosed as illustrative forms of implementing the claims.
This application is a continuation of and claims priority to U.S. application Ser. No. 16/117,132, filed on Aug. 30, 2018 and entitled “System and Method for Configuring Audio Signals to Compensate for Acoustic Changes of the Ear,” which is a non-provisional of and claims priority to Provisional Application No. 62/607,704 filed on Dec. 19, 2017 and entitled “System for Configuring Audio Signals to Compensate for Acoustic Changes of the Ear,” the entirety of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
9271666 | Stirnemann | Mar 2016 | B2 |
20140161267 | Saltykov | Jun 2014 | A1 |
20170251297 | Darlington | Aug 2017 | A1 |
20190191236 | Rauschmayer | Jun 2019 | A1 |
Entry |
---|
Elements of Acoustical Engineering, by H. F. Olson, 1940, pp. 49-76, Chapter 4. |
Non Final Office Action dated Oct. 7, 2019 for U.S. Appl. No. 16/117,132 “System and Method for Configuring Audio Signals to Compensate for Acoustic Changes of the Ear” Rauschmayer, 7 pages. |
Number | Date | Country | |
---|---|---|---|
20200007973 A1 | Jan 2020 | US |
Number | Date | Country | |
---|---|---|---|
62607704 | Dec 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16117132 | Aug 2018 | US |
Child | 16569002 | US |