This application is based upon and claims the benefit of priority from Japanese Patent Application No. 2008-035268, filed Feb. 15, 2008, the entire contents of which are incorporated herein by reference.
1. Field
One embodiment of the present invention relates to an apparatus for cancelling resonance in the outer-ear canals and a method of cancelling resonance in the outer-ear canals.
2. Description of the Related Art
When a person is listening to music through an earphone or a headphone, resonance may develop between the eardrum and the earphone or the headphone. In this case, the music sounds strange to the listener. Various systems have been developed, which cancel such resonance. (See, for example, Jpn. Pat. Appln. KOKAI Publication No. 2000-92589, paragraph 0047 and FIGS. 1 and 2; Jpn. Pat. Appln. KOKAI Publication No. 2002-209300, paragraph 0040 and FIG. 1; and Jpn. Pat. Appln. KOKAI Publication No. 9-187093, paragraph 0024 and FIG. 2).
Jpn. Pat. Appln. KOKAI Publication No. 2000-(hereinafter referred to as Publication 1) discloses a technique of finding the position of an acoustic image outside a listener's head. FIGS. 2(a) and 2(b) of Publication 1 illustrate the principle of finding the position of the acoustic image outside the head. More precisely, FIG. 2(a) explains how sound coming from a speaker is picked up, and FIG. 2(b) explains how a twin earphone or a stereophonic headphone catches sound. In FIG. 2(a), reference numeral 101 denotes a sound-source signal, reference numeral 103 designates a speaker, and reference numeral 102 denotes two microphones set in the outer-ear canals, respectively. In FIG. 2(b), reference numeral 104 designates an earphone or a headphone, reference numeral 105 denotes a digital filter. Note that suffix L in HRTFL and suffix R in HRTFR stand for “left” and “right” respectively.
The principal of finding the position of the acoustic image outside the head lies in electrically formulate a transfer function identical to the transfer function for sound traveling to the listener's eardrum from a sound source that exists outside the listener's head.
However, it is difficult for an electric signal emanating from a living body to pick up the vibration the eardrum are undergoing as sound waves. Hence, the transfer function of the electric signal traveling to the eardrum can hardly be measured accurately from the sound-source signal 101 shown in FIG. 2(a). This is why the listener sets small microphones 102 in his or her outer-ear canals, respectively, and the transfer function of the electric signal, i.e., head related transfer functions (HRTFs) in the left and right ears, is measured from the sound-source signal 101 that has been input to the speaker 103 by using these microphones 102.
The speaker 103 has a specific frequency characteristic. The true transfer function of the electric signal traveling from the input of the speaker 103 to the microphones 102 is therefore given as HRTF/SPTF, where SPTF is the transfer function for the speaker 103.
In the system of FIG. 2(b) of Publication 1, the twin earphone or stereophonic headphone 104 may be used to provide a transfer function that is equivalent to function HRTF/SPTF. To provide this transfer function, the transfer function of a signal traveling from the earphone or headphone 104 to the microphones 102 set in the outer-ear canals, i.e., ear-canal transfer function (ECTF), is measured. If the product of this transfer function ECTF and the transfer function of the digital filter 105 is equal to the transfer function HRTF/SPTF, aural signal identical to the speaker signals can be reproduced at the microphones 102 set in the outer-ear canals.
In the system disclosed in Publication 1, an ex-head sound-image locating means of the type shown in FIG. 5 is used to measure the outer-ear canal transfer function, i.e., transfer function attained while the listener is wearing the earphone or headphone 104. The outer-ear canal transfer function thus measured is corrected by using an adaptive equalization filter.
Microphones 3 that pick up the sound in the outer-ear canals are formed integral with the speakers of the earphone or headphone, as is illustrated in FIG. 1 of Publication 1. A digital filter 11 is used, which stores an impulse response having transfer function HRTF/SPTF that has been measured by such a configuration as shown in FIG. 2(a) of Publication 1.
A band-pass filter 13 is provided, for the following reason. An adaptive filter 12 and the transfer function ECTF are connected in series, and the output of this series circuit may be an impulse. In this case, the transfer function of the adaptive filter 12 is inverse to the function ECTF, i.e., 1/ECTF. However, the function ECTF pertains to both a speaker 1 and the microphones 3 and therefore attenuates outside a specific band. Hence, the transfer function of the adaptive digital filter 12, which is inverse to the transfer function ECTF, attains a large gain outside the specific band.
The tap coefficient or impulse response of the adaptive digital filter 12 can therefore be stably acquired if the result of the convolution performed on the impulse responses of the filter 12 and ECTF is regarded as the impulse response of the band-pass filter 13. In other words, if the band of the band-pass filter 13 is narrower than that of the adaptive digital filter 12, a subtracter 14 will cancel the ex-band part of the transfer function of the adaptive digital filter 12. As a result, a stable solution can be obtained.
In the system disclosed in Publication 1, an adaptive equalization filter is used to correct the outer-ear canal transfer function. In order to correct this transfer function accurately, the microphones 3 must exhibit flat frequency characteristic within the band. This is because the music will sound strange at the eardrum if the adaptive digital filter 12 generates an inverse transfer function from the transfer function ECTF that pertains to the characteristic of the microphones 3. Further, the position of the microphones 3 is important and should therefore be carefully determined. If the microphones 3 are located at the eardrums, no problems will arise. If the microphones 3 are located at the distal ends of the twin earphone or headphone (not at the ends of the outer-ear canals), however, it will pick up sound not at the nodes of a standing sound wave. Consequently, the microphones 3 will acquire such a characteristic that they catch sound at the dips of the standing sound wave. The music will inevitably sound strange to the listener.
Jpn. Pat. Appln. KOKAI Publication No. 2002-209300 (hereinafter referred to as Publication 2) discloses a technique of cancelling the influence of standing waves formed in a twin earphone or headphone and at the listener's eardrum. To cancel the standing waves, the vibration signal emanating from either eardrum should be measured to determine the sound-transfer characteristic in the outer-ear canals. It is difficult, however, to set microphones at the eardrums to detect the vibration signals in the vicinity of the eardrums. In the technique disclosed in Publication 2, the microphones are set at the eardrums of a pseudo-head, in order to measure the outer-ear ear canal transfer function. Based on the characteristic measured, a filter is designed, which can cancel the standing wave that extends from either eardrum and the earphone or headphone.
However, the length and acoustic impedance of outer-ear canals differ, from person to person. The transfer function in the outer ears therefore differs, on the individual basis. It follows that the position where resonance frequency is attained differs, on individual basis, too. Further, the resonance frequency is attained at a position in the left ear, and at a different position in the right ear. The outer-ear canal transfer function should therefore be corrected in accordance with the physical characteristics of the ears of each person. Hence, the characteristic determined by using the pseudo-head can hardly serve to manufacture a filter that proves satisfactory to all users. In view of this, filters of different characteristics may be prepared so that the user may select one that he or she finds best. Here arises a problem. The user can hardly select a filter he or she thinks the best for him or her. Moreover, the filter the user selects can scarcely work flawlessly.
Jpn. Pat. Appln. KOKAI Publication No. 9-187093 (hereinafter referred to as Publication 3) discloses a system that has an electro-acoustic converting means and a resonance-frequency component reducing means connected to the input of the electro-acoustic converting means. The resonance-frequency component reducing means is configured to reduce a resonance-frequency component of a frequency near the resonance frequency in human ears. Thus, the means prevents a decline in the hearing ability of the user who habitually listens to laud music through an earphone or a headphone. That is, the resonance-frequency component reducing means prevents the sound level of the resonance frequency in the ears from increasing excessively. The resonance-frequency component reducing means is an electrical circuit that has a resister, to which a parameter for reducing the resonance-frequency component detected is set. However, no parameters are specified in Publication 3. Methods of determining such a parameter are known in the art. One method is to use a filter inverse to the resonance data actually acquired as described in Publication 1. Another method is to provide a filter similar to the data acquired by, for example, a parametric equalizer. These methods are, however, disadvantageous in the following respects.
1) Since microphones cannot be located at the eardrums, the characteristics of the ears cannot be accurately measured. If the inverse filter designed on the basis of the characteristics measured is subjected to convolution, the resulting sound will be degraded in quality.
2) Many parameters are applied, rendering the tuning extremely difficult. Desirable characteristics may not be attained in some cases. Even if desirable characteristics are attained, it will be very difficult to determine the phase accurately.
As has been described, the conventional apparatus for rectifying resonance in the outer-ear canals cannot easily rectify the resonance in accordance with the structure of the outer-ear canals of each person.
A general architecture that implements the various feature of the invention will now be described with reference to the drawings. The drawings and the associated descriptions are provided to illustrate embodiments of the invention and not to limit the scope of the invention.
Various embodiments according to the invention will be described hereinafter with reference to the accompanying drawings. In general, according to one embodiment of the invention, an apparatus for cancelling resonance in an outer-ear canal, comprises an outer-ear canal model comprising attenuator modules representing reflection coefficients of an earphone or headphone and an eardrum, and a delay module having a delay time corresponding to a distance between the earphone or headphone and the eardrum; an inverse-filter forming unit configured to form an inverse filter of the outer-ear canal model; and a convolution module configured to perform convolution on an impulse response from the inverse filter and a sound-source signal.
According to an embodiment,
As shown in
If the microphone 12 is arranged not at the end of the outer-ear canal, the characteristic it detects will differ from those shown in
The microphone 24 may be arranged in the earphone or headphone 28 or located remote from the earphone or headphone 28. In either case, the microphone 24 must be so positioned that no dips may exist at the peak frequency (i.e., resonance frequency).
In Block 34, the audio signal is converted from a time domain to a frequency domain. In Block 36, resonance peaks are detected on the frequency axis. In view of the frequency characteristic shown in
Two correction filters are formed for the left and right ears, respectively, so that dips may be formed at peak frequencies in order to cancel the resonance peaks for the left and right ears (Block 38). The correction filters may be formed by a parametric equalizer or a graphic equalizer. In this embodiment, a model is used to form the correction filters, as will be explained later in detail.
In Block 40, the correction-filter forming module 14 generates tap coefficients of correction filters for the left and right ears, respectively, and then supplies the tap coefficients, either directly or via the memory 18, to the convolution module 16.
The convolution module 16 performs convolution on the data items transferred from the correction-filter forming module 14 or memory 18 and the left- and right sound-source signals. (Note that the data items are the two tap coefficients representing impulse responses of the left and right ears, respectively). The convolution module 16 therefore generates a left-ear signal and a right-ear signal, each no long having a resonance component.
Thus, two filters are formed, which cancel the resonance peaks detected in the outer-ear canals of the listener. Then, the tap coefficients representing the impulse responses of the left and right ears are set in the convolution module 16. The left and right sound-source signals are then subjected to convolution. As a result, the frequency peaks shown in
So far described is a case where two microphones are arranged in the left and right modules of an earphone or headphone and detect the characteristics of the left and right ears, and two correction filters are formed for the left and right ears, respectively. Nonetheless, the characteristic of only one ear may be detected, and the correction filter formed based on this characteristic may be applied to both the left sound-source signal and the right sound-source signal.
The process of forming such correction filters may be performed every time an audio player, for example, is activated, or every time the user instructs. Alternatively, this process may be performed when the audio player is activated after a time the user set by the user has elapsed.
As described above, the microphone 12 for detecting the characteristics of the outer-ear canals, the correction-filter forming module 14, and the convolution module 16 for performing convolution on the sound-source signals constitute an integrated module. Nonetheless, these components 12, 14 and 16 need not be integrated. For example, the sound-source signals the microphone 12 picks up may be taken into an apparatus such as a personal computer (PC). If this is the case, the personal computer execute software, forming correction filters.
To play back the music, the convolution module 16 may be incorporated in the audio player and corrects the left-ear and right-ear signals in real time, thus playing back the music. Alternatively, the PC may execute software, thereby to correct the sound-source signals, and the signals thus corrected may then be transferred to the audio player.
In the apparatus for canceling the resonance in the outer-ear canals, shown in
How the correction-filter forming module 14 shown in
The sound-wave propagation model thus configured provides such acoustic characteristics of the outer-ear canal as illustrated in
Next, an inverse filter is formed based on a model shown in
Assume that the filter module 74 has the outer-ear-canal acoustic characteristic shown in
The process described above is performed for both the left ear and the right ear. Two correction filters can thereby be formed for the left and right ears, respectively.
A method of improving accuracy of measuring the characteristic will be described. In the model of
As is known in the art, the polymer constituting the eardrum exhibits elasticity that is low mainly at low frequencies and increases as the frequency rises. This is why the model of
As a result, the resonance at a low band is suppressed as seen from the amplitude and phase characteristics of the outer-ear canal, obtained from the model of
The use of the model of
The positions where the correction-filter forming module 14 and convolution module 16, both shown in
The correction-filter forming module 14 and convolution module 16 may be incorporated in an audio player 90. In this case, the tap coefficient generated in the correction-filter forming module 14 is stored in the memory 18, and the sound-source signal read from a flash memory (not shown) or a hard disk (not shown) is corrected in the convolution module 16 and is then output to an earphone or headphone 94. Alternatively, the sound-source signal may be corrected before it is downloaded and may then be stored in a memory (not shown). The correction-filter forming module 14 and convolution module 16 may be incorporated in a remote controller 92 or the earphone or headphone 94. In either case, the microphone 12 is fixed to the earphone or headphone 20 as is illustrated in
As has been explained thus far, this embodiment detects the resonance frequency from the frequency characteristics of the user's outer-ear canals, acquired by the microphones arranged at given positions in the outer-ear canals. A sound-wave propagation model comprises attenuator modules representing the reflection coefficient of the earphone or headphone and the reflection coefficient of the eardrum, and delay modules having a delay time corresponding to the distance between the earphone or headphone and the eardrum. The time corresponding to the distance between an eardrum and an earphone or headphone, which has been obtained from the resonance frequency detected, is set in the delay times of the delay modules. Using this model, an inverse filter module is adaptively equalized (identified). The inverse filter module corrects the frequency characteristic of a sound-source signal, thereby accurately cancelling the resonance specific to the acoustic characteristics of outer-ear canals of any user.
If inverse filter module formed not on the basis of the data acquired without using such a model is employed to cancel the resonance, the resonance frequency cannot be accurately measured because the microphones cannot be arranged at the eardrum. When resonance is cancelled, using this model, the sound quality will be degraded.
Moreover, a high-pass filter module may be added to the above-mentioned model in order to impart the frequency dependency of acoustic impedance. In this case, an inverse filter module can be provided, which has amplitude and phase characteristics having no dips in the low band. This inverse filer module can reduce the quality degradation of the sound.
Generally, a parametric equalizer may be used to form an inverse filter module. In this case, however, the inverse filter module may fail to have desirable characteristic, because the tuning is difficult to accomplish due to the many parameters involved. Even if the inverse filter module exhibits desirable characteristics, it can hardly reflect the phase accurately. Consequently, the phase data inevitably assumes an unnatural state (undergoing an extraordinary phase rotation) when the resonance is cancelled. Nevertheless, the model according to the present embodiment can acquire accurate phase data, as well.
While certain embodiments of the inventions have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel methods and systems described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of the methods and systems described herein may be made without departing from the spirit of the inventions. The various modules of the systems described herein can be implemented as software applications, hardware and/or software modules, or components on one or more computers, such as servers. While the various modules are illustrated separately, they may share some or all of the same underlying logic or code. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the inventions.
Number | Date | Country | Kind |
---|---|---|---|
2008-035268 | Feb 2008 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
6480610 | Fang et al. | Nov 2002 | B1 |
6658122 | Westermann et al. | Dec 2003 | B1 |
7668325 | Puria et al. | Feb 2010 | B2 |
20080123890 | Peng | May 2008 | A1 |
20080137878 | Killion et al. | Jun 2008 | A1 |
20080159554 | Sung et al. | Jul 2008 | A1 |
Number | Date | Country |
---|---|---|
53-42721 | Apr 1978 | JP |
05-083797 | Apr 1993 | JP |
09-185383 | Jul 1997 | JP |
09-187093 | Jul 1997 | JP |
10-294997 | Nov 1998 | JP |
2000-050395 | Feb 2000 | JP |
2000-092589 | Mar 2000 | JP |
2001-285998 | Oct 2001 | JP |
2002-209300 | Jul 2002 | JP |
2008-177798 | Jul 2008 | JP |
Number | Date | Country | |
---|---|---|---|
20090208027 A1 | Aug 2009 | US |