The present invention relates to a sound reproducing apparatus for reproducing a sound by using an in-ear earphone.
A sound reproducing apparatus using an in-ear earphone is compact, highly portable, and useful. On the other hand, since wearing an earphone in an ear blocks an ear canal, there arises a problem that the sound is slightly muffled and that it is difficult to obtain a spacious sound.
For example, let it be assumed that the ear canal of an ear is represented by a simple cylindrical model. When not wearing an earphone in the ear, the cylinder is closed at the eardrum side and is open at the entrance side of the ear, that is, one end of the cylinder is open and the other end is closed ((a) in
One of techniques to solve the above problem is a conventional sound reproducing apparatus which corrects a resonance frequency characteristic of an ear canal to reproduce a sound, thereby realizing a listening state equivalent to that in the case of not wearing an earphone (in the case where the ear canal is not blocked), even when, actually, wearing the earphone in the ear (for example, see Patent Document 1).
Moreover, there is a conventional acoustic-field reproducing apparatus which automatically measures a head-related transfer function of a listening person with use of an in-ear transducer used for both a microphone and an earphone, and convolves an inputted signal with the measured head-related transfer function of the listening person, and which allows the listening person to receive the convolved signal via the in-ear transducer used for both a microphone and an earphone (for example, see Patent Document 2). The conventional acoustic-field reproducing apparatus realizes, through the above processing, the effect of allowing an unspecified listening person to obtain excellent feeling of localization of a plurality of sound sources present in all directions.
However, the conventional sound reproducing apparatus disclosed in Patent Document 1 has a problem that a characteristic of a pseudo head is used for a characteristic of ear-canal correction.
In addition, the conventional acoustic-field reproducing apparatus disclosed in Patent Document 2 measures a head-related transfer function between a speaker and each ear of the listening person, from an input from the speaker and an output from the in-ear transducer used for both a microphone and an earphone. In addition, it is disclosed that, since a point where the measurement is performed coincides with a point where a sound is reproduced, an optimum head-related transfer function can be measured. However, there is a problem that, since an earphone normally has a speaker for reproduction directed toward the inside of an ear, the microphone itself is an obstacle, and therefore a head-related transfer function cannot properly be measured.
Therefore, an object of the present invention is to provide a sound reproducing apparatus capable of realizing a listening state which is suitable for an earphone for listening and is equivalent to a listening state in the case where the ear canal is not blocked even when wearing the earphone, by obtaining a filter for correcting a characteristic of an ear canal of an individual with use of an earphone used for listening and convolving a sound source signal with the filter.
The present invention is directed to a sound reproducing apparatus reproducing sound by using an in-ear earphone. In order to achieve the above object, one aspect of a sound reproducing apparatus of the present invention comprises a measurement signal generating section, a signal processing section, an analysis section, and an ear-canal correction filter processing section.
The measurement signal generating section generates a measurement signal. The signal processing section outputs the measurement signals from an in-ear earphone to an ear canal of a listening person by using a speaker function, and measures, with the in-ear earphone, the signals reflected by an eardrum of the listening person by using a microphone function of the in-ear earphone, both in a state where the in-ear earphone is worn in the ear of the listening person and in a state where the in-ear earphone is not worn in the ear of the listening person. The analysis section analyzes the signals measured in the two states by the signal processing section, and obtains an ear-canal correction filter. The ear-canal correction filter processing section convolves the sound source signal with the ear-canal correction filter obtained by the analysis section, when sound is reproduced from a sound source signal.
The signal processing section may measure a signal in a state where the in-ear earphone is attached to an ear-canal simulator which simulates a characteristic of an ear canal, instead of the state where the in-ear earphone is not worn in the ear of the listening person. In addition, if the analysis section stores a standard ear-canal correction filter which is measured in advance by using the ear-canal simulator which simulates a characteristic of an ear canal, the analysis section can correct the standard ear-canal correction filter and obtain an ear-canal correction filter, based on the signal measured in the state where the in-ear earphone is worn in the ear of the listening person.
It is preferable that the standard ear-canal correction filter is stored as a parameter of an IIR filter. In addition, the analysis section may perform processing on a characteristic obtained through the measurement, only within a range of frequencies causing a change in a characteristic of the ear canal. The range of frequencies causing a change in a characteristic of the ear canal is, for example, from 2 kHz to 10 kHz.
In addition, an HRTF processing section for convolving the sound source signal with a predetermined head-related transfer function may further be provided at a preceding stage of the ear-canal correction filter processing section. Alternatively, an HRTF processing section for convolving the sound source signal convolved with the ear-canal correction filter with a predetermined head-related transfer function, may further be provided at a subsequent stage of the ear-canal correction filter processing section. Alternatively, the analysis section may store a predetermined head-related transfer function and obtain an ear-canal correction filter convolved with the head-related transfer function. Alternatively, the analysis section may calculate a simulation signal for a state where the in-ear earphone is not worn in the ear of the listening person by performing resampling processing on the signal measured by the signal processing section in the state where the in-ear earphone is worn in the ear of the listening person. Typically, the measurement signal is an impulse signal.
According to the present invention, a characteristic of an ear canal of an individual is measured by using an earphone used for listening, and thereby an optimum ear-canal correction filter can be obtained. Thus, a listening state which is suitable for an earphone for listening and is equivalent to a listening state in the case where the ear canal is not blocked, can be realized, even when wearing an earphone.
Firstly, an outline of each component of the sound reproducing apparatus 100 according to the first embodiment will be described.
The measurement signal generating section 101 generates a measurement signal. The measurement signal generated by the measurement signal generating section 101, and a sound source signal which has passed through the ear-canal correction filter processing section 109, are inputted to the signal switching section 102, and the signal switching section 102 outputs one of the inputted signals by switching therebetween in accordance with a reproduction mode or a measurement mode described later. The D/A conversion section 103 converts a signal outputted by the signal switching section 102 from digital to analog. The amplification section 104 amplifies the analog signal outputted by the D/A conversion section 103. The distribution section 105 supplies the amplified signal outputted by the amplification section 104 to the earphone 110, and supplies a signal to be measured when the earphone 110 is operated as a microphone to the microphone amplification section 106. The earphones 110 are worn in both ears of a listening person as a pair of in-ear earphones. The microphone amplification section 106 amplifies the measured signal outputted by the distribution section 105. The A/D conversion section 107 converts the amplified signal outputted by the microphone amplification section 106 from analog to digital. The analysis section 108 analyzes the converted amplified signal to obtain an ear-canal correction filter. The ear-canal correction filter processing section 109 performs convolution processing on the sound source signal with the ear-canal correction filter obtained by the analysis section 108.
Next, operation of the sound reproducing apparatus 100 according to the first embodiment will be described.
The sound reproducing apparatus 100 executes processing in the measurement mode for calculating the ear-canal correction filter to be given to the ear-canal correction filter processing section 109 by using the measurement signal, before executing processing in the reproduction mode for performing sound reproduction based on the sound source signal.
1. Measurement Mode
First, the sound reproducing apparatus 100 is set to the measurement mode by a listening person. When the sound reproducing apparatus 100 is set to the measurement mode, the signal switching section 102 switches a signal path so as to connect the measurement signal generating section 101 to the D/A conversion section 103. Next, the listening person wears a pair of the earphones 110 in the ears (state shown by (a) in
When the measurement is started, the measurement signal generating section 101 generates a predetermined measurement signal. For the measurement signal, an impulse signal exemplified in
Next, the listening person removes the pair of earphones 110 from both ears. At this time, a content inducing the listening person to remove the earphones 110 may be displayed on, e.g., the display (not shown) of the sound reproducing apparatus 100. After removing the pair of earphones 110 from both ears, a measurement is started by, for example, the listening person pressing a measurement start button. Note that, both ears of the listening person and the pair of earphones 110 in the state where the earphones 110 are not worn, have a positional relationship in which the earphones 110 do not have contacts with the ears and in which a measurement signal outputted from the earphone 110 can be conducted into the ear canals (state shown by (b) in
In the above state, the measurement signal is outputted from the pair of earphones 110, passes through the ear canal to be reflected by the eardrum, and returns to the earphones 110. The earphone 110 measures the measurement signal which has returned. The signal (hereinafter, referred to as unwearing-state signal) measured by the earphone 110 is outputted via the distribution section 105, the microphone amplification section 106, and the A/D conversion section 107, to the analysis section 108, and is stored.
On the other hand, another method for measuring the unwearing-state signal is a method using an ear-canal simulator which simulates an ear canal. The ear-canal simulator 121 is a measuring instrument having a cylindrical shape with a length of about 25 mm and a diameter of about 7 mm (
The order in which the wearing-state signal and the unwearing-state signal are measured may be reversed.
The FFT processing section 114 performs fast Fourier transform (FFT) processing on the wearing-state signal and the unwearing-state signal which are outputted from the A/D conversion section 107, to transform them to signals in frequency domain, respectively. The memory section 115 stores the two signals in frequency domain obtained through the FFT processing. The coefficient calculation section 116 reads out the two signals stored in the memory section 115, and subtracts the unwearing-state signal from the wearing-state signal to obtain a difference therebetween as a coefficient. The coefficient represents a conversion from a state of wearing the earphone 110 to a state (unwearing state) of not wearing the earphone 110.
The coefficient obtained by the coefficient calculation section 116 is data in frequency domain. Therefore, the IFFT processing section 117 performs inverse fast Fourier transform (IFFT) processing on the coefficient in frequency domain obtained by the coefficient calculation section 116 to transform the coefficient to a coefficient in time domain. The coefficient in time domain obtained through the transformation by the IFFT processing section 117 is given as an ear-canal correction filter to the ear-canal correction filter processing section 109.
In the case where the ear-canal correction filter processing section 109 performs convolution processing in frequency domain, the coefficient in frequency domain obtained by the coefficient calculation section 116 may directly be given to the ear-canal correction filter processing section 109 without the IFFT processing section 117 performing IFFT processing. Note that, in this case, an FFT length of the FFT processing section 114 needs to be the same as an FFT length used in the ear-canal correction filter processing section 109.
In addition, the FFT processing section 114 may perform FFT processing immediately after a measurement is started (the measurement signal is generated), or may exclude the beginning part of the measurement signal (cause delay) to perform FFT processing, as shown in
2. Reproduction Mode
After giving the ear-canal correction filter to the ear-canal correction filter processing section 109 in the measurement mode, a sound source signal is reproduced as follows.
The sound reproducing apparatus 100 is set to the reproduction mode by the listening person. When the sound reproducing apparatus 100 is set to the reproduction mode, the signal switching section 102 switches a signal path so as to connect the ear-canal correction filter processing section 109 to the D/A conversion section 103. Next, the listening person wears the pair of earphones 110 in the ears, and then a measurement is started by, for example, the listening person pressing a measurement start button.
When a reproduction of the sound source signal is started, the sound source signal is inputted to the ear-canal correction filter processing section 109, and the ear-canal correction filter processing section 109 convolves the sound source signal with the ear-canal correction filter given by the analysis section 108. By performing the convolution processing, an acoustic characteristic equivalent to that in the case of not wearing the earphone 110 (where the ear canal is not blocked) can be obtained, even when wearing the earphone 110. The convolved sound source signal is outputted from the pair of earphones 110 worn in the ears of the listening person, via the signal switching section 102, the D/A conversion section 103, the amplification section 104, and the distribution section 105.
As described above, the sound reproducing apparatus 100 according to the first embodiment of the present invention measures a characteristic of an ear canal of an individual by using the earphone 110 used for listening, and thereby can obtain an optimum ear-canal correction filter. Thus, a listening state which is suitable for the earphone 110 for listening and is equivalent to a listening state in the case where the ear canal is not blocked, can be realized, even when wearing the earphone 110 in the ear.
In the first embodiment, a configuration including the microphone amplification section 106 and the A/D conversion section 107 is used. However, in the case where the sound reproducing apparatus 100 has an ANC (active noise cancel) function, the ANC function can used as both the microphone amplification section 106 and the A/D conversion section 107.
As shown in
When a reproduction of the sound source signal is started in the reproduction mode, the sound signal is inputted to the HRTF processing section 212. The HRTF processing section 212 convolves the sound source signal with a head-related transfer function (HRTF) which is set in advance. By using the head-related transfer function, a sound image which makes the listening person feel as if listening through a speaker can be listened to even if using the earphone 110. The sound source signal convolved with the head-related transfer function is inputted to the ear-canal correction filter processing section 109 from the HRTF processing section 212, and then the ear-canal correction filter processing section 109 convolves the sound source signal with the ear-canal correction filter given by the analysis section 108.
As described above, the sound reproducing apparatus 200 according to the second embodiment of the present invention enhances accuracy of control of three-dimensional sound-field reproduction, and can realize an out-of-head sound localization in a more natural state, in addition to providing the effects of the first embodiment.
Note that, the order in which the ear-canal correction filter processing section 109 and the HRTF processing section 212 are arranged may be reversed.
The sound reproducing apparatus 300 according to the third embodiment shown in
A filter in time domain outputted from the IFFT processing section 117 is inputted to the convolution processing section 318. The HRTF storage section 319 stores in advance a filter coefficient of a head-related transfer function corresponding to a direction in which localization should be performed. The convolution processing section 318 convolves the ear-canal correction filter inputted from the IFFT processing section 117 with the filter coefficient of the head-related transfer function stored in the HRTF storage section 319. The filter convolved by the convolution processing section 318 is given, as an ear-canal correction filter which includes a head-related transfer function characteristic, to the ear-canal correction filter processing section 109.
In the case where the ear-canal correction filter processing section 109 performs convolution processing in frequency domain, the coefficient in frequency domain obtained by the coefficient calculation section 116 may be convolved with the filter coefficient of the head-related transfer function stored in the HRTF storage section 319 without the IFFT processing section 117 performing IFFT processing. Note that, in this case, an FFT length of the FFT processing section 114 needs to be the same as an FFT length used in the ear-canal correction filter processing section 109.
As described above, the sound reproducing apparatus 300 according to the third embodiment of the present invention enhances accuracy of control of three-dimensional sound-field reproduction, and can realize an out-of-head sound localization in a more natural state, in addition to providing the effects of the first embodiment.
Moreover, in the sound reproducing apparatus 300 according to the third embodiment of the present invention, since sound localization processing using the head-related transfer function is performed in the analysis section 308, an amount of operation performed on the sound source signal in the reproduction mode can be reduced in comparison with the sound reproducing apparatus 200 according to the second embodiment.
The sound reproducing apparatus 400 according to the fourth embodiment shown in
The sound reproducing apparatus 400 according to the fourth embodiment measures only the wearing-state signal in the measurement mode. The analysis section 408 obtains an ear-canal correction filter based on the wearing-state signal by the following process.
The FFT processing section 414 performs fast Fourier transform processing on the wearing-state signal outputted from the A/D conversion section 107, to transform the wearing-state signal to a signal in frequency domain. The memory section 415 stores the wearing-state signal in frequency domain obtained through the FFT processing. The coefficient calculation section 416 reads out the wearing-state signal stored in the memory section 415, and analyzes the frequency component of the wearing-state signal to obtain frequencies of a peak and a dip.
The frequencies of the peak and the dip are resonance frequencies of the ear canal. The resonance frequencies can be specified from the wearing-state signal measured in a state where the earphone 110 is worn in the ear. Note that, among resonance frequencies, a range of frequencies causing high resonances which require ear canal correction is from 2 kHz to 10 kHz, with a length of the ear canal taken into consideration. Therefore, upon the calculation of a peak and a dip, an amount of operation can be reduced by calculating only those within the above range of frequencies.
The standard ear-canal correction filter storage section 420 stores parameters of the standard ear-canal filter and the standard ear-canal correction filter which are measured in a state where a particular earphone is attached to an ear-canal simulator which simulates an ear canal of a standard person. Each of the standard ear-canal filter and the standard ear-canal correction filter is formed by an IIR filter. The IIR filter includes a center frequency F, a gain G, and a transition width Q as parameters. The coefficient calculation section 416 reads out the parameters of the standard ear-canal filter from the standard ear-canal correction filter storage section 420, after calculating the frequencies of the peak and the dip of a measured frequency characteristic. The coefficient calculation section 416 corrects the center frequencies F to the corresponding frequencies of the peak and the dip.
After the correction of the standard ear-canal correction filter is completed, the coefficient calculation section 416 converts the standard ear-canal correction filter from a filter for an IIR filter to a filter for an FIR filter, and gives the standard ear-canal correction filter to the ear-canal correction filter processing section 109. In the case where the ear-canal correction filter is formed by an IIR filter, an IIR filter coefficient may be calculated from parameters of the IIR filter and may be given to the ear-canal correction filter processing section 109.
As described above, the sound reproducing apparatus 400 according to the fourth embodiment of the present invention corrects a peak frequency and a dip frequency of the standard ear-canal correction filter based on a measured wearing-state signal. Thus, the effects of the first embodiment can be realized with a small number of measurements. The correction method of the fourth embodiment can be applied to the second and third embodiments in a similar manner.
The sound reproducing apparatus 500 according to the fifth embodiment shown in
The sound reproducing apparatus 500 according to the fourth embodiment measures only the wearing-state signal in the measurement mode. The analysis section 508 obtains an ear-canal correction filter based on the wearing-state signal by the following process.
The resampling processing section 518 performs resampling processing on a wearing-state signal outputted from the A/D conversion section 107. For example, when a sampling frequency for the wearing-state signal is 48 kHz, the same processing as conversion to 24 kHz is performed. This processing means that, since a resonance frequency of a resonance characteristic in the case where one end is closed is equal to ½ of a resonance frequency of a resonance characteristic in the case where both ends are closed, a frequency characteristic in the case where one end is closed is calculated in a simulated manner by converting, to ½, a frequency characteristic measured in a state where both ends are closed.
The FFT processing section 514 performs fast Fourier transform (FFT) processing on the wearing-state signal outputted from the A/D conversion section 107, and on the unwearing-state simulation signal on which resampling processing has been performed by the resampling processing section 518, to transform them to signals in frequency domain, respectively. The memory section 115 stores the two signals in frequency domain obtained through the FFT processing. The coefficient calculation section 116 reads out the two signals stored in the memory section 115, and subtracts the unwearing-state simulation signal from the wearing-state signal to obtain a difference therebetween as a coefficient. The coefficient represents a conversion from a state of wearing the earphone 110 to a state (unwearing state) of not wearing the earphone 110.
As described above, the sound reproducing apparatus 500 according to the fifth embodiment of the present invention performs resampling processing on the wearing-state signal to obtain an unwearing-state simulation signal. Thus, the effects of the first embodiment can be realized with a small number of measurements. The correction method of the fifth embodiment can be applied to the second and third embodiments in a similar manner.
Processings executed in the measurement modes described in the first to fifth embodiments is typically executed via a personal computer (PC) 501 as shown in
Thus, if the processings in the measurement modes can be executed by using the PC 501, there is no need for the sound reproducing apparatuses 100 to 500 to have functions of executing the processing in the measurement modes.
A sound reproducing apparatus of the present invention is applicable to a sound reproducing apparatus or the like which performs sound reproduction by using an in-ear earphone, and particularly, is useful, e.g., when it is desired to realize a listening state equivalent to that in the case where the ear canal is not blocked, even when wearing the earphone in the ear.
Number | Date | Country | Kind |
---|---|---|---|
2008-102275 | Apr 2008 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2009/001574 | 4/3/2009 | WO | 00 | 12/8/2009 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2009/125567 | 10/15/2009 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6658122 | Westermann et al. | Dec 2003 | B1 |
6687377 | Voix et al. | Feb 2004 | B2 |
7082205 | Westermann | Jul 2006 | B1 |
7313241 | Hamacher et al. | Dec 2007 | B2 |
7715577 | Allen et al. | May 2010 | B2 |
7953229 | Saito et al. | May 2011 | B2 |
7957549 | Chiba et al. | Jun 2011 | B2 |
8050421 | Fukuda et al. | Nov 2011 | B2 |
8081769 | Fukuda et al. | Dec 2011 | B2 |
8111849 | Tateno et al. | Feb 2012 | B2 |
20040196991 | Iida et al. | Oct 2004 | A1 |
Number | Date | Country |
---|---|---|
5-199596 | Aug 1993 | JP |
2000-92589 | Mar 2000 | JP |
2002-209300 | Jul 2002 | JP |
2003-102099 | Apr 2003 | JP |
2008-177798 | Jul 2008 | JP |
Number | Date | Country | |
---|---|---|---|
20100177910 A1 | Jul 2010 | US |