The present invention contains subject matter related to Japanese Patent Application JP 2004-147600 filed in the Japanese Patent Office on May 18, 2004, the entire contents of which are incorporated herein by reference.
1. Field of the Invention
The present invention relates to a method and apparatus for sound pickup using a plurality of microphones, and to a method and apparatus for sound reproduction using loudspeakers or headphones based on a sound signal picked up by the sound pickup method and apparatus.
2. Description of the Related Art
Binaural recording is one sound pickup method for recording sound from sound sources placed in an acoustic space while maintaining information about the direction of incoming sound.
Another sound pickup method other than binaural recording for recording sound from sound sources while maintaining information about the direction of incoming sound is to use multiple directional microphones as pickup microphones.
In the multi-directional microphone method, a plurality of, e.g., three, directional microphones are placed with their directional ranges covering different areas for individually picking up sound from different areas. In the sound reproduction side, a plurality of, e.g., three, loudspeakers are placed similarly to the pickup areas for reproducing and outputting sound to a listener.
If the microphones have insufficient directional selectivity, a matrix operation is performed on sound signals output from the microphones to obtain sharper directivity, resulting in high spatial resolution during sound reproduction. A directional characteristic is produced by a matrix operation, and therefore an omni-directional microphone may be used as a pickup microphone.
Such a multi-directional microphone method (including the use of omni-directional microphones, in which a directional characteristic is produced by a matrix operation) is advantageous over the binaural method to give a natural auditory sensation to a listener who changes his/her head orientation during sound reproduction as if he/she changes his/her head orientation in the recording site.
Japanese Unexamined Patent Application Publication No. 2002-271885 discloses a microphone system having three pairs of microphones placed around a reference microphone, in which sound signals output from the microphones are subjected to digital signal processing to control the directional characteristics of the microphones.
Japanese Unexamined Patent Application Publication No. 9-70094 discloses a headphone apparatus that detects motion of the head of a listener who wears a headphone and that processes multi-channel sound signals depending upon the orientation of the listener's head to localize the sound image outside the listener's head in front of or behind the listener.
With recent compact recording devices, high-performance signal processors, large-capacity recording media, etc., it is common to record sound, such as environmental sound, using a plurality of microphones together with information about the direction of incoming sound in an informal, simple, unconscious manner to experience the recorded sound later and to share the experience with others through communication over a network.
In such a recording scene, the orientation of the microphones may frequently be changed during sound pickup.
In either binaural recording or multi-directional microphone recording, when sound is picked up using a plurality of microphones, the orientation of the microphones should not be changed during sound pickup because the change in orientation contributes to listener's confusion about auralization of the sound field during sound reproduction.
Specifically, in a multi-directional microphone system, as shown in
For sound pickup using a plurality of microphones, therefore, it is necessary to fix the microphones during sound pickup, which is not suitable for the informal recording described above.
It is therefore desirable to suppress the listener's confusion about auralization of the sound field during sound reproduction, which is caused by a change in orientation of microphones during sound pickup, when sound is picked up by a multi-directional microphone system and is transmitted to the sound reproduction side and the sound is reproduced using loudspeakers or headphones in the sound reproduction side.
A sound pickup method according to an embodiment of the present invention includes the steps of picking up sound using a plurality of microphones, the plurality of microphones being arranged so that directivity axes of the microphones differ from each other or functioning as a plurality of microphones having directivities in different directions by performing a calculation on sound signals output from the plurality of microphones, detecting rotation of the plurality of microphones, processing the sound signals output from the plurality of microphones according to the detected rotation so that a change in orientation of each of the microphones is canceled, and outputting the processed output sound signals.
A sound pickup and reproduction method according to another embodiment of the present invention includes the steps of picking up sound using a plurality of microphones, the plurality of microphones being arranged so that directivity axes of the microphones differ from each other or functioning as a plurality of microphones having directivities in different directions by performing a calculation on sound signals output from the plurality of microphones, detecting rotation of the plurality of microphones, transmitting information about the detected rotation and the sound signals output from the microphones, receiving the transmitted rotation information and sound signals output from the microphones, and processing the received sound signals output from the microphones according to the received rotation information so that a change in orientation of each of the microphones is canceled.
In the multi-directional microphone system, unlike a two-channel stereo system or a binaural system, incoming sounds from different areas are picked up on an area basis.
The sound pickup method utilizes the feature of the multi-directional microphone system described above to process sound signals output from microphones according to the detected rotation of the microphones so that a change in orientation of each of the microphones is canceled, and transmits the processed output sound signals to the sound reproduction side. Thus, the confusion of a listener about auralization of the sound field in the sound reproduction side, which is caused by a change in orientation of the microphones during sound pickup, can be suppressed.
The sound pickup and reproduction method also utilizes the feature of the multi-directional microphone system described above to transmit sound signals output from microphones and rotation information about the detected rotation of the microphones to the sound reproduction side. In the sound reproduction side, the sound signals output from the microphones are processed according to the rotation information so that a change in orientation of each of the microphones in the sound pickup side is canceled. Thus, the confusion of a listener about auralization of the sound field in the sound reproduction side, which is caused by a change in orientation of the microphones during sound pickup, can be suppressed.
Therefore, when sound picked up using a multi-directional microphone system is transmitted to the sound reproduction side and the sound is reproduced using loudspeakers or a headphone in the sound reproduction side, the confusion of a listener about auralization of the sound field, which is caused by a change in orientation of the microphones during sound pickup, can be suppressed in the sound reproduction side.
The sound pickup and reproduction system shown in
Specifically, the microphones 1a, 1b, and 1c are unidirectional or omni-directional microphones, and are arranged radially with intervals of 120 degrees off-axis. The loudspeakers 4a, 4b, and 4c are also arranged with intervals of 120 degrees around the listener 5.
A synthesis processor 13 converts output sound signals Sa, Sb, and Sc from the microphones 1a, 1b, and 1c into digital sound signals, which are then subjected to digital signal processing, and transmits the resulting signals to the sound reproduction side. In the sound reproduction side, the digital sound signals are subjected to digital signal processing by a reproduction processor 15, and are then converted into analog sound signals Ua, Ub, and Uc. The analog sound signals Ua, Ub, and Uc are supplied to the loudspeakers 4a, 4b, and 4c.
The output sound signals Sa, Sb, and Sc may be transmitted by, for example, exchanging them in real-time wirelessly or via lines or by recording them onto a recording medium and reading them from the recording medium. The listener 5 may be identical to or different from a user carrying out recording.
The sound pickup and reproduction system shown in
The rotation detector 11 is, for example, a rotation angular speed sensor. The rotation detector 11 calculates the integral of the output signal from the rotation detector 11 to determine a rotation angle of the multi-microphone device 2, or the microphones 1a, 1b, and 1c, and adds the rotation angle to the initial azimuth, thereby determining the azimuth of the multi-microphone device 2, or the microphones 1a, 1b, and 1c.
Alternatively, the rotation detector 11 may be a geomagnetic sensor or a gravity sensor. In this case, the rotation detector 11 can directly determine the azimuth of the multi-microphone apparatus 2, or the microphones 1a, 1b, and 1c.
In the sound pickup and reproduction system shown in
In the sound pickup apparatus shown in
Output sound signals Sa, Sb, and Sc from the microphones 1a, 1b, and 1c are amplified by sound amplification circuits 22a, 22b, and 22c, and are then converted into digital sound data Da, Db, and Dc by analog-to-digital (A/D) converters 23a, 23b, and 23c, respectively.
The digital sound data Da, Db, and Dc are recorded in a recording medium 25, e.g., a disc medium, by a recorder 24.
Output data Dr from the gyro 26 indicating the azimuth of the microphones 1a, 1b, and 1 is recorded in a memory 27.
In order to associate the sound data Da, Db, and Dc recorded in the recording medium 25 with the azimuth data Dr recorded in the memory 27, the recorder 24 is controlled by timing information from a timing generator 28 to record the sound data Da, Db, and Dc in the recording medium 25. The timing information from the timing generator 28 is recorded in the memory 27 together with the azimuth data Dr.
The timing information may be formed of various time codes obtained by the recorder 24 or various synchronization signals.
An empty track on the recording medium 25 may be used as the memory 27. The recording medium 25 and the memory 27 may integrally be formed by a single hard disk.
If the microphones 1a, 1b, and 1c have insufficient directional characteristics or if omni-directional microphones are used as the microphones 1a, 1b, and 1c, as shown in
In
For example, the directivity processor 29 performs calculations given by the equations below to determine the sound data Ea, Eb, and Ec:
Ea=Db+Dc−Da Eq. 1(a)
Eb=Da+Dc−Db Eq. 1(b)
Ec=Da+Db−Dc Eq. 1(c)
The calculation given by Eq. 1(b) allows, for example, a unidirectional (cardioid) microphone having a directional characteristic pattern shown in
This directional processing may be performed after the unprocessed sound data Da, Db, and Dc are recorded in the recording medium 25 and read from the recording medium 25.
During sound pickup, if a recording operator 7 changes his/her head orientation by looking around, the multi-microphone microphone device 2 and the gyro 26 rotate to cause a change in orientation of the microphones 1a, 1b, and 1c. The change of orientation is detected by the gyro 26.
The multi-microphone device 2 may have a configuration shown in
In the sound reproduction apparatus shown in
The retrieval of the sound data Da, Db, and Dc from the recording medium 25 and the reading of the azimuth data Dr from the memory 27 are controlled by the controller 52 based on the timing information recorded in the memory 27 so that the retrieval of the sound data Da, Db, and Dc from the recording medium 25 and the reading of the azimuth data Dr from the memory 27 are performed at the same timing as those in the sound pickup processing.
The computation processor 61 performs processing so that the sound data Da, Db, and Dc cancel a change in orientation of the microphones 1a, 1b, and 1c in the sound pickup side based on the azimuth data Dr in the manner described below.
Processed sound data Ta, Tb, and Tc are converted into analog sound signals by digital-to-analog (D/A) converters 54a, 54b, and 54c, and the converted three-channel sound signals are amplified by sound amplification circuits 55a, 55b, and 55c, respectively. The amplified sound signals Ua, Ub, and Uc are supplied to the loudspeakers 4a, 4b, and 4c, respectively.
The loudspeakers 4a, 4b, and 4c are arranged with intervals of 120 degrees around the listener 5 in a similar fashion to the arrangement of the main directivity axes of the microphones 1a, 1b, and 1c in the sound pickup apparatus shown in
Adder circuits 66a, 66b, and 66c calculate additions given by the equations below to determine the processed sound data Ta, Tb, and Tc:
Ta=Kaa×Da+Kba×Db+Kca×Dc Eq. 2(a)
Tb=Kab×Da+Kbb×Db+Kcb×Dc Eq. 2(b)
Tc=Kac×Da+Kbc×Db+Kcc×Dc Eq. 2(c)
The coefficient generation circuit 65 changes the values of the coefficients Kaa, Kab, Kac, Kba, Kbb, Kbc, Kca, Kcb, and Kcc depending upon the value of the azimuth data Dr, that is, the rotation direction and the amount of rotation (rotation angle) of the microphones 1a, 1b, and 1c in the sound pickup side, so that a change in orientation of the microphones 1a, 1b, and 1c is canceled.
Thus, the sound data that allows for cancellation of a change in orientation of the microphones 1a, 1b, and 1c in the sound pickup side can be obtained as the sound data Ta, Tb, and Tc. The confusion of the listener 5 about auralization of the sound field can therefore be suppressed.
As described above with reference to
The sound reproduction apparatus shown in
However, if sound is simply reproduced using a headphone, a sound image is localized in the listener's head, which produces an unnatural auditory sensation.
In headphone reproduction, therefore, it is desirable to perform out-of-head localization processing for localizing the sound image outside the listener's head using the so-called HRTF (Head-Related Transfer Function) technique and processing for generating a sound field that produces a sensation like the listener is surrounded by loudspeakers in the manner shown in
In the sound reproduction apparatus shown in
The digital filters 72a, 72b, and 72c convolve an impulse response from the position of a virtual loudspeaker at which the sound image is to be localized to the left ear of the listener 5 with the sound data Ta, Tb, and Tc, respectively. The digital filters 73a, 73b, and 73c convolve an impulse response from the position of a virtual loudspeaker at which the sound image is to be localized to the right ear of the listener 5 with the sound data Ta, Tb, and Tc, respectively.
Thus, even in headphone reproduction, an auditory sensation like acoustics being reproduced using loudspeakers in the manner shown in
In the sound reproduction apparatus shown in
Specifically, as shown in
The digital filters 92a, 92b, and 92c convolve an impulse response from the position of a virtual loudspeaker at which the sound image is to be localized to the left ear of the listener 5 with the sound data Da, Db, and Dc, respectively. The digital filters 93a, 93b, and 93c convolve an impulse response from the position of a virtual loudspeaker at which the sound image is to be localized to the right ear of the listener 5 with the sound data Da, Db, and Dc, respectively. At the same time, the coefficient generation circuit 94 in the reproduction processor 91 changes the coefficients of the digital filters 92a, 93a, 92b, 93b, 92c, and 93c depending upon the value of the azimuth data Dr, that is, the rotation direction and the amount of rotation (rotation angle) of the microphones 1a, 1b, and 1c in the sound pickup side, so that a change in orientation of the microphones 1a, 1b, and 1c in the sound pickup side is canceled.
The large difference between loudspeaker reproduction and headphone reproduction is an auditory sensation produced when a listener moves his/her head to the right and left. When the listener shakes his/her head in loudspeaker reproduction, a natural auditory sensation like the listener is shaking his/her head in the recording site is produced. On the other hand, when the listener shakes his/her head in headphone reproduction, the sound field also moves, which does not produce an auditory sensation like the listener is in the recording site.
It is therefore desirable to perform headphone reproduction while performing an operation to overcome this problem.
The sound reproduction apparatus shown in
Azimuth data Dj output from the gyro 85 is supplied to the reproduction processor 91 together with the sound data Da, Db, and Dc retrieved from the recording medium 25 by the player 51 shown in
Specifically, as shown in
The digital filters 92a, 92b, and 92c convolve an impulse response from the position of a virtual loudspeaker at which the sound image is to be localized to the left ear of the listener 5 with the sound data Da, Db, and Dc, respectively. The digital filters 93a, 93b, and 93c convolve an impulse response from the position of a virtual loudspeaker at which the sound image is to be localized to the right ear of the listener 5 with the sound data Da, Db, and Dc, respectively. At this time, the coefficient generation circuit 94 in the reproduction processor 91 changes the coefficients of the digital filters 92a, 93a, 92b, 93b, 92c, and 93c depending upon the values of the azimuth data Dr and Dj, that is, the rotation direction and the amount of rotation (rotation angle) of the microphones 1a, 1b, and 1c in the sound pickup side and the direction and the amount of head movement (rotation angle) of the listener 5 in the sound reproduction side, so that both a change in orientation of the microphones 1a, 1b, and 1c in the sound pickup side and a movement of the sound field caused by head movement of the listener 5 in the sound reproduction side are canceled.
Therefore, the confusion of the listener 5 about auralization of the sound field in the sound reproduction side, which is caused by a change in orientation of the microphones 1a, 1b, and 1c in the sound pickup side, can be suppressed. Moreover, the sound field can also be prevented from moving along with head movement when the listener 5 shakes his/her head, which thus produces a natural auditory sensation like the listener 5 is shaking his/her head in the recording site, like loudspeaker reproduction.
In the first embodiment, the multi-microphone device 2 mounted on the head of the recording operator 7 rotates when the recording operator 7 changes the orientation of his/her head during sound pickup to cause a change in orientation of the microphones 1a, 1b, and 1c. In some cases, the recording operator 7 changes not only the orientation of his/her head but also the orientation of his/her body itself during sound pickup, for example, when the recording operator 7 turns to the right or left on a street during sound pickup and when the recording operator 7 rides on a vehicle, such as a roller coaster, and the vehicle turns during sound pickup.
In such cases, acoustic surroundings to be recorded change, and environmental views or landscape views to be taken using a video camera also change at the same time. In such cases, it is desirable not to cancel a change in orientation of the microphones 1a, 1b, and 1c.
A case in which a change in orientation of microphones caused by a rotation of the operator's body is not canceled will be described with reference to
Azimuth data Dt output from the gyro 89 and azimuth data Dr output from the gyro 26 are supplied to a difference detection circuit 36, and the difference detection circuit 36 determines difference data Ds between the azimuth data Dt and Dr.
The difference data Ds indicates the rotation direction and the amount of rotation (rotation angle) of the head of the recording operator 7 on which the multi-microphone device 2 and the gyro 26 are mounted with respect to the body of the recording operator 7 on which the gyro 89 is mounted.
When the azimuth data Dr and Dt change in the manner shown in
In the sound pickup apparatus shown in
In the second embodiment, therefore, a change in orientation of the microphones 1a, 1b, and 1c is not canceled when the change of orientation is caused by a rotation of the body of the recording operator 7. Only when the orientation of the microphones 1a, 1b, and 1c changes with respect to the body of the recording operator 7, the change of orientation is canceled.
The sound data picked up by microphones may be transmitted directly to a remote user (listener) without recording it in a recording medium, or the picked-up sound data may be recorded in a recording medium and reproduced from the recording medium before it is transmitted to a remote user (listener) via a network. The sound data may be transmitted to a plurality of users. The sound data may be compressed and encoded before it is transmitted.
While three pickup microphones are used in the foregoing embodiments, two pickup microphones or four or more pickup microphones may be used.
A plurality of microphones are not necessarily integrally formed or integrally rotated. A system for detecting rotation of the individual microphones and processing output sound data according to the detected rotation may be used. In this case, for example, these microphones may be worn separately by a plurality of operators.
It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Number | Date | Country | Kind |
---|---|---|---|
P2004-147600 | May 2004 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
4817153 | Fernandez | Mar 1989 | A |
5325436 | Soli et al. | Jun 1994 | A |
6021206 | McGrath | Feb 2000 | A |
6259795 | McGrath | Jul 2001 | B1 |
7515992 | Sawada et al. | Apr 2009 | B2 |
Number | Date | Country |
---|---|---|
63-051800 | Mar 1988 | JP |
4-132468 | May 1992 | JP |
08-084286 | Mar 1996 | JP |
09-070094 | Mar 1997 | JP |
10-023600 | Jan 1998 | JP |
11-262097 | Sep 1999 | JP |
2000-004493 | Jan 2000 | JP |
2000-165718 | Jun 2000 | JP |
2001-169309 | Jun 2001 | JP |
2002-271885 | Sep 2002 | JP |
2003-264900 | Sep 2003 | JP |
2003-319499 | Nov 2003 | JP |
2005-124090 | May 2005 | JP |
2005-159731 | Jun 2005 | JP |
2005-277832 | Oct 2005 | JP |
2006-503526 | Jan 2006 | JP |
Number | Date | Country | |
---|---|---|---|
20050259832 A1 | Nov 2005 | US |