The present disclosure relates to a sound field control device, a sound field control method, a program, a sound field control system and a server.
Conventionally, as described in Patent Literatures 1 to 3 listed below, for example, there has been proposed a device for correcting sound volume, delay and directional characteristics of a speaker depending on a position of a viewer and providing the viewer with optimum sound even at a position off a front position.
Patent Literature 1: JP 2005-049656A
Patent Literature 2: JP 2007-214897A
Patent Literature 3: JP 2010-206451A
In reproduction by a speaker, when a viewer auditions at a position which is off an assumed viewing position (normally, a position having an equal distance from all speakers, that is to say, a front position), balance of volume or timing of a sound arriving from each of the speakers is off, the sound quality deteriorates, or a normal position is displaced. In addition, there is a problem that virtual sound source reproduction effect is also lost if the viewer moves.
However, it is difficult for technologies described in Patent Literatures 1 to 3 to optimally adjust the virtual sound source reproduction since the technologies only assume adjustment of the sound volume, a delay amount, or the directional characteristics and give no consideration to size or orientation of a head.
In addition, if a display target object which is a sound source moves when a user plays a game on a mobile device or a tablet, there may arise a sense of discomfort between movement of the display target object and a sound that the user listens to.
Hence, it is needed to optimally adjust the virtual sound source reproduction.
According to the present disclosure, there is provided a sound field control device including a display target object position information acquisition unit for acquiring position information of a display target object corresponding to a sound source, and a virtual sound source position control unit for controlling a virtual sound source position on the basis of position information of the display target object.
Further, it is possible to further include a transmission unit for transmitting, to an external computer, at least the position information of the display target object, and a reception unit for receiving, from the external computer, a virtual sound source reproduction correction factor computed on the basis of the position information of the display target object or information generated on the basis of the virtual sound source reproduction correction factor.
Further, the transmission unit may transmit, to the external computer, sound data together with the position information of the display target object, and the reception unit may receive, from the external computer, sound data that is obtained by correcting the sound data with the virtual sound source reproduction correction factor computed on the basis of the position information of the display target object.
Further, it is possible to further include a viewer position information acquisition unit for acquiring position information of a viewer, and the virtual sound source position control unit may control the virtual sound source position on the basis of the position information of the display target object and the position information of the viewer.
Further, the viewer position information acquisition unit may acquire the position information of the viewer from information obtained by imaging.
Further, it is possible to further include a transmission unit for transmitting, to the external computer, the position information of the display target object and the position information of the viewer, and a reception unit for receiving, from the external computer, a virtual sound source reproduction correction factor computed on the basis of the position information of the display target object and the position information of the viewer or information generated on the basis of the virtual sound source reproduction correction factor.
Further, the transmission unit may transmit, to the external computer, sound data together with the position information of the display target object and the position information of the viewer, and the reception unit may receive, from the external computer, sound data which is obtained by correcting the sound data with the virtual sound source reproduction correction factor computed on the basis of the position information of the display target object and the position information of the viewer.
According to the present disclosure, there is provided a sound field control device including acquiring position information of a display target object corresponding to a sound source, and controlling a virtual sound source position on the basis of the position information of the display target object.
According to the present disclosure, there is provided a program for causing a computer to function as means for acquiring position information of a display target object corresponding to a sound source, and means for controlling a virtual sound source position on the basis of the position information of the display target object.
According to the present disclosure, there is provided a sound field control system including a client terminal including a display target object position information acquisition unit for acquiring position information of a display target object corresponding to a sound source, a transmission unit for transmitting the position information of the target object to an external computer, and a reception unit for receiving, from the external computer, a virtual sound source reproduction correction factor computed on the basis of the position information of the target object; and the external computer including a reception unit for receiving the position information of the display target object, a virtual sound source reproduction correction factor computation unit for computing the virtual sound source reproduction correction factor on the basis of the position information of the display target object, and a transmission unit for transmitting, to the client terminal, the virtual sound source reproduction correction factor or information generated on the basis of the virtual sound source reproduction correction factor.
According to the present disclosure, there is provided a server including the external computer including a reception unit for receiving, from a client terminal, position information of a display target object corresponding to a sound source, a virtual sound source reproduction correction factor computation unit for computing the virtual sound source reproduction correction factor on the basis of the position information of the display target object, and a transmission unit for transmitting, to the client terminal, the virtual sound source reproduction correction factor or information generated on the basis of the virtual sound source reproduction correction factor.
According to the present disclosure, there is provided a sound field control method including acquiring, by a client terminal, position information of a display target object corresponding to a sound source, transmitting, by the client terminal, the position information of the target object to an external computer, receiving, by the external computer, the position information of the display target object, computing, by the external computer, the virtual sound source reproduction correction factor on the basis of the position information of the display target object, and transmitting, by the external computer, to the client terminal, the virtual sound source reproduction correction factor or information generated on the basis of the virtual sound source reproduction correction factor.
According to the present disclosure, there is provided a sound field control device including a position information acquisition unit for acquiring position information of a viewer from information obtained by imaging, and a virtual sound source position control unit for controlling a virtual sound source position on the basis of the position information.
The virtual sound source position control unit may control the virtual sound source position in a manner that a normal position of a sound image is fixed irrespective of a position of the viewer.
The virtual sound source position control unit may control the virtual sound source position in a manner that a normal position of a sound image relatively moves according to a position of the viewer.
The virtual sound source position control unit may control the virtual sound source position by changing a head transfer function, on the basis of the position information.
The virtual sound source position control unit may control the virtual sound source position, on the basis of the position information, by smoothly changing a factor before a position of the viewer changes to a factor after the position of the viewer changes.
The virtual sound source position control unit may control the virtual sound source position, on the basis of the position information, when movement of the viewer exceeds a predetermined value.
It is possible to further include a control unit for controlling sound volume, a delay amount of sound, or a directional characteristic, on the basis of the position information.
It is possible to further include an imaging unit for acquiring the position information of the viewer.
It is possible to further include a posture information acquisition unit for acquiring posture information, and the virtual sound source position control unit may control the virtual sound source position, on the basis of the position information and the posture information.
The position information acquisition unit may acquire, from another device including an imaging unit for imaging the viewer, information obtained by the imaging.
According to the present disclosure, there is provided a sound field control method including acquiring position information of a viewer, and controlling a virtual sound source position on the basis of the position information.
According to the present disclosure, there is provided a program for causing a computer to function as means for acquiring position information of a viewer, and means for controlling a virtual sound source position on the basis of the position information.
According to the present disclosure, there is provided a sound field control system including an imaging device for imaging a viewer, and a sound field control device including a position information acquisition unit for acquiring position information of the viewer from information obtained from the imaging device, and a virtual sound source position control unit for controlling a virtual sound source position on the basis of the position information.
According to the present disclosure, virtual sound source reproduction can be optimally adjusted.
Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the appended drawings. Note that, in this specification and the drawings, elements that have substantially the same function and structure are denoted with the same reference signs, and repeated explanation is omitted.
Note that a description will be given in the following order:
1. First Embodiment
1.1. Appearance example of a sound field control device
1.2. Configuration example of a sound field control unit
1.3. Configuration example of a sound field adjustment processing unit
1.4. Processing in the sound field control device
1.5. Positional relationship between a viewer and a sound output unit
1.6. Processing in a virtual sound source reproduction correction unit
1.7. Processing in a sound volume correction/change unit
1.8. Processing in a delay amount correction/change unit
1.9. Processing in a virtual sound source reproduction correction/change unit and a directional characteristic correction/change unit
1.10. Specific configuration example of the sound field control device
2. Second Embodiment
2.1. Overview of a second embodiment
2.2. Processing to be performed in a virtual sound source reproduction correction/change unit of the second embodiment
3. Third embodiment
3.1. Overview of a third embodiment
3.2. Configuration example of the third embodiment
The imaging unit 102 images a face and a body of the viewer (user) listening to the sound. The viewing position computation unit 104 computes a position of the viewer and orientation of the face from an image obtained from the imaging unit 102. Note that the imaging unit 102 (and the viewing position computation unit 104) may be provided in a separate device from a device in which the sound field control device 100 is provided. A sound source is inputted into the sound control unit 106. The sound control unit 106 processes the sound so that good sound quality, normal position, and virtual sound source reproduction (virtual surround) effect can be obtained, depending on a position of the viewer. The sound output unit 108 is a speaker for outputting the sound controlled by the sound control unit 106.
[1.2. Configuration Example of a Sound Field Control Unit]
The factor change determination unit 110 determines whether or not to change a factor on the basis of an image of a viewer imaged by the imaging unit 102. If the factor change determination unit 110 updates the factor every time the viewer moves only slightly or moves his or her face slightly, it is likely that a change in a tone color when a factor is updated cannot be ignored. Thus, the factor change determination unit 110 does not change a factor if motion is small. The factor change determination unit 110 makes a determination to change the factor when there is a significant (more than predetermined) change in the viewer position, which is then stabilized. In this case, the factor computation unit 112 computes an optimal sound field processing factor depending on the changed viewer position.
The factor change/sound field adjustment processing unit 114 performs sound field adjustment processing while changing the factor. The factor change/sound field adjustment processing unit 114 performs the sound field adjustment processing, while making a factor change from a factor corresponding to a previous viewer position to a factor of a current viewer position which is newly computed by the factor computation unit 112. Then, the factor change/sound field adjustment processing unit 114 smoothly changes the factor so that noise such as a sound interruption does not occur.
In addition, while the factor is being changed, the factor is not reset even if the sound control unit 106 receives a new position information computation result sent from the viewing position computation unit 104. For this reason, the factor is not changed more than is necessary, and timing of when position information is sent from the viewing position detection unit 104 does not have to be synchronous with timing of the sound processing.
On the one hand, when the viewer position does not change and if the factor change determination unit 110 determines not to change the factor, the sound field adjustment processing unit 116 performs regular sound field adjustment processing appropriate for the viewing position. The normal sound field adjustment processing corresponds to processing in step S32 in
[1.3. Configuration Example of a Sound Field Adjustment Processing Unit]
In the following, a configuration of the sound field adjustment processing unit 116 will be described.
If a viewer position is displaced from an assumed viewing position (assumed auditioning position), the sound volume correction unit 122, the delay amount correction unit 124, and the directional characteristic correction unit 126 correct sound volume difference, arrival time difference, and a change in frequency characteristics of a sound arriving from each speaker, which are generated due to the displacement. The sound volume correction unit 122 corrects the sound volume difference, the delay amount correction unit 124 corrects the arrival time difference, and the directional characteristic correction unit 126 corrects the change in the frequency characteristics. Now, in many cases, the assumed viewing position (assumed viewing position) is a center position of right and left speakers of a television or audio system and the like, that is, a front of the television or audio system.
The sound volume correction unit 122 corrects sound volume on the basis of a viewer position acquired from the viewing position computation unit 104 so that the sound volume reaching the viewer from each speaker is equal. Sound volume A is proportional to a distance r, from each speaker to the center of a viewer's head and the following expression is true. In the following expression, Atti is a distance between the assumed auditioning position and the speaker.
Atti=ri/ro
Based on the viewer position acquired from the viewing position computation unit 104, the delay amount correction unit 124 corrects a delay amount so that time to reach the viewer from each speaker is equal. The delay amount ti of each speaker is expressed by the following expression where the distance from each speaker to the center of the viewer's head is ri and the largest ri is rmax. However, c is sound velocity.
ti=(rmax−ri)/c
Based on the viewer position acquired from the viewing position computation unit 104, the directional characteristic correction unit 126 corrects the frequency characteristic of the directional characteristics of each speaker that is changed due to the displacement of the viewing position to a characteristic at the assumed viewing position. The corrected frequency characteristic Ii is obtained by the following expression where the frequency characteristic of a speaker i at the assumed viewing position is Hi and the frequency characteristic at the viewing position is Gi.
Ii=Hi/Gi
In the following, processing in the directional characteristic correction unit 126 will be described in more detail.
As can be seen from
Thus, the directional characteristics of the speaker are measured, an equalizer which may correct any effect of the directional characteristics is computed in advance, and equalizer processing is performed depending on detected direction information θh, θv, that is, orientation of the speaker main body to the listener. This enables implementation of well-balanced reproduction that does not rely on the orientation of the speaker to the listener.
As an example of a correction filter, a correction filter S can be obtained by the following expression where a frequency characteristic at an ideal viewing position is Hideal and a characteristic at a position away therefrom is H.
S=Hideal/H
In the following, a configuration of the factor change/sound field adjustment unit 114 in
Basic processing in the factor change/sound field adjustment unit 114 is similar to the virtual sound source reproduction correction unit 120, the sound volume correction unit 122, the delay amount correction unit 124, and the directional characteristic correction 126 in
[1.4. Processing in the Sound Field Control Device]
In the following, processing in the sound field control device 100 according to the embodiment will be described.
In addition, in step S20, it is determined based on a factor in-transition flag whether or not factor change processing is in transition. If the factor change processing is in transition (the factor in-transition flag is set), the process proceeds to step S22 where the factor transition processing is continuously performed. The factor transition process in step S22 corresponds to the processing of the factor change/sound field adjustment unit 114 described in
Following step S22, the process proceeds to step S24. In step S24, it is determined whether or not the factor transition has ended. If the factor transition has ended, the process proceeds to step S26 where the factor in-transition flag is released. Following step S24, the process returns to START. On the one hand, if the factor transition has not ended in step S24, the process returns to START without releasing the factor in-transition flag.
In addition, in step S20, if the factor is not in transition (the factor in-transition flag is released), the process proceeds to step S28. In step S28, based on a result of the position change smoothing in step S12, it is determined whether or not the viewing position has changed. If the viewing position has changed, the process proceeds to step S30. In step S30, a target factor is changed and the factor in-transition flag is set. Following step S30, the process proceeds to step S32 where normal processing is performed.
On the one hand, in step S28, if the viewing position has not changed, the process proceeds to the normal processing in step S32 without setting the factor in-transition flag. Following step S32, the process returns to START.
[1.5. Positional Relationship Between a Viewer and a Sound Output Unit]
If processing of the sound volume correction unit 122, the delay amount correction unit 124 and the directional characteristic correction unit 126 corrects the sound volume difference, the arrival time difference, and the change in the frequency characteristic, respectively, in the sounds reaching from respective speakers, the sounds are adjusted so that they have equal values to a case in which the left (L) sound output unit 108 in
However, only with the processing of the sound volume correction unit 122, the delay amount correction unit 124 and the directional characteristic correction unit 126, the virtual sound source reproduction effect cannot be adequately corrected because an angular aperture of the speaker, a distance between the speaker and the viewer, and orientation of the viewer's face change. Thus, the virtual sound source reproduction correction/change unit 130 according to the embodiment makes a correction so as to obtain the virtual sound source reproduction effect.
[1.6. Processing in a Virtual Sound Source Reproduction Correction Unit]
The virtual sound source reproduction correction unit 120 changes each parameter for the virtual sound source reproduction. Main parameters include a head transfer function, direct sound, a delay amount in crosstalk and the like. That is, a change in the head transfer function due to a change in the angular aperture of the speaker (sound volume correction unit 122), the distance between the speaker and the viewer, the orientation of the viewer's face is corrected. In addition, in a case where a sound source is actually placed at the virtual sound source position, the virtual sound source reproduction correction unit 120 can address the change in the orientation of the viewer's face by making a correction to a difference in the direct sound and the delay amount in crosstalk.
In the following, a method for creating a head transfer function and a method for switching the head transfer function depending on a viewer position by the virtual sound source reproduction correction unit 120 of the first embodiment will be described.
(1) Measurement of a Head Transfer Function
As shown in
(2) Computation of a Virtual Sound Source Reproduction Correction Factor
For example, computation of a virtual sound source reproduction correction factor at a viewing position 1 in
Note that in the above expressions,
In addition, since it can be considered in an approximate manner that SPL and SPR are corrected to an equal distance/identical angle by the sound volume correction unit, the delay amount correction unit, and the directional characteristic correction unit, approximation can be performed such as H1LL=H1RR and H1LR=H1RL. Therefore, as shown below, the virtual sound source reproduction correction factor can be determined from a smaller number of tables.
(3) Switching of Head Transfer Functions
For example, in
Note that for a reason similar to the above, approximation can be performed such as H2LL=H2RR and H2LR=H2RL. Therefore, as shown below, the virtual sound source reproduction correction factor can be determined from a smaller number of tables.
In addition, processing of the sound volume correction unit 122, the delay amount correction unit 124, and the directional characteristic correction unit 126 can be considered as a change in head transfer functions. However, when a correction is made only with the head transfer functions, data of the head transfer functions corresponding to each position must be held, which thus extends the tone. Therefore, it is preferred to divide the head transfer functions into respective parts.
[1.7. Processing in a Sound Volume Correction/Change Unit]
As shown in
Att=AttCurr+αt
[1.8. Processing in a delay amount correction/change unit]
As shown in
As shown in
The addition unit 132d adds the past delay amount AttCurr outputted from the attenuator 134b to the new delay amount AttTrgt outputted from the attenuator 134c. This enables a smooth change from the past delay amount AttCurr to the new delay amount AttTrgt as time elapses.
[1.9. Processing in a Virtual Sound Source Reproduction Correction/Change Unit and a Directional Characteristic Correction/Change Unit]
As shown in
As shown in
The addition unit 130e adds the past signal AttCurr outputted from the attenuator 130c to the new signal AttTrgt outputted from the attenuator 132s. This enables a smooth change from the past signal AttCurr to the new signal AttTrgt as time elapses.
Similarly, as shown in
[1.10. Specific Configuration Example of the Sound Field Control Device]
With the above configuration, a viewer can obtain the appropriate virtual sound source reproduction effect and feel an appropriate normal position or spatial expanse.
Note that it is also possible to perform correction processing for a plurality of persons by using a plurality of speakers. In the case of the plurality of persons, performing the virtual sound source reproduction correction, in particular, is effective.
As described above, according to the first embodiment, since each parameter is changed for the virtual sound source reproduction on the basis of a viewer position, the virtual sound source reproduction effect can be obtained irrespective of a viewing position, thereby making it possible to feel an appropriate normal position or spatial expanse.
In addition, provision of the viewing position computation unit 104 for real-time detecting positional relationships among and angles of a viewer and a plurality of speakers enables real-time detection of a change in the positional relationships among the plurality of speakers and the viewer. Then, based on a computation result from the viewing position computation unit 104, a positional relationship of each of the plurality of speakers with respect to the viewer is computed. Since a sound signal output parameter is set for each of the plurality of speakers from the computation result, the sound signal output parameter can be set in response to a real-time change in the positional relationships of the plurality of speakers and the viewer. With this, even when the viewer moves, sound volume, a delay, a directional characteristic, and a head transfer function of a sound from each speaker can be modified to provide the viewer with optimal sound state and virtual sound source reproduction effect.
In addition, since a factor is changed when a computation result of the viewing position computation unit 104 changes more than a predetermined amount, and when the computation result is stabilized for a predetermined period of time or longer, alleviation of a sense of discomfort due to excessive factor changing or the control efficiency can be improved.
Furthermore, since a factor is smoothly changed so that no discontinuous waveform is generated, noise does not occur. Thus, it is possible to follow a change in a viewing position without causing a sense of discomfort and continuously provide an appropriate sound field real time.
In addition, since a sound image normal position, which is a target of the virtual sound source reproduction, can be freely changed, the sound image normal position can be dynamically changed, such as fixing the sound image to a space, for example.
(2. Second Embodiment)
[2.1. Overview of a Second Embodiment]
In the following, a second embodiment of the present disclosure will be described. In the first embodiment as described above, the configuration for making a correction so that the virtual sound source reproduction effect can be maintained when a viewing position is displaced is shown. Specifically, as shown in
In contrast to this, the second embodiment shows an example in which the virtual sound source reproduction effect is positively changed in response to a change of a viewer position. Specifically, as shown in
A configuration of a sound field control device 100 according to the second embodiment is similar to
[2.2. Processing to be Performed in a Virtual Sound Source Reproduction Correction/Change Unit of the Second Embodiment]
In the following, a method for creating a head transfer function and a method for switching the head transfer function depending on a viewer position in the second embodiment will be described.
Now, when the viewer moves to a viewing position 2, unlike the embodiment 1, a position of the virtual sound source relative to the viewer considerably changes. Thus, it is essential to change from H1L, H1R to H2L, H2R.
As described above, according to the second embodiment, since the virtual sound source reproduction correction/change unit 130 performs processing so that a normal position of a sound image is maintained absolutely to a space, a viewer can have a perception of moving in the space by move in that space.
(3. Third Embodiment)
[3.1. Overview of a Third Embodiment]
In the following, a third embodiment of the present disclosure will be described. As shown in
[3.2. Configuration Example of the Third Embodiment]
With this, even when a camera is not mounted on the device 300 or a function is turned off (OFF), for example, posture of the device can be computed from the gyro sensor and a viewing position can be expected. Therefore, based on a viewing position, sound field correction processing similar to the first embodiment can be performed. A specific configuration of a sound control unit 106 is similar to the first embodiment as shown in
(4. Fourth Embodiment)
In the following, a fourth embodiment of the present disclosure will be described.
(5. Fifth Embodiment)
In the following, a fifth embodiment of the present disclosure will be described. As described above, the imaging unit 102 (and the viewing position computation unit 104) in the first embodiment may be provided in a separate device from a device in which a sound field control device 100 is provided. The fifth embodiment illustrates a configuration in which an imaging unit 102 is provided in a separate device from a device in which a sound field control device 100 is provided.
In the system of
(6. Sixth Embodiment)
In the following, a sixth embodiment of the present disclosure will be described. The sixth embodiment illustrates a case in which a normal position of a sound changes real time by manipulation of a user, such as a case in which a game is played on a personal computer or a tablet and the like.
When a user plays a game, a position of a sound source may move with a position of a display target object (display object) on a screen. For example, when a display target object such as a character, a car, an airplane and the like moves on the screen, a sense of reality can be enhanced by moving the position of the sound source of the display target object as the display target object moves. Also, when the display target object is displayed in three dimensions, the sense of reality can be enhanced by moving the position of the sound field accompanying movement of the display target object in a three-dimensional direction.
Such a movement of the display target object occurs as the game progresses or also occurs as a result of manipulation of the user.
In the case of a game, similar to
In this manner, when a normal position of a sound changes real time, an appropriate HRTF is dynamically computed considering a relative position of the virtual sound source position, in addition to information on the viewer (user) position and a reproduced sound source position. Since the virtual sound source position SPv changes real time in
A sound control unit 106 performs control similar to the first embodiment. Now, a virtual sound source reproduction correction unit 120 included in the sound control unit 106 sequentially changes HL(t) and HR(t) as time elapses with the above mathematical expression, on the basis of the position of the virtual sound source computed by the virtual sound source position computation unit 144, to compute the virtual sound source reproduction correction factor. With this, the position of the virtual sound source can be changed real time, depending on the position of the display target object.
As described above, according to the sixth embodiment, in such a case as a game in which a display target object moves while generating sound, a position of the virtual sound source can be changed real time with a position of the display target object. Therefore, a sound field with a sense of reality depending on a position of a display target object can be provided.
(7. Seventh Embodiment)
In the following, a seventh embodiment of the present disclosure will be described. As described in the sixth embodiment, when a virtual sound source position is controlled depending on a position of a display target object of a game, for example, a volume of computation by a CPU increases. Thus, load becomes too heavy for a CPU incorporated in a tablet, a smart phone and the like, and some cases in which desired control cannot be performed are also assumed. Therefore, it is more preferable to implement the sixth embodiment described above with the cloud computing described in the fourth embodiment. The seventh embodiment illustrates a case in which content of processing in such a preferred case is changed, depending on processing speed of the server (cloud computer 500) and the client (device 400), throughput of the client.
In the next step S34, the device 400 transmits a request for processing to the cloud computer 500. Now, the device 400 transmits sound data and information such as a viewer position, a sound source position, virtual sound source position information and the like to the cloud computer 500, requesting the cloud computer to perform processing.
The cloud computer 500 performs the processing according to the method for processing notified by the device 400 in step S30. In the next step S36, the cloud computer 500 transmits a reply to the request for processing to the device 400. In step S36, the cloud computer 500 sends back to the device 400 sound data after processing, or a reply on a factor necessary for the processing and the like.
For example, when a transmission rate with the cloud computer 500 is relatively fast although CPU capacity of the device 400 is insufficient, in step S34, the device 400 transmits metadata such as sound data, the viewer position, the sound source position, a virtual sound source position and the like to the cloud computer 500. Then, the device 400 requests the cloud computer 500 to select an appropriate HRTF from a volume of database, perform the virtual sound source reproduction processing, and return sound data after processing to the device 400. In step S36, the cloud computer 500 transmits the sound data after processing to the device 400. This enables higher precision, rich sound source processing with low CPU capacity in the device 400.
On the one hand, if the CPU capacity of the device 400 is sufficient, in step S34, the device 400 transmits the position information or only a difference thereof to the cloud computer 500. Then, in response to the request from the device 400, in step S36, the cloud computer 500 sends back to the device 400 the appropriate factor such as an HRTF and the like from the volume of database, and the virtual sound source reproduction processing is performed on the side of the client. In addition, the device 400 can make a faster response by preloading to the cloud computer 500 supplementary data for predicting position information such as HRTF data in the neighborhood of the position information or information on a difference of position information transmitted previously, rather than transmitting the position information itself such as a current viewer position, sound source position or virtual sound source position and the like in step S34.
In the case in which (1) an amount of characteristic of a HRTF is transmitted, rather than the cloud computer 500 sequentially transmitting to the device 400 a HRTF computed from position information and the like, a HRTF is transmitted once, and subsequently, a difference to the HRTF transmitted last time, an amount of change, is transmitted. With this, a transmission quantity can be minimized after the HRTF is transmitted once, thereby enabling reduction of the transmission band. On the one hand, since the device 400 sequentially computes a HRTF on the basis of the difference and the amount of change, the load on the CPU of the device 400 increases.
In the case in which (2) a HRTF is transmitted, the cloud computer 500 sequentially transmits a HRTF computed from the position information and the like to the device 400. In this case, since the HRTF is transmitted every time, the transmission band becomes larger than the case in (1). On the one hand, since the device 400 can sequentially receive the HRTF itself from the cloud computer 500, the load on the CPU of the device 400 is smaller than the case in (1).
In the case in which (3) information of an HRTF in which a sound source is convolved is transmitted, the cloud computer 500 sequentially transmits to the device 400 information (sound information) of a HRTF computed from position information and the like into which a sound source is further convolved. Specifically, the cloud computer 500 performs processing to the sound control unit 106 of the sound field control device 100. In this case, since an amount of information to be transmitted from the cloud computer 500 to the device 400 increases, the transmission band is larger than (1) and (2). On the one hand, since the device 400 can output sound by directly using the received information, the load on the CPU of the device 400 is smallest.
Information on which processing in (1) to (3) is performed is included in the notification of the method for processing that the device 400 transmits in step S30 of
In the following, a case in which the sound field control device 100 is a head tracking headphone will be described.
An input analog sound signal Ai, which is supplied to a terminal 611 and corresponds to a signal of a sound source 605, is converted to a digital sound signal Di by an A/D converter 621, and the digital sound signal Di is supplied to a signal processing unit 630.
As a unit including software (processing program) by a dedicated DSP (Digital Signal Processor) and the like or a hardware circuit, the signal processing unit 630 functionally consist of digital filters 631, 632, a time difference setting circuit 638, and a level difference setting circuit 639, and supplies the digital sound signal Di from the A/D converter 621 to the digital filers 631 and 632.
The digital filters 631 and 632 convolve impulse responses which correspond to transfer functions HLc and HRc reaching a left ear 1L and a right ear 1R of a listener 1 from the sound source 605, and consist of FIR filters, for example.
Specifically, in the digital filters 631 and 632, respectively, a sound signal supplied to input terminals is sequentially delayed by cascade-connected delay circuits for a delay time having a sampling period τ thereof, the sound signal supplied to the input terminals and the output signal of each delay circuit are multiplied by a factor of an impulse response in each multiplication circuit, the output signal of each multiplication circuit is sequentially added in each adder circuit, and the sound signal after filtering is obtained at the output terminal.
Sound signals L1 and R1 which are outputs of these digital filters 631 and 632 are supplied to the time difference setting circuit 638, and sound signals L2 and R2 which are outputs of the time difference setting circuit 638 are supplied to the level difference setting circuit 639. Sound signals L3 and R3 which are outputs of the level difference setting circuit 639 are D/A converted by D/A converters 641R, 641L and supplied to speakers 603R, 603L by way of elements 642R, 642L.
In the configuration described above, orientation of a face of the user wearing the headphone 600 can be detected from information obtained from a gyro sensor that the headphone is equipped with. This enables a virtual sound source position to be controlled, depending on the orientation of the headphone 600. For example, control can be performed so that the virtual sound source position does not change when the orientation of the headphone 600 changes. With this, the user wearing the headphone 600 can recognize that sound is generated from a same position even if the face of the user turns, which thus can enhance a sense of reality. In addition, the configuration for controlling the virtual sound source position on the basis of the information obtained from the gyro sensor can be made similar to the third embodiment.
(8. Eighth Embodiment)
In the following, an eighth embodiment of the present disclosure will be described. In the eighth embodiment, when a sound field control device 100 is incorporated in a small device such as a smart phone, a virtual sound source is reproduced through the use of a ultrasonic speaker. In the small device such as the smart phone, since spacing between right and left speakers is narrow, it is difficult to cancel crosstalk in which right and left sounds are mixed. In such a case, use of the ultrasonic speaker in the small device such as the smart phone enables cancellation of the crosstalk.
(9. Ninth Embodiment)
In the following, a ninth embodiment of the present disclosure will be described. The ninth embodiment describes a case in which a sound source is configured in a device separate from a device for sensing a viewer's position or orientation such as a camera or an ultrasonic sensor, a gyro sensor and the like.
The preferred embodiments of the present disclosure have been described above with reference to the accompanying drawings, whilst the technical scope of the present disclosure is not limited to the above examples, of course. A person skilled in the art may find various alterations and modifications within the scope of the appended claims, and it should be understood that they will naturally come under the technical scope of the present invention.
Additionally, the present technology may also be configured as below.
a display target object position information acquisition unit for acquiring position information of a display target object corresponding to a sound source; and
a virtual sound source position control unit for controlling a virtual sound source position on the basis of position information of the display target object.
a transmission unit for transmitting, to an external computer, at least the position information of the display target object; and
a reception unit for receiving, from the external computer, a virtual sound source reproduction correction factor computed on the basis of the position information of the display target object or information generated on the basis of the virtual sound source reproduction correction factor.
wherein the transmission unit transmits, to the external computer, sound data together with the position information of the display target object, and
wherein the reception unit receives, from the external computer, sound data that is obtained by correcting the sound data with the virtual sound source reproduction correction factor computed on the basis of the position information of the display target object.
a viewer position information acquisition unit for acquiring position information of a viewer,
wherein the virtual sound source position control unit controls the virtual sound source position on the basis of the position information of the display target object and the position information of the viewer.
a transmission unit for transmitting, to the external computer, the position information of the display target object and the position information of the viewer; and
a reception unit for receiving, from the external computer, a virtual sound source reproduction correction factor computed on the basis of the position information of the display target object and the position information of the viewer or information generated on the basis of the virtual sound source reproduction correction factor.
wherein the transmission unit transmits, to the external computer, sound data together with the position information of the display target object and the position information of the viewer, and
wherein the reception unit receives, from the external computer, sound data which is obtained by correcting the sound data with the virtual sound source reproduction correction factor computed on the basis of the position information of the display target object and the position information of the viewer.
acquiring position information of a display target object corresponding to a sound source; and
controlling a virtual sound source position on the basis of the position information of the display target object.
means for acquiring position information of a display target object corresponding to a sound source; and
means for controlling a virtual sound source position on the basis of the position information of the display target object.
a client terminal including
the external computer including
the external computer including
acquiring, by a client terminal, position information of a display target object corresponding to a sound source;
transmitting, by the client terminal, the position information of the target object to an external computer;
receiving, by the external computer, the position information of the display target object;
computing, by the external computer, the virtual sound source reproduction correction factor on the basis of the position information of the display target object; and
transmitting, by the external computer, to the client terminal, the virtual sound source reproduction correction factor or information generated on the basis of the virtual sound source reproduction correction factor.
a position information acquisition unit for acquiring position information of a viewer from information obtained by imaging; and
a virtual sound source position control unit for controlling a virtual sound source position on the basis of the position information.
a control unit for controlling sound volume, a delay amount of sound, or a directional characteristic, on the basis of the position information.
an imaging unit for acquiring the position information of the viewer.
a posture information acquisition unit for acquiring posture information,
wherein the virtual sound source position control unit controls the virtual sound source position, on the basis of the position information and the posture information.
acquiring position information of a viewer; and
controlling a virtual sound source position on the basis of the position information.
means for acquiring position information of a viewer; and
means for controlling a virtual sound source position on the basis of the position information.
an imaging device for imaging a viewer; and
a sound field control device including
Number | Date | Country | Kind |
---|---|---|---|
2012-003266 | Jan 2012 | JP | national |
2012-158022 | Jul 2012 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2012/083078 | 12/20/2012 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2013/105413 | 7/18/2013 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6490359 | Gibson | Dec 2002 | B1 |
20030091204 | Gibson | May 2003 | A1 |
20040240686 | Gibson | Dec 2004 | A1 |
20060064300 | Holladay | Mar 2006 | A1 |
20070011196 | Ball et al. | Jan 2007 | A1 |
20090252379 | Kondo et al. | Oct 2009 | A1 |
20100150355 | Kon | Jun 2010 | A1 |
20100323793 | Andall | Dec 2010 | A1 |
20100328423 | Etter | Dec 2010 | A1 |
20110007915 | Park | Jan 2011 | A1 |
20110235807 | Hayashi et al. | Sep 2011 | A1 |
20130121515 | Hooley | May 2013 | A1 |
Number | Date | Country |
---|---|---|
9-90963 | Apr 1997 | JP |
2005-49656 | Feb 2005 | JP |
2005-295181 | Oct 2005 | JP |
2006-94315 | Apr 2006 | JP |
2007-81928 | Mar 2007 | JP |
2007-214897 | Aug 2007 | JP |
2010-206451 | Sep 2010 | JP |
Entry |
---|
Extended European Search Report issued Jul. 20, 2015 in Patent Application No. 12865517.2. |
International Search Report issued Jan. 29, 2013 in PCT/JP2012/083078. |
Combined Chinese Office Action and Search Report issued Dec. 2, 2015 in Patent Application No. 201280066052.8 (with English language translation). |
Japanese Office Action issued May 10, 2016 in Patent Application No. 2013-553232 (without English Translation). |
Number | Date | Country | |
---|---|---|---|
20140321680 A1 | Oct 2014 | US |